BuildShipBBuildShip
Powered by
anywhichwayA
BuildShip•2y ago•
1 reply
anywhichway

inference speed on Groq using Llama 3 70B through Buildship

The inference speed on Groq using Llama 3 70B through Buildship seems to vary dramatically. At times I get 9 and 10 second delays that I do not get when interacting with Groq directly. It has the feel of cold starting somewhere, but I am on the Pro plan where pricing page says no cold starts. This may be true of other inference engines, I do not know.
BuildShip banner
BuildShipJoin
Join founders, builders, devs using Low-code / No-code to create backend, APIs, scheduled Jobs, automation, AI workflows
7,971Members
Resources
Recent Announcements

Similar Threads

Was this page helpful?

Similar Threads

Groq (or custom code) on buildship
thiwaTthiwa / ❓・buildship-help
2y ago
FlutterFlow transcribe recorded Audio through BuildShip API
MickMMick / ❓・buildship-help
12mo ago
can we build zapier functionalities through buildship?
SudheerSSudheer / ❓・buildship-help
2y ago
replicate api for Llama 3 not working
HocusPocusHHocusPocus / ❓・buildship-help
2y ago