@derekpankaew yes I'm sure this would have been solid for the Yolo models you were running at Next Fitness! With models that small too they would have scaled up in the matter of a couple seconds too :)
This looks seriously impressive. I'll be trying it out soon. Seems like this will be a huge step up over Google Cloud Run in terms of speed. I see on your roadmap that you're planning on moving to beefier GPUs in the future. Which GPUs are you running now?
Also, from a technical perspective, I assume this works by moving models from CPU to GPU at inference time? Trying to wrap my head around how you're getting such fast cold starts.
@gregpriday Thanks for the support!
The roadmap is now! We used to only do T4 GPUs, but now we also support A100 GPUs which are yielding faster cold boots + inference + download speeds.
To understand how banana works it may be easier to think of us as a compiler company. When you send us a model we do stuff under-the-hood to make it run faster/cheaper. CPU/GPU memory hacks are definitely involved (how we load memory, where, when). A key point is none of our optimizations affect model outputs. This means we don't do weight quantization dynamic layer/node pruning which yield way smaller/faster models but does affect output.
I have ideas that involve ML, and seeing products being launched that make it easier for our dev community to deploy and run encourages me to take my ideas seriously. Thank you! I shall give it a try for sure
Are there any metrics around the cold boot? Does it depend on the model size, etc? And what does the quota look like on the max number of concurrency, model size, etc?
@eric_j1 Yes! Cold boots vary based on model size, but a GPTJ model (which takes 20 minutes to load to GPU usually) comes live on our platform in 10 seconds. Most customers see 1-5s cold boots right now.
We just upped max concurrency! We're provisioned for the average model (8gb GPU RAM) to spike to 200x concurrency, but do have a soft cap of 10 to prevent customers from accidentally overscaling. We can adjust that for anyone who needs more :)
Been using Banana in production for a good bit now, nothing but great things to share! Few notes:
- Product is insanely good. Specifically, we use it for indexing jobs requiring a good bit of GPU compute. These jobs are huge, sometimes involving up to 1M inferences of a large NL model. Banana is perfect for this use case, as we can burst up to 10+ GPUs, only pay for the compute we use, and quickly scale back down to near zero.
- Team is very strong, super responsive to questions and are experts at deploying & scaling ML models. We often get advice and recommendations from their team on how to best do something, and it's been really appreciated!
- Lastly, velocity / speed of iteration has been ridiculous. They're moving really quick, have an ambitious roadmap, and ship new features and improvements daily. It's been really cool to watch.
Would highly recommend anyone check them out!
@gallantlabs Thank you for your kind words & support, Morgan! Fantastic having you as a customer and inspiring seeing your progress as a team! Always a message away :)
gpu coldstarts is a tough problem to solve! glad banana is taking on this challenge! definitely super cool product! especially for hobby projects that require gpus and you don't want to pay an arm and an leg to host a demo
Product Hunt