# GPUtopia/Arguflow bounty integration ## Requirements - [X] OpenAI compliant API over an LLM thru [LocalAI](https://github.com/arguflow/LocalAI) running on gputopia network - [X] Text-embedding model running in [Arguflow embedding server](https://github.com/arguflow/arguflow/tree/main/embedding-server) on gputopia network - [ ] ~10-minute deployment of Arguflow on top of GPUTopia compute ## Bounty amount ## Proposed path forward for gputopia - advanced routing and model preloading - always picking handler that have the model loaded - sending out preloads in anticipation of future requests - check fastembed/onyx vs torch for exe size and performance esp parallelism, one of these should work - bge-large-en-v1.5 - ember-v1 - gte-large - stella-base-en-v2 ## For Arguflow - please test our api endpoint (sign up get an api token, and try it out) - send feeback and requests until it works nicely