Hello, everyone.
I'm a solo founder based in Argentina, and I created neuralrouting.io. It's an LLM router designed to evaluate prompt complexity and select the most cost-effective model.
Instead of sending everything to GPT-4, it routes simpler queries to cheaper models and caches similar requests so you don’t pay twice.
I also built a Shadow Engine that runs cheaper models in parallel to benchmark quality, filter PII, rate limit, and detect agent loops. It’s still early and only supports OpenAI and Groq right now.
No one’s really using it yet, so I’m looking for devs or small teams working with LLMs who want to break it, give feedback, and tell me what’s missing.
If you wanna try it out or ask questions, check it out at neuralrouting.io or just reply here!
Would love to hear your thoughts.