9 comments

  • Oras an hour ago

    The problem is well articulated and nice story for both cofounders.

    One thing I don’t get is why would anyone use a direct service that does the same thing as others when there are services such as openrouter where you can use the same model from different providers? I would understand if your landing page mentioned fine-tuning only and custom models, but just listing same open source models, tps and pricing wouldn’t tell me how you’re different from other providers.

    I remember using banana.dev a few years ago and it was very clear proposition that time (serverless GPU with fast cold start)

    I suppose positioning will take multiple iterations before you land on the right one. Good luck!

    • 2uryaa 11 minutes ago

      Hey Oras, thank you for the feedback! I think we definitely could list on OpenRouter but as you point out, our end goal is to host finetuned models for individuals. The IonRouter product is mostly to showcase our engine. In the backend, we are multiplexing finetuned and open source models on a homogenous fleet of GPUs. So if you feel no performance difference on our cloud, we're already proving what we set out to show.

      I do think we will lean harder into the hosting of fine-tuned models though, this is a good insight.

  • GodelNumbering 2 hours ago

    As an inference hungry human, I am obviously hooked. Quick feedback:

    1. The models/pricing page should be linked from the top perhaps as that is the most interesting part to most users. You have mentioned some impressive numbers (e.g. GLM5~220 tok/s $1.20 in · $3.50 out) but those are way down in the page and many would miss it

    2. When looking for inference, I always look at 3 things: which models are supported, at which quantization and what is the cached input pricing (this is way more important than headline pricing for agentic loops). You have the info about the first on the site but not 2 and 3. Would definitely like to know!

    • 2uryaa 9 minutes ago

      Thank you for the feedback! I think we will definitely redo the info on the frontpage to reorg and show quantizations better. For reference, Kimi, GLM, Minimax are NVFP4. The rest are FP8. But I will make this more obvious on the site itself.

  • reactordev an hour ago

    “Pricing is per token, no idle costs: GPT-OSS-120B is $0.02 in / $0.095 out, Qwen3.5-122B is $0.20 in / $1.60 out. Full model list and pricing at https://ionrouter.io.”

    Man you had me panicking there for a second. Per token?!? Turns out, it’s per million according to their site.

    Cool concept. I used to run a Fortune 500’s cloud and GPU instances hot and ready were the biggest ask. We weren’t ready for that, cost wise, so we would only spin them up when absolutely necessary.

  • nylonstrung an hour ago

    Unless I misunderstood it seems like this is trailing the pareto frontier in cost and speed.

    Compare to providers like Fireworks and even with the openrouter 5% charge it's not competitive

  • cmrdporcupine an hour ago

    Very cool, I see that "Deploy your finetunes, custom LoRAs, or any open-source model on our fleet." is "Book a call" -- any sense of what pricing will actually look like here, since this seems like it's kind of where your approach wins out, the ability to swap in custom model easier/cheaper?

    Just curious how close we are to a world where I can fine tune for my (low volume calls) domain and then get it hosted. Right now this is not practical anywhere I've seen, at the volumes I would be doing it at (which are really hobby level).

  • erichocean an hour ago

    > what would make this actually useful for you?

    A privacy policy that's at least as good as Vertex.ai at Google.

    Otherwise it's a non-starter at any price.

    • Oras an hour ago

      What's unique about Vertex's privacy policy?