Hacker Newsnew | past | comments | ask | show | jobs | submit | veunes's commentslogin

This is the inevitable evil of the man in the middle. OpenRouter by definition decrypts your traffic to route it to the provider (OpenAI, Anthropic). Technically, they can read everything The problem is that for the Enterprise segment, this is a showstopper. No bank or hospital will route data through an aggregator that openly states it classifies prompts via Google API (even sampled ones). This confirms that OpenRouter remains a tool for indie hackers and researchers, not for serious B2B

I'm not surprised. Roleplay means endless sessions with huge context (character history, world, previous dialogues). On commercial APIs (OpenAI/Anthropic), that long-context costs a fortune. On OpenRouter, many OSS models, especially via providers like DeepInfra or Fireworks, cost pennies or are even free, like some Free-tier models. The RP community is very price-sensitive, so they massively migrate to cheap OSS models via aggregators. It skews the stats but highlights a real niche for cheap inference

Yeah, using an API aggregator to run a 7B model is economically strange if you have even a consumer GPU. OpenRouter captures the cream of complex requests (Claude 3.5, o1) that you can't run at home. But even for local hosting, medium models are starting to displace small ones because quantization lets you run them on accessible hardware, and the quality boost there is massive. So the "Medium is the new Small" trend likely holds true for the self-hosted segment as well.

The 4x growth in prompt length is a fundamental shift. We've quickly moved from "Q&A" mode to "upload full context and analyze" mode.

This completely changes infrastructure requirements: KV-caching becomes a necessity, and prefill time becomes a critical metric, often more important than generation speed. That's exactly why models with cheap long context (Gemini, DeepSeek) are winning the race against "smarter" but expensive models. Inference economics are now dictated by context length


That engine had character

What's tricky is that even tiny improvements in fuel economy or emissions can justify a redesign when you're building at scale

Plenty of engines hit the 20-year mark, but not many do it while still powering everything from family sedans to track-ready Type Rs

The V6 2GR is about to do it for Toyota. Subaru had the EJ going strong for 20 years. Honda has this. Nissan had a few contenders.. there's a list: https://en.wikipedia.org/wiki/Ward%27s_10_Best_Engines

I think the article's broader point still stands: Honda built a platform with enough foresight and flexibility that it could be continuously refined rather than scrapped and replaced every few years

True, but this is actually common for engine manufacturers. Most manufacturers have an engine series that spans multiple decades.

> rather than scrapped and replaced every few years

This doesn’t really happen for mainstream engines. Maybe for specialty and exotic engines, but not the engines you see powering the commuter cars and trucks on the road. Engine development is expensive. Nobody is scrapping and replacing their bread and butter engine design every few years.


What's impressive isn't just the longevity, but how gracefully it's evolved over two decades while still feeling relevant in today's turbocharged, emissions-strangled landscape

The own shovels for own mines strategy has a hidden downside: isolation. NVIDIA sells shovels to everyone - OpenAI, Meta, xAI, Microsoft - and gets feedback from the entire market. They see where the industry is heading faster than Google, which is stewing in its own juices. While Google optimizes TPUs for current Google tasks (Gemini, Search), NVIDIA optimizes GPUs for all possible future tasks. In an era of rapid change, the market's hive mind usually beats closed vertical integration.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: