Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yeah, JSON mode in Ollama, which isn’t even the full llama.cpp grammar functionality, performs better than OpenAI for me at this point. I don’t understand how they can be raking in billions of dollars and can’t even get this basic stuff right.


llama.cop Performs better than what?

3.5 Turbo, one of the 4.0 models, API or App?


JSON mode and function-calling with a JSON schema in the OpenAI API.


Right, but which model?

It makes a huge difference.


I’ve been using OpenChat 3.5 1210 most recently. Before that, Mistral-OpenOrca. Both return JSON more consistently than gpt-3.5-turbo.


gpt-3.5-turbo is not the benchmark


I don’t know what point you’re trying to make. They also return JSON more consistently than gpt-4, but I don’t use that because it’s overkill and expensive for my text extraction tasks.


Because people have different interests and want to hear your results for different reasons.

Some want to consider results relative to cost, and some are interested only in how it compares to SOTA.


I mean, sure, but the parent should also just explicitly state what it is they were asking or claiming. I’ve answered every question asked. Making vague declarations about something not being “the benchmark,” while not stating what you think “the benchmark” should be, is unhelpful.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: