Yes but if you don’t know the answer by the time the light goes on (the question is finished read), you will never get in. And if you buzz in without knowing the answer you will lose points. So you have to know the answer before the light goes , then be ready to buzz as soon as eligible. Jeopardy is a good example.
I struggle to get o1 (or any chatgpt model) is getting it to stick to a context.
e.g. I will upload a pdf or md of an library's documentation and ask it to implement something using those docs, and it keeps on importing functions that don't exist and aren't in the docs. When I ask it where it got `foo` import from, it says something like, "It's not in the docs, but I feel like it should exist."
Maybe I should give o1 pro a shot, but claude has never done that and building mostly basic crud web3 apps, so o1 feels like it might be overpriced for what I need.
I don't think any reasonable person is using "lines of code" as a hard metric for anything -- as you've noted, the quality and impact of code isn't proportional to its length.
Lines of code serves as a directional heuristic at best, but that's ok.
There's something you don't know that it may know and you want to see what it knows. This is like just a sentence or maybe a few both input and output. All the other talk about this model vs that model vs agents vs rag vs prompt engineering is all about practitioner worries. Keep in mind the thing is probably wrong as you would with any of them. Or that they are subliminally telling you something wrong you may accidentally repeat in front of someone at some time where it really matters and you're going to let everyone and yourself down. Which is current state of all of these things, so, if you're not building them, or are an NLP specialist working with multidisciplinary researchers on a specific goal of pushing research, then these things all have the same utility at the end of the day. Some of the most short sighted systems advice seems to just spill out of Claude unsolicited, so, whatever the big models are up to isn't entirely helpful in my opinion. Hopefully they'll be pressured to reveal their prompts and other safety measures.