Hacker Newsnew | past | comments | ask | show | jobs | submit | momo_O's commentslogin

Jeopardy is a bad example because you're required to wait until the end of the question before buzzing in, or else you are penalized.


Yes but if you don’t know the answer by the time the light goes on (the question is finished read), you will never get in. And if you buzz in without knowing the answer you will lose points. So you have to know the answer before the light goes , then be ready to buzz as soon as eligible. Jeopardy is a good example.


I struggle to get o1 (or any chatgpt model) is getting it to stick to a context.

e.g. I will upload a pdf or md of an library's documentation and ask it to implement something using those docs, and it keeps on importing functions that don't exist and aren't in the docs. When I ask it where it got `foo` import from, it says something like, "It's not in the docs, but I feel like it should exist."

Maybe I should give o1 pro a shot, but claude has never done that and building mostly basic crud web3 apps, so o1 feels like it might be overpriced for what I need.


I don't think any reasonable person is using "lines of code" as a hard metric for anything -- as you've noted, the quality and impact of code isn't proportional to its length.

Lines of code serves as a directional heuristic at best, but that's ok.


> if you can make a reasonable guess as to the total contained information about your topic within the given size of model

Can you expand a bit on how you might do this?


There's something you don't know that it may know and you want to see what it knows. This is like just a sentence or maybe a few both input and output. All the other talk about this model vs that model vs agents vs rag vs prompt engineering is all about practitioner worries. Keep in mind the thing is probably wrong as you would with any of them. Or that they are subliminally telling you something wrong you may accidentally repeat in front of someone at some time where it really matters and you're going to let everyone and yourself down. Which is current state of all of these things, so, if you're not building them, or are an NLP specialist working with multidisciplinary researchers on a specific goal of pushing research, then these things all have the same utility at the end of the day. Some of the most short sighted systems advice seems to just spill out of Claude unsolicited, so, whatever the big models are up to isn't entirely helpful in my opinion. Hopefully they'll be pressured to reveal their prompts and other safety measures.


Deep learning requires attention span.

I fear that shifting to a TikTok style education is just another lead towards an Idiocracy.


How is the workaround structured?


How old were you when you transitioned?


> Crypto might well be the biggest vehicle for financial fraud the world has ever seen.

How do you measure this? If by “total value defrauded” then fiat has to be #1.


sigh

What point are you trying to drive here?


Source?


> Other folks got money out of Madoff during his scam years, and the trustee says “I’m taking that money back to pay back others.”

I wonder if things like the political donations could be clawed back.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: