Hacker Newsnew | past | comments | ask | show | jobs | submit | whoaoweird's commentslogin

I moved to Linux after Win10 stopped receiving updates. It's WAY better than I expected it would be. Highly recommend people making a cut over. (I used Bazzite, but there's other options out there.)


It's wild to see you downvoted. Only about 10% of blind people know braille. There are many more people who have visual impairments but are not blind. Braille is not a universal solution (though I would rather have it than not have it).


But you don't need to know braille to learn how the most common bills are marked.

Just like you don't need to know Japanese to count the exact amount of yen bills.


Hackernews is filled with you healthy 20 year olds who do not understand that many will by the age of 30 have reduced vision.

Society is not supposed to be engineered for young single healthy 20 year old males.


Chiming in to complain that a good, working solution to a problem just doesn't happen to solve ALL PROBLEMS is just banality or perhaps pedantry. Unless it was also proposing an alternative that might do better...

Braille on money also doesn't help dyslexic quadrplegics with dysesthesia... Checkmate.


I think that's an extremely ungenerous read here. The thread is about how different size bills and different color bills solve a lot of problems with people who have low vision. Adding braille solves the same problem, but for a subset of people that different sized/color bills solves.

If you have a good, working solution that's widely used worldwide, and someone suggests a worse solution that works for fewer people, it's more than fair to point out that "your solution is worse, less common, and works for fewer people".

Your last sentence is a low effort strawman, I'm not sure why you felt it necessary to include.


It was interesting to see how often the OpenAI model changed the face of the child. Often the other two models wouldn't, but OpenAI would alter the structure of their head (making it rounder), eyes (making them rounder), or altering the position and facing of the children in the background.

It's like OpenAI is reducing to some sort of median face a little on all of these, whereas the other two models seemed to reproduce the face.

For some things, exactly reproducing the face is a problem -- for example in making them a glass etching, Gemini seemed unwilling to give up the specific details of the child's face, even though that would make sense in that context.


It looks to me like OpenAI's image pipeline takes an image as input, derives the semantic details, and then essentially regenerates an entirely new image based on the "description" obtained from the input image.

Even Sam Altman's "Ghiblified" twitter avatar looks nothing like him (at least to me).

Other models seem much more able to operate directly on the input image.


You can see this in the images of the Newton: in GPT's versions, the text and icons are corrupted.


Isn't this from the model working o. really low res images, and then bein uppscalef afterwards?



This is inherent in the architecture of chatgpt. It's a unified model: text, images, etc all become tokenized input. It's similar to re-encoding your image in a lossy format, the format is just the black box of chatgpt's latent space.

This leads to incredibly efficient, dense semantic consistency because every object in an image is essentially recreated from (intuitively) an entire chapter of a book dedicated to describing that object's features.

However, it loses direct pixel reference. For some things that doesn't matter much, but humans are very discerning regarding faces.

Chatgpt is architecturally unable to reproduce exactly the input pixels - they're always encoded into tokens, then decoded. This matters more for subjects for which we are sensitive to detail loss, like faces.


Encoding/decoding tokens doesn't automatically mean lossy. Images, at least in term of raw pixels can be a very inefficient form of storing information from information theoretic perspective.

Now, the difficulty is in achieving an encoding/decoding scheme that is both: information efficient AND semantically coherent in latent space. Seems like there is a tradeoff here.


I've noticed that OpenAI modifies faces on a regular basis. I was using it to try and create examples of different haircuts and the face would randomly turn into a different face -- similar but noticeably changed. Even when I prompted to not modify the face, it would do it regardless. Perhaps part of their "safety" for modifying pictures of people?


I had thought it was a deliberate choice to avoid potential abuse, however Sora put an end to that line of thinking.


It's also changing scene features too. Like removed background trees.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: