Hacker Newsnew | past | comments | ask | show | jobs | submit | more lvl155's commentslogin

This is purely price gouging because these rams are not ECC and server grade.


The article references the original coverage which talks to this:

> Despite server-grade RDIMM memory and HBM being the main attractions for hardware manufacturers building AI servers, the entire memory industry, including DDR5, is being affected by price increases. The problem for consumers is that memory manufacturers are shifting production prioritization toward datacenter-focused memory types and producing less consumer-focused DDR5 memory as a result.

But I'm sure the hysteria around that isn't helping prices come back down either.


Except when you have datacenters also building racks with desktop hardware. I believe that was hetzner?


I really don’t understand the hype around Gemini. Opus/Sonnet/GPT are much better for agentic workflows. Seems people get hyped for the first few days. It also has a lot to do with Claude code and Codex.


Gemini is a lot more bang for the buck. It's not just cheaper per token, but with the subscription, you also get e.g. a lot more Deep Research calls (IIRC it's something like 20 per day) compared to Anthropic offerings.

Also, Gemini has that huge context window, which depending on the task can be a big boon.


Google deep research writes way too much useless fluff though, like introduction to the industry etc.


I'm completely the opposite. I find Gemini (even 2.5 Pro) much, much better than anything else. But I hate agentic flows, I upload the full context to it in aistudio and then it shines - anything agentic cannot even come close.


I recently wrote a small CLI tool for scanning through legacy codebases. For each file, it does a light parse step to find every external identifier (function call, etc...), reads those into the context, and then asks questions about the main file in question.

It's amazing for trawling through hundreds of thousands of lines of code looking for a complex pattern, a bug, bad style, or whatever that regex could never hope to find.

For example, I recently went through tens of megabytes(!) of stored procedures looking for transaction patterns that would be incompatible with read committed snapshot isolation.

I got an astonishing report out of Gemini Pro 3, it was absolutely spot on. Most other models barfed on this request, they got confused or started complaining about future maintainability issues, stylistic problems or whatever, no matter how carefully I prompted them to focus on the task at hand. (Gemini Pro 2.5 did okay too, but it missed a few issues and had a lot of false positives.)

Fixing RCSI incompatibilities in a large codebase used to be a Herculean task, effectively a no-go for most of my customers, now... eminently possible in a month or less, at the cost of maybe $1K in tokens.


If this is a common task for you, I'd suggest instead using an LLM to translate your search query into CodeQL[1], which is designed to scan for semantic patterns in a codebase.

1. https://codeql.github.com/


+1 - Gemini is consistently great at SQL in my experience. I find GPT 5 is about as good as gemini 2.5 pro (please treat is as praise). Haven't had a chance to put Gemini 3 to a proper sql challenge yet.


Is there any chance you'd be willing to share that tool? :)


It's a mess vibe coding combined with my crude experiments with the new Microsoft Agent Framework. Not something that's worth sharing!

Also, I found that I had to partially rewrite it for each "job", because requirements vary so wildly. For example, one customer had 200K lines of VBA code in an Access database, which is a non-trivial exercise to extract, parse, and cross-reference. Invoking AI turned out to be by far the simplest part of the whole process! It wasn't even worth the hassle of using the MS Agent Framework, I would have been better off with plain HTTPS REST API calls.


I think you're both correct. Gemini is _still_ not that good at agentic tool usage. Gemini 3 has gotten A LOT better, but it still can do some insane stupid stuff like 2.5


Personally my hype is for the price, especially for Flash. Before Sonnet 4.5 was competitive with Gemini 2.5 Pro, the latter was a much better value than Opus 4.1.


with gemini you have to spend 30 minutes deleting hundreds of useless comments littered in the code that just describe what the code itself does


The comments would improve code quality because it's a way for the LLM to use a scratchpad to perform locally specific reasoning before writing the proceeding code block, which would be more difficult for the LLM to just one shot.

You could write a postprocessing script to strip the comments so you don't have to do it manually.


I haven't had a comment generated for 3.0 pro at all unless specified.


Do what’s best for your son. There’s nothing that will overcome that guilt. Not saying what’s best for your son is for you to reveal the truth now or later. That’s entirely situational and only you know.


I don't think it's fair to consider only the son in this. I'd say do what's best for both your son and your wife. And to me that's pretty clearly telling his wife the truth here.


I think this is right. Many parents could think of choosing to opt out because raising a family can be difficult sometimes, but it’s a responsibility people have taken. Take it seriously. It’s a bit chicken to say “I’m out” whatever the reason. See them through their formative years then do your thing.


I completely agree. Life as a parent is only about oneself insofar as looking after yourself is good for your family.


children mimic the actions and lifestyle of their parents

being a role model means demonstrating how to be happy

an existence proof of a good and happy life is a powerful thing

people learn at a young age that words are cheap. advice doesn't cut it


The kid will be fine. Couples separate or get divorced all the time for all kinds of reasons. “I have two loving parents who are now separated” doesn’t appear on the shortlist of bad things that can happen to kids.


It’s because their cogs is basically same. They are vertically integrated.


Just a small note worth mentioning: McD do not buy things from anyone. They literally do it all from scratch. Vertically integrated.


You just listed eggs, pork and chicken. They will run you way over $0.50 a plate. Even if you shop at absolute dirt cheap groceries, it’s more like $4-5 per plate when you factor in all the costs not including labor.


It’s always surprising to me how so many people believe in ideals. Finding love. Living happily thereafter. These are ideals and you’d be truly lucky to have it in your life. One in a billion. Life is ephemeral. We are innately fickle beings shifting from one equilibrium to another. Yet we long for a mirage of permanence.


Judging by the follow up that reveals he was cheating on his lover as well and had been doing it for years he seems to agree with your last two lines. He got the most out of life, but hurt his lover, wife and daughter in the process


It’s obviously part of the human condition to yearn for the ideal/divine/essential.


You don’t find love: you build it together.


This problem is so much worse when you look at server mobo configurations that basically jumped from 8 to 12 slots. Meaning you need 50% more sticks to saturate versus Epyc 7003/2. I was hoping to build a Genoa-X server for compute and ram cost just went bonkers (that’s on a nearly 2-yo system). I decided to shelve that idea.


Design is hard because models change almost on a weekly basis. Quality abruptly falls off or drastically changes without announcements. It’s like building a house without proper foundation. I don’t know how many tokens I wasted because of this. I want to say 30% of the cost and resources this year for me.


Let’s just say what it is: devs are too constrained to jump ship right now. It’s a massive land grab and you are not going to spend time tinkering with CUDA alternatives when even a six-month delay can basically kill your company/organization. Google and Apple are two companies with enough resources to do it. Google isn’t because they’re keeping it proprietary to their cloud. Apple still have their heads stuck in sand barely capable of fixing Siri.


Google has their own TPUs so they don’t have any vendor lock-in issues at all.

OpenAI OTOH is big enough that the vendor lock-in is actually hurting them, and them making that massive deal with AMD may finally push the needle for AMD and improve things in the ecosystem to make AMD a smooth experience.


Having your own ASIC comes with a huge sunk cost. One gets advantages from that, but it's still a lock, just a lock of a different color. But with Google money and manpower, management can probably pursue both paths in parallel and not care.


Google’s TPU’s are not powering Gemini or whatever X equivalent LLM you want to compare to.


This isn't true. Gemini is trained and run almost entirely on TPUs. Anthropic also uses TPUs for inference, see, e.g., https://www.anthropic.com/news/expanding-our-use-of-google-c... and https://www.anthropic.com/engineering/a-postmortem-of-three-.... OpenAI also uses TPUs for inference at least in some measure: https://x.com/amir/status/1938692182787137738?t=9QNb0hfaQShW....


I can assure you that most internal ML teams are using TPUs both for training and inference, they are just so much easier to get. Whatever GPUs exist are either reserved for Google Cloud customers, or loaned temporarily to researchers who want to publish easily externally reproducible results.



They are, even Apple famously uses Google Cloud for their cloud based AI stuff solely because of Apple not wanting to buy NVidia.

Google Cloud does have a lot of NVidia, but that’s for their regular cloud customers, not internal stuff.


What is powering Gemini?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: