Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I asked it to analyze my tennis serve. It was just dead wrong. For example, it said my elbow was bent. I had to show it a still image of full extension on contact, then it admitted, after reviewing again, it was wrong. Several more issues like this. It blamed it on video being difficult. Not very useful, despite the advertisements: https://x.com/sundarpichai/status/1990865172152660047


The default FPS it's analyzing video at is 1, and I'm not sure the max is anywhere near enough to catch a full speed tennis serve.


Ah, I should have mentioned it was a slow motion video.

> The default FPS it's analyzing video at is 1

Source?


https://ai.google.dev/gemini-api/docs/video-understanding#cu...

"By default 1 frame per second (FPS) is sampled from the video."


OK, I just used https://gemini.google.com/app, I wonder if it's the same there.


I’ve never seen such a huge delta between advertised capabilities and real world experience. I’ve had a lot of very similar experiences to yours with these models where I will literally try verbatim something shown in an ad and get absolutely garbage results. Do these execs not use their own products? I don’t understand how they are even releasing this stuff.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: