Gemini 3 Deep Think (blog.google)
494 points by tosh 6 hours ago
lukebechtel 6 hours ago
Arc-AGI-2: 84.6% (vs 68.8% for Opus 4.6)
Wow.
https://blog.google/innovation-and-ai/models-and-research/ge...
raincole 3 hours ago
Even before this, Gemini 3 has always felt unbelievably 'general' for me. It can beat Balatro (ante 8) with text description of the game alone[0]. Yeah, it's not an extremely difficult goal for humans, but considering:
1. It's an LLM, not something trained to play Balatro specifically
2. Most (probably >99.9%) players can't do that at the first attempt
3. I don't think there are many people who posted their Balatro playthroughs in text form online
I think it's a much stronger signal of its 'generalness' than ARC-AGI. By the way, Deepseek can't play Balatro at all.
ankit219 8 minutes ago
Agreed. Gemini 3 Pro for me has always felt like it has had a pretraining alpha if you will. And many data points continue to support that. Even as flash, which was post trained with different techniques than pro is good or equivalent at tasks which require post training, occasionally even beating pro. (eg: in apex bench from mercor, which is basically a tool calling test - simplifying - flash beats pro). The score on arc agi2 is another datapoint in the same direction. Deepthink is sort of parallel test time compute with some level of distilling and refinement from certain trajectories (guessing based on my usage and understanding) same as gpt-5.2-pro and can extract more because of pretraining datasets.
(i am sort of basing this on papers like limits of rlvr, and pass@k and pass@1 differences in rl posttraining of models, and this score just shows how "skilled" the base model was or how strong the priors were. i apologize if this is not super clear, happy to expand on what i am thinking)
tl 10 minutes ago
Per BalatroBench, gemini-3-pro-preview makes it to round (not ante) 19.3 ± 6.8 on the lowest difficulty on the deck aimed at new players. Round 24 is ante 8's final round. Per BalatroBench, this includes giving the LLM a strategy guide, which first-time players do not have. Gemini isn't even emitting legal moves 100% of the time.
ebiester 2 hours ago
It's trained on YouTube data. It's going to get roffle and drspectred at the very least.
silver_sun 2 hours ago
Google has a library of millions of scanned books from their Google Books project that started in 2004. I think we have reason to believe that there are more than a few books about effectively playing different traditional card games in there, and that an LLM trained with that dataset could generalize to understand how to play Balatro from a text description.
Nonetheless I still think it's impressive that we have LLMs that can just do this now.
mjamesaustin an hour ago
gilrain an hour ago
winstonp 3 hours ago
DeepSeek hasn't been SotA in at least 12 calendar months, which might as well be a decade in LLM years
cachius 3 hours ago
tehsauce 20 minutes ago
How does it do on gold stake?
dudisubekti 3 hours ago
But... there's Deepseek v3.2 in your link (rank 7)
littlestymaar 3 hours ago
> . I don't think there are many people who posted their Balatro playthroughs in text form online
There are *tons* of balatro content on YouTube though, and it makes absolutely zero doubt that Google is using YouTube content to train their model.
sdwr 3 hours ago
acid__ 2 hours ago
> Most (probably >99.9%) players can't do that at the first attempt
Eh, both myself and my partner did this. To be fair, we weren’t going in completely blind, and my partner hit a Legendary joker, but I think you might be slightly overstating the difficulty. I’m still impressed that Gemini did it.
nubg 5 hours ago
Weren't we barely scraping 1-10% on this with state of the art models a year ago and it was considered that this is the final boss, ie solve this and its almost AGI-like?
I ask because I cannot distinguish all the benchmarks by heart.
modeless 4 hours ago
François Chollet, creator of ARC-AGI, has consistently said that solving the benchmark does not mean we have AGI. It has always been meant as a stepping stone to encourage progress in the correct direction rather than as an indicator of reaching the destination. That's why he is working on ARC-AGI-3 (to be released in a few weeks) and ARC-AGI-4.
His definition of reaching AGI, as I understand it, is when it becomes impossible to construct the next version of ARC-AGI because we can no longer find tasks that are feasible for normal humans but unsolved by AI.
beklein 3 hours ago
mapontosevenths 2 hours ago
hmmmmmmmmmmmmmm 3 hours ago
fishpham 5 hours ago
Yes, but benchmarks like this are often flawed because leading model labs frequently participate in 'benchmarkmaxxing' - ie improvements on ARC-AGI2 don't necessarily indicate similar improvements in other areas (though it does seem like this is a step function increase in intelligence for the Gemini line of models)
layer8 4 hours ago
jstummbillig 4 hours ago
XenophileJKO 4 hours ago
olalonde 4 hours ago
verdverm 5 hours ago
Here's a good thread over 1+ month, as each model comes out
https://bsky.app/profile/pekka.bsky.social/post/3meokmizvt22...
tl;dr - Pekka says Arc-AGI-2 is now toast as a benchmark
Aperocky 4 hours ago
mNovak 3 hours ago
I'm excited for the big jump in ARC-AGI scores from recent models, but no one should think for a second this is some leap in "general intelligence".
I joke to myself that the G in ARC-AGI is "graphical". I think what's held back models on ARC-AGI is their terrible spatial reasoning, and I'm guessing that's what the recent models have cracked.
Looking forward to ARC-AGI 3, which focuses on trial and error and exploring a set of constraints via games.
causal 3 hours ago
Agreed. I love the elegance of ARC, but it always felt like a gotcha to give spatial reasoning challenges to token generators- and the fact that the token generators are somehow beating it anyway really says something.
throw310822 3 hours ago
The average ARC AGI 2 score for a single human is around 60%.
"100% of tasks have been solved by at least 2 humans (many by more) in under 2 attempts. The average test-taker score was 60%."
modeless 2 hours ago
imiric an hour ago
colordrops 3 hours ago
Wouldn't you deal with spatial reasoning by giving it access to a tool that structures the space in a way it can understand or just is a sub-model that can do spatial reasoning? These "general" models would serve as the frontal cortex while other models do specialized work. What is missing?
amelius 6 minutes ago
causal 3 hours ago
mnicky 5 hours ago
Well, fair comparison would be with GPT-5.x Pro, which is the same class of a model as Gemini Deep Think.
aeyes 4 hours ago
https://arcprize.org/leaderboard
$13.62 per task - so we need another 5-10 years for the price to run this to become reasonable?
But the real question is if they just fit the model to the benchmark.
onlyrealcuzzo 3 hours ago
Why 5-10 years?
At current rates, price per equivalent output is dropping at 99.9% over 5 years.
That's basically $0.01 in 5 years.
Does it really need to be that cheap to be worth it?
Keep in mind, $0.01 in 5 years is worth less than $0.01 today.
willis936 3 hours ago
golem14 an hour ago
A grad student hour is probably more expensive…
re-thc 3 hours ago
What’s reasonable? It’s less than minimum hourly wage in some countries.
willis936 2 hours ago
igravious 4 hours ago
That's not a long time in the grand scheme of things.
throwup238 4 hours ago
culi 2 hours ago
Yes but with a significant (logarithmic) increase in cost per task. The ARC-AGI site is less misleading and shows how GPT and Claude are not actually far behind
saberience 4 hours ago
Arc-AGI (and Arc-AGI-2) is the most overhyped benchmark around though.
It's completely misnamed. It should be called useless visual puzzle benchmark 2.
It's a visual puzzle, making it way easier for humans than for models trained on text firstly. Secondly, it's not really that obvious or easy for humans to solve themselves!
So the idea that if an AI can solve "Arc-AGI" or "Arc-AGI-2" it's super smart or even "AGI" is frankly ridiculous. It's a puzzle that means nothing basically, other than the models can now solve "Arc-AGI"
CuriouslyC 4 hours ago
The puzzles are calibrated for human solve rates, but otherwise I agree.
saberience 4 hours ago
karmasimida 5 hours ago
It is over
baal80spam 5 hours ago
I for one welcome our new AI overlords.
logicprog 4 hours ago
Is it me or is the rate of model release is accelerating to an absurd degree? Today we have Gemini 3 Deep Think and GPT 5.3 Codex Spark. Yesterday we had GLM5 and MiniMax M2.5. Five days before that we had Opus 4.6 and GPT 5.3. Then maybe two weeks I think before that we had Kimi K2.5.
i5heu 3 hours ago
I think it is because of the Chinese new year. The Chinese labs like to publish their models arround the Chinese new year, and the US labs do not want to let a DeepSeek R1 (20 January 2025) impact event happen again, so i guess they publish models that are more capable then what they imagine Chinese labs are yet capable of producing.
woah an hour ago
Singularity or just Chinese New Year?
r2vcap 33 minutes ago
Please use the term “Lunar New Year” instead of “Chinese New Year,” as the lunar calendar is a respected tradition in many Asian countries. For example, both California and New York use the term “Lunar New Year” in their legislation.
rfoo 9 minutes ago
zzrush 8 minutes ago
phainopepla2 18 minutes ago
aliston 3 hours ago
I'm having trouble just keeping track of all these different types of models.
Is "Gemini 3 Deep Think" even technically a model? From what I've gathered, it is built on top of Gemini 3 Pro, and appears to be adding specific thinking capabilities, more akin to adding subagents than a truly new foundational model like Opus 4.6.
Also, I don't understand the comments about Google being behind in agentic workflows. I know that the typical use of, say, Claude Code feels agentic, but also a lot of folks are using separate agent harnesses like OpenClaw anyway. You could just as easily plug Gemini 3 Pro into OpenClaw as you can Opus, right?
Can someone help me understand these distinctions? Very confused, especially regarding the agent terminology. Much appreciated!
logicprog 2 hours ago
> Also, I don't understand the comments about Google being behind in agentic workflows.
It has to do with how the model is RL'd. It's not that Gemini can't be used with various agentic harnesses, like open code or open claw or theoretically even claude code. It's just that the model is trained less effectively to work with those harnesses, so it produces worse results.
re-thc 3 hours ago
There are hints this is a preview to Gemini 3.1.
rogerkirkness 4 hours ago
Fast takeoff.
redox99 3 hours ago
There's more compute now than before.
bpodgursky 3 hours ago
Anthropic took the day off to do a $30B raise at a $380B valuation.
IhateAI 3 hours ago
Most ridiculous valuation in the history of markets. Cant wait to watch these compsnies crash snd burn when people give up on the slot machine.
andxor 2 hours ago
kgwgk 3 hours ago
jascha_eng 2 hours ago
brokencode 3 hours ago
They are using the current models to help develop even smarter models. Each generation of model can help even more for the next generation.
I don’t think it’s hyperbolic to say that we may be only a single digit number of years away from the singularity.
lm28469 3 hours ago
I must be holding these things wrong because I'm not seeing any of these God like superpowers everyone seem to enjoy.
brokencode 2 hours ago
sekai an hour ago
> I don’t think it’s hyperbolic to say that we may be only a single digit number of years away from the singularity.
We're back to singularity hype, but let's be real: benchmark gains are meaningless in the real world when the primary focus has shifted to gaming the metrics
brokencode an hour ago
rob-wagner 29 minutes ago
I’ve been using Gemini 3 Pro on a historical document archiving project for an old club. One of the guys had been working on scanning old handwritten minutes books written in German that were challenging to read (1885 through 1974). Anyways, I was getting decent results on a first pass with 50 page chunks but ended up doing 1 page at a time (accuracy probably 95%). For each page, I submit the page for a transcription pass followed by a translation of the returned transcription. About 2370 pages and sitting at about $50 in Gemini API billing. The output will need manual review, but the time savings is impressive.
xnx 5 hours ago
Google is absolutely running away with it. The greatest trick they ever pulled was letting people think they were behind.
wiseowise 3 hours ago
Their models might be impressive, but their products absolutely suck donkey balls. I’ve given Gemini web/cli two months and ran away back to ChatGPT. Seriously, it would just COMPLETELY forget context mid dialog. When asked about improving air quality it just gave me a list of (mediocre) air purifiers without asking for any context whatsoever, and I can list thousands of conversations like that. Shopping or comparing options is just nonexistent. It uses Russian propaganda sources for answers and switches to Chinese mid sentence (!), while explaining some generic Python functionality. It’s an embarrassment and I don’t know how they justify 20 euro price tag on it.
mavamaarten 2 hours ago
I agree. On top of that, in true Google style, basic things just don't work.
Any time I upload an attachment, it just fails with something vague like "couldn't process file". Whether that's a simple .MD or .txt with less than 100 lines or a PDF. I tried making a gem today. It just wouldn't let me save it, with some vague error too.
I also tried having it read and write stuff to "my stuff" and Google drive. But it would consistently write but not be able to read from it again. Or would read one file from Google drive and ignore everything else.
Their models are seriously impressive. But as usual Google sucks at making them work well in real products.
davoneus 2 hours ago
sequin an hour ago
How can the models be impressive if they switch to Chinese mid-sentence? I've observed those bizarre bugs too. Even GPT-3 didn't have those. Maybe GPT-2 did. It's actually impressive that they managed to botch it so badly.
Google is great at some things, but this isn't it.
chermanowicz 2 hours ago
It's so capable at some things, and others are garbage. I uploaded a photo of some words for a spelling bee and asked it to quiz my kid on the words. The first word it asked, wasn't on the list. After multiple attempts to get it to start asking only the words in the uploaded pic, it did, and then would get the spellings wrong in the Q&A. I gave up.
gokhan an hour ago
Agreed on the product. I can't make Gemini read my emails on GMail. One day it says it doesn't have access, the other day it says Query unsuccessful. Claude Desktop has no problem reaching to GMail, on the other hand :)
kilroy123 2 hours ago
Sadly true.
It is also one of the worst models to have a sort of ongoing conversation with.
HardCodedBias 3 hours ago
Their models are absolutely not impressive.
Not a single person is using it for coding (outside of Google itself).
Maybe some people on a very generous free plan.
Their model is a fine mid 2025 model, backed by enormous compute resources and an army of GDM engineers to help the “researchers” keep the model on task as it traverses the “tree of thoughts”.
But that isn’t “the model” that’s an old model backed by massive money.
Ozzie_osman 3 hours ago
Peacetime Google is not like wartime Google.
Peacetime Google is slow, bumbling, bureaucratic. Wartime Google gets shit done.
nutjob2 3 hours ago
OpenAI is the best thing that happened to Google apparently.
taurath 18 minutes ago
RationPhantoms 2 hours ago
lern_too_spel 2 hours ago
Wartime Google gave us Google+. Wartime Google is still bumbling, and despite OpenAI's numerous missteps, I don't think it has to worry about Google hurting its business yet.
kenjackson 3 hours ago
But wait two hours for what OpenAI has! I love the competition and how someone just a few days ago was telling how ARC-AGI-2 was proof that LLMs can't reason. The goalposts will shift again. I feel like most of human endeavor will soon be just about trying to continuously show that AI's don't have AGI.
kilpikaarna 2 hours ago
> I feel like most of human endeavor will soon be just about trying to continuously show that AI's don't have AGI.
I think you overestimate how much your average person-on-the-street cares about LLM benchmarks. They already treat ChatGPT or whichever as generally intelligent (including to their own detriment), are frustrated about their social media feeds filling up with slop and, maybe, if they're white-collar, worry about their jobs disappearing due to AI. Apart from a tiny minority in some specific field, people already know themselves to be less intelligent along any measurable axis than someone somewhere.
7777332215 3 hours ago
Soon they can drop the bioweapon to welcome our replacement.
nutjob2 3 hours ago
"AGI" doesn't mean anything concrete, so it's all a bunch of non-sequiturs. Your goalposts don't exist.
Anyone with any sense is interested in how well these tools work and how they can be harnessed, not some imaginary milestone that is not defined and cannot be measured.
kenjackson 3 hours ago
amunozo 5 hours ago
Those black nazis in the first image model were a cause of inside trading.
naasking 3 hours ago
Google is still behind the largest models I'd say, in real world utility. Gemini 3 Pro still has many issues.
Razengan 4 hours ago
Gemini's UX (and of course privacy cred as with anything Google) is the worst of all the AI apps. In the eyes of the Common Man, it's UI that will win out, and ChatGPT's is still the best.
ainch 20 minutes ago
I find Gemini's web page much snappier to use than ChatGPT - I've largely swapped to it for most things except more agentic tasks.
xnx 3 hours ago
Google privacy cred is ... excellent? The worst data breach I know of them having was a flaw that allowed access to names and emails of 500k users.
laurex 3 hours ago
bitpush 3 hours ago
Razengan 3 hours ago
alexpotato 3 hours ago
> Gemini's UX ... is the worst of all the AI apps
Been using Gemini + OpenCode for the past couple weeks.
Suddenly, I get a "you need a Gemini Access Code license" error but when you go to the project page there is no mention of this or how to get the license.
You really feel the "We're the phone company and we don't care. Why? Because we don't have to." [0] when you use these Google products.
PS for those that don't get the reference: US phone companies in the 1970s had a monopoly on local and long distance phone service. Similar to Google for search/ads (really a "near" monopoly but close enough).
jonathanstrange 4 hours ago
You mean AI Studio or something like that, right? Because I can't see a problem with Google's standard chat interface. All other AI offerings are confusing both regarding their intended use and their UX, though, I have to concur with that.
ergonaught 3 hours ago
xnx 3 hours ago
wiseowise 3 hours ago
uxhoiuewfhhiu an hour ago
Gemini is completely unusable in VS Code. It's rated 2/5 stars, pathetic: https://marketplace.visualstudio.com/items?itemName=Google.g...
Requests regularly time out, the whole window freezes, it gets stuck in schizophrenic loops, edits cannot be reverted and more.
It doesn't even come close to Claude or ChatGPT.
dfdsf2 5 hours ago
Trick? Lol not a chance. Alphabet is a pure play tech firm that has to produce products to make the tech accessible. They really lack in the latter and this is visible when you see the interactions of their VP's. Luckily for them, if you start to create enough of a lead with the tech, you get many chances to sort out the product stuff.
dakolli 4 hours ago
You sound like Russ Hanneman from SV
s-kymon 4 hours ago
sigmar 6 hours ago
Here is the methodologies for all the benchmarks: https://storage.googleapis.com/deepmind-media/gemini/gemini_...
The arc-agi-2 score (84.6%) is from the semi-private eval set. If gemini-3-deepthink gets above 85% on the private eval set, it will be considered "solved"
>Submit a solution which scores 85% on the ARC-AGI-2 private evaluation set and win $700K. https://arcprize.org/guide#overview
gs17 5 hours ago
Interestingly, the title of that PDF calls it "Gemini 3.1 Pro". Guess that's dropping soon.
sigmar 5 hours ago
I looked at the file name but not the document title (specifically because I was wondering if this is 3.1). Good spot.
edit: they just removed the reference to "3.1" from the pdf
josalhor 4 hours ago
staticman2 4 hours ago
That's odd considering 3.0 is still labeled a "preview" release.
ainch 18 minutes ago
WarmWash 4 hours ago
The rumor was that 3.1 was today's drop
losvedir 4 hours ago
riku_iki 4 hours ago
> If gemini-3-deepthink gets above 85% on the private eval set, it will be considered "solved"
They never will do on private set, because it would mean its being leaked to google.
Scene_Cast2 3 hours ago
It's a shame that it's not on OpenRouter. I hate platform lock-in, but the top-tier "deep think" models have been increasingly requiring the use of their own platform.
raybb 3 hours ago
OpenRouter is pretty great but I think litellm does a very good job and it's not a platform middle man, just a python library. That being said, I have tried it with the deep think models.
imiric an hour ago
Part of OpenRouter's appeal to me is precisely that it is a middle man. I don't want to create accounts on every provider, and juggle all the API keys myself. I suppose this increases my exposure, but I trust all these providers and proxies the same (i.e. not at all), so I'm careful about the data I give them to begin with.
octoberfranklin 2 minutes ago
simianwords 5 hours ago
OT but my intuition says that there’s a spectrum
- non thinking models
- thinking models
- best of N models like deep think an gpt pro
Each one is of a certain computational complexity. Simplifying a bit, I think they map to - linear, quadratic and n^3 respectively.
I think there are certain class of problems that can’t be solved without thinking because it necessarily involves writing in a scratchpad. And same for best of N which involves exploring.
Two open questions
1) what’s the higher level here, is there a 4th option?
2) can a sufficiently large non thinking model perform the same as a smaller thinking?
futureshock 2 hours ago
I think step 4 is the agent swarm. Manager model gets the prompt and spins up a swarm of looping subagents, maybe assigns them different approaches or subtasks, then reviews results, refines the context files and redeploys the swarm on a loop till the problem is solved or your credit card is declined.
simianwords 2 hours ago
i think this is the right answer
edit: i don't know how this is meaningfully different from 3
NitpickLawyer 5 hours ago
> best of N models like deep think an gpt pro
Yeah, these are made possible largely by better use at high context lengths. You also need a step that gathers all the Ns and selects the best ideas / parts and compiles the final output. Goog have been SotA at useful long context for a while now (since 2.5 I'd say). Many others have come with "1M context", but their usefulness after 100k-200k is iffy.
What's even more interesting than maj@n or best of n is pass@n. For a lot of applications youc an frame the question and search space such that pass@n is your success rate. Think security exploit finding. Or optimisation problems with quick checks (better algos, kernels, infra routing, etc). It doesn't matter how good your pass@1 or avg@n is, all you care is that you find more as you spend more time. Literally throwing money at the problem.
mnicky 5 hours ago
> can a sufficiently large non thinking model perform the same as a smaller thinking?
Models from Anthropic have always been excellent at this. See e.g. https://imgur.com/a/EwW9H6q (top-left Opus 4.6 is without thinking).
simianwords 5 hours ago
its interesting that opus 4.6 added a paramter to make it think extra hard.
anematode 19 minutes ago
It found a small but nice little optimization in Stockfish: https://github.com/official-stockfish/Stockfish/pull/6613
Previous models including Claude Opus 4.6 have generally produced a lot of noise/things that the compiler already reliably optimizes out.
Decabytes 2 hours ago
Gemini has always felt like someone who was book smart to me. It knows a lot of things. But if you ask it do anything that is offscript it completely falls apart
dwringer an hour ago
I strongly suspect there's a major component of this type of experience being that people develop a way of talking to a particular LLM that's very efficient and works well for them with it, but is in many respects non-transferable to rival models. For instance, in my experience, OpenAI models are remarkably worse than Google models in basically any criterion I could imagine; however, I've spent most of my time using the Google ones and it's only during this time that the differences became apparent and, over time, much more pronounced. I would not be surprised at all to learn that people who chose to primarily use Anthropic or OpenAI models during that time had an exactly analogous experience that convinced them their model was the best.
esafak 2 hours ago
I'd rather say it has a mind of its own; it does things its way. But I have not tested this model, so they might have improved its instruction following.
vkazanov 2 hours ago
Well, one thing i know for sure: it reliably misplaces parentheses in lisps.
esafak an hour ago
jetter 3 hours ago
it is interesting that the video demo is generating .stl model. I run a lot of tests of LLMs generating OpenSCAD code (as I have recently launched https://modelrift.com text-to-CAD AI editor) and Gemini 3 family LLMs are actually giving the best price-to-performance ratio now. But they are very, VERY far from being able to spit out a complex OpenSCAD model in one shot. So, I had to implement a full fledged "screenshot-vibe-coding" workflow where you draw arrows on 3d model snapshot to explain to LLM what is wrong with the geometry. Without human in the loop, all top tier LLMs hallucinate at debugging 3d geometry in agentic mode - and fail spectacularly.
mchusma 2 hours ago
Hey, my 9 year old son uses modelrift for creating things for his 3d printer, its great! Product feedback: 1. You should probably ask me to pay now, I feel like i've used it enough. 2. You need a main dashboard page with a history of sessions. He thought he lost a file and I had to dig in the billing history to get a UUID I thought was it and generate the url. I would say naming sessions is important, and could be done with small LLM after the users initial prompt. 3. I don't think I like the default 3d model in there once I have done something, blank would be better.
We download the stl and import to bambu. Works pretty well. A direct push would be nice, but not necessary.
gundmc 2 hours ago
Yes, I've been waiting for a real breakthrough with regard to 3D parametric models and I don't think think this is it. The proprietary nature of the major players (Creo, Solidworks, NX, etc) is a major drag. Sure there's STP, but there's too much design intent and feature loss there. I don't think OpenSCAD has the critical mass of mindshare or training data at this point, but maybe it's the best chance to force a change.
lern_too_spel 2 hours ago
If you want that to get better, you need to produce a 3d model benchmark and popularize it. You can start with a pelican riding a bicycle with working bicycle.
Metacelsus 6 hours ago
According to benchmarks in the announcement, healthily ahead of Claude 4.6. I guess they didn't test ChatGPT 5.3 though.
Google has definitely been pulling ahead in AI over the last few months. I've been using Gemini and finding it's better than the other models (especially for biology where it doesn't refuse to answer harmless questions).
CuriouslyC 4 hours ago
Google is way ahead in visual AI and world modelling. They're lagging hard in agentic AI and autonomous behavior.
throwup238 5 hours ago
The general purpose ChatGpt 5.3 hasn’t been released yet, just 5.3-codex.
neilellis 5 hours ago
It's ahead in raw power but not in function. Like it's got the worlds fast engine but one gear! Trouble is some benchmarks only measure horse power.
NitpickLawyer 5 hours ago
> Trouble is some benchmarks only measure horse power.
IMO it's the other way around. Benchmarks only measure applied horse power on a set plane, with no friction and your elephant is a point sphere. Goog's models have always punched over what benchmarks said, in real world use @ high context. They don't focus on "agentic this" or "specialised that", but the raw models, with good guidance are workhorses. I don't know any other models where you can throw lots of docs at it and get proper context following and data extraction from wherever it's at to where you'd need it.
scarmig 2 hours ago
> especially for biology where it doesn't refuse to answer harmless questions
Usually, when you decrease false positive rates, you increase false negative rates.
Maybe this doesn't matter for models at their current capabilities, but if you believe that AGI is imminent, a bit of conservatism seems responsible.
Davidzheng 5 hours ago
I gather that 4.6 strengths are in long context agentic workflows? At least over Gemini 3 pro preview, opus 4.6 seems to have a lot of advantages
verdverm 5 hours ago
It's a giant game of leapfrog, shift or stretch time out a bit and they all look equivalent
nkzd 4 hours ago
Google models and CLI harness feels behind in agentic coding compared OpenAI and Antrophic
simianwords 6 hours ago
The comparison should be with GPT 5.2 pro which has been used successfully to solve open math problems.
aliljet 3 hours ago
The problem here is that it looks like this is released with almost no real access. How are people using this without submitting to a $250/mo subscription?
andxor 2 hours ago
People are paying for the subscriptions.
tootie an hour ago
I gather this isn't intended a consumer product. It's for academia and research institutions.
siva7 4 hours ago
I can't shake of the feeling that Googles Deep Think Models are not really different models but just the old ones being run with higher number of parallel subagents, something you can do by yourself with their base model and opencode.
Davidzheng 4 hours ago
And after i do that, how do i combine the output of 1000 subagents into one output? (Im not being snarky here, i think it's a nontrivial problem)
tifik 3 hours ago
The idea is that each subagent is focused on a specific part of the problem and can use its entire context window for a more focused subtask than the overall one. So ideally the results arent conflicting, they are complimentary. And you just have a system that merges them.. likely another agent.
mattlondon 4 hours ago
You just pipe it to another agent to do the reduce step (i.e. fan-in) of the mapreduce (fan-out)
It's agents all the way down.
jonathanstrange 3 hours ago
Start with 1024 and use half the number of agents each turn to distill the final result.
sinuhe69 4 hours ago
I'm pretty certain that DeepMind (and all other labs) will try their frontier (and even private) models on First Proof [1].
And I wonder how Gemini Deep Think will fare. My guess is that it will get half the way on some problems. But we will have to take an absence as a failure, because nobody wants to publish a negative result, even though it's so important for scientific research.
zozbot234 4 hours ago
The 1st proof original solutions are due to be published in about 24h, AIUI.
neilellis 5 hours ago
Less than a year to destroy Arc-AGI-2 - wow.
Davidzheng 5 hours ago
I unironically believe that arc-agi-3 will have a introduction to solved time of 1 month
ACCount37 2 hours ago
Not very likely?
ARC-AGI-3 has a nasty combo of spatial reasoning + explore/exploit. It's basically adversarial vs current AIs.
etyhhgfff 4 hours ago
The AGI bar has to be set even higher, yet again.
dakolli 3 hours ago
wow solving useless puzzles, such a useful metric!
esafak an hour ago
modeless 3 hours ago
It's still useful as a benchmark of cost/efficiency.
XCSme 4 hours ago
But why only a +0.5% increase for MMMU-Pro?
kingstnap 2 hours ago
Its possibly label noise. But you can't tell from a single number.
You would need to check to see if everyone is having mistakes on the same 20% or different 20%. If its the same 20% either those questions are really hard, or they are keyed incorrectly, or they aren't stated with enough context to actually solve the problem.
It happens. Old MMLU non pro had a lot of wrong answers. Simple things like MNIST have digits labeled incorrect or drawn so badly its not even a digit anymore.
kenjackson 3 hours ago
Everyone is already at 80% for that one. Crazy that we were just at 50% with GPT-4o not that long ago.
saberience 4 hours ago
It's a useless meaningless benchmark though, it just got a catchy name, as in, if the models solve this it means they have "AGI", which is clearly rubbish.
Arc-AGI score isn't correlated with anything useful.
Legend2440 an hour ago
It's correlated with the ability to solve logic puzzles.
It's also interesting because it's very very hard for base LLMs, even if you try to "cheat" by training on millions of ARC-like problems. Reasoning LLMs show genuine improvement on this type of problem.
jabedude 4 hours ago
how would we actually objectively measure a model to see if it is AGI if not with benchmarks like arc-AGI?
WarmWash 3 hours ago
simonw 4 hours ago
The pelican riding a bicycle is excellent. I think it's the best I've seen.
tasuki 40 minutes ago
Tbh they'd have to be absolutely useless at benchmarkmaxxing if they didn't include your pelican riding a bicycle...
nickthegreek 2 hours ago
I routinely check out the pelicans you post and I do agree, this is the best yet. It seemed to me that the wings/arms were such a big hangup for these generators.
Manabu-eo 4 hours ago
How likely this problem is already on the training set by now?
simonw 4 hours ago
If anyone trains a model on https://simonwillison.net/tags/pelican-riding-a-bicycle/ they're going to get some VERY weird looking pelicans.
suddenlybananas 3 hours ago
throwup238 4 hours ago
For every combination of animal and vehicle? Very unlikely.
The beauty of this benchmark is that it takes all of two seconds to come up with your own unique one. A seahorse on a unicycle. A platypus flying a glider. A man’o’war piloting a Portuguese man of war. Whatever you want.
recursive 4 hours ago
zarzavat 4 hours ago
You can always ask for a tyrannosaurus driving a tank.
verdverm 4 hours ago
I've heard it posited that the reason the frontier companies are frontier is because they have custom data and evals. This is what I would do too
enraged_camel 3 hours ago
Is there a list of these for each model, that you've catalogued somewhere?
throwup238 4 hours ago
The reflection of the sun in the water is completely wrong. LLMs are still useless. (/s)
margalabargala 4 hours ago
It's not actually, look up some photos of the sun setting over the ocean. Here's an example:
throwup238 3 hours ago
deron12 4 hours ago
It's worth noting that you mean excellent in terms of prior AI output. I'm pretty sure this wouldn't be considered excellent from a "human made art" perspective. In other words, it's still got a ways to go!
Edit: someone needs to explain why this comment is getting downvoted, because I don't understand. Did someone's ego get hurt, or what?
gs17 4 hours ago
It depends, if you meant from a human coding an SVG "manually" the same way, I'd still say this is excellent (minus the reflection issue). If you meant a human using a proper vector editor, then yeah.
fvdessen 3 hours ago
dfdsf2 4 hours ago
Indeed. And when you factor in the amount invested... yeah it looks less impressive. The question is how much more money needs to be invested to get this thing closer to reality? And not just in this instance. But for any instance e.g. a seahorse on a bike.
saberience 4 hours ago
Do you have to still keep trying to bang on about this relentlessly?
It was sort of humorous for the maybe first 2 iterations, now it's tacky, cheesy, and just relentless self-promotion.
Again, like I said before, it's also a terrible benchmark.
jeanloolz an hour ago
I'll agree to disagree. In any thread about a new model, I personally expect the pelican comment to be out there. It's informative, ritualistic and frankly fun. Your comment however, is a little harsh. Why mad?
Davidzheng 4 hours ago
Eh, i find it more of a not very informative but lighthearted commentary
simonw 3 hours ago
It being a terrible benchmark is the bit.
dfdsf2 4 hours ago
Highly disagree.
I was expecting something more realistic... the true test of what you are doing is how representative is the thing in relation to the real world. E.g. does the pelican look like a pelican as it exists in reality? This cartoon stuff is cute but doesnt pass muster in my view.
If it doesn't relate to the real world, then it most likely will have no real effect on the real economy. Pure and simple.
chriswarbo 4 hours ago
I disagree. The task asks for an SVG; which is a vector format associated with line drawings, clipart and cartoons. I think it's good that models are picking up on that context.
In contrast, the only "realistic" SVGs I've seen are created using tools like potrace, and look terrible.
I also think the prompt itself, of a pelican on bicycle, is unrealistic and cartoonish; so making a cartoon is a good way to solve the task.
peaseagee 4 hours ago
The request is for an SVG, generally _not_ the format for photorealistic images. If you want to start your own benchmark, feel free to ask for a photorealistic JPEG or PNG of a pelican riding a bicycle. Could be interesting to compare and contrast, honestly.
Legend2440 an hour ago
I'm really interested in the 3D STL-from-photo process they demo in the video.
Not interested enough to pay $250 to try it out though.
ramshanker 5 hours ago
Do we get any model architecture details like parameter size etc.? Few months back, we used to talk more on this, now it's mostly about model capabilities.
Davidzheng 5 hours ago
I'm honestly not sure what you mean? The frontier labs have kept arch as secrets since gpt3.5
willis936 2 hours ago
At the very least gemini 3's flyer claims 1T parameters.
vessenes 5 hours ago
Not trained for agentic workflows yet unfortunately - this looks like it will be fantastic when they have an agent friendly one. Super exciting.
dakolli 4 hours ago
Its really weird how you all are begging to be replaced by llms, you think if agentic workflows get good enough you're going to keep your job? Or not have your salary reduced by 50%?
If Agents get good enough it's not going to build some profitable startup for you (or whatever people think they're doing with the llm slot machines) because that implies that anyone else with access to that agent can just copy you, its what they're designed to do... launder IP/Copyright. Its weird to see people get excited for this technology.
None of this good. We are simply going to have our workforces replaced by assets owned by Google, Anthropic and OpenAI. We'll all be fighting for the same barista jobs, or miserable factory jobs. Take note on how all these CEOs are trying to make it sound cool to "go to trade school" or how we need "strong American workers to work in factories".
BeetleB an hour ago
> Its really weird how you all are begging to be replaced by llms, you think if agentic workflows get good enough you're going to keep your job? Or not have your salary reduced by 50%?
The computer industry (including SW) has been in the business of replacing jobs for decades - since the 70's. It's only fitting that SW engineers finally become the target.
timeattack an hour ago
I agree with you and have similar thoughts (maybe, unfortunately for me). I personally know people who outsource not just their work, but also their life to LLMs, and reading their exciting comments makes me feel a mix of cringe, fomo and dread. But what is the engame for me and you likes, when we finally would be evicted from our own craft? Stash money while we still can, watching 'world crash and burn', and then go and try to ascend in some other, not yet automated craft?
dakolli 18 minutes ago
sgillen 2 hours ago
I think a lot of people assume they will become highly paid Agent orchestrators or some such. I don't think anyone really knows where things are heading.
ergonaught 3 hours ago
Most folks don't seem to think that far down the line, or they haven't caught on to the reality that the people who actually make decisions will make the obvious kind of decisions (ex: fire the humans, cut the pay, etc) that they already make.
blibble 27 minutes ago
newswasboring 2 hours ago
You don't hate AI, you hate capitalism. All the problems you have listed are not AI issues, its this crappy system where efficiency gains always end up with the capital owners.
OtomotO 4 hours ago
Or we just end capitalism.
French revolution style.
shrugs
dakolli 4 hours ago
uxhoiuewfhhiu an hour ago
Dirak 2 hours ago
Praying this isn't another Llama4 situation where the benchmark numbers are cooked. 84.6% on Arc-AGI is incredible!
jonathanstrange 5 hours ago
Unfortunately, it's only available in the Ultra subscription if it's available at all.
ismailmaj 4 hours ago
top 10 elo in codeforces is pretty absurd
andrewstuart 4 hours ago
Gemini was awesome and now it’s garbage.
It’s impossible for it to do anything but cut code down, drop features, lose stuff and give you less than the code you put in.
It’s puzzling because it spent months at the head of the pack now I don’t use it at all because why do I want any of those things when I’m doing development.
I’m a paid subscriber but there’s no point any more I’ll spend the money on Claude 4.6 instead.
halapro 3 hours ago
I never found it useful for code. It produced garbage littered with gigantic comments.
Me: Remove comments
Literally Gemini: // Comments were removed
andrewstuart 3 hours ago
It would make more sense to me if it had never been awesome.
mortsnort 2 hours ago
ergonaught 3 hours ago
It seems to be adept at reviewing/editing/critiquing, at least for my use cases. It always has something valuable to contribute from that perspective, but has been comparatively useless otherwise (outside of moats like "exclusive access to things involving YouTube").
m3kw9 4 hours ago
Gemini 3 Pro/Flash is stuck in preview for months now. Google is slow but they progress like a massive rock giant.
okokwhatever 4 hours ago
I need to test the sketch creation a s a p. I need this in my life because learning to use Freecad is too difficult for a busy person like me (and frankly, also quite lazy)
sho_hn 4 hours ago
FWIW, the FreeCAD 1.1 nightlies are much easier and more intuitive to use due to the addition of many on-canvas gizmos.
syntaxing 5 hours ago
Why a Twitter post and not the official Google blog post… https://blog.google/innovation-and-ai/models-and-research/ge...
dang 4 hours ago
Just normal randomness I suppose. I've put that URL at the top now, and included the submitted URL in the top text.
meetpateltech 5 hours ago
The official blog post was submitted earlier (https://news.ycombinator.com/item?id=46990637), but somehow this story ranked up quickly on the homepage.
verdverm 5 hours ago
@dang will often replace the post url & merge comments
HN guidelines prefer the original source over social posts linking to it.
aavci 5 hours ago
Agreed - blog post is more appropriate than a twitter post
HardCodedBias 3 hours ago
Always the same with Google.
Gemini has been way behind from the start.
They use the firehose of money from search to make it as close to free as possible so that they have some adoption numbers.
They use the firehose from search to pay for tons of researchers to hand hold academics so that their non-economic models and non-economic test-time-compute can solve isolated problems.
It's all so tiresome.
Try making models that are actually competitive, Google.
Sell them on the actual market and win on actual work product in millions of people lives.
dperhar 4 hours ago
Does anyone actually use Gemini 3 now? I cant stand its sleek salesy way of introduction, and it doesnt hold to instructions hard – makes it unapplicable for MECE breakdowns or for writing.
copperx 4 hours ago
I do. It's excellent when paired with an MCP like context7.
throwa356262 4 hours ago
I dont agree, Gemini 3 is pretty good, even the Lite version.
dperhar 4 hours ago
What do you use it for and why? Genuinely curious
jeffbee 3 hours ago
It indeed departs from instructions pretty regularly. But I find it very useful and for the price it beats the world.
"The price" is the marginal price I am paying on top of my existing Google 1, YouTube Premium, and Google Fi subs, so basically nothing on the margin.