ChatGPT Images 2.0 (openai.com)
819 points by wahnfrieden 15 hours ago
Livestream: https://openai.com/live/
System card: https://deploymentsafety.openai.com/chatgpt-images-2-0/chatg...
lionkor an hour ago
Every cent you spend on this, remember: The people who made this possible are not even getting a millionth of a cent for every billion USD made with it (they are getting nothing). Same with code; that code you spent years pouring over, fixing, etc. is now how these companies make so much money and get so much investment. It's like open source, except you get shafted.
barnabee an hour ago
A lot of people here aren't going to like it, but the only reasonable way out I can see is to eventually socialise ownership and control of AI.
I don't see an alternative that isn't really bad.
master-lincoln an hour ago
Seize the means of production!
barnabee an hour ago
user34283 31 minutes ago
I figure capitalism may soon become obsolete. But I don’t think this speculation is going to make for interesting discussion on here.
I find the technical discussion more interesting and could do without some of the moral grandstanding in the comments.
sp_c 42 minutes ago
I don't understand why everyone is all up and arms about Images / Art being generated by AI, but when it comes to code... well who cares? The people who made all the code training data are also getting nothing!
Potentially the one difference is that developers invented this and screwed themselves, whereas artists had nothing to do with AI.
r5109 22 minutes ago
Rob Pike cares. In other places apart from HN there is more resistance. Perceived lack of resistance has multiple reasons:
- Criticism of AI is discouraged or flagged on most industry owned platforms.
- The loudest pro-AI software engineers work for companies that financially benefit from AI.
- Many are silent because they fear reprisals.
- Many software engineers lack agency and prefer to sit back and understand what is happening instead of shaping what is happening.
- Many software engineers are politically naive and easily exploited.
Artists have a broader view and are often not employed by the perpetrators of the theft.
lwhi 4 minutes ago
The same developers who fed the machine, didn't make the machine.
Your comparison is incorrect.
happymellon 28 minutes ago
> Potentially the one difference is that developers invented this and screwed themselves
Hopefully you mean developers invented this and screwed over other developers.
How many folks working on the code at OpenAI have meaninfully contributed to Open Source? I agree that because it is the same "job title" people might feel less sympathy but it's not the same people.
sandworm101 29 minutes ago
Because artists generally own thier material (with exceptions at the very high end) whereas professional coders have generally abandoned ownership by seeding it as "work product" to thier employers. Copy my drawings and you steal from me, a person. Copy a bit of code or a texture pack from a game and you steal from whatever private equity owns that game studio. Private equity doesnt have feelings to hurt.
billynomates 27 minutes ago
pawelduda 8 minutes ago
People who provided training material for AI images, received payment in likes and shares
lwhi 5 minutes ago
Is this satire?
bradley13 35 minutes ago
If you put stuff on the internet, people (and machines) can see it. How do you think human artists learn? By looking at other people's artwork. AI can do exactly the same thing.
As for code: All of my code is open source. I don't care if people (or machines) learn from it. In fact, as a teacher, I sincerely hope that they do!
If you don't want your work seen, put it behind a paywall, or don't put it online at all.
lwhi 3 minutes ago
A very basic point of view. If you can't see how you're being disingenuous, there's no point in having a conversation with you.
deepvibrations 16 minutes ago
Not a fair comparison... A model can ingest a countless number works in day and reproduce stylistic fingerprints on demand, at zero marginal cost. How are the people it learned from meant to compete with that?
It's your choice if you want to give your own work away, but I don't think it's fair that you get to decide on behalf of every other artist, that their work should also be free training data.
Do you want all musicians and artists to put their work behind paywalls? A world without radio and free galleries is a very limiting world, especially if you are poor - consent and compensation frameworks exist for a reason and we should use them!
rolymath 35 minutes ago
That's fine for me. As someone who can't draw or design for shit, I am getting effectively millions of dollars worth of artist time for $20/month.
The solution is to socialize AI, not ban it.
minimaxir 11 hours ago
So during my Nano Banana Pro experiments I wrote a very fun prompt that tests the ability for these image generation models to follow heuristics, but still requires domain knowledge and/or use of the search tool:
Create a 8x8 contiguous grid of the Pokémon whose National Pokédex numbers correspond to the first 64 prime numbers. Include a black border between the subimages.
You MUST obey ALL the FOLLOWING rules for these subimages:
- Add a label anchored to the top left corner of the subimage with the Pokémon's National Pokédex number.
- NEVER include a `#` in the label
- This text is left-justified, white color, and Menlo font typeface
- The label fill color is black
- If the Pokémon's National Pokédex number is 1 digit, display the Pokémon in a 8-bit style
- If the Pokémon's National Pokédex number is 2 digits, display the Pokémon in a charcoal drawing style
- If the Pokémon's National Pokédex number is 3 digits, display the Pokémon in a Ukiyo-e style
The NBP result is here, which got the numbers, corresponding Pokemon, and styles correct, with the main point of contention being that the style application is lazy and that the images may be plagiarized: https://cdn.bsky.app/img/feed_fullsize/plain/did:plc:oxaerni...Running that same prompt through gpt-2-image high gave an...interesting contrast: https://cdn.bsky.app/img/feed_fullsize/plain/did:plc:oxaerni...
It did more inventive styles for the images that appear to be original, but:
- The style logic is by row, not raw numbers and are therefore wrong
- Several of the Pokemon are flat-out wrong
- Number font is wrong
- Bottom isn't square for some reason
Odd results.
MrManatee 2 hours ago
Prompts like this feel like it's using the wrong abstraction. The "obvious" thing to do with something like this would be to generate some code that generates the image and then run that code.
Inspired by this, I tried something much simpler. I asked it to draw 12 concentric circles. With three tries it always drew 10 instead. https://chatgpt.com/share/69e87d08-5a14-83eb-9a3b-3a8eb14692...
dvt 9 hours ago
This is an amazing test and it's kinda' funny how terrible gpt-2-image is. I'd take "plagiarized" images (e.g. Google search & copy-paste) any day over how awful the OpenAI result is. Doesn't even seem like they have a sanity checker/post-processing "did I follow the instructions correctly?" step, because the digit-style constraint violation should be easily caught. It's also expensive as shit to just get an image that's essentially unusable.
the_arun 7 hours ago
This is from Gemini - https://lens.usercontent.google.com/banana?agsi=CmdnbG9iYWw6...
fblp 6 hours ago
anshumankmr 7 hours ago
that is interesting cause I feel gpt-image-1 did have that feature.
(source: https://chatgpt.com/share/69e83569-b334-8320-9fbf-01404d18df...)
weird-eye-issue 6 hours ago
hyperadvanced 7 hours ago
I wouldn’t say it’s terrible. I wouldn’t say it’s a huge step forward in terms of quality compared to what I’ve seen before from AI
AussieWog93 2 hours ago
For what it's worth, NBP made some mistakes too.
Artistic oddities aside (why are the 8-bit sprites 16-bit, why do the charcoal drawings have colour, why does the art of specifically the Gen 1 Pokemon look so off.), 271 is Lombre, not Lotad.
Palmik an hour ago
I do not think this is a good prompt or useful benchmark, but nonetheless, it seems to work better for me: https://chatgpt.com/share/69e88a94-ded8-8395-b5dc-abceb2f44d...
vincentbuilds 4 hours ago
banana Pro gets the logic and punts on the art; gpt-2-image gets the art and punts on the logic. Feels like instruction-following and creativity sit on opposite ends of the same slider.
dieortin 35 minutes ago
This feels incredibly AI generated
pfortuny an hour ago
Just try a 23-sided plane convex polygon.
rrr_oh_man 9 hours ago
Why would you consider this a good prompt?
minimaxir 9 hours ago
Because both Nano Banana Pro and ChatGPT Images 2.0 have touted strong reasoning capabilities, and this particular prompt has more objective, easy-to-validate criteria as opposed to the subjective nature of images.
I have more subjective prompts to test reasoning but they're your-mileage-may-vary (however, gpt-2-image has surprisingly been doing much better on more objective criteria in my test cases)
razorbeamz 2 hours ago
Neither of them drew them in an 8-bit style either. It's way too many colors.
dodslaser an hour ago
Maybe they're so advanced they learned to write to the palette registers mid-scanline.
Razengan 5 hours ago
Even a few months ago, ChatGPT/Sora's image generation performed better than Gemini/Nano Banana for certain weird prompts:
Try things like: "A white capybara with black spots, on a tricycle, with 7 tentacles instead of legs, each tentacle is a different color of the rainbow" (paraphrased, not the literal exact prompt I used)
Gemini just globbed a whole mass of tentacles without any regards to the count
m3kw9 5 hours ago
Prob a very unscientific way to test an image model. This would me likely because they have the reasoning turned down and let its instant output takeover
minimaxir 5 hours ago
There's no good scientific way to test a closed-source model with both nondeterministic and subjective output.
This example image was generated using the API on high, not the low reasoning version. (it is slow and takes 2 minutes lol)
crustaceansoup 4 hours ago
If the results are quantifiable/objective and repeatable it's scientific, how is it not scientific?
The reasoning amount is part of the evaluation isn't it?
TeMPOraL 3 hours ago
This is the best kind of science there is: direct, empirical test.
parasti 4 hours ago
A great technical achievement, for sure, but this is kind of the moment where it enters uncanny valley to me. The promo reel on the website makes it feel like humans doing incredible things (background music intentionally evokes that emotion), but it's a slideshow of computer generatated images attempting to replicate the amazing things that humans do. It's just crazy to look at those images and have to consciously remind myself - nobody made this, this photographed place and people do not exist, no human participated in this photo, no human traced the lines of this comic, no human designer laid out the text in this image. This is a really clever amalgamation machine of human-based inputs. Uncanny valley.
qnleigh 2 hours ago
No this is what life looks like on the other side of the uncanny valley. The images don't look creepy because they look artificial or wrong. They're a reminder of a creepy new reality where our eyes can no longer tell us what's real.
rambojohnson an hour ago
Online.
freakynit 4 minutes ago
Collection of some amazing prompts and corresponding images: https://gpt2-image-showcase.pagey.site/
Credits: https://github.com/magiccreator-ai/awesome-gpt-image-2-promp...
simonw 15 hours ago
I've been trying out the new model like this:
OPENAI_API_KEY="$(llm keys get openai)" \
uv run https://tools.simonwillison.net/python/openai_image.py \
-m gpt-image-2 \
"Do a where's Waldo style image but it's where is the raccoon holding a ham radio"
Code here: https://github.com/simonw/tools/blob/main/python/openai_imag...Here's what I got from that prompt. I do not think it included a raccoon holding a ham radio (though the problem with Where's Waldo tests is that I don't have the patience to solve them for sure): https://gist.github.com/simonw/88eecc65698a725d8a9c1c918478a...
simonw 14 hours ago
I just got a much better version using this command instead, which uses the maximum image size according to https://github.com/openai/openai-cookbook/blob/main/examples...
OPENAI_API_KEY="$(llm keys get openai)" \
uv run 'https://raw.githubusercontent.com/simonw/tools/refs/heads/main/python/openai_image.py' \
-m gpt-image-2 \
"Do a where's Waldo style image but it's where is the raccoon holding a ham radio" \
--quality high --size 3840x2160
https://gist.github.com/simonw/88eecc65698a725d8a9c1c918478a... - I found the raccoon!I think that image cost 40 cents.
makira 14 hours ago
Fed into a clear Claude Code max effort session with : "Inspect waldo2.png, and give me the pixel location of a raccoon holding a ham radio.". It sliced the image into small sections and gave:
"Found the raccoon holding a ham radio in waldo2.png (3840×2160).
- Raccoon center: roughly (460, 1680)
- Ham radio (walkie-talkie) center: roughly (505, 1650) — antenna tip around (510, 1585)
- Bounding box (raccoon + radio): approx x: 370–540, y: 1550–1780
It's in the lower-left area of the image, just right of the red-and-white striped souvenir umbrella, wearing a green vest. "
Which is correct!cwillu 14 hours ago
M3L0NM4N 11 hours ago
prmoustache 3 hours ago
Funny how it can look convincing from far away but once you zoom in you find out most characters have a mix of leprosy and skin cancer.
wewtyflakes 11 hours ago
A startling number of people either have no arms, one arm, a half of an arm, or a shrunken arm; how odd!
rattlesnakedave 8 hours ago
ehnto 2 hours ago
cozzyd 8 hours ago
globular-toast 4 hours ago
davebren 14 hours ago
The faces...that's nice that it turned a kid's book into an abomination
Filligree 9 hours ago
vaulstein 6 hours ago
keithnz 7 hours ago
jdironman 7 hours ago
The real NFTs where the images we generated along the way
louiereederson 14 hours ago
The people in this image remind me of early this person does not exist, in the best way
dfee 11 hours ago
gpt5 11 hours ago
I tried it on the ChatGPT web UI and it also worked, although the ham radio looks like a handbag to me.
luxpir 4 hours ago
djmips 5 hours ago
mirekrusin 9 hours ago
Can it generate non halloween version though?
This lower-is-better danse macabre, nightmares inducing ratio feels like interesting proxy for models capability.
ireadmevs 14 hours ago
I found it on the 2nd image! On the 1st one not yet...
dzhiurgis 6 hours ago
Cost me < 1 cents - https://elsrc.com/elsrc/waldo/wojak.jpg
And this medium quality, high resolution https://elsrc.com/elsrc/waldo/10_wojaks.jpg was 13cents
p.s. aaaand that's soft launch my SaaS above, you can replace wojak.jpg with anything you want and it will paint that. It's basically appending to prompt defined by elsrc's dashboard. Hopefully a more sane way to manage genai content. Be gentle to my server, hn!
Barbing 6 hours ago
>I think that image cost 40 cents.
Kinda made me sad assuming the author didn't license anything to OpenAI.
I recognize it could revert (99% of?) progress if all the labs moved to consent-based training sets exclusively, but I can't think of any other fair way.
$.40 does not represent the appropriate value to me considering the desirability of the IP and its earning potential in print and elsewhere. If the world has to wait until it’s fair, what of value will be lost? (I suppose this is where the big wrinkle of foreign open weight models comes in.)
rafram 6 hours ago
makira 15 hours ago
> though the problem with Where's Waldo tests is that I don't have the patience to solve them for sure
I see an opportunity for a new AI test!
vunderba 13 hours ago
There have already been several attempts to procedurally generate Where’s Waldo? style images since the early Stable Diffusion days, including experiments that used a YOLO filter on each face and then processed them with ADetailer.
It's a difficult test for genai to pass. As I mentioned in a different thread, it requires a holistic understanding (in that there can only be one Waldo Highlander style), while also holding up to scrutiny when you examine any individual, ordinary figure.
simonw 14 hours ago
I've actually been feeding them into Claude Opus 4.7 with its new high resolution image inputs, with mixed results - in one case there was no raccoon but it was SURE there was and told me it was definitely there but it couldn't find it.
marricks 10 hours ago
Like... this has things that AI will seemingly always be terrible at?
At some point the level of detail is utter garbo and always will be. An artist who was thoughtful could have some mistakes but someone who put that much time into a drawing wouldn't have:
- Nightmarish screaming faces on most people
- A sign that points seemingly both directions, or the incorrect one for a lake and a first AID tent that doesn't exist
- A dog in bottom left and near lake which looks like some sort of fuzzy monstrosity...
It looks SO impressive before you try to take in any detail. The hand selected images for the preview have the same shit. The view of musculature has a sternocleidomastoid with no clavicle attachment. The periodic table seems good until you take a look at the metals...
We're reconfiguring all of our RAM & GPUs and wasting so much water and electricity for crappier where's Waldos??
p1esk 9 hours ago
AI will seemingly always be ...
You do realize that the whole image generation field is barely 10 years old?
I remember how I was able to generate mnist digits for the first time about 10 years ago - that seemed almost like magic!
halamadrid 6 hours ago
Really hard to look at these images given how not human like the humans are. A few are ok, but a lot are disfigured or missing parts and its hard to find a raccoon in here.
vova_hn2 9 hours ago
Thanks for the image, I will see their faces in my nightmares.
vunderba 9 hours ago
This happens all too frequently when you ask a GenAI model to create an image with a large crowd especially a “Where’s Waldo?” style scenes, where by definition you’re going to be examining individual faces very closely.
hackable_sand 6 hours ago
What about the faces of the people ChatGPT killed?
pants2 14 hours ago
The second 4K image definitely has a raccoon on the left there! Nice.
nerdsniper 9 hours ago
That is a devilishly difficult prompt for current diffusion tasks. Kudos.
ritzaco 14 hours ago
haha took me a while to notice that one of the buildings is labelled 'Ham radio'
ElFitz 14 hours ago
Damn. There’s a fun game app to make here ^^
dymk 10 hours ago
Is there? The moment you look closely at the puzzle (which is... the whole point of Where's Waldo), you notice all the deformities and errors.
ElFitz 3 hours ago
amelius 2 hours ago
arealaccount 14 hours ago
I see the raccoon
tptacek 14 hours ago
5.4 thinking says "Just right of center, immediately to the right of the HAM RADIO shack. Look on the dirt path there: the raccoon is the small gray figure partly hidden behind the woman in the red-and-yellow shirt, a little above the man in the green hat. Roughly 57% from the left, 48% from the top."
(I don't think it's right).
ritzaco 14 hours ago
I tried
> please add a giant red arrow to a red circle around the raccoon holding a ham radio or add a cross through the entire image if one does not exist
and got this. I'm not sure I know what a ham radio looks like though.
https://i.ritzastatic.com/static/ffef1a8e639bc85b71b692c3ba1...
jackpirate 14 hours ago
simonw 13 hours ago
davecahill 8 hours ago
neom 11 hours ago
Here is my regular "hard prompt" I use for testing image gen models:
"A macro close-up photograph of an old watchmaker's hands carefully replacing a tiny gear inside a vintage pocket watch. The watch mechanism is partially submerged in a shallow dish of clear water, causing visible refraction and light caustics across the brass gears. A single drop of water is falling from a pair of steel tweezers, captured mid-splash on the water's surface. Reflect the watchmaker's face, slightly distorted, in the curved glass of the watch face. Sharp focus throughout, natural window lighting from the left, shot on 100mm macro lens."
google drive with the 2 images: https://drive.google.com/drive/folders/1-QAftXiGMnnkLJ2Je-ZH...
Ran a bunch both on the .com and via the api, none of them are nearly as good as Nano Banana.
(My file share host used to be so good and now it's SO BAD, I've re-hosted with them for now I'll update to google drive link shortly)
jcattle 3 hours ago
I mean, your prompt is basically this skit: https://www.youtube.com/watch?v=BKorP55Aqvg ("The Expert" 7 red lines: all strictly perpendicular, some with green ink some with transparent ink)
I couldn't imagine the image you were describing. I've listed some of the red lines with green ink I've noticed in your prompt:
Macro Close Up - Sharp throughout
Focus on tiny gear - But also on tweezers, old watchmakers hand, water drop?
Work on the mechanism of the watch (on the back of the watch) - but show the curved glass of the watch face which is on the front
This is the biggest. Even if the mechanism is accessible from the front, you'd have to remove the glass to get to it. It just doesn't make sense and that reflects in the images you get generated. There's all the elements, but they will never make sense because the prompt doesn't make sense.
fc417fc802 2 hours ago
The last point (reflection by front glass versus mechanism access so no front glass) is the only issue I see with it. Other than that I can easily visualize an image that satisfies the prompt. I think that the general idea is a good one because it's satisfable while having multiple competing requirements that impose geometric constraints on the scene without providing an immediate solution to said constraints as well as requiring multiple independent features (caustics, reflections, fluid dynamics, refraction, directional lighting) that are quite complicated to get right.
To illustrate that there aren't any contradictions (other than the final bit about the reflection in the glass). Consider a macro shot showing partial hands, partial tweezers, and pocket watch internals. That's much is certainly doable. Now imagine the partial left hand holding a half submerged pocket watch, fingertips of right hand holding front half of tweezers that are clasping a tiny gear, positioned above the work piece with the drop of water falling directly below. Capture the watchmaker's perspective. I could sketch that so an image model capable of 3D reasoning should have no trouble.
It's precisely the sort of scene you'd use to test a raytracer. One thing I can immediately think to add is nested dielectrics. Perhaps small transparent glass beads sitting at the bottom of the dish of water with the edge of the pocket watch resting on them, make the dish transparent glass, and place the camera level with the top of the dish facing forward?
https://blog.yiningkarlli.com/2019/05/nested-dielectrics.htm...
A second thing I can think to add is a flame. Perhaps place a tealight candle on the far side of the dish, the flame visible through (and distorted by) the water and glass beads?
jcattle 2 hours ago
rrr_oh_man 9 hours ago
Why would you consider this a good prompt?
brynnbee 8 hours ago
My observations have been that image generation is especially challenged when asked to do things that are unusual. The fewer instances of something happening it has to train on, the worse it tends to be. Watch repair done in water fits that well - is there a single image on the internet of someone repairing a watch that is partially submerged in water? It also tends to be bad at reflections and consistency of two objects that should be the same.
the_lucifer 10 hours ago
Looks like your image host has rate limited viewing the shared images, wanted to give you a heads up
neom 10 hours ago
Thanks, I need to get off Zight, they used to be such an nice option for fast file share but they've really suffered some of the worst enshittification I've seen yet.
pb7 10 hours ago
Links are broken.
waynesonfire 10 hours ago
So.. sign up. "Get Sight for free". Ads everywhere bro.
TrackerFF 24 minutes ago
This is the first model I've used for mockups where I feed reference images, and they truly look real and good enough for pro use. I'm impressed.
madrox 11 hours ago
This seems like a great time to mention C2PA, a specification for positively affirming image sources. OpenAI participates in this, and if I load an image I had AI generate in a C2PA Viewer it shows ChatGPT as the source.
Bad actors can strip sources out so it's a normal image (that's why it's positive affirmation), but eventually we should start flagging images with no source attribution as dangerous the way we flag non-https.
Learn more at https://c2pa.org
debazel 7 hours ago
> but eventually we should start flagging images with no source attribution as dangerous the way we flag non-https.
Yes, lets make all images proprietary and locked behind big tech signatures. No more open source image editors or open hardware.
henry-j 7 hours ago
C2PA is actually an open protocol, à la SMTP. the whole spec is at https://spec.c2pa.org/, available for anyone to implement.
Melatonic 2 hours ago
Why would the image itself have to be proprietary to have some new piece of metadata attached to it ?
mdasen 10 hours ago
> Bad actors can strip sources out
I think the issue is that it's not just bad actors. It's every social platform that strips out metadata. If I post an image on Instagram, Facebook, or anywhere else, they're going to strip the metadata for my privacy. Sometimes the exif data has geo coordinates. Other times it's less private data like the file name, file create/access/modification times, and the kind of device it was taken on (like iPhone 16 Pro Max).
Usually, they strip out everything and that's likely to include C2PA unless they start whitelisting that to be kept or even using it to flag images on their site as AI.
But for now, it's not just bad actors stripping out metadata. It's most sites that images are posted on.
henry-j 7 hours ago
There’s actually a part of the NY state budget right now (TEDE part X, for my law nerds) that’d require social media companies to preserve non-PII provenance metadata and surface it to the user, if the uploaded image has it.
linkedin already does this--- see https://www.linkedin.com/help/linkedin/answer/a6282984, and X’s “made with ai” feature preserves the metadata but doesn’t fully surface it (https://www.theverge.com/ai-artificial-intelligence/882974/x...)
madrox 10 hours ago
You're implying social platforms aren't bad actors ;)
In seriousness, social platforms attributing images properly is a whole frontier we haven't even begun to explore, but we need to get there.
woadwarrior01 11 hours ago
Yeah, OpenAI has been attaching C2PA manifests to all their generated images from the very beginning. Also, based on a small evaluation that I ran, modern ML based AI generated image detectors like OmniAID[1] seem to do quite well at detecting GPT-Image-2 generated images. I use both in an on-device AI generated image detector that I built.
paradoxyl 3 hours ago
What a dystopian, pro-tyranny ask. Horrifying.
swalsh 12 hours ago
Been using the model for a few hours now. I'm actually reall impressed with it. This is the first time i've found value in an image model for stuff I actually do. I've been using it to build powerpoint slides, and mockups. It's CRAZY good at that.
johnwheeler 8 hours ago
Yeah, it's funny. I would expect to see more enthusiasm versus just basic run-of-the-mill, "oh, there it is". Leave it to the HN crowd. This is incredible. I don't even like OpenAI.
pembrook 2 hours ago
HN is engineer heavy so its a bunch of people who spend their days looking at code. If it's not a coding model they'll likely never use it.
To the average HN'er, images and design are superfluous aesthetic decoration for normies.
And for those on HN who do care about aesthetics, they're using Midjourney, which blows any GPT/Gemini model out of the water when it comes to taste even if it doesn't follow your prompt very well.
The examples given on this landing page are stock image-esque trash outside of the improvements in visual text generation.
c16 9 minutes ago
That video seems like it was made for the tiktok generation. Slow down.
skybrian 11 hours ago
This time it passed the piano keyboard test:
https://chatgpt.com/s/m_69e7ffafbb048191b96f2c93758e3e40
But it screwed up when attempting to label middle C:
https://chatgpt.com/s/m_69e8008ef62c8191993932efc8979e1e
Edit: it did fix it when asked.
vunderba 10 hours ago
When NB 2 came out I actually had to increase the difficulty of the piano test - reversing the colors of all the accidentals and the naturals, and it still managed it perfectly.
justani 4 hours ago
I have a few cases where nano banana fails all the time, even gpt image 2 is failing.
A 3 * 3 cube made out of small cubes, with a small 2 * 2 cube removed from it - https://chatgpt.com/share/69e85df6-5840-83e8-b0e9-3701e92332...
Create a dot grid containing a rectangle covering 4 dots horizontally and 3 dots vertically - https://chatgpt.com/share/69e85e4b-252c-83e8-b25f-416984cf30...
One where Nano banana fails but gpt image 2 worked: create a grid from 1 to 100 and in that grid put a snake, with it's head at 75 and tail at 31 - https://chatgpt.com/share/69e85e8b-2a1c-83e8-a857-d4226ba976...
teruakohatu 4 hours ago
> A 3 * 3 cube made out of small cubes, with a small 2 * 2 cube removed from it - https://chatgpt.com/share/69e85df6-5840-83e8-b0e9-3701e92332...
It is a little ambiguous (what exactly is a "3x3 cube") but I tried a bunch of variations and I simply could not get any Gemini models to produce the right output.
sigmoid10 an hour ago
You can do it, but it takes two steps. Code is generally better to create such strict geometry (even from ambiguous prompts), while the image diffusion model is great for tuning style and lighting.
https://chatgpt.com/share/69e88b5c-8628-83eb-8851-f587ef2c95...
schneehertz 9 hours ago
Generating a 4096x4096 image with gemini-3.1-flash-image-preview consumes 2,520 tokens, which is equivalent to $0.151 per image.
Generating a 3840x2160 image with gpt-image-2 consumes 13,342 tokens, which is equivalent to $0.4 per image.
This model is more than twice as expensive as Gemini.
strangescript 9 hours ago
this is apples to oranges, the flash version version a full version
this thing is like 5x better than flash at fine grain detail
ac29 9 hours ago
Google's naming might be misleading, currently 3.1 flash image outperforms the available pro version (3.0 pro) on most benchmarks: https://deepmind.google/models/model-cards/gemini-3-1-flash-...
altcognito 9 hours ago
.40 cents for high quality output is insanely cheap
it is only going to get cheaper
eclipticplane 9 hours ago
porphyra 11 hours ago
The improvement in Chinese text rendering is remarkable and impressive! I still found some typos in the Chinese sample pic about Wuxi though. For example the 笼 in 小笼包 was written incorrectly. And the "极小中文也清晰可读" section contains even more typos although it's still legible. Still, truly amazing progress. Vastly better than any previous image generation model by a large margin.
Lucasoato 9 hours ago
Is this even better than Chinese models? I suppose they focus much more on that aspect, simply because their training data might include many more examples of Chinese text.
Ladioss 2 hours ago
Maybe they just use Qwen Image under the hood ;p
AltruisticGapHN an hour ago
This is insanely good. But wow, prompting to get any one of these images is way more complicated than prompting Claude Code. There is a ton of vocabulary that comes with it relating to the camera, the lighting, the mood etc.
amunozo 14 hours ago
This is not as exciting as previous models were, but it is incredibly good. I am starting to think that expressing thoughts in words clearly is probably the most important and general skill of the future.
aulin 5 hours ago
Well that was probably the most important general skill even before this.
sigmoid10 2 hours ago
Perhaps for managers. But for everyone actually doing something, you used to need technical proficiency with tools. Now AI is becoming the universal tool.
bamboozled 41 minutes ago
In other words, communication is an important skill.
echelon 12 hours ago
> I am starting to think that expressing thoughts in words clearly is probably the most important and general skill of the future.
Without question.
AI will be indistinguishable from having a team. Communicating clearly has always and will always mattered.
This, however, is even stronger. Because you can program and use logic in your communications.
We're going to collectively develop absolutely wild command over instruction as a society. That's the skill to have.
adamhartenz 8 hours ago
How can AI be the amazing thing you say it is, but also too stupid to understand unless you get really good at communicating. Wouldn't better AI just mean it understands your ramblings better?
pickleRick243 8 hours ago
jstanley 3 hours ago
raincole 8 hours ago
yreg 10 hours ago
On the other hand LLMs are getting very good at understanding poorly constructed instructions as well.
So being able to express oneself clearly in a structured way may not be such an edge.
amunozo 5 hours ago
____tom____ 14 hours ago
No mention of modifying existing images, which is more important than anything they mentioned.
I think we all know the feeling of getting an image that is ok, but needs a few modifications, and being absolutely unable to get the changes made.
It either keeps coming up with the same image, or gives you a completely new take on the image with fresh problems.
Anyone know if modification of existing images is any better?
Anything better that OpenAI?
frmersdog 10 hours ago
Image editing program -> different versions of the image, each with some but not all of the elements you want, on each layer -> mask out the parts you don't need/apply mask, fill with black, soft brush with white the parts you want back in. Copy flattened/merged, drop it back into the image model, keep asking for the changes. As long as each generation adds in an element you want, you can build a collage of your final image.
user34283 3 hours ago
It's the first thing I tried, because Nano Banana 2 deteriorates the output with each turn, becoming unusable with just a few edits.
ChatGPT Images 2.0 made it unusable at the first turn. At least in the ChatGPT app editing a reference image absolutely destroyed the image quality. It perfectly extracted an illustration from the background, but in the process basically turned it from a crisp digital illustration into a blurry, low quality mess.
tomjen3 14 hours ago
There was an Edit button in one of the images in the livestream
dktp 14 hours ago
One interesting thing I found comparing OpenAI and Gemini image editing is - Gemini rejects anything involving a well known person. Anything. OpenAI is happy to edit and change every time I tried
I have a sideproject where I want to display standup comedies. I thought I could edit standup comedy posters with some AI to fit my design. Gemini straight up refuses to change any image of any standup comedy poster involving a well know human. OpenAI does not care and is happy to edit away
Melatonic 14 hours ago
How does it determine they are well known and not just similar looking?
yreg 10 hours ago
Gemini often rejects photos of random people (even ones it generated itself) because it thinks they look too similar to some well known person.
dktp 14 hours ago
I don't know tbh. I've tried it on 10-20 various level of famous standups and Gemini refuses every time
Just for testing, I just tried this https://i.ytimg.com/vi/_KJdP4FLGTo/sddefault.jpg ("Redesign this image in a brutalist graphic design style"). Gemini refuses (api as well as UI), OpenAI does it
arjie 13 hours ago
Melatonic 13 hours ago
BoorishBears 7 hours ago
There are models specifically for detecting well known people https://docs.aws.amazon.com/rekognition/latest/dg/celebritie...
vunderba 10 hours ago
Are you using Google Gemini directly? I've found the Vertex API seems to be significantly less strict.
sanex 8 hours ago
Having the launch website just scrollable generated images is so slick. I love this.
gverrilla 5 hours ago
You can click the images too, to see the prompt that got them gen'ed.
overgard 11 hours ago
Pretty mixed feelings on this. From the page at least, the images are very good. I'd find it hard to know that they're AI. Which I think is a problem. If we had a functioning congress, I wonder if we might end up with legislation that these things need to be watermarked or otherwise made identifiable as AI generated..
I also don't like that these things are trained on specific artist's styles without really crediting those artists (or even getting their consent). I think there's a big difference between an individual artist learning from a style or paying it homage, vs a machine just consuming it so it can create endless art in that style.
niek_pas an hour ago
Maybe I'm stupid and naive but I just don't really see how any of this is _fundamentally_ different from Photoshop. Trusting the images you're looking at on the internet has been impossible for a long time. That's why we have institutions and social relations we place trust in instead.
kansface 9 hours ago
> If we had a functioning congress, I wonder if we might end up with legislation that these things need to be watermarked or otherwise made identifiable as AI generated..
Not a lawyer, but that reads as compelled speech to me. Materially misrepresenting an image would be libel, today, right?
overgard 7 hours ago
Well, considering that AI generated content can't be copyrighted (afaik at least), I think we're in very different legal territory when it comes to AI creating things. While it's true that deepfakes could be considered libel.. good luck prosecuting that if you can't even figure out where the image came from.
The problem is it's all too easy to generate - you can't really do much about an individual piece of slop because there's so much of it. I think we need a way to filter this stuff, societally.
bryanhogan 10 hours ago
Trying to watermark or otherwise label them as AI generated is a lost fight, we should assume every image and video we see online may be AI generated.
rootusrootus 9 hours ago
This helps the segment of society that is interested in applying critical thinking to what they see. I am not sure that is anything like a majority or even a significant plurality. It seems like just about every image or video gets accused of being AI these days, but predictably the accusations depend on the ideology of the accuser.
apsurd 10 hours ago
You might be onto something. I find every image unsettling. they're very good no doubt, but maybe it disturbs me because all of it is a complete copy of what someone else created. I know, I know, there is no pure invention. That's not what i mean. Humans borrow from other humans all the time. There's a humanity in that! A machine fully repurposing a human contribution as some kind of new creation, iono i'm old, it's weird and i don't like it.
Maybe i'm just bloviating also.
drstewart 44 minutes ago
>If we had a functioning congress, I wonder if we might end up with legislation that these things need to be watermarked or otherwise made identifiable as AI generated..
Can you name any countries that you think are functioning, and what their laws are on watermarked AI images?
rambojohnson an hour ago
Just tried it and got six fingers and half a thumb on a simple portrait. Mickey Mouse stuff.
squidsoup 11 hours ago
Are camera manufacturers working on signed images? That seems like the only way our trust in any digital media doesn't collapse entirely.
user34283 19 minutes ago
Yes, I think they have been for years. C2PA Content Credentials are supported in cameras and some phones already today.
randyrand 8 hours ago
Signed images don’t get you much. You can just hardwire the image sensor to a computer and sign raw pixels.
Barbing 4 hours ago
Is the situation brighter for a company who owns the hardware and the software, for Apple?
Taking a picture of an AI generated image aside, theoretically could Apple attest to origin of photos taken in the native camera app and uploaded to iCloud?
Fascinating, by the way, thank you!
wiseowise an hour ago
Make cameras tamper resistant, like POS terminals.
Nition 8 hours ago
Ultimately even with that tech, you can still take a photo of an AI generated scene. Maybe coupled with geolocation data in the signature or something it might work.
Barbing 4 hours ago
Any thoughts on attempted multiple camera/360 camera solutions? Can make it cost prohibitive to generate exceptional fakes… for a little while
Kind of like showing the proctor around your room with your webcam before starting the exam.
—
I think legacy media stands a chance at coming back as long as they maintain a reputation of deeply verifying images, not being fooled.
petesergeant 6 hours ago
I see signing chains as the way to go here. Your camera signs an image, you sign the signed image, your client or editor signs the image you signed etc etc. Might finally have a use for blockchain.
elAhmo an hour ago
I am super out of the loop here, what happened with Dall-E?
lossyalgo 11 hours ago
Someone remind me again why this is a good idea to be able to create perfect fake images?
wiseowise an hour ago
Something, something democratization. Because having a skill is inherently oppressive nowadays.
VA1337 39 minutes ago
So is it better than nano-banana after all?
rambojohnson an hour ago
Just tried it and got the usual six fingers, and half a thumb. What are they actually iterating on with these models by now…
aledevv 28 minutes ago
Only vintage-style images?
bensyverson 15 hours ago
I caught the last minute of this—was it just ChatGPT Images 2.0?
punty 15 hours ago
It appears so!
minimaxir 15 hours ago
yes
PDF_Geek 2 hours ago
The free tier for ChatGPT feels pretty much nerfed at this point. I’m barely getting 10 prompts in before it drops me down to the basic model. The restrictions are getting ridiculous. Is anyone else seeing this?
baalimago 3 hours ago
"Benchmarks" aside, do anyone actually use these image models for anything?
razorbeamz 2 hours ago
Here in Japan every fucking food truck uses them for pictures of their menu, which really pisses me off because it's not representative of their food at all.
medlazik 3 hours ago
Look around? It's everywhere. Try talking to a graphic designer looking for a job theses days. Companies didn't wait for these tools to be good to start using them.
croisillon 3 hours ago
MAGA to show how terrible Europe is ;)
codebolt 3 hours ago
Anyone test it out for generating 2D art for games? Getting nano banana to generate consistent sprite sheets was seemingly impossible last time i tried a few months ago.
nickandbro 11 hours ago
200+ points in Arena.ai , that's incredible. They are cleaning house with this model
moralestapia 11 hours ago
point delta (from 2nd) not total
nickandbro 11 hours ago
hahahacorn 14 hours ago
One of the images in the blog (https://images.ctfassets.net/kftzwdyauwt9/4d5dizAOajLfAXkGZ7...) is a carbon copy of an image from an article posted Mar 27, 2026 with credits given to an individual: https://www.cornellsun.com/article/2026/03/cornell-accepts-5...
Was this an oversight? Or did their new image generation model generate an image that was essentially a copy of an existing image?
arjie 14 hours ago
That has to be the wrong stock image included or something, bloody hell.
magick image-l.webp image-r.jpg -compose difference -composite -auto-level -threshold 30% diff.png
It's practically all dark except for a few spots. It's the same image just different size compression whatever. I can't find it in any stock image search, though. Surely it could not have memorized the whole image at that fidelity. Maybe I just didn't search well enough.Melatonic 13 hours ago
Or the image was generated with AI in the first place and a test for Images 2.0
IsTom 12 hours ago
arjie 13 hours ago
recitedropper 14 hours ago
This is hilarious. Seems like kind of a random image for a model to memorize, but it could be.
There is definitely enough empirical validation that shows image models retain lots of original copies in their weights, despite how much AI boosters think otherwise. That said, it is often images that end up in the training set many times, and I would think it strange for this image to do that.
Regardless, great find.
Nition 8 hours ago
I feel it's too much of a perfect match to be generated from the model's memory. It's pixel perfect. Gotta be a mistake.
minimaxir 14 hours ago
Given the recency of that image, it is unlikely it is in the training data and therefore I would go with oversight.
ajam1507 8 hours ago
The image is likely older than the article given this picture from over a year ago.
afro88 3 hours ago
JimsonYang 10 hours ago
> you can make your own mangas
No you can’t.
You still have the studio ghibili look from the video. The issue of generating manga was the quality of characters, there’s multiple software to place your frame.
But I am hopeful. If I put in a single frame, can it carry over that style for the next images? It would be game changing if a chat could have its own art style
thelucent 13 hours ago
It seems to still have this gpt image color that you can just feel. The slight sepia and softness.
honzaik 13 hours ago
I was just wondering about that. Did they embrace it as a “signature look”? it cant be accidental, right?
GaryBluto 10 hours ago
It's definitely not accidental but I'm not completely sure whether or not it is simply a "tell" or watermark or an attempt to foster brand association.
dymk 10 hours ago
It's the Stranger Things nostalgia filter. Almost all the sample pictures they had looked like they were vaguely from the 90s-00s era.
Oras 11 hours ago
My test for image models is asking it to create an image showing chess openings. Both this model and Banana pro are so bad at it.
While the image looks nice, the actual details are always wrong, such as showing pawns in wrong locations, missing pawns, .. etc.
Try it yourself with this prompt: Create a poster to show opening game for Queen's Gambit to teach kids to play chess.
lxgr 11 hours ago
It almost nailed it for me (two squares have both white and black color). All pieces and the position look correct.
tempaccount5050 11 hours ago
What move? Who's turn is it? Declined or accepted? Garbage in, garbage out.
bogtap82 11 hours ago
In some cases I would agree with this, but image model releases including this one are beginning to incorporate and market the thinking step. It is not a reach at this point to expect the model to take liberties in order to deliver a faithful and accurate representation of your request. A model could still be accurate while navigating your lack of specificity.
timacles 10 hours ago
Kasparov vs Karpov ‘87 Olympiad. Move 6
dudul 11 hours ago
What do you mean? Parent clearly describes the Queen's Gambit. 1.d4 d5 2.c4 There is no room for ambiguity here.
kuboble 7 hours ago
tezza 3 hours ago
I've rushed out my standardised quality check images for gpt-image-2:
https://generative-ai.review/2026/04/rush-openai-gpt-image-2...
I've done a series over all the OpenAI models.
gpt-image-2 has a lot more action, especially in the Apple Cart images.
RigelKentaurus 13 hours ago
If every single image on their blog was generated by Images 2.0 (I've no reason to believe that's not the case), then wow, I'm seriously impressed. The fidelity to text, the photorealism, the ability to show the same character in a variety of situations (e.g. the manga art) -- it's all great!
platinumrad 10 hours ago
Why do all of the cartoons still look like that? Genuinely asking.
orthoxerox an hour ago
That was my reaction as well. Either they have decided than LLMs have this "house style" for stylized 2D art and we should deal with it, or no amount of prompting can get rid of it.
modeless 12 hours ago
Can it generate transparent PNGs yet?
alasano 11 hours ago
Previous gpt image models could (when generating, not editing) but gpt-image-2 can't.
Noticed it earlier while updating my playground to support it
lxgr 11 hours ago
Works for me, but really weirdly on iOS: Copying to clipboard somehow seems to break transparency; saving to the iOS gallery does not. (And I’ve made sure to not accidentally depend on iOS’s background segmentation.)
vunderba 9 hours ago
OpenAI’s API docs are frustratingly unclear on this. From my experience, you can definitely generate true transparent PNG files through the ChatGPT interface, including with the new GPT-Image-2 model, but I haven’t found any definitive way to do the same thing via the API.
jumploops 6 hours ago
Looks like analog clocks work well enough now, however it still struggles with left-handed people.
Overall, quite impressed with its continuity and agentic (i.e. research) features.
dazhbog 11 hours ago
Yay, let's burn the planet computing more slopium..
naseemali925 4 hours ago
Its amazingly good at creating UI mockups. Been trying this to create UI mockups for ideas.
mvkel 7 hours ago
I wonder if this confirms version 1 of some kind of "world model."
It has an unprecedented ability to generate the real thing (for example, a working barcode for a real book)
BohdanPetryshyn an hour ago
Am I the only one for whom videos in OpenAI releases never load? Tried both Chrome and Safari
franze 10 hours ago
the tragedy of image generating ai is that it is used to massively create what already exists instead of creating something truly unique - we need ai artists - and yeah, they will not be appreciated
franze 10 hours ago
so yeah a smart move of openai would be to sponsor artists - provokant ones, junior ones, with nothing to lose - but that cell in the spreadsheet will be too small to register and will prop. never happen
vunderba 8 hours ago
I decided to run gpt-image-2 on some of the custom comics I’ve come up with over the years to see how well it would do, since some of them are pretty unusual. Overall, I was quite impressed with how faithful it adhered to the prompts given that multi-panel stuff has to maintain a sense of continuity.
Was surprised to see it be able to render a decent comic illustrating an unemployed Pac-Man forced to find work as a glorified pie chart in a boardroom of ghosts.
etothet 12 hours ago
I would love to see prompt examples that created the images on the announcement page.
DauntingPear7 10 hours ago
You can by changing the view before the gallery
james2doyle 10 hours ago
In the next round of ChatGPT advertisements, if they don’t use AI generated images, then that means they don’t believe in their own product right?
muyuu 12 hours ago
I wonder if this will be decent at creating sprite frame animations. So far I've had very poor results and I've had to do the unthinkable and toil it out manually.
vunderba 10 hours ago
I created this little demo of an animated sprite sheet using generative AI. It's not great, but it is passable.
muyuu 9 hours ago
Looks good to me. Would be nice to see the process. I'm having trouble with parts of the stride when the far leg is ahead. Doing 8-directional isometric right now.
freedomben 11 hours ago
I had exactly the same thought! I've got a game I've been wanting to build for over a decade that I recently started working on. The art is going to be very challenging however, because I lack a lot of those skills. I am really hoping the AI tools can help with that.
Is anyone doing this already who can share information on what the best models are?
gizmodo59 11 hours ago
Use the imagegen skill in codex and ask it to create sprites. It works really well.
muyuu 10 hours ago
freedomben 10 hours ago
ZeWaka 11 hours ago
It's still bad.
jcattle 3 hours ago
Can we talk about how jarring the announcement video is?
AI generated voice over, likely AI generated script (You see, this model isn't just generating images, it's thinking!). From what it looks like only the editing has some human touch to it?
It does this Apple style announcement which everyone is doing, but through the use of AI, at least for me, it falls right into the uncanny valley.
StefanBatory 2 hours ago
Do you think those working at ChatGPT have ever wondered how they are contributing to dismantling democracy and ensuring nothing is true by now? The ultimate technological postmodernism.
wiseowise an hour ago
They’re too busy counting cash. Most of them are what? 30 something to 50? By the time democracy is dismantled they’ll be living in their protected mansions.
lifeisstillgood 11 hours ago
Pretty much all of the kerfuffle over AI would go away of it was accurately priced.
After 2008 and 2020 vast (10s of trillions) amounts of money has been printed (reasonably) by western gov and not eliminated from the money supply. So there are vast sums swilling about - and funding things like using massively Computationally intensive work to help me pick a recipie for tonight.
Google and Facebook had online advertising sewn up - but AI is waaay better at answering my queries. So OpenAI wants some of that - but the cost per query must be orders of magnitude larger
So charge me, or my advertisers the correct amount. Charge me the right amount to design my logo or print an amusing cat photo.
Charge me the right cost for the AI slop on YouTube
Charge the right amount - and watch as people just realise it ain’t worth it 95% of the time.
Great technology - but price matters in an economy.
fizlebit 4 hours ago
Scrolling through those images it just feels like intellectual theft on a massive scale. The only place I think you're going to get genuinely new ideas is from humans. Whether those humans use AI or not I don't care, but the repetitive slop of AI copying the creative output of humans I don't find that interesting. Call me a curmudgeon. I guess humans also create a lot of derivative slop even without AI assistance. If this leads somehow to nicer looking user interfaces and architecture maybe that is good thing. There are a lot of ugly websites, buildings and products.
kanodiaayush 11 hours ago
It stands out to me that this page itself is wonderful to go through (the telling of the product through model generated images).
cyberjunkie 6 hours ago
Looks like AI and I look away from any image generated by a LLM. It's my easy internal filter to weed out everything that isn't art.
dakiol 12 hours ago
> On the flip side, there are hundreds of ways that these tools cause genuine harm, not just to individuals but to entire systems.
Yeah, agree. I think it's the first time I'm asking myself: Ok, so this new cool tech, what is it good for? Like, in terms of art, it's discarded (art is about humans), in terms of assets: sure, but people is getting tired of AI-generated images (and even if we cannot tell if an image is AI-generated, we can know if companies are using AI to generate images in general, so the appealing is decreasing). Ads? C'mon that's depressing.
What else? In general, I think people are starting to realize that things generated without effort are not worth spending time with (e.g., no one is going to read your 30-pages draft generated by AI; no one is going to review your 500 files changes PR generated by AI; no one is going to be impressed by the images you generate by AI; same goes for music and everything). I think we are gonna see a Renaissance of "human-generated" sooner rather than later. I see it already at work (colleagues writing in slack "I swear the next message is not AI generated" and the like)
lucaslazarus 11 hours ago
> I think it's the first time I'm asking myself: Ok, so this new cool tech, what is it good for?
I feel like this is something people in the industry should be thinking about a lot, all the time. Too many social ills today are downstream of the 2000s culture of mainstream absolute technoöptimism.
Vide. Kranzberg's first law--“Technology is neither good nor bad; nor is it neutral.”
runarberg 11 hours ago
Completely unrelated, but I am curious about your keyboard layout since you mistyped ö instead of - these two symbols are side by side in the Icelandic layout, and the ö is where - in the English (US) layout. As such this is a common type-o for people who regularly switch between the Icelandic and the English (US) layout (source: I am that person). I am curious whether more layouts where that could be common.
bulletsvshumans 11 hours ago
heisenzombie 11 hours ago
lxgr 11 hours ago
I can’t design wallpapers/stickers/icons/…, but I can describe what I want to an image generation model verbally or with a source photo, and the new ones yield pretty good results.
For icons in particular, this opens up a completely new way of customizing my home screen and shortcuts.
Not necessary for the survival of society, maybe, but I enjoy this new capability.
latexr 11 hours ago
So we get a fresh new cheap way to spread propaganda and lies and erode trust all across society while cementing power and control for a few at the top, and in return get a few measly icons (as if there weren’t literally thousands of them freely available already) and silly images for momentaneous amusement?
What a rotten exchange.
SamuelAdams 11 hours ago
jll29 10 hours ago
thesmtsolver2 10 hours ago
camillomiller 11 hours ago
Is that worth the cost of this technology? Both in terms of financial shenanigans and its environmental cost?
subroutine 11 hours ago
Legend2440 11 hours ago
vrc 11 hours ago
3dsnano 11 hours ago
Gigachad 11 hours ago
This is where I’m at. If you can’t be bothered to write/make it, why would I be bothered to read or review it?
tempaccount5050 11 hours ago
Because I'm not an artist and can't afford to pay one for whatever business I have? This idea that only experts are allowed to do things is just crazy to me. A band poster doesn't have to be a labor of love artisanal thing. Were you mad when people made band posters with MS word instead of hiring a fucking typesetter? I just don't get it.
overgard 11 hours ago
AkBKukU 11 hours ago
squidsoup 11 hours ago
Arch485 11 hours ago
Peritract 8 hours ago
jll29 10 hours ago
swader999 11 hours ago
Gigachad 11 hours ago
reaperducer 11 hours ago
Jtarii 10 hours ago
satisfice 10 hours ago
zulban 11 hours ago
Nobody can be bothered to make my cat out of Lego and the size of mount Everest but if an AI did I'd sure love to see it.
Your quip is pithy but meaningless.
Gigachad 11 hours ago
loudandskittish 10 hours ago
Exactly how I feel. There is already more art, movies, music, books, video games and more made by human beings than I can experience in my lifetime. Why should I waste any time on content generated by the word guessing machine?
atleastoptimal 11 hours ago
The issue is that the signalling makes sense when human generated work is better than AI generated. Soon AI generated work will be better across the board with the rare exception of stuff the top X% of humans put a lot of bespoke highly personalized effort into. Preferring human work will be luxury status-signalling just like it is for clothing, food, etc.
dilDDoS 11 hours ago
I'm probably in a weird subgroup that isn't representative of the general public, but I've found myself preferring "rough" art/logos/images/etc, basically because it signals a human put time into it. Or maybe not preferring, but at least noticing it more than the generally highly refined/polished AI artwork that I've been seeing.
appplication 10 hours ago
james2doyle 10 hours ago
I think "better" is doing a lot of heavy lifting in this argument. Better how?
Is an AI generated photo of your app/site going to be more accurate than a screenshot? Or is an AI generated image of your product going to convey the quality of it more than a photo would?
I think Sora also showed that the novelty of generating just "content" is pretty fleeting.
I would be interested to see if any of the next round of ChatGPT advertisements use AI generated images. Because if not, they don’t even believe in their own product.
masswerk 10 hours ago
The issue being, it's not an expression of anything. Merely like a random sensation, maybe some readable intent, but generic in execution, which isn't about anything even corporate art should be about. Are we going to give up on art, altogether?
Edit: One of the possible outcomes may be living in a world like in "Them" with glasses on. Since no expression has any meaning anymore, the message is just there being a signal of some kind. (Generic "BUY" + associated brand name in small print, etc.)
ragequittah 10 hours ago
fwipsy 11 hours ago
Only novel art is interesting. AI can't really do novel. It's a prediction algorithm; it imitates. You can add noise, but that mostly just makes it worse. It can be used to facilitate original stuff though.
But so many people want to make art, and it's so cheap to distribute it, that art is already commoditized. If people prefer human-created art, satisfying that preference is practically free.
atleastoptimal 10 hours ago
idiotsecant 10 hours ago
paulddraper 11 hours ago
"Artisanal art" as it were.
vinyl7 10 hours ago
The goal of art isn't to be perfect or as realistic as possible. The goal of art is to express, and enjoy that unique expression.
davebren 10 hours ago
> Preferring human work will be luxury status-signalling just like it is for clothing, food, etc.
What? Those items are luxuries when made by humans because they are physical goods where every single item comes with a production and distribution cost.
strulovich 11 hours ago
Here’s one example:
I just recently used for image generation to design my balcony.
It was a great way to see design ideas imagined in place and decide what to do.
There are many cases people would hire an artist to illustrate an idea or early prototype. AI generated images make that something you can do by yourself or 10x faster than a few years ago.
dwd 10 hours ago
Did the same for my front garden.
Not withstanding a few code violations, it generated some good ideas we were then able to tweak. The main thing was we had no idea of what we wanted to do, but seeing a lot of possibilities overlaid over the existing non-garden got us going. We were then able to extend the theme to other parts of the yard.
tecoholic 10 hours ago
100%. A picture is worth a thousand words only when it conveys something. I love to see the pictures from my family even when they are taken with no care to quality or composition but I would look at someone else’s (as in gallery/exhibitions) only when they are stunning and captured beautifully. The medium is only a channel to communicate.
Also, this can’t be real. How many publications did they train this stuff on and why are there no acknowledgment even if to say - we partnered with xyz manga house to make our model smarter at manga? Like what’s wrong with this company?
_the_inflator 11 hours ago
We need to flip the script. AI is trying to do marketing: add “illegal usage will lead to X” is a gateway to spark curiosity. There is this saying that censoring games for young adults makes sure that they will buy it like crazy by circumventing the restrictions because danger is cool.
There is nothing that cannot harm. Knives, cars, alcohol, drugs. A society needs to balance risks and benefits. Word can be used to do harm, email, anything - it depends on intention and its type.
_the_inflator 11 hours ago
I see your point but reconsider: we will and need to see. Time will tell and this is simply economics: useful? Yes, no.
I started being totally indifferent after thinking about my spending habits to check for unnecessary stuff after watching world championships for niche sports. For some this is a calling for others waste. It is a numbers game then.
Havoc 10 hours ago
>and even if we cannot tell if an image is AI-generated, we can know if companies are using AI to generate images in general, so the appealing is decreasing
Is that true? Don't think I'd get tired of images that are as good as human made ones just because I know/suspect there may have been AI involved
youdots 11 hours ago
The technically (in both senses) astonishing and amazing output is not far off from some of the qualities of real advertising: Staged, attention grabbing, artificially created, superficially demanded, commercially attractive qualities. These align, and lots of similarities in the functions and outcomes of these two spheres come to mind.
simonw 11 hours ago
I think there's real value to be had in using this for diagrams.
Visual explanations are useful, but most people don't have the talent and/or the time to produce them.
This new model (and Nano Banana Pro before it) has tipped across the quality boundary where it actually can produce a visual explanation that moves beyond space-filling slop and helps people understand a concept.
I've never used an AI-generated image in a presentation or document before, but I'm teetering on the edge of considering it now provided it genuinely elevates the material and helps explain a concept that otherwise wouldn't be clear.
mwcampbell 10 hours ago
Are there any models that are specifically trained to produce diagrams as SVG? I'd much prefer that to diffusion-based raster image generation models for a few reasons:
- The usual advantages of vector graphics: resolution-independence, zoom without jagged edges, etc.
- As a consequence of the above, vector graphics (particularly SVG) can more easily be converted to useful tactile graphics for blind people.
- Vector graphics can more practically be edited.
twobitshifter 8 hours ago
resters 11 hours ago
This is the key point. In my view it's just like anything else, if AI can help humans create better work, it's a good thing.
I think what we'll find is that visual design is no longer as much of a moat for expressing concepts, branding, etc. In a way, AI-generated design opens the door for more competition on merits, not just those who can afford the top tier design firm.
lol_me 11 hours ago
yeah I'm not sure I'm in agreement that we can hand-wave assets and ads as entire classes of valuable content
swader999 11 hours ago
I tend to share your same view. But is there really a line like you describe? Maybe AI just needs to get a few iterations better and we'll all love what it generates. And how's it really any different than any Photoshop computer output from the past?
JumpCrisscross 11 hours ago
> What else?
I used to have an assistant make little index-card sized agendas for gettogethers when folks were in town or I was organising a holiday or offsite. They used to be physical; now it's a cute thing I can text around so everyone knows when they should be up by (and by when, if they've slept in, they can go back to bed). AI has been good at making these. They don't need to be works of art, just cute and silly and maybe embedded with an inside joke.
pesus 9 hours ago
I'm not seeing how it takes more than 5 minutes to type up an itinerary. If you want to make it cute and silly, just change up the font and color and add some clip art.
If this is the best use case that exists for AI image generation, I'm only further convinced the tech is at best largely useless.
JumpCrisscross 8 hours ago
jll29 10 hours ago
You are kidding, right?
It's good that my friends don't make a coffee date feel like a board meeting (with an agenda shared by post 14 working days ahead of the meeting, form for proxy voting attached).
reaperducer 11 hours ago
I don't care how many times you write "cute," having my vacation time programmed with that level of granularity and imposed obligation sounds like the definition of "dystopian."
If I got one of your cute schedule cards while visiting you, I'd tear it up, check into a cheap motel, and spend the rest of my vacation actually enjoying myself.
Edit: I'm not an outlier here. There have even been sitcom episodes about overbearing hosts over-programming their guests' visits, going back at least to the Brady Bunch.
JumpCrisscross 11 hours ago
gustavus 11 hours ago
I'm working on an edutech game. Before I would've had much less of a product because I don't have the budget to hire an artist and it would've been much less interactive but because of this I'm able to build a much more engaging experience so that's one thing. For what it's worth.
NikolaNovak 11 hours ago
While I agree with you, hacker news audience is not in the middle of the bell curve.
I get this sounds elitist - but tremendous percentage of population is happily and eagerly engaging with fake religious images, funny AI videos, horrible AI memes, etc. Trying to mention that this video of puppy is completely AI generated results in vicious defense and mansplaining of why this video is totally real (I love it when video has e.g. Sora watermarks... This does not stop the defenders).
I agree with you that human connection and artist intent is what I'm looking for in art, music, video games, etc... But gawd, lowest common denominator is and always has been SO much lower than we want to admit to ourselves.
Very few people want thoughtful analysis that contradicts their world view, very few people care about privacy or rights or future or using the right tool, very few people are interested in moral frameworks or ethical philosophy, and very few people care about real and verifiable human connection in their "content" :-/
Peritract 8 hours ago
HN is absolutely not more critical of AI output than the norm.
It's been true for various technologies that HN (and tech audiences in general) have a more nuanced view, but AI flips the script on that entirely. It's the tech world who are amazed by this, producing and being delighted by endless blogposts and 7-second concept trailers.
ryandrake 9 hours ago
I recently shoulder-surfed a family member scrolling away on their social media feed, and every single image was obvious AI slop. But it didn't matter. She loved every single one, watched videos all the way through, liked and commented on them... just total zombie-consumption mode and it was all 100% AI generated. I've tried in the past pointing out that it's all AI generated and nothing is real, and they simply don't care. People are just pac-man gobbling up "content". It's pretty sad/scary.
slibhb 10 hours ago
> Like, in terms of art, it's discarded (art is about humans)
If a work of art is good, then it's good. It doesn't matter if it came from a human, a neanderthal, AI, or monkeys randomly typing.
Jtarii 10 hours ago
The connection with the artist, directly, or across space and time, is a critical part of any artwork. It is one human attempting to communicate some emotional experience to another human.
When I watch a Lynch film I feel some connection to the man David Lynch. When I see a AI artwork, there is nothing to connect with, no emotional experience is being communicated, it is just empty. It's highest aspiration is elevator music, just being something vaguely stimulating in the background.
papa_bear 10 hours ago
Provenance is part of the work. If a roomful of monkeys banged out something that looked like anything, I'd absolutely hang it on my wall. I would not say the same for 99% of AI generated art.
avaer 10 hours ago
Whether art is considered good is in practice highly contextual. One of those contexts is who (what) made it.
papichulo2023 11 hours ago
Seems good enough to generate 2D sprites. If that means a wave of pixel-art games I count it as a net win.
I dont think gamers hate AI, it is just a vocal miniority imo. What most people dislike is sloppy work, as they should, but that can happen with or without AI. The industry has been using AI for textures, voices and more for over a decade.
vunderba 10 hours ago
> Seems good enough to generate 2D sprites.
It’s really not. That's actually a pet peeve of mine as someone who used to spent a lot of time messing with pixel art in Aseprite.
Nobody takes the time to understand that the style of pixel art is not the same thing as actual pixel art. So you end up with these high-definition, high-resolution images that people try to pass off as pixel art, but if you zoom in even a tiny bit, you see all this terrible fringing and fraying.
That happens because the palette is way outside the bounds of what pixel art should use, where proper pixel art is generally limited to maybe 8 to 32 colors, usually.
There are plenty of ways to post-process generative images to make them look more like real pixel art (square grid alignment, palette reduction, etc.), but it does require a bit more manual finesse [1], and unfortunately most people just can’t be bothered.
loudandskittish 11 hours ago
There are already more games being released on Steam than anyone can keep up with, I'm not sure how adding another "wave" on top of it helps.
tiagod 11 hours ago
AI for textures for over a decade? What AI?
papichulo2023 11 hours ago
Thonn 11 hours ago
Are you kidding? I think I see more vitriol for AI in gaming communities than anywhere else. To the point where steam now requires you to disclose its usage
papichulo2023 11 hours ago
NetOpWibby 11 hours ago
The Human Renaissance is something I've been thinking of too and I hope it comes to pass. Of course, I feel like societally, things are gonna get worse for a lot of folks. You already see it in entire towns losing water or their water becoming polluted.
You'd think these kickbacks leaders of these towns are getting for allowing data centers to be built would go towards improving infrastructure but hah, that's unrealistic.
WTF is that unrealistic? SMH
Lerc 10 hours ago
>You already see it in entire towns losing water or their water becoming polluted
Do you have any references for such cases? I have seen talk of such thing at risk, but I am unaware of any specific instances of it occuring
NetOpWibby 5 hours ago
underlipton 11 hours ago
>Like, in terms of art, it's discarded (art is about humans)
I dunno how long this is going to hold up. In 50 years, when OpenAI has long become a memory, post-bubble burst, and a half-century of bitrot has claimed much of what was generated in this era, how valuable do you think an AI image file from 2023 - with provenance - might be, as an emblem and artifact of our current cultural moment, of those first few years when a human could tell a computer, "Hey, make this," and it did? And many of the early tools are gone; you can't use them anymore.
Consider: there will never be another DallE-2 image generation. Ever.
RIMR 11 hours ago
My only actual use of image or video AI tools is self-entertainment. I like to give it prompts and see the results it gives me.
That's it. I can't think of a single actual use case outside of this that isn't deliberately manipulative and harmful.
colechristensen 11 hours ago
>In general, I think people are starting to realize that things generated without effort are not worth spending time with
Agreed mostly, BUT
I'm building tools for myself. The end goal isn't the intermediate tool, they're enabling other things. I have a suspicion that I could sell the tools, I don't particularly want to. There's a gap between "does everything I want it to" and "polished enough to justify sale", and that gap doesn't excite me.
They're definitely not generated without effort... but they are generated with 1% of the human effort they would require.
I feel very much empowered by AI to do the things I've always wanted to do. (when I mention this there's always someone who comes out effectively calling me delusional for being satisfied with something built with LLMs)
iLoveOncall 11 hours ago
Porn and memes. Obviously. This is all that Stable Diffusion has been used for since it was released.
ArchieScrivener 11 hours ago
I completely disagree, this replaces art as a job. Why does human art need monetary feedback to be shared? If people require a paycheck to make art then it was never anything different than what Ai generated images are.
As for advertising being depressing - its a little late to get up on the high horse of anti-Ads for tech after 2 decades of ad based technology dominating everything. Go outside, see all those bright shiny glittery lights, those aren't society created images to embolden the spirit and dazzle the senses, those are ads.
North Korea looks weird and depressing because the don't have ads. Welcome to the west.
tomrod 11 hours ago
AI loopidity rearing it's head. Just send the bullet points that we all want anyway, right?! Stop sending globs of text and other generated content!
agnishom 10 hours ago
I don't know how this benefits humanity. In what way was ChatGPT Images 1.0 not already good enough? Perhaps some new knowledge was created in the process?
tomchui157 3 hours ago
Img2+ seed dance 2 = image AGI
Melatonic 14 hours ago
Can it generate anything high resolution at increased cost and time? Or is it always restricted?
jwpapi 10 hours ago
Why is it all so asian?
twobitshifter 8 hours ago
Having 60% of the world’s population might do that.
XCSme 10 hours ago
Oh wow, scrolling through the page on mobile makes me dizzy
RyanJohn 7 hours ago
Oh my god, it's very nice!
dahuangf 2 hours ago
good job
apparent 10 hours ago
I find the video to be very annoying. Am I supposed to freeze frame 4x per second to be able to see whether the images are actually good? I've never before felt stressed watching a launch video.
Havoc 9 hours ago
Yeah same. At first I thought they're using it to conceal quality, but pausing it they do actually look really good, so strange choice.
Maybe it's meant to convey pace & hype
apparent 9 hours ago
Maybe so, but to me it conveys a headache.
bitnovus 15 hours ago
great obfuscation idea - hidden message on a grain of rice
ibudiallo 11 hours ago
And here I was proud of myself, having taught my mom and her friends how to discern real from fakes they get on WhatsApp groups. Another even more powerful tool for scammers. I'm taking a break.
bananaflag 4 hours ago
I told my mom not to believe anything unless she trusts the source. The way people always did with text.
XorNot 11 hours ago
IMO you're fighting the wrong battle: there'll always be a new model.
But the broader concept of fake news and the manufactured nature of media and rhetoric is much more relevant - e.g. whether or not something's AI is almost immaterial to the fact that any filmed segment does not have to be real or attributed to the correct context.
Its an old internet classic just to grab an image and put a different caption on it, relying on the fact no one can discern context or has time to fact check.
gfody 11 hours ago
there's something funny going on with the live stream audio
szmarczak 14 hours ago
Wow, the difference between AI and non-AI images collapses. I hate the future where I won't be able to tell the difference.
Flere-Imsaho 14 hours ago
I wake up everyday, read the tech news, and usually see some step change in AI or whatever. It's wild to think I'm living through such a massive transformation in my lifetime. The future of tech is going to be so different from when I was born (1980), I guess this is how people born in 1900 felt when they got to see man land on the moon?
> Wow, the difference between AI and non-AI images collapses. I hate the future where I won't be able to tell the difference.
Image generation is now pretty much "solved". Video will be next. Perhaps things will turn out the same as chess: in that even though chess was "solved" by IBM's Deep Blue, we still value humans playing chess. We value "hand made" items (clothes, furniture) over the factory made stuff. We appreciate & value human effort more than machines. Do you prefer a hand-written birthday card or an email?
toraway 14 hours ago
"Solved" seems a tad overstated if you scroll up to Simonw's Where's Waldo test with deformed faces plus a confabulated target when prompted for an edit to highlight the hidden character with an arrow.
Flere-Imsaho 13 hours ago
abraxas 14 hours ago
As someone born in 1975 I always felt until the last couple of years that I had been stuck in a long period of stagnation compared to an earlier generation. My grandmother who was born in the 1910s got to witness adoption of electricity, mass transit, radio, television, telephony, jet flights and even space exploration before I was born.
Feels like now is a bit of a catchup after pretty tepid period that was most of my life.
cubefox 11 hours ago
dag100 14 hours ago
Chess exists solely for the sake of the humans playing it. Even if machines solved chess, people would rather play chess against a person than a machine because it is a social activity in a way. It's like playing tennis versus a person compared to tennis against a wall.
Photographs, videos, and digital media in general, in contrast, are used for much, much more than just socializing.
gekoxyz 14 hours ago
Well, for some of these images for the first time I can't tell that they are AI generated
mcfry 7 hours ago
How hard is it to have a video player with a fucking volume toggle?
esafak 15 hours ago
rqa129 14 hours ago
Thanks, all displayed images look horrible and artificial. This will fail like Sora.
gekoxyz 14 hours ago
Hard disagree on this, I was coming here to comment that this is the first time I really can't tell that some of the photos are AI generated.
furyofantares 14 hours ago
I felt the same, particularly with the diagrams / magazines anyway.
I don't think it'll fail like Sora though. gpt-image-1.5 didn't fail.
livinglist 11 hours ago
Denial is real…
QuantumGood 13 hours ago
Your single other comment is simplistic hyperbole as well, so this is presumably a bot account.
bitnovus 15 hours ago
No gpt-5.5
wahnfrieden 13 hours ago
Thursday
dzonga 11 hours ago
for video game assets this is massive.
but in general though - will people believe in anything photographic ?
imagine dating apps, photographic evidence.
I'm guessing we're gonna reach a point where - you fuck up things purposely to leave a human mark.
telman17 5 hours ago
> for video game assets this is massive.
Storefronts like Steam require disclosing use of AI assets for art. In most indie dev spaces, devs are scolded for using AI art in their games. I wonder if this perspective will change in a few years.
squidsoup 11 hours ago
> but in general though - will people believe in anything photographic ?
Hopefully film makes a come back.
andai 11 hours ago
lol at the fake handwritten homework assignment. Know your customer!
OutOfHere 9 hours ago
ChatGPT image generation is and has been horrific for the simple reason that it rejects too many requests. This hasn't changed with the new model. There are too many legal non-adult requests that are rejected, not only for edits, but also for original image generation. I'd rather pay to use something that actually works.
davikr 10 hours ago
It definitely lost the characteristic slop look.
irishcoffee 11 hours ago
This is so stupid. As a free OSS tool it’s amazing. Paying money for this is fucking stupid. How blind are we all to now before this tech?
rqa129 14 hours ago
Can it generate Chibi figures to mask the oligarchy's true intentions on Twitter and make them more relatable?
volkk 14 hours ago
the guys presenting are probably all like 25x smarter than I am but good god, literally 0 on screen presence or personality.
sho_hn 14 hours ago
That's a trained skill, and they presumably have focused on other skills.
brcmthrowaway 14 hours ago
Yeah, skills to make them a cool 10mn a year
volkk 14 hours ago
eh, i don't think personalities are trained. on screen presence for sure, but you'd see right through it IRL.
dymk 10 hours ago
OsrsNeedsf2P 11 hours ago
I liked it that way, felt more authentic to see the noobs
E-Reverance 14 hours ago
I think its endearing
Aethelwulf 12 hours ago
didn't think that sam guy was that bad
minimaxir 15 hours ago
HN submission for a direct link to the product announcement which for some reason is being penalized by the HN algorithm: https://news.ycombinator.com/item?id=47853000
dang 10 hours ago
(We eventually merged the threads hither)
simonw 14 hours ago
Suggest renaming this to "OpenAI Livestream: ChatGPT Images 2.0"
dang 12 hours ago
(We've since merged the threads and moved the livestream link to the toptext)
I_am_tiberius 14 hours ago
or "How we make money with your images 2.0".
sho_hn 14 hours ago
In 5 years and 3 months between DALL-E and Images 2.0 we've managed to progress from exuberant excitement to jaded indifference.
nba456_ 11 hours ago
Who's 'we'? Speak for yourself!
kibibu 12 hours ago
Because we are all seeing the harm these tools are being used for.
It's just another step into hell.
welder 10 hours ago
Introducing DeepFakes 2.0 /s
zb3 15 hours ago
Image generation? Hmm, would be cool if OpenAI also made a video-generation model someday..
incognito124 14 hours ago
If only there was a social network with solely AI generated videos, I would pay literal money for it...
allenbina 7 hours ago
If I may address this with both skepticism and curiosity, why. I think I speak for everyone when I say I would pay to go back to facebook 2018. No algorithm, no ai.
Bigpet 5 hours ago
biosubterranean 11 hours ago
Oh no.
ai4thepeople 11 hours ago
Each day when my AI girlfriend wakes me up and shows me the latest news, I feel: This is it! We are living in a revolution!
Never before in history did humanity have the possibility of seeing a picture of a pack of wolves! The dearth of photographs has finally been addressed!
I told my AI girlfriend that I will save money to have access to this new technology. She suggested a circular scheme where OpenAI will pay me $10,000 per year to have access to this rare resource of 21th century daguerreotype.
green_wheel 6 hours ago
Well artists, you guys had a good run thank you for your service.
manishfp 5 hours ago
Goated release tbh. The text work inside the images are nice
aliljet 15 hours ago
I am hopeful that OpenAI will potentially offer clarity on their loss-leading subscription model. I'd prefer to know the real cost of a token from OpenAI as opposed to praying the venture-funded tokens will always be this cheap.
tkgally 9 hours ago
I had it produce a two-page manga with Japanese dialogue. Nearly perfect:
https://www.gally.net/temp/20260422-chatgpt-images-2-example...
prvc 6 hours ago
I hope they will consider releasing DALL-E 2 publicly, now that there has been so much progress since it was unveiled. It had a really nice vibe to it, so worth preserving.
andy_ppp 6 hours ago
Yes, I’ve always thought of AI companies as sentimental. They will definitely do this :-/
prvc 5 hours ago
That's why I want it; their motives for doing it, should they decide to, would presumably be different.
Danox 9 hours ago
Sam Altman in his meeting with Tim Cook two and a half years ago give me money. I think it’ll take $150 billion dollars, Tim Cook well here’s what we’re going to do, this is what I think it’s worth…
Later Google tried the same thing, Apple we will give you a $1 billion dollar a year refund, what’s changed in two and a half years?