Claude now creates interactive charts, diagrams and visualizations (claude.com)
130 points by adocomplete 4 hours ago
captainbland 3 hours ago
I feel like this is a feature which improves the perceived confidence of the LLM but doesn't do much for correctness of other outputs, i.e. an exacerbation of the "confidently incorrect" criticism.
kemayo 2 hours ago
It's a mismatch with our intuition about how much effort things take.
If there's humans involved, "I took this data and made a really fancy interactive chart" means that you put a lot more work into it, and you can probably somewhat assume that this means some more effort was also put into the accuracy of the data.
But with the LLM it's not really very much more work to get the fancy chart. So the thing that was a signifier of effort is now misleading us into trusting data that got no extra effort.
(Humans have been exploiting this tendency to trust fancy graphics forever, of course.)
outlore 7 minutes ago
I suspect chain of thought while building the chart will improve the overall correctness of the answer
ipython 20 minutes ago
Already happened. :)
https://www.reddit.com/r/dataisugly/comments/1mk5wdb/this_ch...
elliotbnvl 2 hours ago
It's a usability / quality of life feature to me. Nothing to do with increasing perceived confidence. I guess it depends on how much you already (dis)trust LLMs.
I'm finding more and more often the limiting factor isn't the LLM, it's my intuition. This goes a way towards helping with that.
programmertote 2 hours ago
A recent LinkedIn post that I came across as an example of people trusting (or learning to trust) AI too much while not realizing that it can make up numbers too: https://www.linkedin.com/posts/mariamartin1728_claude-wrote-...
P.S. Credit to the poster, she posted a correction note when someone caught the issue: https://www.linkedin.com/posts/mariamartin1728_correction-on...
vunderba 2 hours ago
A similar thing happened when Google started really pushing generating flowcharts as a use-case with Nano Banana. A slick presentation can distract people from the only thing that really matters - the accuracy of the underlying data.
Angostura an hour ago
As a slightly different tack, I’ve been using Copilot to generate flowcharts from some of the fiendishly complex (and badly written) standard operating procedures we have at work.
People find them quite easy to check - easier than the raw document. My angle with teams is use these to check your processes. If the flow is wrong it’s either because the LLM has screwed up, or because the policy is wrong/badly written. It’s usually the latter. It’s a good way to fix SOPs
nerdjon 2 hours ago
This was my first thought as well, all this does is further remove the user from seeing the chat output and instead makes it appear as if the information is concretely reliable.
I mean is it really that shocking that you can have an LLM generate structured data and shove that into a visualizer? The concern is if is reliable, which we know it isnt.
ericmcer 2 hours ago
The further they can get people from the reality of `This just spits out whatever it thinks the next token will be` the more they can push the agenda.
j45 2 hours ago
Its' a reasonable concern. Often it can be mitigated by prompting in a manner that invokes research and verification instead of defaulting to a corpus.
Passive questions generate passive responses.
mikkupikku 2 hours ago
I agree. Maybe next they'll add emotionally evocative music, with swelling orchestral bits when you reach the exciting climate of the slop.
darepublic 2 hours ago
When I ask chatgpt to create a mermaid diagram for me it regularly will add new lines to certain labels that will break the parse. If you then feed the parse error back to it the second version is always correct And it seems to exactly know the problem. There are some other examples where it will almost always get it wrong the first time but right if nudged to correct itself. I wonder what the underlying cause is
overfeed 2 hours ago
> I wonder what the underlying cause is
It responds with the statistically most probable text based on its training data, which happens to be different with the errors vs without. I suspect high-fidelity diagramming requires a different attention architecture from the common ones used in sentence-optimized models.
ar0b an hour ago
"Prompt Repetition Improves Non-Reasoning LLMs " - https://arxiv.org/pdf/2512.14982
What instance of ChatGPT are you doing that with? (Reasoning?)
darepublic 14 minutes ago
Observed from 5.2, on chatgpt.com. earlier versions did worse.. as in, they might take a few prompts to generate a parseable syntax. Newer versions just usually deliver one unparseable version then get it right second try. Likely I could prompt engineer to one shot but I think I would always need the specific warning about newlines.
deckar01 an hour ago
Mermaid is really bad about cutting off text after spaces, so you have to insert <br>s everywhere. I’m guessing this is getting rendered instead of escaped by your interface. Or just lost in translation at the tokenizer.
stefan_ 2 hours ago
Today I asked Claude to create me a squidward looking out the window meme and it started generating HTML & CSS to draw squidward in a style best described as "4 year old preschooler". Not quite it yet.
czk 3 hours ago
I tried the periodic table in their examples using sonnet 4.6 on the $20/mo plan. After a few minutes Claude told me it reached the max message length and bailed. I pressed continue and eventually it generated the table, but it wasn't inline, it was a jsx artifact, and I've now hit my daily usage limit.
data-ottawa 2 hours ago
I’m intermittently getting artifacts vs the new visuals api, depending on which version of the Claude app I use. iOS/iPadOS apps are not yet supporting the visualization API, and I don’t see an app-store update yet.
fixxation92 4 hours ago
I find it absolutely mindblowing to witness the rate at which Anthropic can ship new features. Only a year ago I couldn't wait to see some sort of Github integration and then it appeared only a week later. Seriously impressive stuff.
throwatdem12311 25 minutes ago
Have you actually used their products? They are janky, full of bugs and barely work half the time.
They write 100% of their code with Claude. Some of their engineers apparently burn over 100k worth of tokens per month.
It’s not surprising they ship fast at all when the product is actually falling apart at the seams and they just vibe code everything.
mchusma 20 minutes ago
I think to the extent they are making a speed v quality tradeoff, I think they are making the right call. 10x speed over quality any day for me. Reminds me of:
"If brute force doesn't work, you aren't using enough of it." - Isaac Arthur
throwatdem12311 10 minutes ago
ericmcer 2 hours ago
They also pay... insane salaries, like double industry average. That coupled with an IPO on the horizon means they probably have their pick of engineers.
brcmthrowaway 30 minutes ago
You have 23 year olds earning $2mn/year, at least it isn't in HFT though!
dheera 2 hours ago
Their interviews are actually very much focused on how fast you can code something that works.
Razengan 3 hours ago
Meanwhile, you still can't Sign in with Apple on the website.
But you can Sign in with Google.
If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Also if you try to copy-paste a prompt from Notes etc into the Claude chat, it gets added as an attachment, so you can't edit the prompt. If you do the four-finger shortcut to paste it as text, it mangles newlines etc.
Why are they so dumb about such basic UX for so long?
radley 3 hours ago
> you still can't Sign in with Apple on the website.
Apple forces developers to offer Sign in with Apple on iOS devices if any other sign in service is used. Apple can't force them to do it on non-Apple platforms.
Wowfunhappy 3 hours ago
> If you signed up with your Apple on the iOS Claude app, to access your account on the computer, you have to open the passwords app and copy your random email address and paste it into the Claude website login.
Isn't this basically Apple's fault? When you signed up, Apple provided a fake email address in leu of your real one. This is great for privacy but means the service has the wrong email.
I'm sure they didn't want to provide an Apple sign in option at all, but it's required by App Store rules.
nerdjon 2 hours ago
tstrimple 2 hours ago
Razengan 2 hours ago
bombela 3 hours ago
Add to the list backtick handling. If you start a backtick block on the claude web chat, you cannot leave it with the keyboard. You are now stuck between the backticks. It is as if they wanted to reproduce Slack misery.
catgirlinspace 3 hours ago
short_sells_poo 3 hours ago
They clearly vibe code a lot (most? all?) of their stuff, and it shows. Elementary features are broken regularly and while I appreciate them trying new features, I'd appreciate it more if existing ones were reliable and promptly fixed if broken.
mikkupikku 2 hours ago
Trouble is, vibe coding refinements and bug fixes works well but probably isn't a good track to promotions at Anthropic (or virtually any other company.)
Gareth321 2 hours ago
I asked it to do some portfolio analysis for me and it created BEAUTIFUL, tabbed, interactive charts UNPROMPTED. This is kind of magical. The charts were not just beautiful, but actually super useful in understanding the data faster. I honestly could not have produced those in a week if you asked me to.
atonse 3 hours ago
Wow, I asked it to build me a simple diagram explaining agile development and it did an amazing job. Wow it felt magical to watch that diagram slowly animating to life.
Like a much prettier version of Mermaid.
Kudos, Anthropic. Geez, this is so nice.
Now I'm going to ask it to draw a diagram of a pelican riding a bicycle, why not?
wuweiaxin 2 hours ago
The artifact output model is more useful than it looks at first. We use Claude in a multi-agent pipeline and discovered that structured artifact outputs reduce parse errors significantly compared to freeform text responses -- the model seems to reason differently when it knows the output will be rendered. Curious whether Anthropic sees similar quality improvements in tool-use tasks when the output has a concrete format constraint.
JoshGG 3 hours ago
This is pretty neat and I am experimenting with it now, but hasn't ChatGPT had capability to create graphs and interact with data for a while? "ChatGPT advanced data analysis" for example. I'm asking in good faith as maybe some of you have been using that and can compare the two and give an informed opinion.
I usually use a lot of other tools for data analysis or write code with Claude code or another LLM to do data analysis and visualization.
article about the ChatGPT charts and graphs https://www.zdnet.com/article/how-to-use-chatgpt-to-make-cha...
Gareth321 3 hours ago
> but hasn't ChatGPT had capability to create graphs and interact with data for a while?
It's pretty bad (for me). I have to use extremely prescriptive language to tell ChatGPT what to create. Even down to the colours in the chart, because otherwise it puts black font on black background (for example). Then I have to specifically tell it to put it in a canvas, and make it interactive, and make it executable in the canvas. Then if I'm lucky I have to hit a "preview" button in the top right and hope it works (it doesn't). I could write several paragraphs telling it to do something like what Claude just demo'd and it wouldn't come close. I'm trying Claude now for financial insights and it's effortless with beautiful UX.
For posterity, Gemini is pretty good with these interactive canvases. Not nearly as good, but FAR better than ChatGPT.
drewda 3 hours ago
When using Claude Code, we often prompt it to draft diagrams in MermaidJS syntax.
Great for summarizing a multi-step process and quick to render with simple tools.
asim 3 hours ago
It was inevitable until the point all apps will disappear and AI will be the entry point for all work. You can see how anything required appear based on a single request. After which world models and other forms of interaction that are more dynamic will make sense and we'll need something that's not a screen.
joshribakoff 3 hours ago
Its a large leap from “we made a config driven diagram tool and trained an llm on that config” to “all apps will disappear”. If you’re predicting such grand claims please be more precise than “AI” which is a term we cant define.
ericmcer 2 hours ago
Yeah an app doesn't "disappear" because you put an AI interface in front of it and then use a bunch of old school programming to parse LLM output and feed that into your old app. 99% of the work is still building the old app.
bogzz 3 hours ago
You're harshing the vibe, man.
abnercoimbre 3 hours ago
gkfasdfasdf 3 hours ago
I would love to know how they built this. Did they use json-render [0], openui [1], or rolled their own?
gavinray 3 hours ago
Right-click the page and inspect the source code?
razerbeans 3 hours ago
Interesting. So if I'm reading this correctly, this is distinctly different than the artifacts that Claude creates? If that's the case, why create it inline as opposed to an artifact? Any time I get a visual, I tend to find them so useful that I _want_ them to be an artifact that I can export and share.
shiftyck 3 hours ago
Claude is broken for me since this was released, prompts are just timing out and stopping after 10 attempts
johsole an hour ago
love to see it, my auto researcher is getting more capable with less effort every release
I_am_tiberius 3 hours ago
Does anyone know which library they use? Or something developed internally?
I_am_tiberius 3 hours ago
Ok, asked it myself: Chart.js
jzig 3 hours ago
Unable to reproduce the recipe image in the iOS app. It first gave a normal text answer. Then when referencing this blog post it produced a wonky HTML artifact.
wuweiaxin 2 hours ago
Reliability has been the real bottleneck for multi-agent setups in production. The hard part isnt getting one agent to do something clever once - its making repeated runs observable and bounded when tools fail halfway through. Idempotency checks, explicit handoff state, and human review gates have mattered more for us than adding another model or another agent role.
alansaber an hour ago
New account spamming LM replies, nice
atonse 3 hours ago
Anyone else able to use Claude with Excel? I've tried adding it to our (very small) Office365 org and it just fails. Been failing for months.
alansaber an hour ago
All the office js integrations are still pretty shitty
mehdibl 3 hours ago
Isn't this mainly a skill injected in the context? Rather a model/platform specific feature?
groby_b an hour ago
Aaand all the way at the bottom, there it is. The first glimpse of what will be an ad carousel.
(Literally nobody needs an image of a cake when asking for a cake recipe)
alansaber an hour ago
What if you've never seen a cake before?
groby_b 23 minutes ago
That's the target group for the feature. You're right. You got me.
HotGarbage 3 hours ago
Interactive slop is still slop.