OpenAI's new GPT‑5.3‑Codex‑Spark is 15 times faster at coding than its predecessor.
See full article...
See full article...
They have redundancy and employ yield improvement techniques so that they can ship a viable product even though portions of each chip are defective.is this really a wafer sized chip? Because that sounds…difficult and expensive…and very difficult. A manufacturing or packaging defect is going to be very expensive when you have to throw the entire wafer away.
Maybe. Wafer-scale "chips" can have enormous amounts of programmable redundancy. This happens even with individual dies in more normal situations.is this really a wafer sized chip? Because that sounds…difficult and expensive…and very difficult. A manufacturing or packaging defect is going to be very expensive when you have to throw the entire wafer away.
uint32_t new = (old >> 22) << 2;uint32_t new = (old >> 20) & 0xFFC;This was inevitable. ASICs were always going to beat GPUs just as they have always beat GPUs for other problem areas. It just takes a decade or so for the algorithms to settle down to the point that you can use a ASIC rather than a more programmable thing like a GPU.Aka: the cancer has spread to a new species of hardware.
Per this Techcrunch article, that $200/mo plan loses money because people are using it too much.$200/month
Wow.
Well somehow you need to create the concept of large income amounts if you want to continue justify massive money burn AI companies are doing.
The trick would be to design the chip such that every part is redundant and can be disabled if there's a defect. Tricky, but certainly not impossible.is this really a wafer sized chip? Because that sounds…difficult and expensive…and very difficult. A manufacturing or packaging defect is going to be very expensive when you have to throw the entire wafer away.
You could design a single control region and multiple compute areas that can be individually fused off. As long as the control region is good, then you can bin the parts according to available compute clusters.Bigger die size means more rejects from the wafer, or a high acceptance of defects. That's.. a huge die size. They can't be making many fully functional chips of that size. There just aren't enough lines at TSMC.
Oh, I can do you one better:ChatGPT: You have a bug in your code. You should be doing this:
They're all still using the extremely-limited space on TSMC's best node.Affordable GPUs soon? Maybe?
please?
I'm sorry, I have to type this through tears of guffawing laughter.Presumably by the time this has been reduced down to robots
It's already hereI'm sorry, I have to type this through tears of guffawing laughter.
When do you figure this will be?
Oh, I can do you one better:
Me: you're failing the unit test.
Claude: let me fix that.
Me: you just fudged the unit test.
Claude: let me fix that.
Me: you just fudged the unit test in a different way. Revert the unit test and fix the code.
Claude: let me fix that.
Me: you just fudged the unit test in yet another different way.
...
What is the cost for a similar program on this iteration compared to the predecessor?OpenAI’s new GPT‑5.3‑Codex‑Spark is 15 times faster at coding than its predecessor.
That's because you asked it to do something it hadn't done before. "Ooh, it can crank out a crypto exchange platform in Rust in 20 minutes!" "Oh, you mean like these 20 open-source ones it is cribbing off of?"The first iteration was very off. I gave it some additional comments, and it was "I see now that I was wrong. ok, let me look up how X works, I see now, let me insert X... I see now that I was wrong. ok, let me look up how X works.. I see now that I was wrong..." after 4 or 5 iterations, I stopped it and wrote it by hand.
-- IF they stop right now and don't train any new models. Yeah, it's that hilarious.If they then want the same margin as Oracle they reach a cost that is ~10x their current revenue.
Side note: tools like ReSharper have been happily doing this, and more, for about two decades now. No LLM required, this is called proper static code analysis. For old folks, it's semantic linting.It can search our codebase / make simple changes - and the "smarter auto-correct" saves some time. Copying functionality and having it auto-insert the new variables into the new debug statement (a bit like having a spreadsheet auto-adjust relatively-indexed cells when copying a formula) IS helpful.
Wafer scale tech has always excited me, one day they will be stacking them multiple cm high with internal cooling channels.
Once a LLM produces a bad output best practice is to scrap that context window or revert it to an earlier state. Model and tool providers should really work to encourage users to do this, or to partially automate the process.Oh, I can do you one better:
Me: you're failing the unit test.
Claude: let me fix that.
Me: you just fudged the unit test.
Claude: let me fix that.
Me: you just fudged the unit test in a different way. Revert the unit test and fix the code.
Claude: let me fix that.
Me: you just fudged the unit test in yet another different way.
...
No shit, this happened -- we actually tried this after Ars reported on this a while back.
In essence, you cannot let the LLM have write access to your unit tests. So you either write them yourself from scratch, or have them be generated and then have to monitor whether they cover whatever the LLM spits out. Which means you have to fully review every line from scratch anyway, and...
Wait, wasn't this bullshit supposed to make me MORE productive?
By Grabthar's Hammer.... .... ... ....
What a savings.
Reminds me of this:Oh, I can do you one better:
Me: you're failing the unit test.
Claude: let me fix that.
Me: you just fudged the unit test.
Claude: let me fix that.
Me: you just fudged the unit test in a different way. Revert the unit test and fix the code.
Claude: let me fix that.
Me: you just fudged the unit test in yet another different way.
is this really a wafer sized chip? Because that sounds…difficult and expensive…and very difficult. A manufacturing or packaging defect is going to be very expensive when you have to throw the entire wafer away.
estimates are 2-3 million per unit for the the CS-3
I award you one "understatement of the week" award. Seriously, 15 times faster! You mean 15 times faster than something like an H200, which costs about 500 times less?not at all cheap
1. LLM shits out a metric shit-ton of code for the low low price of a nice carOnce a LLM produces a bad output best practice is to scrap that context window or revert it to an earlier state. Model and tool providers should really work to encourage users to do this, or to partially automate the process.
I don’t know why you’re being downvoted. You’re right.I'd like to know what actual serious work can be done with a 128K token window. I don't have anything in our code base that would qualify -- outside of trivial services that a monkey could maintain anyway.
Hell, that kind of window barely allows for our main database schema for crying out loud.
Personally I use VSCode and Gemini 3 pro for my "i'd usually delegate this to a junior developer" prompts. I give it the right files for context and some plain english psuedo code to do a small task at a time while I'm thinking about the next step. I think most serious professionals use the models in this way. The "magic bullet, I one shotted another TODO list or Minesweeper clone" stuff is non-sense, ignore it, but as for helping accelerate your work the tooling they've implemented in VSCode is pretty mature at this point and I get good results from Gemini. Think of it as your pair programmer who is eager to do the grunt work while you figure out the big picture. We work primarily in PHP/Laravel on the backend and React-Typescript-Tailwinds on the front end and get few hallucinations and good results without needing to be super verbose in the prompting at this point. Our backend monolith is quite big but I always give it a head start by including the relevant files in my prompt.I'd like to know what actual serious work can be done with a 128K token window. I don't have anything in our code base that would qualify -- outside of trivial services that a monkey could maintain anyway.
Hell, that kind of window barely allows for our main database schema for crying out loud.
ya the last company that tried this couldn't make it work but that was like the pentium erais this really a wafer sized chip? Because that sounds…difficult and expensive…and very difficult. A manufacturing or packaging defect is going to be very expensive when you have to throw the entire wafer away.
Even Cerebras' first-gen chips had the ability to map around broken tiles. They aren't getting perfect yield, but all they need is enough tiles to be a competitive product. It's not like Nvidia is getting a lot of perfect chips with their reticle-limited GPUs, either, so being able to make viable products from less-than-perfect parts is crucial.kind of impressed cerebras actually worked, AI researchers tried to do this back decades ago but couldn't make the chips. the yield was atrocious