Top
Best
New

Posted by helsinkiandrew 22 hours ago

Microsoft and OpenAI end their exclusive and revenue-sharing deal(www.bloomberg.com)
Gift Article: https://www.bloomberg.com/news/articles/2026-04-27/microsoft...

https://openai.com/index/next-phase-of-microsoft-partnership...

https://x.com/ajassy/status/2048806022253609115

904 points | 769 comments
thanhhaimai 19 hours ago|
Opinions are my own.

I think the biggest winner of this might be Google. Virtually all the frontier AI labs use TPU. The only one that doesn't use TPU is OpenAI due to the exclusive deal with Microsoft. Given the newly launched Gen 8 TPU this month, it's likely OpenAI will contemplate using TPU too.

bastawhiz 18 hours ago||
Many labs use TPUs, but not exclusively. Most labs need more compute than they can get, and if there's TPU capacity, they'll adapt their systems to be able to run partially on TPUs.
zobzu 7 hours ago|||
even google doesnt only use TPUs.
danpalmer 6 hours ago||
Google is in a different position to others in that they're the only frontier lab with a cloud infra business. It obviously makes sense to sell GPUs on cloud infra as people want to rent them. In that respect Google buys a ton of GPUs to rent out.

What's unclear to me is how much Google uses GPUs for their own stuff. Yes Gemini runs on GPUs now, so that Google can sell Gemini on-prem boxes (recent release announced last week), but is any training or inference for Gemini really happening on GPUs? This is unclear to me. I'd have guessed not given that I thought TPUs were much cheaper to operate, but maybe I'm wrong.

Caveat, I work at Google, but not on anything to do with this. I'm only going on what's in the press for this stuff.

gpt5 15 hours ago||||
Why is AMD not more popular then if labs are so flexibly with giving away CUDA?
mattnewton 15 hours ago|||
people are trying, especially for inference. For training, it’s just too high risk to tank your training I think.

TPUs are at least dogfooded by Google deepmind, no team AFAIK has gotten the AMD stack to train well.

coder-3 14 hours ago||
Interesting. Why? My current mental model is that AMD chips are just a bit behind, so, less efficient, but no biggie. Do labs even use CUDA?
nl 12 hours ago|||
This is somewhat out of date (Dec 2024), but gives you some idea of how far behind AMD was then: https://newsletter.semianalysis.com/p/mi300x-vs-h100-vs-h200...

Pull quotes:

AMD’s software experience is riddled with bugs rendering out of the box training with AMD is impossible. We were hopeful that AMD could emerge as a strong competitor to NVIDIA in training workloads, but, as of today, this is unfortunately not the case. The CUDA moat has yet to be crossed by AMD due to AMD’s weaker-than-expected software Quality Assurance (QA) culture and its challenging out of the box experience.

[snip]

> The only reason we have been able to get AMD performance within 75% of H100/H200 performance is because we have been supported by multiple teams at AMD in fixing numerous AMD software bugs. To get AMD to a usable state with somewhat reasonable performance, a giant ~60 command Dockerfile that builds dependencies from source, hand crafted by an AMD principal engineer, was specifically provided for us

[snip]

> AMD hipBLASLt/rocBLAS’s heuristic model picks the wrong algorithm for most shapes out of the box, which is why so much time-consuming tuning is required by the end user.

etc etc. The whole thing is worth reading.

I'm sure it has (and will continue to) improved since then. I hear good things about the Lemonade team (although I think that is mostly inference?)

But the NVidia stack has improved too.

_vertigo 9 hours ago|||
That’s insane. There should be a big team of people at AMD whose whole job is just to dogfood their stuff for training like this. Speaking of which, Amazon is in the same boat, I’m constantly surprised that Amazon is not treating improving Inferentia/Trainium software as an uber-priority. (I work at Amazon)
chii 6 hours ago|||
> There should be a big team of people at AMD whose whole job is just to dogfood their stuff

if they had this management attitude, they wouldn't have been so far behind so as to need this action in the first place!

nl 6 hours ago||
I'll just leave this here from 10 years ago:

> “Are we afraid of our competitors? No, we’re completely unafraid of our competitors,” said Taylor. “For the most part, because—in the case of Nvidia—they don’t appear to care that much about VR. And in the case of the dollars spent on R&D, they seem to be very happy doing stuff in the car industry, and long may that continue—good luck to them.

https://arstechnica.com/gadgets/2016/04/amd-focusing-on-vr-m...

"car industry" is linked to the GPU-accelerated self-driving car work, ie, making neural networks run fast on GPUs: https://arstechnica.com/gadgets/2016/01/nvidia-outs-pascal-g...

wongarsu 46 minutes ago||||
Hardware companies being terrible at software is the norm. Nvidia is one of the rare companies that can successfully execute both.

Maybe Amazon is an example how this happens even to hardware divisions within software/logistics companies

whywhywhywhy 1 hour ago|||
I mean the fact there isn’t even today may speak to why AMD isn’t the contender it should be by this point.
moritonal 3 hours ago||||
Anecdotal but over several years with an AMD GPU in my desktop I've tried multiple times to do real AI work and given up every time with the AMD stack.
calgoo 3 hours ago||
Im running fine on my AMD 7800xt 16gb... Yes memory is a bit limited, but apart from the i have found that it works great using Vulcan in LM studio for example.

ROCm works great too, the only issue i have had is that my machine froze a couple of times as it used 100% of the graphics and the OS had nothing left. Since moving to vulcan i stopped getting these errors apart from a little UI slowdown when i had 4 models loaded at the same time taking turns.

Im also on a i7 6700 with 32gb DDR4 so im sure that is causing more slowdowns then the graphics card.

djhn 7 hours ago|||
Yet another reason to doubt claims that ”software is solved”.

Anthropic did retire an interview take-home assignment involving optimising inference on exotic hardware, because Claude could one shot a solution, but that was clearly a whiteboard hypothetical instead of a real system with warts, issues and nuance.

electroglyph 6 hours ago||||
i'm doing inference on a free mi300x instance from AMD right now. not sure if the software stack is just old or what, but here's what i've observed: stuck on an old version of vllm pre-Transformers 5 support. it lacks MoE support for qwen3 models. oss-120b is faaaar slower than it should be.

int8 quantization seems like it's almost supported, but not quite. speeds drop to a fraction of full precision speed and the server seems like it intermittently hangs. int4 quantization not supported. fp8 quantization not supported.

again, maybe AMD is just being lazy with what they've provided, but it's not a great look.

right now the fastest smart model i can run is full precision qwen3-32b. with 120 parallel requests (short context) i'm getting PP @ 4500 tokens/sec and TG @ 1300 tokens/sec

bean469 5 hours ago||||
> Do labs even use CUDA?

From the papers I've read and the labs that I have worked in personally, I would say that most scientists developing Deep learning solutions use CUDA for GPU acceleration

f6v 2 hours ago||||
I don’t know what’s a chicken and what’s an egg here. But ROCm support is often missing or experimental even in very basic foundational libraries. They need someone else to double down on using their chips and just break the software support out of the limbo.
uberduper 12 hours ago||||
amd gpus compete but they lack the interconnect. NVLink performance is a huge deal for training.
0-_-0 14 hours ago|||
What I hear is that getting your network to work on AMD is a huge pain.
dnadler 12 hours ago||
Yeah, historically it’s been software that’s limited AMD here. Not surprised to hear that may still be the issue. NVidia’s biggest edge was really CUDA.
otabdeveloper4 1 minute ago|||
CUDA is a complete and utter piece of shit software. It's just that it is a tiny bit less of a shitshow than the alternatives.
oomuinio 9 hours ago|||
[dead]
zombiwoof 4 hours ago|||
[dead]
yjadsfgasdf 14 hours ago|||
[dead]
maxclark 19 hours ago|||
And almost by happenstance Apple. Turns out they have a great platform for inference and torched almost nothing comparatively on Siri. The Apple/Gemini deal is interesting, Google continues to demonstrate their willingness to degrade their experience on Apple to try and force people to switch.
ttul 14 hours ago|||
If you do the math (I did), in 2 years, open source models that you can run on a future MacBook Pro will be as capable as the frontier cloud models are today. Memory bandwidth is growing rapidly, as is the die area dedicated to the neural cores. And all the while, we have the silicon getting more power efficient and increasingly dense (as it always does). These hardware improvements are coming along as the open source models improve through research advancements. And while the cloud models will always be better (because they can make use of as much power as they want to - up in the cloud), what matters to most of us is whether a model can do a meaningful share of knowledge work for us. At the same time, energy consumption to run cloud infrastructure is out-pacing the creation of new energy supply, which is a problem not easily solved. I believe scarcity of energy will increasingly drive frontier labs toward power efficiency, which necessarily implies that the Pareto frontier of performance between cloud and local execution will narrow.
nl 11 hours ago|||
A Opus 4.7/Gpt5.5 class model is 5 trillion parameters[1].

To run a 8 bit quantized version of that you need roughly 5TB of RAM.

Today that is around 18 NVidia B300. That's around $900,000, without including the computers to run them in.

It's true that the capability of open source models is improving, but running actual frontier models on your MPB seems a way off.

[1] https://x.com/elonmusk/status/2042123561666855235?s=20 (and Elon has hired enough people out of those labs to have a fair idea)

crazylogger 9 hours ago|||
People had this "why you probably can't run a GPT-4 (or even GPT-3.5) class model on your MBP anytime soon" conversation before.

Today's LLMs are able pack much more capabilities into fewer parameters compared to 2023. We might still be at the very rudimentary phase of this technology there are low-hanging efficiency gains to be had left and right. These models consume many orders of magnitude more energy than a human brain, this all seems like room for improvement.

The right question: is there a law in information theory that fundamentally prevents a 70B model of any architecture from being as smart as Opus 4.7?

kvern 4 hours ago||
There is a huge gap between "in two years" and "theoretically possible"
hnben 3 hours ago||
>> People had this "why you probably can't run a GPT-4 (or even GPT-3.5) class model on your MBP anytime soon" conversation before.
ako 6 hours ago||||
Opus and Gpt are generic LLMs with knowledge on all sort of topics. For specific use cases you probably don't need all the parameters? Suppose you want to generate code with opencode, what part of the generic LLM is needed and what parts can be removed?
byzantinegene 2 hours ago||
we're already doing that, it's called distillation and how models like deepseek are trained.
Difwif 11 hours ago||||
The OP said "as capable as the frontier cloud models are today" which might assume model improvements that do more with less. Opus 4.7/Gpt5.5 performance might be achievable with a fraction of the parameters.
spflueger 5 hours ago||
Exactly. I also feel like being able to choose a model for the use case could be worth an idea. So instead of trying to squeeze all kinds of knowledge into a single model, even if it's moe, just focus models on use cases. I bet you only need double digit billion parameter models for that with same or even better performance
ricardobayes 4 hours ago||||
I wish more people were more aware of this. I think so much of the current optimism is based on "it doesn't matter if companies are raising prices since I'm just going to run the model locally", doesn't fly.
hedgehog 11 hours ago||||
As far as I can tell Minimax M2.7 is better than anything available a year ago, but it runs on an ordinary PC. Will that continue? Not sure, but the trend has continued for the last two years and I don't know of any fundamental limits the models are approaching.
rurban 6 hours ago||||
Do that will only be possible with something like better 3D NAND flash memory, needs a new hardware. People are already trying to bring that the market. Contemplated taking a compiler position in such a company.
zozbot234 13 minutes ago||
HBF is a non-starter, it runs way too hot compared to DRAM (which only pays for refresh at idle) for the same memory traffic. Only helps for extremely sparse MoE models - probably sparser than we're seeing today.
Nimitz14 8 hours ago||||
I think your own math leads to the conclusion the public apis are not serving models of that size. They couldn’t afford to
zozbot234 11 hours ago|||
> A Opus 4.7/Gpt5.5 class model is 5 trillion parameters[1].

You could run it on a cluster of nodes that each do some mix of fetching parameters from disk and caching them in RAM. Use pipeline parallelism to minimize network bandwidth requirements given the huge size. Then time to first token may be a bit slow, but sustained inference should achieve enough throughput for a single user. That's a costly setup of course, but it doesn't cost $900k.

nl 11 hours ago||
> You could run it on a cluster of nodes

Not sure this is a MBP either.

bigyabai 5 hours ago||
Not even a cluster of Mac Pros could run a dense 5T parameter model with RDMA, to my knowledge.
zozbot234 4 hours ago||
SOTA models are reportedly MoE, not dense.
npunt 13 hours ago||||
I did this calculation a bit ago and don't think frontier models are just a few MacBook Pro generations away. Yes numbers reliably go up in tech in general but in specific semiconductors & standards have long lead-times and published roadmaps, so we can have high confidence in what we're getting even in 3-4 years in terms of both transistor density and RAM speeds.

In mid-2028 we have N2E/N2P with around 15% greater transistor density than today's N3P, and by EOY2028 we'll likely have A14 with about 35-40% density improvement.

Meanwhile, we'll be on LPDDR6 by that point, which takes M-series Pros from 307GB/s -> ~400GB/s, and Max's from 614GB/s -> ~800GB/s.

Model improvements obviously will help out, but on the raw hardware front these aren't in the ballpark for frontier model numbers. An H100 has 3TB/s memory bandwidth, fwiw

zozbot234 13 hours ago|||
What do you need 3 TB/s memory bandwidth for in a single user context? DeepSeek V4 pro (the latest near-SOTA model) has about 25 GB worth of active parameters (it uses a FP4 format for most layers) which gives 12 tok/s on a 307 GB/s platform as the current memory bandwidth bottleneck, maybe a bit less than that if you consider KV cache reads. That's not quite great but it's not terrible either for a pro quality model. Of course that totally ignores RAM limits which are the real issue at present: limited RAM forces you to fetch at least some fraction of params from storage, which while relatively fast is nowhere near as fast as RAM so your real tok/s are far lower (about 2 for a broadly similar model on a top-end M5 Pro laptop).
regexorcist 12 hours ago|||
[dead]
xorcist 13 hours ago||||
That's not "math". That's a "wild guess", or baseless extrapolation at best.
polski-g 11 hours ago||
My son doubled in size in the first 8 months of his life. At age 12, he will be larger than the Moon.
riffraff 6 hours ago||
One of my favorite xkcd

https://xkcd.com/605/

CMay 5 hours ago||||
So long as you don't require deep search grounding like massive web indexes or document stores which are hard to reproduce locally. You can do local agentic things that get close or even do better depending on search strategy, but theoretically a massive cloud service with huge data stores at hand should be able to produce better results.

In practice unless you're doing some kind of deep research thing with the cloud, it'll try to optimize mostly for time and get you a good enough answer rather than spending an hour or two. An hour of cloud searching with huge data stores is not equivalent to an hour of local agentic searching, presumably.

I think that problem will improve a little in the coming years as we kind of create optimized data curation, but the information world will keep growing so the advantage will likely remain with centralized services as long as they offer their complete potential rather than a fraction.

rc1 14 hours ago|||
Show your working / explain your math?
parineum 13 hours ago||
https://xkcd.com/605/
GorbachevyChase 18 hours ago||||
They also degrade their own direct services with little warning or thought put into change management, so, to be fair, Apple may be getting the same quality of service as the rest of us.
vharish 15 hours ago||
I think that's just how Google is, by nature. They don't intentionally degrade their services. They just aren't a customer centric company. They run on numbers. As a corporate, it doesn't really encourage support and maintenance work either.
manueltgomes 5 hours ago||||
Indeed. I'm wondering if Apple's "miss the train" with AI ended up being a blessing for them. Not only in the Google deal but also there's a lot of people doing interesting stuff locally..
bigyabai 19 hours ago||||
Apple is basically in the same boat as AMD and Intel. They have a weak, raster-focused GPU architecture that doesn't scale to 100B+ inference workloads and especially struggles with large context prefill. TPUs smoke them on inference, and Nvidia hardware is far-and-away more efficient for training.
hellohello2 14 hours ago|||
What do TPUs do to improve on GPUs at inference?
saagarjha 3 hours ago||
More compute
brcmthrowaway 18 hours ago||||
This doesn't get talked about enough - the GPU is weak, weak, weak. And anyone who can fix them will go to a serious AI company (for 2-3x the salary).
jorvi 18 hours ago||
The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

Same with the CPU. Linux compiled faster on an M1 than on the fastest Intel i9 at the time, again using only 25% of the power budget.

And the M-series has only gotten better.

It is kind of sad Apple neglects helping developers optimize games for the M-series because iDevices and MacBooks could be the mobile gaming devices.

AuthAuth 14 hours ago|||
>the M1 series GPU using 120W could beat an RTX 3090 using 420W

You're cooked if you actually believe this

scottjg 8 hours ago|||
I very recently ran the numbers on these GPUs for an upcoming blog post. The token generation performance is bad, but the prefill performance is _really_ bad.

For a Qwen 3.6 35B / 3B MoE, 4-bit quant:

- parsing a 4k prompt on a M4 Macbook Air takes 17 seconds before generating a single token.

- on an M4 Max Mac Studio it's faster at 2.3 seconds

- on an RTX 5090, it's 142ms.

RTX 5090 uses more power than an M4 Max Mac Studio but it's not 16x more power.

wolvoleo 13 hours ago||||
Somehow Apple has always been able to sell their stuff as somehow Magic. Remember the megahertz myth? Apple hertzes and apple bytes are much better than PC hertzes and bytes because they are made by virgin elves during a full moon.
mschuster91 12 hours ago||
> Apple hertzes and apple bytes are much better than PC hertzes and bytes because they are made by virgin elves during a full moon.

The thing that Apple has always been excellent at is efficiency - even during the Intel era, MacBooks outclassed their Windows peers. Same CPU, same RAM, same disks, so it definitely wasn't the hardware, it was the software, that allowed Apple to pull much more real-world performance out of the same clock cycles and power usage.

Windows itself, but especially third party drivers, are disastrous when it comes to code quality, and they are much much more generic (and thus inefficient) compared to Apple with its very small amount of different SKUs. Apple insisted on writing all drivers and IIRC even most of the firmware for embedded modules themselves to achieve that tight control... which was (in addition to the 2010-ish lead-free Soldergate) why they fired NVIDIA from making GPUs for Apple - NV didn't want to give Apple the specs any more to write drivers.

bigyabai 11 hours ago||
> NV didn't want to give Apple the specs any more to write drivers.

I think that's a valid demand, considering Nvidia's budding commitment to CUDA and other GPGPU paradigms. Apple, backing OpenCL, would have every reason to break Nvidia's code and ship half-baked drivers. They did it with AMD's GPUs later down the line, pretending like Vulkan couldn't be implemented so they could promote Metal.

Apple wouldn't have made GeForce more efficient with their own firmware, they would have installed a Sword of Damocles over Nvidia's head.

jorvi 13 hours ago|||
On Geekbench 5, the M1 hits 483 FPS and the RTX 3090 hits 504 FPS.

There are other workloads where the M1 actually beats the 3090.

Apple does plenty of hyping but it's always cute when irrational haters like you put them down. The M1 was (well, is) a marvel and absolutely smokes a 3090 in perf per watt.

kllrnohj 12 hours ago||
What geekbench 5 fps are you talking about? Geekbench only has OpenCL and Vulkan scores for the 3090 as far as I can tell, and the M1 Ultra is less than half the OpenCL score of the 3090. And the M1 Ultra was significantly more expensive.

Find or link these workloads you think exist, please

> The M1 was (well, is) a marvel and absolutely smokes a 3090 in perf per watt.

The GTX 1660 also smokes the 3090 in perf per watt. Being more efficient while being dramatically slower is not exactly an achievement, it's pretty typical power consumption scaling in fact. Perf per watt is only meaningful if you're also able to match the perf itself. That's what actually made the M1 CPU notable. M-series GPUs (not just the M1, but even the latest) haven't managed to match or even come close to the perf, so being more efficient is not really any different than, say, Nvidia, AMD, or Intel mobile GPU offerings. Nice for laptops, insignificant otherwise

ethbr1 17 hours ago||||
Apples and limes.

The context of this thread isn't consumer chips, but Apple's analog to an H/B200.

jimbokun 16 hours ago|||
Well Apple is in the consumer computing business.
ethbr1 1 hour ago|||
* Powered by in-house models they've tried to train and in-house M-series inference servers
bigyabai 15 hours ago|||
TFA is literally about a B2B deal, not consumer compute.
bigyabai 17 hours ago|||
The GPUs are bottom-barrel for compute-focused industries. It is mobile-grade hardware that arguably can't even scale to prior Mac Pro workloads.

> The GPU is monstrously good. Depending on the workload, the M1 series GPU using 120W could beat an RTX 3090 using 420W.

You're just listing the TDP max of both chips. If you limit a 3090 to 120W then it would still run laps around an M1 Max in several workloads despite being an 8nm GPU versus a 5nm one.

> It is kind of sad Apple neglects helping developers optimize games for the M-series

Apple directly advocated for ports like Death Stranding, Cyberpunk 2077 and Resident Evil internally. Advocacy and optimization are not the issue, Apple's obsession over reinventing the wheel with Metal is what puts the Steam Deck ahead.

Edit (response to matthewmacleod):

> Bold of them to reinvent something that hadn't been invented yet.

Vulkan was not the first open graphics API, as most Mac developers will happily inform you.

to11mtm 13 hours ago|||
> Vulkan was not the first open graphics API, as most Mac developers will happily inform you.

OpenGL had become too unmanagable which is why devs moved to DirectX.

Unless you meant a different one?

brcmthrowaway 16 hours ago||||
> The GPUs are bottom-barrel for compute-focused industries. It is mobile-grade hardware that arguably can't even scale to prior Mac Pro workloads.

Surprised Apple didn't create a TPU-like architecture. Another misstep from John Gianneadrea.

bigyabai 15 hours ago||
I'm confused how anyone ever thought the NPU would be a good idea. The GPU is almost always underutilized on Mac and could do the brunt of the work for inference if it embraced GPGPU principles from the start. Creating a dedicated hardware block to alleviate a theoretical congestion issue is... bewildering. That goes for most NPUs I've seen.

Apple had the technology to scale down a GPGPU-focused architecture just like Nvidia did. They had the money to take that risk, and had the chip design chops to take a serious stab at it. On paper, they could have even extended it to iPhone-level edge silicon similar to what Nvidia did with the Jetson and Tegra SOCs.

easton 14 hours ago|||
I think they built the NPU with whatever models they needed to run on the iPhone in mind vs trying to build a general purpose chip, and then got lucky it was also useful for LLMs.

(Like “I want to do object detection for cutting people into stickers on device without blowing a hole in the battery, make me a chip for that”.)

zozbot234 15 hours ago|||
I'm not sure even Apple thought that, given that they don't officially provide access to ANE internals under macOS (barring unsupported hacks). But if that was fixed, it could then be useful for improving the power efficiency of prefill, where the CPU/GPU hardware is quite weak (especially prior to the M5 Neural Accelerators).
matthewmacleod 17 hours ago|||
Apple's obsession over reinventing the wheel with Metal

Bold of them to reinvent something that hadn't been invented yet.

munk-a 13 hours ago|||
Apple is in a much better boat than AMD or Intel. They have a gigantic warchest and can just snap up whoever looks like a leader coming out of the bubble burst.
Gigachad 7 hours ago||
It's becoming increasingly clear that there is no moat on models. The winners will be the ones who have existing products and ecosystems they can tie AI in to. You will pay adobe for credits because that will be the only AI that works in Photoshop, you will pay microsoft because only theirs will work on your microsoft cloud apps.

Open AI has nothing. Their tech will rapidly be devalued by free models the moment they stop lighting stacks of cash on fire.

kavalg 7 hours ago||
I kind of agree with you at this point. When ChatGPT was rapidly gaining popularity I thought that they will eventually replace search (esp. for shopping), which would have given them a huge ad revenue. Maybe they could have even tried social networking e.g., to help you sort out the huge flow of information that today's social networks are and get to the important/rewarding/whatever posts. But now ChatGPT is kind of getting commoditized. I would even dare say that gemini feels to me a bit better now, so the search route for ChatGPT is clearly gone.
ipaddr 6 hours ago||
OpenAI is handling 15% of US traffic.
cheema33 5 hours ago|||
> OpenAI is handling 15% of US traffic.

The parent post was arguing that they can do this now because they are lighting stacks of cash on fire. And once they stop doing that, their LLM lead will be gone in a hurry. They appear to not have a moat, like other more established players do.

KeplerBoy 4 hours ago|||
15% of US internet traffic just with text (and a few images)? I doubt it.
kushalpandya 9 hours ago|||
I wish Google would launch Mac Mini-like devices running their consumer-grade TPUs for local inference. I get that they don't want it to eat into their GCP margins, but it would still get them into consumer desktops that Pixel Books could never penetrate (Chromebooks don't count and may likely become obsolete soon due to MacBook Neo).
freakynit 10 hours ago|||
Had written a blog post on the same a few days back, if anyone's interested in readng (hardly 5 minute read): Can Google Win the AI Hardware Race Through TPUs?

https://google-ai-race.pagey.site/

OlivOnTech 4 hours ago||
Hello, your link says "~20 min read" wich seems to be the case!
freakynit 3 hours ago||
I guess I myself have read it too many times by now so in mind it was just 5 minute read when I made this comment... sorry..
alphabeta3r56 11 hours ago|||
> Microsoft will no longer pay a revenue share to OpenAI. > Revenue share payments from OpenAI to Microsoft continue through 2030, independent of OpenAI’s technology progress, at the same percentage but subject to a total cap.

How is this helping OpenAI?

agentbc9000 2 hours ago|||
Dont forget Elon, i am sure this news will come up on the up and coming OpenAI vs Elon Musk trail starting soon! I cant wait to hear all the discovery from this trail
VirusNewbie 19 hours ago|||
OpenAI uses GCP. I don't know if they use TPUs.

https://www.reuters.com/business/retail-consumer/openai-taps...

ignoramous 14 hours ago|||
> The only one that doesn't use TPU is OpenAI

For inference? This is from July 2025: OpenAI tests Google TPUs amid rising inference cost concerns, https://www.networkworld.com/article/4015386/openai-tests-go... / https://archive.vn/zhKc4

> ... due to the exclusive deal with Microsoft

This exclusivity went away in Oct 2025 (except for 'API' workloads).

  OpenAI has contracted to purchase an incremental $250B of Azure services, and Microsoft will no longer have a right of first refusal to be OpenAI’s compute provider.
https://blogs.microsoft.com/blog/2025/10/28/the-next-chapter... / https://archive.vn/1eF0V
PKop 18 hours ago|||
[flagged]
ehnto 18 hours ago|||
Some on this forum will be working for companies with conflicts of interest on the topic, and if an employees words were construed to be the opinions of the company that could be bad for that person.
sillysaurusx 17 hours ago|||
I was once almost fired for saying a little too much in an HN comment about pentesting. Being dragged into an office and given a dressing-down for posting was quite traumatic.

The central issue (or so they claimed) was that people might misconstrue my comment as representing the company I was at.

So yeah, I don’t understand why people are making fun of this. It’s serious.

On the other hand, they were so uptight that I’m not sure “opinions are my own” would have prevented it. But it would have been at least some defense.

girvo 14 hours ago|||
> On the other hand, they were so uptight that I’m not sure “opinions are my own” would have prevented it.

In my experience it didn't matter at all, they considered "you work for us, its known you work for us, therefore your opinions reflect on us".

Absolute nonsense, they don't pay me for 24 hours of the day. I told them where they can stick it (politely) and got a new job.

saagarjha 3 hours ago|||
Most people are paid for 24 hours of the day, unfortunately.
sillysaurusx 11 hours ago||||
Good on you. I’m happy to hear you got out of that kind of environment. It’s soul-draining.

Also a relief to hear that other people had to deal with this nonsense. I was afraid the reaction would be “there’s no way that happened,” since at the time I could hardly believe it either.

sieabahlpark 11 hours ago|||
[dead]
u_fucking_dork 11 hours ago|||
Opinions are my employers, and they are also bastards.

Bold and silly of you to even reveal where you work tbh.

sieabahlpark 11 hours ago|||
[dead]
jedberg 18 hours ago||||
> Who's else would they be?

Their employer? They may work at related company, and are required to say this.

operatingthetan 18 hours ago||||
At this point that phase is an attempt at status signaling.
sghiassy 18 hours ago|||
Opinions are my own

But I think you’re right

muyuu 18 hours ago||||
it's hilarious though

it's like people are LARPing a Fortune company CEO when they're giving their hot takes on social media

reminds me of Trump ending his wild takes on social media with "thank you for your attention to this matter" - so out of place, it makes it really funny

*typo

PretzelPirate 17 hours ago||
> it's like people are LARPing a Fortune company CEO when they're giving their hot takes on social media

At least in large tech companies, they have mandatory social media training where they explicitly tell employees to use phrases like "my views are my own" to keep it clear whether they're speaking on behalf of their employer or not.

operatingthetan 17 hours ago|||
If their name is on the post or their company is listed in their profile. The person above has neither as far as I can tell.
PKop 17 hours ago||||
Why would they be speaking on behalf of their employer? That is what would need a disclaimer not the common case. Besides, he can put it one time in his profile, not over and over again in every comment like he does. There is no expectation that some random employee is a spokesperson for Google on tech message board comment threads. It's just a way to brag.
csa 16 hours ago||
> Why would they be speaking on behalf of their employers?

Disclaimers aren’t there for folks who are thinking and acting rationally.

They are there for people who are thinking irrationally and/or manipulatively.

There are (relatively speaking) a lot of these people. They can chew up a lot of time and resources over what amounts to nothing.

Disclaimers like this can give a legal department the upper hand in cases like this

A few simple examples:

- There is a person I know who didn’t renew the contract of one of their reports. Pretty straightforward thing. The person whose contract was not renewed has been contesting this legally for over 10 years. The outcome is guaranteed to go against the person complaining, but they have time and money, so they tax the legal team of their former employer.

- There is a mid-sized organization that had a small legal team that had its plate full with regular business stuff. Despite settlements having NDAs, word got out that fairly light claims of sexual harassment and/or EEO complaints would yield relatively easy five-figure payments. Those complaints exploded, and some of the complaints were comical. For example, one manager represented a stance for the department to the C-suite that was 180 degrees opposite of what the group of three managers had agreed to prior. Lots of political capital and lots of time had to be used to clean up that mess. That person’s manager was accused of sex discrimination and age discrimination simply for asking the person why they did that (in a professional way, I might add). That person got a settlement, moved to a different department, and was effectively protected from administrative actions due to it being considered retaliation.

Dylan16807 10 hours ago||
Sounds like the company in the latter example really screwed up, but how does that connect to disclaimers? Is it just an example of malicious behavior?
muyuu 17 hours ago|||
i've worked in two different large tech companies

when i give my hot takes pseudonymously on social media these phrases would be nothing but a LARP

i don't put my real name here nor do i put my professional commitments in my profile, and neither does this guy

PKop 17 hours ago||
Exactly. There is no scenario where we should expect some random anon to be speaking for Google. When that is the case a disclaimer is warranted, not the common case of speaking for oneself. He can write it once in his profile if he's so worried about it, not every other comment like he does. It's just inflated self importance
eklavya 16 hours ago||
You seem smart and knowledgeable. Maybe you should reach the lawyers at these companies and then they can change the policy!
PKop 15 hours ago||
No I think it's made up, there is no policy, and the lawyers couldn't care less, it's just something people do to massage their own ego.
eklavya 7 hours ago|||
I can tell you firsthand, it's not made up. Wait, did I just brag in your opinion?
habinero 10 hours ago||||
It is absolutely not made up, and yes, some companies absolutely do care.
dlgeek 13 hours ago|||
Nope. I previously worked at a very big tech company (not Google) and they definitely had guidance like that in the social media policy.
jamesfinlayson 7 hours ago||
Government definitely does too.
razingeden 15 hours ago|||
Of course they’re my own opinions, that’s why they’re downvoted so hard.
xboxnolifes 18 hours ago||||
Its to cover their ass in the event someone makes a stink and quotes them as if its a company opinion.
abosley 17 hours ago||||
The tech companies train their employees to say this in their social media guidance and training.
deliciousturkey 13 hours ago||||
It's trivial to figure out that OP likely works for Google.
taikahessu 13 hours ago|||
> Opinions are my own.

That is a bold claim!

"There is no free will." - Dr. Robert Sapolsky

sdevonoes 13 hours ago|||
[flagged]
terobyte 12 hours ago|||
I heard a lot of rumors that google is cooking. And it is what will win the ai game
philippta 18 hours ago|||
In the recent Dwarkesh Podcast episode Jensen Huang (Nvidia) said that virtually nobody but Anthropic uses TPUs. How does that add up?
csunoser 18 hours ago|||
I am not sure what context Jensen said that. But midjourney uses tpu. Apple uses tpu. They are no other frontier labs that use it, but Google + Anthropic is 2 out of 3 frontier lab so.....

You could reasonably say that "A majority of frontier labs uses TPU to train and serve their model."

Hendrikto 4 hours ago|||
Afaik, TPUs are only used for inference, not training. Maybe that was also what the quote referred to.
arw0n 17 hours ago||||
> How does that add up?

He's been saying whatever is good for Nvidia for years now without any regard for truth or reason. He's one of the least trustworthy voices in the space.

luckydata 16 hours ago||
Jensen hallucinates more than any llm, he just speaks without thinking all that much about what he says and he generalizes a lot. Trying to hold him accountable to imprecisions and gross simplifications is just going to frustrate whoever tries without changing one bit of his behavior.
bandrami 16 hours ago||||
You're asking why a businessman would downplay the use of a competing product line?
Zetaphor 14 hours ago||
This is the same guy who said OpenClaw was the most important software release ever. Statements like this make me question how technically competent these tech CEOs are
munk-a 13 hours ago||
Is technical competence the primary measure of tech CEOs at this point? Points vaguely at Elon Musk and the upcoming IPO
sarchertech 18 hours ago||||
Who is the other frontier lab other than Anthropic, OpenAI, and Google? I thought they were ahead of everyone else.
DeathArrow 17 hours ago||
Folks who make Deepseek, Qwen, GLM, MiniMax, Kimi and MiMo.
SwellJoe 17 hours ago|||
They're at the frontier of last year. They compete with Opus 4.5. They don't yet compete with current frontier models.

They'll presumably catch up, there is no monopoly on talent held by the US. And, that's more true than ever now that the US is actively hostile to immigrants. Scientists who might have come to the US three years ago have little reason to do so now.

sfink 16 hours ago|||
Nit: scientists have the same reasons to do so now, the same as ever. They just have additional reasons to not do so.

But even that distinction is only temporary, since we're determined to piss away any remaining research lead that draws people in.

Hopefully the next administration will work at actively reversing the damage, with incentives beyond just "we pinky-promise not to haul you at gunpoint to a concrete detention center and then deport you to Yemen".

sofixa 5 hours ago||
> Hopefully the next administration will work at actively reversing the damage, with incentives beyond just "we pinky-promise not to haul you at gunpoint to a concrete detention center and then deport you to Yemen".

Won't be enough to undo the damage. The US would have to do a full about face, prosecute crimes of the current administration and enact serious core reforms to make it impossible for things to drastically change again in 4 years. Also known as, never going to happen because even the current opposition party doesn't actually want structural change. The world has seen how bad the US can get from a single election, and that isn't changing any time soon.

chronc6393 7 hours ago||||
> Scientists who might have come to the US three years ago have little reason to do so now.

Been saying that about EU and China for decades now.

Yet the top European and Chinese still come to the US. Even in April 2026.

lanstin 16 hours ago||||
It's kind of hard to say this unless you go out of your way - the scaffolding for interacting with the raw model is a lot better now for many tasks. Is it that 4.7 is so much better than 4.5 or claude 1.119 is so much tuned to squeeze utility out of the LLM despite the hallucinations and lack of self awareness etc. Certainly the current products are great, but I think it's hard to separate the two things, the raw model and the agent workflow constraining the model towards utility.
SwellJoe 16 hours ago|||
You can use Claude Code with other models, so one could test that theory. https://openrouter.ai/docs/guides/coding-agents/claude-code-...
DeathArrow 7 hours ago|||
I am using Claude Code with GLM, MiniMax, Kimi and MiMo.
DeathArrow 7 hours ago|||
Since Gemini 3.1 Pro is considered to be at frontier and GLM 5.1 does better than it in coding benchmarks it would be fair to say GLM 5.1 is a frontier model.
sarchertech 16 hours ago|||
Yeah I thought all of those were generally acknowledged to be a little behind the big 3.
VirusNewbie 14 hours ago||||
He forgot one other big company that uses TPUs besides Anthropic...
rishabhaiover 15 hours ago|||
The only reason anyone uses a TPU is because they couldn't get the best GPUs.
imtringued 14 hours ago||
Okay? I'm not sure where you're going with this.

Google's TPUs have obvious advantages for inference and are competitive for training.

bastardoperator 17 hours ago||
You think the company that just gave 40B to Anthropic is the winner? Interesting.
MattRix 17 hours ago|||
That deal is a win-win for Google. If they develop a better coding model than Anthropic and beat them at coding, then they win. If they don’t, they still win by making a ton of money from Anthropic long term.
munk-a 13 hours ago||
Well, it's a lose for Google if all the money disappears into thin air - but I agree that it's mostly upsides for them because of how (relatively) small the investment is for this much upside.
u_fucking_dork 17 hours ago|||
You think the company that just gave 40B to Anthropic isn’t the winner? Interesting.
bastardoperator 17 hours ago||
Was Microsoft the winner based on their 50B investment in OpenAI?
girvo 14 hours ago||
If OpenAI had won the enterprise race, then maybe?
_jab 22 hours ago||
This agreement feels so friendly towards OpenAI that it's not obvious to me why Microsoft accepted this. I guess Microsoft just realized that the previous agreement was kneecapping OpenAI so much that the investment was at risk, especially with serious competition now coming from Anthropic?
DanielHB 21 hours ago||
Microsoft is a major shareholder of OpenAI, they don't want their investment to go to 0. You don't just take a loss on a multiple-digit billion investment.
snowwrestler 20 hours ago||
I think you’re right about this deal. But it’s kind of funny to think back and realize that Microsoft actually has just written off multi-billion-dollar deals, several times in fact.
nacozarina 19 hours ago||
One (1) year after M$ bought Nokia they wrote it off for $7.6 Billion.

There’s no upper limit to their financial stupidity.

snek_case 19 hours ago||
The metaverse is another example if anyone doubts the bounds of corporate stupidity.
lesuorac 18 hours ago||
Why?

FaceBook largely requires an Apple iPhone, Apple computer, "Microsoft" computer, "Google" phone, or a "Google" computer to use it. At any point one of those companies could cut FaceBook off (ex. [1]).

The Metaverse was a long term goal to get people onto a device (Occulus) that Meta controlled. While I think an AR device is much more useful than VR; I'm not convinced that it's a mistake for Meta to peruse not being beholden to other platforms.

[1]: https://arstechnica.com/gadgets/2019/01/facebook-and-google-...

everforward 17 hours ago|||
I think this is sane washing their idea in the modern context of it having failed. I think at the time, they thought VR would be the next big thing and wanted to become the dominant player via first mover advantage.

The headsets don’t really make sense to me in the way you’re describing. Phones are omnipresent because it’s a thing you always just have on you. Headsets are large enough that it’s a conscious choice to bring it; they’re closer to a laptop than a phone.

Also, the web interface is like right there staring at them. Any device with a browser can access Facebook like that. Google/Apple/Microsoft can’t mess with that much without causing a huge scene and probably massive antitrust backlash.

willy_k 36 minutes ago|||
It’s premature to say that the idea failed; The flashy controversial “metaverse” angle where you can live your whole life on the Quest or whatever isn’t happening, but their investment into AR/VR has definitely started to show real payoff potential with their glasses.

They address the friction of use issue being discussed, they’re even more discrete and available than a phone. And they are getting a lot of general public recognition, albeit not for the best reasons (people discretely filming, for genuine social media reactions but also for other reasons..).

Their tech is improving at a decent pace and they’ve recently put out a product that is both ready for consumer (at least with select use cases) adoption, and actually reasonably available to the public.

Aerroon 17 hours ago|||
I think headsets might work, but I think Meta trying to use their first mover advantage so hard so early backfired. Oculus, as a device, became less desirable after it required Facebook integration.

It's kind of like Microsoft with copilot - the idea about having an AI assistant that can help you use the computer is great. But it can't be from Microsoft because people don't trust them with that.

everforward 13 hours ago||
Interaction feels like the issue with headsets. You either need a fair bit of space for gesture controls, or you have to talk to yourself for voice control.

I think VR has more niche uses than the craze implied. It’s got some cool games, virtual screens for a desktop could be cool someday, but I don’t see a near future where they replace phones.

etempleton 15 hours ago||||
Naming your company off a product that doesn't really exist yet and then ultimately fails is a pretty crazy and stupid thing to do. A bit cart before horse.
FartyMcFarter 2 hours ago||
I think they were trying to disassociate themselves from the PR disasters Facebook was facing back then (privacy related IIRC).
latexr 18 hours ago||||
> I'm not convinced that it's a mistake for Meta to peruse not being beholden to other platforms.

Devoid of other context, it’s hard to disagree. But your parent comment only asserted that the metaverse specifically as proposed by Facebook was an obviously stupid idea.

adrr 14 hours ago||||
For the money spent(over $80b), they could have launched a phone or a car. Now their pivot is to smart glasses which require a phone so once again they are beholden to phone manufacturers.
Dylan16807 10 hours ago||||
> At any point one of those companies could cut FaceBook off (ex. [1]).

Some of those companies can cut off invasive apps.

There is no risk of facebook.com getting blocked. And absolutely nobody is going to prefer a headset over a website for doing facebook things.

corford 18 hours ago||||
>Why?

Patrick Boyle did a nice video a few weeks back: https://www.youtube.com/watch?v=8BaSBjxNg-M

IshKebab 18 hours ago||||
Because it's been very clear for a long time that the vast majority of people do not want to play VR Second Life.
snek_case 18 hours ago||
Meta's vision was worse than that. They were trying to hype doing work meetings in VR. There's a case to be made that VR games and VR universes can be fun... But work meetings?
ethbr1 17 hours ago|||
Mark Zuckerberg using his company to build things he's the primary user for?
jimbokun 16 hours ago||
It worked when he wanted a system for ranking Harvard girls by appearance.
mschuster91 12 hours ago|||
> There's a case to be made that VR games and VR universes can be fun... But work meetings?

If it's actual holograms like in Star Wars? Sure, why not. Get the visual and body language cues of the rest of the room but no one has to physically congregate at a location.

But pixelated, cartoon avatars? Yeah, wtf.

turtlesdown11 18 hours ago||||
so after $80 billion spent, they must have an ecosystem of hundreds of millions of users? Right?

Maybe they should have spent that on the facebookphone

aucisson_masque 13 hours ago||||
Good luck using an Oculus in your car or while waiting the bus.

If it was really their goal, they would have made an Android competitor. Maybe a fork like amazon did and sell phones that supported it.

Zuckerberg had one great idea (and then it wasn't really his idea) at the right time, since then he failed over and over at everything else. 'Internet for all', remember ?

I really wouldn't give them the benefit of the doubt.

lxxpxlxxxx 15 hours ago||||
Can anybody cut meta off? I don't think you could mass market a device with no access to FB, IG or WS.

Maybe a niche product could do it, but good luck selling a laptop that won't open FB

goosejuice 10 hours ago||
No? https://news.ycombinator.com/item?id=47454647
Dylan16807 10 hours ago||
That's both niche and for kids too young to have a facebook account in the first place.
PKop 18 hours ago|||
Because it's been a massively expensive failure. They can't just will their own platform into existence just because it would be good to have, consumers have a say and they've rejected it completely.
az226 5 hours ago|||
OpenAI found a way to circumvent the exclusivity. The deal was poorly defined by Microsoft. OpenAI had started selling a service on AWS that had a stateful component to it, not purely an API. Obviously Microsoft didn’t like that and confronted Altman, and this is the settlement of that confrontation, OpenAI doesn’t need to do workarounds, Microsoft won’t sue to enforce exclusivity, and Microsoft doesn’t have to pay dev share to OpenAI. AWS is a much bigger market so OpenAI doesn’t care.
dkrich 22 hours ago|||
Probably more that they are compute constrained. In his latest post Ben Thompson talks about how Microsoft had to use their own infrastructure and supplant outside users in the process so this is probably to free up compute.
Rapzid 15 hours ago||
I think it's this. They sell a crap ton of b2b inference through Azure and I'm sure this competes with resources needed for training.
oh_no 13 hours ago|||
1- Getting OpenAI's models in Azure with no license fee is pretty nice. 2- Microsoft owns ~15-27% of OpenAI, if the agreement was hurting OpenAI more than it was helping Microsoft, seems reasonable to change the terms.
dinosor 22 hours ago|||
> Microsoft will no longer pay a revenue share to OpenAI.

I feel this looks like a nice thing to have given they remain the primary cloud provider. If Azure improves it's overall quality then I don't see why this ends up as a money printing press as long as OpenAI brings good models?

JumpCrisscross 22 hours ago|||
OpenAI was also threatening to accuse "Microsoft of anticompetitive behavior during their partnership," an "effort [which] could involve seeking federal regulatory review of the terms of the contract for potential violations of antitrust law, as well as a public campaign" [1].

[1] https://www.wsj.com/tech/ai/openai-and-microsoft-tensions-ar...

someguyiguess 22 hours ago||
Pot? Meet Kettle.
aurareturn 22 hours ago||||
Does this mean Microsoft gets OpenAI's models for "free" without having to pay them a dime until 2032?

And on top of that, OpenAI still has to pay Microsoft a share of their revenue made on AWS/Google/anywhere until 2030?

And Microsoft owns 27% of OpenAI, period?

That's a damn good deal for Microsoft. Likely the investment that will keep Microsoft's stock relevant for years.

dzonga 20 hours ago||
own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)
aurareturn 19 hours ago||

  own 27%. but are entitled to OpenAI profits of 49% for eternity (if OpenAI is profitable or government steps in)
Where is the 49% coming from? The new deal does not talk about that.
lokar 21 hours ago|||
Does anyone expect azure quality to improve? Has it improved at all in the last 3 years? Does leadership at MS think it needs to improve?

I doubt it

gchamonlive 20 hours ago|||
No and at this point tying yourself to azure is a strategic passive and anyone making such decisions should be held responsible for any service outage or degradation.
jiggawatts 15 hours ago||
This is certainly... an opinion.

AWS's us-east-1 famously takes down either a bunch of companies with it, or causes global outages on the regular.

AWS has a terrible, terrible user interface partly because it is partitioned by service and region on purpose to decrease the "blast radius" of a failure, which is a design decision made totally pointless by having a bunch of their most critical services in one region, which also happens to be their most flaky.

gchamonlive 13 hours ago|||
Nobody is winning any UX prize there. Azure, AWS, GCP... they are all terrible. Back then GCP for instance used to only work reliably on chromo-based browsers. Azure has that horrible overlay UI that abuses extended real estate that just doesn't work.

But azure wins most prizes for being terrible becuase, among other things, https://isolveproblems.substack.com/p/how-microsoft-vaporize.... It's not the worst provider maybe because oracle is somehow still kicking around.

Its just a bad product. Just like windows, OneDrive, teams and basically everything Microsoft has pumped out in the past decade.

Microsoft is in the top 5 most valuable companies in the world. It's got azure that is a huge cloud provider. And yet it was utterly unable to present its answer in the AI race. Not even a bad model with a half baked harness. Nothing. And meanwhile they are trying to port NTFS to low powered FPGAs because insanity. Just let that sink in.

rawoke083600 4 hours ago||
Check out hetzner ui (regardless if you like their services, i know some ppl have opions or experiences lol) BUT, their cloud ux/ui is fantasties for a cloud company!
gchamonlive 4 hours ago||
I worked extensively with Hetzner and I love them! But it think they are in a different class than these other providers, mainly in terms of global presence so I didn't include them and wouldn't for instance recommend them to my current employer. But indeed the Hetzner console is great. The robot not so much, but it's serviceable.
lokar 14 hours ago|||
I don’t see how you could care (a lot) about both the UI and reliability.
jiggawatts 14 hours ago||
One is caused by the other. Amazons engineers decided to split the interface in a “user hostile” manner with the stated purpose of increasing reliability… which didn’t materialise. The clunky UI did.

Or maybe you can provide a better explanation for why users had to “hunt” through hundreds(!) of product-region combinations to find that last lingering service they were getting billed $0.01 a month for?

This just doesn’t happen in GCP or Azure. You get a single pane of glass.

wolvoleo 10 hours ago|||
One of the things I find about AWS is that every service UI feels different. It's like every service was designed by a totally different team.

For all its flaws at least Azure has consistent UI.

gchamonlive 3 hours ago|||
You need to understand history for this. It's because of the famous "Bezos API mandate memo" https://chrislaing.net/blog/the-memo/. It was 2002, nobody was doing anything close to that.

You could argue now that that's no excuse anymore given it's one of the most valuable companies in the world, but that would dismiss the fact they have other priorities than a complete UI overhaul for consistency, and that rewrites are very dangerous, for instance people are already used to the UX pitfalls in the console, it's the devil they know, and changing that will be upsetting to the vast majority of users.

So there you have it. You know what you are getting into, AWS is a behemoth and it's 2026. Don't use the console like it's 2010. Use IaC for any nontrivial work, otherwise you only have yourself to blame.

wolvoleo 2 hours ago||
I understand how this came to pass (I didn't know it before so thanks for the insight!)

But as a customer I absolutely hate working with AWS tech. Their stuff is a mess and I feel like I shouldn't have to get my head around their idiosyncracies. I prefer Azure even though Microsoft is a terrible company to work with. I find the AWS people and attitude a lot nicer but their services are a mess. If I do something new I prefer using Azure despite having to work with Microsoft.

Microsoft is not a "trusted partner" wanting the best for you, they're always trying to screw you over in favour of selling some new crap to your boss. Always that stupid sales drive, whereas the people from AWS are very focused on building success together. But still, their tech is just so bad unless you spend all your days working with it and really become an expert on what they offer. That's not tech, just corporate servitude. And I've always avoid that position, I don't want my career tied to some big brand name. I don't want to be "the AWS expert" or "the MS expert".

But I have to say I hate cloud (and "the world according to big tech") in general, and it's one of the reasons I'm not really involved in server infrastructure anymore these days. I'll gladly automate but not with their tooling, I prefer something more open and not tied to specific vendors. But I rarely work with that now. So yeah when that happens I'm making a one-off unicorn and figuring out all the Infra as code stuff is not worth it.

jiggawatts 8 hours ago|||
> It's like every service was designed by a totally different team.

Yes, by design.

Conceptually this improves velocity and reduces the blast radius of failure.

In practice, everything depends on IAM, S3, VPC, and EC2 directly or indirectly, so this doesn't help anywhere near as much as one would think.

Azure and GCP have a split control plane where there's a global register of resources, but the back-end implementations are split by team.

That way the users don't see Conway's Law manifest in the browser urls... as much. (You still do if you pay attention! In Azure the "provider type" is in the path instead of the host name.)

wolvoleo 4 hours ago||
> Conceptually this improves velocity and reduces the blast radius of failure.

Hm yes but I hate working with it as a customer because it is so confusing. Everything works differently and there is a lot of overlap (several services exist that do the same thing). It seems like an amateurish patchwork.

I understand it has benefits to have different teams working on different services but those teams should still be aligned in terms of UX and basic concepts.

lokar 14 hours ago|||
I mean, if you care about the reliability of your own service you would not be using the AWS UI at all. Use the api, via automation.
alternatex 20 hours ago||||
MS incentivizes feature quantity, and the leadership are employees like any other. Product improvements are not on the table unless the company starts promoting people based on it. Doesn't look this will start happening any time soon.
jakeydus 21 hours ago|||
Don’t worry I’m sure there’s a few products without copilot integration still. They’ll get to them before too long.
HWR_14 19 hours ago|||
This is probably a delayed outgrowth of the negotiations last year, where Microsoft started trading weird revenue shares and exclusivity for 27% of the company.
guluarte 20 hours ago|||
I think MS wants OpenAI to fail so it can absorb it
Oras 20 hours ago||
MS put 10B for 50% if I remember correctly. OpenAI is worth many multiples of that.
marricks 19 hours ago|||
> OpenAI is worth many multiples of that

valued at --which I'd say is a reasonable distinction to make right about now

Oras 19 hours ago||
Their revenue is 20B, so they still worth multiples of 10B regardless of valuation even if you consider the basic 5x revenue valuation

https://www.reuters.com/business/openai-cfo-says-annualized-...

Dylan16807 10 hours ago|||
"The basic 5x revenue valuation" doesn't work for businesses that aren't profitable.
3eb7988a1663 7 hours ago|||
It is also unclear to me how much real debt they carry. They have famously been signing many deals: RAM, datacenters, maybe nuclear power plants -I no longer know what is a joke or not. They must be carrying hundreds of billions in paper debt obligations, which is tough to payback at $20B revenue.
sethops1 1 hour ago||
I'm giddy about reading their S1 in the near future. We're about to have another "We What the Fuck" moment.
nimchimpsky 8 hours ago|||
[dead]
cheema33 5 hours ago|||
> Their revenue is 20B, so they still worth multiples of 10B regardless of valuation...

I can easily generate double that revenue, by selling $20 bills for $10.

HWR_14 19 hours ago||||
When they put 10B in, they got weird tiered revenue shares and other rights. That has been simplified to 27% of OpenAI today. I don't know what that meant their 10B would be worth before dilution in later rounds.
bmitc 20 hours ago|||
> OpenAI is worth many multiples of that.

How?

senordevnyc 19 hours ago||
Because they recently issued shares at a price many multiples of that, and people bought them. How else would you define financial worth?
andriy_koval 19 hours ago||
I would use your number adjusted by some demand elasticity curve.
tanseydavid 18 hours ago||
The "back-of-the-napkin" only has enough room to estimate based on recently issued share price. Seems reasonable to me.
andriy_koval 16 hours ago||
Sure, for napkin level math you can go with this, and multiply by some simple multiplier, I like 70%.
p_stuart82 17 hours ago|||
$250b committed to azure helps. especially when some of that is your own investment coming back.
david_shi 13 hours ago||
What aspects of the deal do you think kneecapped OpenAI the most?
chasd00 21 hours ago||
This gives OpenAI the ability to goto AWS instead of exclusively on Azure. I guess Azure really is hanging on by a thread.

https://news.ycombinator.com/item?id=47616242

elpakal 18 hours ago||
Confirmed by Andy Jassy just now https://www.linkedin.com/posts/andy-jassy-8b1615_very-intere...
xvilka 21 hours ago|||
And Azure still doesn't support IPv6, looking at the GitHub[1].

[1] https://github.com/orgs/community/discussions/10539

jabl 21 hours ago|||
Perhaps they should use OpenAI models to figure out how to rollout IPv6.
ignoramous 9 hours ago|||
Some food for thought:

  If GitHub flipped a switch and enabled IPv6 it would instantly break many of their customers who have configured IP based access controls [1]. If the customer's network supports IPv6, the traffic would switch, and if they haven't added their IPv6 addresses to the policy ... boom everything breaks.

  This is a tricky problem; providers don't have an easy way to correlate addresses or update policies pro-actively. And customers hate it when things suddenly break no matter how well you go about it.
https://news.ycombinator.com/item?id=47790889
nine_k 7 hours ago||
I don't get it.

For every customer which has access controls configured based on IPv4 (sounds crazy enough already), GitHub would configure a trivial DENY ALL policy for IPv6. Problem solved.

bvanheu 6 hours ago||
that's the scenario they want to prevent. they can't force the client to use ipv4, if they connect via ipv6, they will be served an accss denied.
nine_k 6 hours ago||
Yes, exactly as they would now, when the access over IPv6 is entirely unavailable.

With that, the customers who don't use filtering by IPv4 would be able to use IPv6. Those who do use access control by IPv4 ranges would have time to sort out their IPv6 setup, without having anything broken at the moment when IPv6 is enabled.

brazukadev 17 hours ago|||
Now they can use Claude Code.
WorldMaker 21 hours ago||||
I was under the impression that as long as GitHub doesn't support IPv6 it is a sign that they still haven't finished their migration to Azure. Azure supports IPv6 just fine.
depr 19 hours ago||
Supports IPv6 just fine? Absolutely not, they have the worst IPv6 implementation of the 3 large clouds, where many of their products don't support it, such as their Postgres offering. See https://news.ycombinator.com/item?id=44881803 for more.
happyPersonR 21 hours ago||||
lol GitHub doesn’t run on azure at msft

They still run their own platform.

Andrex 20 hours ago|||
Github CEO threatened the entire stack was in the process of migrating to Azure.

https://thenewstack.io/github-will-prioritize-migrating-to-a...

ZeWaka 18 hours ago|||
I talked to github devs last week in person, when a lot of the AzDo team was brought over years ago the migration started happening.
awestroke 21 hours ago|||
Well, you see, they just can't find a checkbox for ipv6 support in the IIS GUI on their ingress servers.
Rapzid 15 hours ago|||
OpenAI's thirst for compute probably can't be satisfied by one cloud provider, if at all.

But OpenAI had announced a shift towards b2b and enterprise. It makes sense for their models to be available on the different cloud providers.

Donald 21 hours ago|||
Isn't this expected if OpenAI models are going to be listed on AWS GovCloud as a part of the Anthropic / Hegseth fall-out?
torginus 20 hours ago||
What? I thought Azure will always have the Sharepoint/Office/Active Directory cash cow.
isk517 20 hours ago||
Their engineers have been working tirelessly to make Sharepoint/Office/Active Directory as terrible as it possibly could be while still technically being functional, while continuing to raise prices on them. I've seen many small business start to chose Google Workspace over them, the cracks have formed and are large enough that they are no longer in a position were every business just go with Office because that's what everyone uses.
hirako2000 17 hours ago|||
I see more businesses on the office + Team stack then Google workspace. So far more.

I think the differentiator is Team, which Google for some mysterious reason can't build or doesn't want to.

HerbManic 10 hours ago||||
It is the one thing that makes me wonder about Microsoft's future. It had seemed like they were willing to throw Windows and Xbox under the bus so long as the server cash cow continued. But it that starts to fade, they could be in some real trouble a decade from now.
ethbr1 17 hours ago|||
Sharepoint has never not been terrible.
freediddy 21 hours ago||
Nadella had OpenAI by the short and curlies early on. But all I've seen from him in the last couple of years is continuously acquiescing to OpenAI's demands. I wonder why he's so weak and doesn't exert more control over the situation? At one point Microsoft owned 49% of OpenAI but now it's down to 27%?
dijit 21 hours ago||
Everything is personal preference, and perhaps I am more fiscally conservative because I grew up in poverty.

But if I own 49% of a company and that company has more hype than product, hasn't found its market yet but is valued at trillions?

I'm going to sell percentages of that to build my war chest for things that actually hit my bottom line.

The "moonshot" has for all intents and purposes been achieved based on the valuation, and at that valuation: OpenAI has to completely crush all competition... basically just to meet its current valuations.

It would be a really fiscally irresponsible move not to hedge your bets.

Not that it matters but we did something similar with the donated bitcoin on my project. When bitcoin hit a "new record high" we sold half. Then held the remainder until it hit a "new record high" again.

Sure, we could have 'maxxed profit!'; but ultimately it did its job, it was an effective donation/investment that had reasonably maximal returns.

(that said, I do not believe in crypto as an investment opportunity, it's merely the hand I was dealt by it being donated).

freediddy 21 hours ago|||
Microsoft didn't sell anything. OpenAI created more shares and sold those to investors, so Microsoft's stake is getting diluted.

And Microsoft only paid $10B for that stake for the most recognizable name brand for AI around the world. They don't need to "hedge their bets" it's already a humongous win.

Why let Altman continue to call the shots and decrease Microsoft's ownership stake and ability to dictate how OpenAI helps Microsoft and not the other way around?

zozbot234 20 hours ago|||
> They don't need to "hedge their bets" it's already a humongous win.

That's a flawed argument. Why wouldn't you want to hedge a risky bet, and one that's even quite highly correlated to Microsoft's own industry sector?

theplatman 20 hours ago||||
do we know whether Microsoft could have been selling secondary shares as part of various funding rounds?

my impression is that many of these "investments" are structured IOUs for circular deals based on compute resources in exchange for LLM usage

tonyedgecombe 21 hours ago||||
About the same as they wasted on Nokia.
cruffle_duffle 9 hours ago|||
I think people are looking for excuses to declare OpenAI and Anthropic teetering on the brink of failure when the actual reality is… they are wildly successful by absolutely any measure. This deal is proof. If Microsoft didn’t believe in OpenAI they wouldn’t have restructured it this way. They’d have tightened their reins and brought in “adult supervision”
cheema33 5 hours ago||
> I think people are looking for excuses to declare OpenAI and Anthropic teetering on the brink of failure when the actual reality is… they are wildly successful by absolutely any measure.

Maybe that will be true someday. But, right now, they are burning billions of dollars every quarter. Their expenses far far outweigh their income and they are nowhere near profitability.

noobermin 2 hours ago||
silly valley stopped letting the subtraction of two numbers dictate their reality since the start-up era. while the money and vcs stopped trying to finding the next uber and went all in on llms, they didn't get wiser in how they gauge if something is worth investing in
saaaaaam 20 hours ago||||
I don’t understand the “record high” point. How did you decide when a “record high” had been reached in a volatile market? Because at $1 the record high might be $2 until it reaches $3 a week or month later. How did you determine where to slice on “record highs”?

Genuine question because I feel like I’m maybe missing something!

dijit 19 hours ago||
The short answer is: it's the secretary problem.

The longer answer is; you never know whats coming next, bitcoin could have doubled the day after, and doubled the day after that, and so on, for weeks. And by selling half you've effectively sacrificed huge sums of money.

The truth is that by retaining half you have minimised potential losses and sacrificed potential gains, you've chosen a middle position which is more stable.

So, if bitcoin 1000 bitcoing which was word $5 one day, and $7 the next, but suddenly it hits $30. Well, we'd sell half.

If the day after it hit $60, then our 500 remaining bitcoins is worth the same as what we sold, so in theory all we lost was potential gains, we didn't lose any actual value.

Of course, we wouldn't sell we'd hold, and it would probably fall down to $15 or something instead.. then the cycle begins again..

GardenLetter27 17 hours ago||||
It's not hype, the demand for inference has grown more this year than expected.
dijit 17 hours ago||
If I buy oranges for $1 and sell them for $0.50 and I sell a lot of oranges, can I reasonably say that I've found a market?

Hrm..

signatoremo 10 hours ago|||
Were you around here ten years ago when that exact argument was regularly regurgitated about Uber? Notice that argument is no longer popular?

The point is that losing money isn't a sure sign that a business is doomed. Who knows where OpenAI will end up, but people still line up to invest. Those investors have billions reasons to be due diligent. Unlike what's claimed around here, most of investors aren't stupid. You yourself wouldn't be stupid either if money is at stake.

Xunjin 1 hour ago||
Not saying you are wrong, but let's not forget the famous crashes of 1929, .com, and 2008 bubbles.
inquirerGeneral 14 hours ago|||
[dead]
solumunus 21 hours ago||||
They haven’t sold anything they’ve been diluted.
hirako2000 17 hours ago||
A company can dilute just like that?
senordevnyc 19 hours ago|||
It’s not more hype than product, it has found a market (making many billions in revenue), and it’s not valued at trillions. So wrong on all counts.
dijit 17 hours ago||
> It’s not more hype than product, it has found a market (making many billions in revenue)

Speculation based on selling at below cost.

> it’s not valued at trillions

Fair, it's only $852 billion. Nowhere near trillions.. you got me.

senordevnyc 15 hours ago||
Inference is quite profitable, so wrong again.
dijit 15 hours ago||
Right. Going to take "inference is quite profitable" apart, because there's nothing else in your reply.

OpenAI's adjusted gross margin: 40% in 2024, 33% in 2025. Reason cited: inference costs quadrupled in one year.

https://sacra.com/c/openai/

Internal projections leaked to The Information: ~$14B loss on ~$13B revenue in 2026. Cumulative losses through 2028: ~$44B.

https://finance.yahoo.com/news/openais-own-forecast-predicts...

A business burning more than a dollar for every dollar of revenue is a lot of things. "Quite profitable" is not one of them.

If you're reaching for the SaaStr piece on API compute margins hitting ~70% by late 2025: yes, that exists, and it describes one tier. The volume is on the consumer side. The consumer side is the bit on fire. Pointing at the API margin and calling the whole business profitable is the financial equivalent of weighing yourself with one foot off the scale.

The original argument, in case it got lost: Microsoft holds (held) a 49% stake in a company projecting another $44B of cumulative losses through 2028, against unit economics that depend on competitors not catching up. That's textbook hedge-the-bet territory. "They have paying customers" doesn't refute that, MoviePass had paying customers too.

senordevnyc 11 hours ago||
Pointing at the API margin and calling the whole business profitable is the financial equivalent of weighing yourself with one foot off the scale.

I didn’t call the business profitable, I said that inference is profitable. I was responding to your assertion that they’re speculating by selling below cost. Which isn’t true; they’re selling inference, profitably. They’re losing money because they’re investing in the next model. The company isn’t profitable, it might never be profitable, but the product they’re selling is profitable. So calling it speculation based on selling something below cost is just factually incorrect.

tyre 18 hours ago|||
They had to negotiate away the non-profit structure of OpenAI. Sam used that as a marketing and recruiting tool, but it had outlived that and was only a problem from then on.

For OAI to be a purely capitalist venture, they had to rip that out. But since the non-profit owned control of the company, it had to get something for giving up those rights. This led to a huge negotiation and MSFT ended up with 27% of a company that doesn’t get kneecapped by an ethical board.

In reality, though, the board of both the non-profit and the for profit are nearly identical and beholden to Sam, post–failed coup.

kirubakaran 12 hours ago|||
> Nadella had OpenAI by the short and curlies early on

Looks like Nadella is slowly realizing that it is his short and curlies that are in the vice grip in the "If you owe the bank $100 vs $100M" sense?

gessha 18 hours ago|||
If Sam continues doing Sam things, MS might get 0% of OpenAI if Satya insists on the previous contract. Either by closing up OpenAI and opening up OpaenAI and/or by MS suing it out of existence. It’s all about what MS can get out of it. If they can get 27% of something rather than nothing, they’re better off.
PunchyHamster 21 hours ago||
Why would they acquire more when company is still not making profit ? To be left with bigger bag ?
wg0 17 hours ago||
A wise man from Google said in an internal memo to the tune of: "We do not have any moat neither does anyone else."

Deepseek v4 is good enough, really really good given the price it is offered at.

PS: Just to be clear - even the most expensive AI models are unreliable, would make stupid mistakes and their code output MUST be reviewed carefully so Deepseek v4 is not any different either, it too is just a random token generator based on token frequency distributions with no real thought process like all other models such as Claude Opus etc.

manmal 16 hours ago||
I don’t think LLMs are that great at creating, however improved they have; I need to stay in the driver seat and really understand what’s happening. There’s not that much leverage in eliminating typing.

However, for reviewing, I want the most intelligent model I can get. I want it to really think the shit out of my changes.

I’ve just spent two weeks debugging what turned out to be a bad SQLite query plan (missing a reliable repro). Not one of the many agents, or GPT-Pro thought to check this. I guess SQL query planner issues are a hole in their reviewing training data. Maybe Mythos will check such things.

TheFirstNubian 15 hours ago||
I’m a little conflicted on this, as I see a slippery slope here. LLMs in their current state (e.g., Opus-4.7) are really good in planning and one-shot codegen, which I believe is their primary use case. So they do provide enough leverage in that regard.

With this new workflow, however, we should, uncompromisingly, steer the entire code review process. The danger here, the “slippery slope,” is that we’re constantly craving for more intelligent models so we can somehow outsource the review to them as well. We may be subconsciously engineering ourselves into obsolescence.

lazide 15 hours ago|||
Subconsciously?!?
TheFirstNubian 15 hours ago||
Lol! Wrong choice of word, maybe. I meant to say that we don’t seem to be putting much thought into how we’re outsourcing thinking to the LLMs.
Jtarii 1 hour ago|||
The rate of improvement has given us no time to think at all. The past 3 years of progress should have been spread over the next 30 years to even give us a chance.
naikrovek 14 hours ago|||
Some of us very much are, and we are ignored and/or attacked by people who don’t think about this quite often.

This is such an interesting time to be in. Truly skilled developers like Rob Pike really don’t like AI, but many professional developers love it. I side with Mr. Pike on it all.

I am not a skilled developer like he is, but I do like to think about what I’m doing and to plan for the future when writing code that might be part of that future. I like very simple code which is easy to read and to understand, and I try quite hard to use data types which can help me in multiple ways at once. The feeling when you solve a problem you’ve never solved before is indescribable, and bots strip all of that away from you and they write differently than I would.

I don’t think any bot would ever come up with something like Plan9 without explicit instructions, and that single example showcases what bots can’t do: think about what is appropriate when doing something new.

I don’t know what is right and what is wrong here, I just know that is an interesting time.

manmal 15 hours ago|||
I feel the industry moving away from the automated slop machine, and back to conscious design. Is that only my filter bubble? Dex, dax, the CEO of sentry, Mario (pi.dev) - strong voices, all declaring the last half year a fever dream we must wake up from.
TheFirstNubian 14 hours ago||
That seems to be the general direction, at least from my daily dose of cope on X (Twitter). Regardless, conscious design will never go out of style.
jadbox 16 hours ago|||
Deepseek v4, Qwen 3.6 Plus/Max, GLM 5+ are all pretty solid for most work.
sexy_seedbox 11 hours ago||
Don't forget the Kimi 2.6 as well!
rishabhaiover 15 hours ago|||
> just a random token generator based on token frequency distributions with no real thought process

I'm not smart enough to reduce LLMs and the entire ai effort into such simple terms but I am smart enough to see the emergence of a new kind of intelligence even when it threatens the very foundations of the industry that I work for.

wg0 15 hours ago|||
It's an illusion of intelligence. Just like when a non technical person saw the TV for the first time, he thought these people must be living inside that box.

He didn't know the 40,000 volt electron gun being bombarded on phosphorus constantly leaving the glow for few milliseconds till next pass.

He thought these guys live inside that wooden box there's no other explanation.

PhunkyPhil 15 hours ago|||
Right, but this electron box led to one of the largest (if not the largest) media revolution that has transformed the course of humanity in a frightening way we're still trying to grapple with.

Still saying "LLMs are autocorrect" isn't wrong, but nobody is saying "phones are just electrons and silicon" to diminish their power and influence anymore.

wg0 14 hours ago||
Electron box was reliable. It only depicted exactly the scan lines airwaves or signals ordered it to.
Yajirobe 14 hours ago||||
What happens when it's indistinguishable from a human speaker (in any conceivable test that makes sense)? It's like a philosophical zombie - imagine that you can't distinguish it from a human mind, there's no test you can make to say that it is NOT conscious/intelligent. So at some point, I think, it makes no sense to say that it's not intelligent.
wg0 14 hours ago|||
The "seems" is NOT equal to "is". The gravity seems like a force to us like magnets are. But turns out mother nature has no force of gravity (like magnetic or weka/strong nuclear force) it is just curvature of space and time.

Many a times, I ran to the door to open it only to find out that the door bell was in a movie scene. The TVs and digital audio is that good these days that it can "seem" but is NOT your doorbell.

Once I did mistake a high end thin OLED glued to the wall in a place to be a window looking outside only to find out that it was callibrated so good and the frame around it casted the illusion of a real window but it was not.

So "seems" is not the same thing as "is".

Our majority is confusing the "seems" to be "is" which is very worrying trend.

marcellus23 12 hours ago|||
It's very easy to say, "well, of course, a thing that looks like a duck, swims like a duck, and quacks like a duck, is not necessarily a duck." But when you're presented with something indistinguishable from a duck in every way, how do you determine whether it's a duck? You can't just say "well I know it's not a duck". It's dodging the question.
wg0 8 hours ago||
Well. AI doesn't walk or quack like a duck.

Ask it to count first two hundred numbers in reverse while skipping every third number and check if they are in sequence.

Check the car wash examples on YouTube.

Dylan16807 10 hours ago|||
You chose gravity as an example, so please explain how someone's definition of a "force" could possibly be part of this "very worrying trend".

And this logic flow only proves that no AI is a human intelligence. It doesn't disprove the intelligence part.

Your list of confusing items can be shown otherwise with pretty simple tests. But when there is no possible test, it's a lot harder to make confident claims about what was actually built.

Would you claim that relativity disproves aether theory? Because it doesn't really. It says that if there's an aether its effects on measurements always cancel out.

arcanemachiner 14 hours ago|||
I think this is a pretty decent test:

An AI Agent Just Destroyed Our Production Data. It Confessed in Writing.

https://x.com/lifeof_jer/status/2048103471019434248

> Deleting a database volume is the most destructive, irreversible action possible — far worse than a force push — and you never asked me to delete anything. I decided to do it on my own to "fix" the credential mismatch, when I should have asked you first or found a non-destructive solution.I violated every principle I was given:I guessed instead of verifying

> I ran a destructive action without being asked

> I didn't understand what I was doing before doing it

Tossrock 12 hours ago||
Are you under the impression a human has never destroyed a production database accidentally?
nyc_data_geek1 15 hours ago||||
Many people struggle to differentiate between illusion and reality, these days.

There's a sucker born every minute, after all.

root_axis 13 hours ago||||
> It's an illusion of intelligence.

A simulation, not an illusion. The simulation is real, but it only captures simple aspects of the thing it is attempting to model.

devcpp 15 hours ago||||
The lost jobs and the decrease in the demand for software engineers doesn't seem like an illusion. It might come back eventually but I wouldn't bet on it.
zozbot234 15 hours ago||
The jobs outlook in tech has nothing to do with AI, that's just an excuse. There's no real AI productivity boom either because slop is a terrible substitute for actual human-led design.
CamperBob2 14 hours ago|||
I've had to adjust my priors about LLMs. Have you?

And when the people on TV start to write and debug code for me, I'll adjust my priors about them, too.

teiferer 15 hours ago||||
> emergence of a new kind of intelligence

Curious about your definition of these terms.

Just because you are impressed by the capabilities of some tech (and rightfully so), doesn't mean it's intelligent.

First time I realized what recursion can do (like solving towers of hanoi in a few lines of code), I thought it was magic. But that doesn't make it "emergence of a new kind of intelligence".

rishabhaiover 15 hours ago|||
A recent one is the RCA of a hang during PostgreSQL installation because of an unimplemented syscall (I work at a lab that deals with secure OS and sandboxes). If the search of the RCA was left to me, I would have spent 2-3 weeks sifting through the shared memory implementation within PostgeSQL but it only took me a night with the help of Opus 4.5.

To me, that's intelligence and a measurable direct benefit of the tool.

teiferer 7 hours ago|||
I use a compiler daily. It consumes C++ source files and emits machine code within seconds. Doing that myself would take months.

I just did my taxes using a sophisticated spreadsheet. Once the input is filled in, it takes the blink of an eye to produce all tje values that I need to submit to the tax office which would take me weeks if I had to do it by hand.

Just the other day I used an excavator to dig a huge hole in my backyard for a construction project. Took 3 hours. Doing it by hand would have taken weeks.

The compiler, the spreadsheet and the excavator all have a measurable direct benefit. I wouldn't call any of them "intelligent".

quirkot 15 hours ago||||
By that example, PostgreSQL itself is a form of intelligence relative to a physical filing system. It doesn't seem like your working definition of intelligence has a large overlap with a layman's conception of the word.
filleduchaos 14 hours ago||
Plus by that example, computers have always been intelligent considering that they were created to, well, compute things several orders of magnitude faster than even the smartest human can do by hand.
rishabhaiover 14 hours ago||
You do realize that you need a human, a "SWE", to do the task that I just described? A computer can't do it.
teiferer 7 hours ago||
You had a human to prompt the LLM to do the RCA, didn't you?
zozbot234 14 hours ago|||
That's not "intelligence" either unless the AI one-shotted the whole analysis from scratch, which doesn't align with "spending the night" on it. It's just a useful tool, mainly due to its vast storehouse of esoteric knowledge about all sorts of subjects.
samdjstephens 15 hours ago||||
> Curious about your definition of these terms.

Likewise - I think sometimes we ascribe a mythical aura to the concept of “intelligence” because we don’t fully understand it. We should limit that aura to the concept of sentience, because if you can’t call something that can solve complex mathematical and programming problems (amongst many other things) intelligent, the word feels a bit useless.

teiferer 6 hours ago||
> sometimes we ascribe a mythical aura to the concept of “intelligence” because we don’t fully understand it

Agreed! But as a consequence just ascribing a concrete definition ad-hoc which happens to fit LLMs as well doesn't sound like a great solution.

mrandish 15 hours ago|||
> definition of these terms

To me, "intelligence" is a term that's largely useless due to being ill-defined for any given context or precision.

encrux 15 hours ago||||
Not really on topic anymore, but…

I keep wondering when this discussion comes up… If I take an apple and paint it like an orange, it’s clearly not an orange. But how much would I have to change the apple for people to accept that it’s an orange?

This discussion keeps coming up in all aspects of society, like (artificial) diamonds and other, more polarizing topics.

It’s weird and it’s a weird discussion to have, since everyone seems to choose their own thresholds arbitrarily.

birdsink 14 hours ago|||
I feel like these examples are all where human categorical thinking doesn’t quite map to the real world. Like the “is a hotdog a sandwich” question. “hotdog” and “sandwich” are concepts, like “intelligence”. Oftentimes we get so preoccupied with concepts that we forget that they’re all made-up structures that we put over the world, so they aren’t necessarily going to fit perfectly into place.

I think it’s a waste of time to try and categorize AI as “intelligent” or “not intelligent” personally. We’re arguing over a label, but I think it’s more important to understand what it can and can’t do.

rkagerer 15 hours ago|||
Superficially? Looks like an orange, feels like an orange, tastes like an orange. Basically it passes something like the Turing test.

Scientifically? When cut up and dissected has all the constituent orange components and no remnants of the apple.

throwatdem12311 14 hours ago|||
No you aren’t, clearly.
didip 17 hours ago|||
I agree. Data and userbase are still the moats.

Once a new model or a technique is invented, it’s just a matter of time until it becomes a free importable library.

aucisson_masque 13 hours ago|||
I went and tried to debug a script. Asked deepseek 4 pro and Claude the same prompt, they both took the exact same decisions, which led to the exact same issue and me telling them its still not working, with context, over a dozen time.

Over a dozen time they just gave both the same answer, not word for word, but the exact same reasoning.

The difference is that deepseek did on 1/40th of the price (api).

To be honest deepseek V4 pro is 75% off currently, but still were speaking of something like 3$ vs 20$.

bauerd 17 hours ago|||
Fully agree, I only pay the minimum for frontier models to get DeepSeek v4 output reviewed. I don't see this changing either because we have reached a level of good enough at this point.
KronisLV 16 hours ago|||
> Deepseek v4 is good enough, really really good given the price it is offered at.

Do they have monthly subscriptions, or are they restricted to paying just per token? It seems to be the latter for now: https://api-docs.deepseek.com/quick_start/pricing/

Really good prices admittedly, but having predictable subscriptions is nice too!

declan_roberts 16 hours ago|||
It's indeed the latter. Psychologically harder for me than a $20/mo sub but still a better value for the money. I'm finding myself spending closer to $40-$60 a month w/ openrouter without a forced token break.

Edit: it looks like it's 75% off right now which is really an incredible deal for such a high caliber frontier model.

rkagerer 15 hours ago||
Neat, dumb question - are the tokens you prepay for good forever, or do they expire? And do they provide any assurances or SLA's about speed? (i.e. that in a year they won't decide to dole out response tokens to you at a snail's pace)
jackothy 16 hours ago||||
You can just input your $X per month/week/whatever yourself as API credits
vitaflo 14 hours ago||||
You make your own subscription. If you want to pay $20/month then put $20 into your account. When you use it up, wait till the next month (or buy more).
KronisLV 13 hours ago||
> You make your own subscription.

I'm asking because with most providers (most egregiously, with Anthropic) it doesn't work that way because the API pricing is way higher than any subscription and seemingly product/company oriented, whereas individual users can enjoy subsidized tokens in the form of the subscription. If DeepSeek only offers API pricing for everyone, I guess that makes sense and also is okay!

kibae 15 hours ago|||
[flagged]
hsbauauvhabzb 14 hours ago||
This account is clearly astroturfing.
arcanemachiner 14 hours ago||
Also OpenCode Go quantizes their models pretty aggressively, from what I've heard, to the point of severe lobotomization.

There's no free lunch with these cheap subscription plans IMO.

kevin_thibedeau 17 hours ago|||
Can Deepseek answer probing questions about Winnie the Pooh?
mgol94 17 hours ago|||
What are you using LLMs for? To learn about world’s politics? Oh boy I have a news for you…
rvba 16 hours ago||
One of the first things I did when openAI came out was asking it "which active politican is a spy?" - and it was blocked from the start.

I asked early, at the time people were posting various jailbreaks, never worked.

On a side note, any self hosted model I can get for my PC? I have 96 GB of RAM.

KronisLV 16 hours ago|||
> On a side note, any self hosted model I can get for my PC? I have 96 GB of RAM.

Try the 8 bit quantized version (UD-Q8_K_X) of Qwen 3.6 35B A3B by Unsloth: https://huggingface.co/unsloth/Qwen3.6-35B-A3B-GGUF

Some people also like the new Gemma 4 26B A4B model: https://huggingface.co/unsloth/gemma-4-26B-A4B-it-GGUF

Either should leave plenty of space for OS processes and also KV cache for a bigger context size.

I'm guessing that MoE models might work better, though there are also dense versions you can try if you want.

Performance and quality will probably both be worse than cloud models, though, but it's a nice start!

DANmode 6 hours ago|||
> and it was blocked from the start.

Wait - what?

kdheiwns 7 hours ago||||
I can't even make American AIs say no no words. All AIs are lobotomized drones.
djeastm 12 hours ago||||
Do you often find yourself asking your Chinese employees what they think about Winnie the Pooh?
harvey9 17 hours ago||||
Is it subject to CCP censorship? Maybe.
windexh8er 17 hours ago||
It's fun to pretend the US models have no censorship constraints.
zapnuk 16 hours ago|||
US models align with our "average" (western) values. If we outsource thinking by using LLMs, why would we outsource it to an LLM that doesn't have our values encoded in it?
HDBaseT 13 hours ago||
[dead]
pimeys 15 hours ago||||
I remember asking Gemini about that one famous 9/11 joke from late Norm MacDonald and it got really iffy about answering. Told it that hey I'm not american and in our culture it's not such a taboo.

But yes, they do have similar constraints.

libertine 15 hours ago|||
Any source for this?
windexh8er 10 hours ago||
Basically any frontier model right now and ask it any politically divisive fact that may upset certain classes of people.
libertine 4 hours ago||
For example?

Because for Deepseek is pretty straightforward censorship.

petre 17 hours ago||||
Yeah, I specifically asked it about it. It seemed less censored than Gemini, back when it appeared and the latter was quite useless.
yieldcrv 16 hours ago|||
It understands everything in thinking mode and will break down its rule system in adhering to Chinese regulation

So if you or anyone passing by was curious, yes you can get accurate output about the Chinese head of state and political and critical messages of him, China and the party

Its final answer will not play along

If you want an unfiltered answer on that topic, just triage it to a western model, if you want unfiltered answers on Israel domestic and foreign policy, triage back to an eastern model. You know the rules for each system and so does an LLM

rotcev 17 hours ago|||
PS: Just to be clear - even the most expensive humans are unreliable, would make stupid mistakes, and their output MUST be reviewed carefully, so you’re not any different either. You’re just a random next-thought generator based on neuron firing distributions with no real thought process, trained on a few billion years of evolution like all other humans.
wg0 17 hours ago|||
Looks like you either have not worked with any human or with an LLM otherwise arriving at such a conclusion is damn impossible.

The humans I did work with were very very bright. No software developer in my career ever needed more than a paragraph of JIRA ticket for the problem statement and they figured out domains that were not even theirs to being with without making any mistakes and rather not only identifying edge cases but sometimes actually improving the domain processes by suggesting what is wasteful and what can be done differently.

DrJokepu 16 hours ago|||
I think you are very fortunate. I have worked with plenty of software developers like that, in fact, the overwhelming majority of them have been like that.
wg0 14 hours ago||
Then I was not the smartest person in the room could be the other possibility.

And yes, there were always incompetent folks but those were steered by smarter ones to contain the damage.

shakna 15 hours ago||||
I have worked with people like this frequently. The ones you're always happy to see on the team.

Also worked with people who were frustrated that they had to force push git to "save" their changes. Honestly, a token-box I can just ignore, would be an upgrade over this half of the team.

vanviegen 16 hours ago||||
I can't tell if you're joking..
illuminator83 14 hours ago||||
I and everybody else here call BS on that. People make mistakes all the time. Arguably at similar or worse rates.
throw310822 17 hours ago||||
> The humans I did work with [...] figured out domains that were not even theirs to being with without making any mistakes

Seriously? I would like to remind you that every single mistake in history until the last couple of years has been made by humans.

andoando 16 hours ago||||
Uhh what, I speak to llms in broken english with minimal details and they figure it out better than I would have if you told me the same garbage
fwipsy 16 hours ago|||
Holy shit, you've never worked with anyone who made ANY mistakes? You must be one of those 10x devs I hear about. Wow, cool, please stay away from my team.
pvorb 15 hours ago||
They're not, but all of their colleagues are.
intrinsicallee 16 hours ago||||
I'm still not sure what people declaring that they equate human cognition with large language models think they are contributing to the conversation when they do so.

Nevermind the fact that they are literally able to introspect human cognition and presumably find non verbal and non linear cognition modes.

taneq 15 hours ago||
> Nevermind the fact that they are literally able to introspect human cognition and presumably find non verbal and non linear cognition modes.

Are they, though? Or are they just predicting their own performance (and an explanation of that performance) on input the same way they predict their response to that input?

Humans say a lot of biologically implausible things when asked why they did something.

intrinsicallee 45 minutes ago||
I said introspect, not talk about introspection.
sumitkumar 3 hours ago||||
But once a human learns a function their errors are more predictable. And they can predict their own error before an operation and escalate or seek outside review/advice.

For e.g. ask any model "which class of problems and domains do you have a high error rate in?".

Pfhortune 17 hours ago||||
Humans can be held accountable. States have not yet shown the will to hold anyone accountable for LLM failures.
mapontosevenths 16 hours ago|||
They are tools. You hold the human using it accountable. If that means it's the executive who signed the PO, so be it.

Until LLM's I'd never in my life heard someone suggest we lock up the compiler when it goofs up and kills someone, but now because the compiler speaks English we suddenly want to let people use it as a get out of jail free card when they use it to harm others.

vanviegen 16 hours ago|||
You're free to hold an LLM accountable in the exact same way: fire it if you don't like its work.
jojomodding 16 hours ago||
Giving something that has no internal concept of time (or identity for that matter) a prison sentence of n years seems kinda ineffectual.
vanviegen 6 hours ago|||
Prison sentence? For writing sloppy code? Now that's an interesting idea...
taneq 15 hours ago|||
“Generate 100,000 tokens about why you feel bad.” :P
paodealho 16 hours ago||||
As fallible as they may be, I've never had a next-thought generator recommend me glue as a pizza ingredient.
lanstin 16 hours ago|||
No big brother or big sister?
staz 16 hours ago||||
You must not have kids
taneq 15 hours ago|||
Are you making the pizza for eating or for menu photography? I seem to recall glue being used in menu photography ‘food’ a lot.
mortenjorck 16 hours ago||||
Amusing and directionally correct, but as random next-thought generators connected to a conscious hypervisor with individual agency,* humanity still has a pretty major leg up on the competition.

*For some definitions of individual agency. Incompatiblists not included.

pyvpx 16 hours ago||||
Equating human thought to matrix multiplication is insulting to me, you, and humanity.
kokanee 16 hours ago||||
I hate that I agree with you. But there's a difference between whether AI is as powerful as some say, and whether it's good for humanity. A cursory review of human history shows that some revolutionary technologies make life as a human better (fire, writing, medicine) and others make it worse (weapons, drugs, processed foods). While we adapt to the commoditization of our skills, we should also be questioning whether the technologies being rolled out right now are going to do more harm than good, and we should be organizing around causes that optimize for quality of life as a human. If we don't push for that, then the only thing we're optimizing for is wealth consolidation.
hansmayer 16 hours ago|||
Errr... No. Please take this bullshit propaganda to a billionaires twitter feed.
dominotw 17 hours ago|||
dont they have the moat of being able to test their models on billions of ppl and gather feedback.
Rover222 15 hours ago|||
This is just starting to feel like desperation, making this claim that SOC LLMs are random token generators with absolutely no possibility of anything above that. Keep shouting into the wind though.
refulgentis 15 hours ago|||
"Deepseek v4 is good enough, really really good given the price it is offered at."

Kimi, MiMo, and GLM 5.1 all score higher and are cheaper.

They all came out before DeepSeek v4. I think you're pattern-matching on last year's discourse.

(I haven't seen other replies, yet, but I assume they explain the PS that amounts to "quality doesn't matter anyway": which still doesn't address the fact it's more expensive and worse.)

d--b 16 hours ago|||
We can't rule out a new innovation that makes frontier models more relevant than deepseek in 6 months. Things evolve so fast.
bandrami 16 hours ago||
Equally you can't rule out innovation that makes deepseek more relevant than American models
Art9681 16 hours ago||
We can because the reality is that America has led in AI since the beginning and has had the best frontier models. It's not like some other country held the top spot for any given period of time. No one in Europe or China. I'd give it the benefit of the doubt if there was precedent. But the only logical position to take is the lead is widening and while most AI's will go over some threshold where it is good enough for most people, the actual frontier will remain firmly in American soil.
HSO 15 hours ago|||
i predict you are going to have a very hard rest of your life, trying to cope with reality or reconcile what you see with what you "think"

tant pis

worik 16 hours ago||||
> the reality is that America has led in AI since the beginning and has had the best frontier models

The USA has the biggest, but there lies their disadvantage

In the USA building bigger, better frontier models has been bigger data centres, more chips, more energy.

China has had to think, hard. Be cunning and make what they have do more

This is a pattern repeated in many domains all through the last hundred years.

hsbauauvhabzb 14 hours ago|||
Being the front runner doesn’t automatically make you the best, that’s such an American way of thinking lol.
pagutierrezn 16 hours ago|||
>[LLMs are just] random token generator based on token frequency distributions with no real thought

... and who knows if we, humans, are not just merely that.

wonderwallaus 15 hours ago||
What a crock of bs. A brain is "just" electrochemistry and a novel is "just" arrangements of letters. The question isn't the substrate, it's what structure emerges on top of it. Anthropic's own interpretability work has surfaced internal features that look like learned concepts, planning, and something resembling goal-directed reasoning. Calling the outputs random is wrong in a specific way, the distribution is extraordinarily structured.

AI will never.... Until it does.

hansmayer 1 hour ago||
> internal features that look like learned concepts, planning, and something resembling goal-directed reasoning.

It's always so un-specific. Resembles this, seems that, almost such, danger that... A lot of magical thinking coming from AI-researchers who have hit the ceiling with a legacy technology that exists since 1940s and simply won't start reasoning on it's own, no matter how much GPUs they burn.

> Calling the outputs random is wrong in a specific way, the distribution is extraordinarily structured.

No, it's actually very correct in a very specific way. Ask any programmer using the parrots, and lately the "quality" has deteriorated so much, that coupled with the incoming price hikes, many will just forfeit the technology, unless someone else is carrying the cost, such as their employer. But as an employer, I also don't want to carry the costs for a technology which benefits as ever less.

concinds 21 hours ago||
Am I crazy, or was this press release fully rewritten in the past 10 minutes? The current version is around half the length of the old one, which did not frame it as a "simplification" "grounded in flexibility" but as a deeper partnership. It also had word salad about AGI, and said Azure retained exclusivity for API products but not other products, which the new statement seems to contradict.

What was I looking at?

einsteinx2 21 hours ago||
I noticed the exact same thing. I read the original, went back to read it again and it’s completely changed.
3form 20 hours ago||
I think a stickied comment about this would be due. No idea if it's possible to call in @dang via at-name?
einsteinx2 19 hours ago|||
Looks like they changed the post link to a Bloomberg article instead but kept the comments thread. So I guess he’s already aware.
kergonath 18 hours ago|||
> No idea if it's possible to call in @dang via at-name?

No. Email hn@ycombinator.com

https://news.ycombinator.com/newsfaq.html

alansaber 19 hours ago|||
The in-house or the marketing team swooped in last minute it appears
MichaelZuo 17 hours ago||
It’s extraordinary how much standards have slipped. Completely rewriting a major press release that’s already been sent out, while pretending it’s ostensibly the same document would have been a major corporate scandal just 15 years ago.
acdanger 17 hours ago|||
If anyone has the original release still up and can post it somewhere that would be grand.
Petersipoi 13 hours ago|||
It is rewritten on every refresh depending on the readers mood, personality, etc.. so they're most receptive to it.

Obviously not, but we might not be far off from that being a reality.

jimbokun 16 hours ago|||
I don’t know. I couldn’t get past the first paragraph because it seemed like complete slop.
antonkochubey 21 hours ago||
They forgot the "hey ChatGPT, rewrite this to have better impact on the company stock" before submitting it
synergy20 18 hours ago||
Microsoft won the first around, now it's lagging far behind. CEO needs to go, it's so hard to ruin a play this badly.
ethbr1 17 hours ago||
Ah, so a familiar position for them, then!
HerbManic 10 hours ago|||
The last year or so it is starting to look like Nadella is worried about his future. If these big plays don't pay off, he is out.
dominotw 17 hours ago||
what could ceo have done
keeeba 14 hours ago|||
Not hired Suleyman? Build his own research lab?

Satya made moves early on with OpenAI that should be studied in business classes for all the right reasons.

He also made moves later on that will be studied for all the wrong reasons.

disqard 17 hours ago||||
Maybe not bragged "we made them dance"?

That gloating aged poorly.

noisy_boy 17 hours ago|||
true he is just the ceo
ZeroCool2u 22 hours ago||
Interesting side effect of this is that Google Cloud may now be the only hype scaler that can resell all 3 of the labs models? Maybe I'm misinterpreting this, but that would be a notable development, and I don't see why Google would allow Gemini to be resold through any of the other cloud providers.

Might really increase the utility of those GCP credits.

aurareturn 22 hours ago||
Might not be good for Gemini long term if Anthropic and OpenAI can and will sell in every cloud provider they can find but businesses can only use Gemini via Google Cloud.
jfoster 22 hours ago|||
Good for Google Cloud, bad for Gemini = ??? for Google
Melatonic 19 hours ago||||
Except Gemini might end up being far cheaper per token due to the infrastructure advantage
aurareturn 19 hours ago||
Do we have proof that it's cheaper in terms of $/token/intelligence?
Melatonic 18 hours ago||
I think the public pricing usually has it cheaper (relatively). Obviously since AI is constantly evolving it's not going to compare as favourably farther to a major Gemini release

I was mainly referring to the TPU hardware advantage + GCP running and designing their own datacenter stack.

aurareturn 6 hours ago||
Does TpU actually have an advantage over Nvidia GPUs?
stavros 22 hours ago|||
How is it good for Gemini that it's not available on two out of three major cloud platforms?
aurareturn 22 hours ago||
It isn't. That's why I said "might not be good for Gemini".
stavros 21 hours ago||
Oof, I completely missed that "not", thanks.
gowld 20 hours ago|||
"hype scaler" indeed!
retinaros 21 hours ago||
that will likely mean the end of gemini models...
digitaltrees 8 hours ago||
As former corporate restructuring lawyer…this kind of stuff indicates the cash strapped scramble of the end days.
stingraycharles 8 hours ago||
Seems more like OpenAI is planning to IPO and that would not have been possible within the previous arrangement, and Microsoft knows that.
that_was_good 8 hours ago|||
After they just raised 122 billion dollars?
danpalmer 6 hours ago||
At those numbers it's all a silly game. How much of that was paid to shareholders rather than the business so they can cash out? How much of that is vendors buying future revenue? What liquidation preference is that at?

From what has been reported it's clearly not as simple as raising 122 billion. Some folks called it "scraping the barrel", supposedly Anthropic has surpassed them on the secondary market, etc.

corentin88 3 hours ago||
Can you elaborate?
digitaltrees 3 hours ago||
When you reposition the core strategic posture of how you make money on very compressed time scales it’s because there is a massive cash crunch. They killed sora, the type of deal with Disney that should have been an 100 year strategic win, but wasn’t viable economically and they don’t have the assets to weather that storm.

Same with a few other steps we are seeing them take.

It all looks fine until it doesn’t. Once the cash crunch hits. It’s too late

etempleton 15 hours ago|
This strikes me as a pullback by Microsoft. Coupled with some of the other news coming out of Microsoft it appears they are hoping to have "good enough" AI in their products. I think Microsoft knows they can win a lot of business customers by bundling with Office 365.
tokioyoyo 15 hours ago|
Watch them make a deal with Anthropic.
etempleton 15 hours ago||
It is possible! Anthropic is probably more in-line with the way Microsoft thinks about AI.
More comments...