Top
Best
New

Posted by bilsbie 6/30/2025

There are no new ideas in AI, only new datasets(blog.jxmo.io)
490 points | 289 commentspage 2
piinbinary 6/30/2025|
AI training is currently a process of making the AI remember the dataset. It doesn't involve the AI thinking about the dataset and drawing (and remembering) conclusions.

It can probably remember more facts about a topic than a PhD in that topic, but the PhD will be better at thinking about that topic.

jayd16 6/30/2025||
Its a bit more complex than that. Its more about baking out the dataset into heuristics that a machine can use to match a satisfying result to an input. Sometimes these heuristics are surprising to a human and can solve a problem in a novel way.

"Thinking" is too broad a term to apply usefully but I would say its pretty clear we are not close to AGI.

tantalor 6/30/2025|||
Maybe that's why PhDs keep the textbooks they use at hand, so they don't have to remember everything.

Why should the model need to memorize facts we already have written down somewhere?

nkrisc 6/30/2025||
> It can probably remember more facts about a topic than a PhD in that topic

So can a notebook.

cadamsdotcom 6/30/2025||
What about actively obtained data - models seeking data, rather than being fed. Human babies put things in their mouths, they try to stand and fall over. They “do stuff” to learn what works. Right now we’re just telling models what works.

What about simulation: models can make 3D objects so why not give them a physics simulator? We have amazing high fidelity (and low cost!) game engines that would be a great building block.

What about rumination: behind every Cursor rule for example, is a whole story of why a user added it. Why not take the rule, ask a reasoning model to hypothesize about why that rule was created, and add that rumination (along with the rule) to the training data. Providing opportunities to reflect on the choices made by their users might deepen any insights, squeezing more juice out of the data.

Centigonal 6/30/2025||
Simulation and embodied AI (putting the AI in a robotic arm or a car so it can try stuff and gather information about the results) are very actively being explored.
cadamsdotcom 6/30/2025||
What about at inference time? ie. in response to a query.

We let models write code and run it. Which gives them a high chance of getting arithmetic right.

Solving the “crossing the river” problem by letting the model create and run a simulation would give a pretty high chance of getting it right.

Centigonal 7/1/2025||
The newest Claude update comes with a python sandbox built right into the API for exactly this reason.

https://docs.anthropic.com/en/docs/agents-and-tools/tool-use...

kevmo314 6/30/2025||
That would be reinforcement learning. The juice is quite hard to squeeze.
cadamsdotcom 6/30/2025||
Agreed for most cases.

Each Cursor rule is a byproduct of tons of work and probably contains lots that can be unpacked. Any research on that?

kevmo314 7/1/2025||
Yeah, at a very high level it's similar to an actor-critic reinforcement learning algorithm. The rule text is a value function and one could build a critic model that takes as input the rule text and the main model's (the actor's) output to produce a reward.

This is easier said than done though because this value function is so noisy it's often hard to learn from it. And also whether or not a response (the model output) matches the value function (the Cursor rules) is not even that easy to grade. It's been easier to train the chain-of-thought style reasoning since one can directly score it via the length of thinking.

This new paper covers some of the difficulties of language-based critic models: https://openreview.net/pdf?id=0tXmtd0vZG

Generally speaking, the algorithm and approach is not new. Being able to do it in a reasonable amount of compute is the new part.

cadamsdotcom 7/1/2025||
Suggestion was even simpler - feed a reasoning model a prompt like “tell me a few reasons a user might’ve created this Cursor rule: {RULE_TEXT}”

Do that for a bunch of rules scraped from a bunch of repos - and you’ve got yourself a dataset for training a new model with - or maybe for fine tuning.

kevmo314 7/1/2025||
Yeah, go for it.
bladecd 7/12/2025||
The only real important thing in AI is data, not infrastructure, not fancy methods.
somebodythere 6/30/2025||
I don't know if it matters. Even if the best we can do is get really good at interpolating between solutions to cognitive tasks on the data manifold, the only economically useful human labor left asymptotes toward frontier work; work that only a single-digit percentage of people can actually perform.
seydor 6/30/2025||
There are new ideas, people are finding new ways to build vision models, which then are applied to language models and vice versa (like diffusion).

The original idea of connectionism is that neural networks can represent any function, which is the fundamental mathematical fact. So we should be optimistic, neural nets will be able to do anything. Which neural nets? So far people have stumbled on a few productive architectures, but it appears to be more alchemy than science. There is no reason why we should think there won't be both new ideas and new data. Biology did it, humans will do it too.

> we’re engaged in a decentralized globalized exercise of Science, where findings are shared openly

Maybe the findings are shared, if they make the Company look good. But the methods are not anymore

JKCalhoun 7/1/2025||
> It’s not crazy to argue that all the underlying mechanisms of these breakthroughs existed in the 1990s, if not before.

That's not super relevant in my mind. It's because they're showing fruit now that will allow research to move forward. And the success, as we know, draws a lot of eyeballs, dollars, resources.

If this path was going to hit a wall, we will hit it more quickly now. If another way is required to move forward, we are more likely to find it now.

sakex 6/30/2025||
There are new things being tested and yielding results monthly in modelling. We've deviated quite a bit from the original multi head attention.
ahmedhawas123 7/1/2025||
I think this is reflective of current state, but does not mean this will be the future. I think there is a lot of innovation to come on revisiting some of the 1990s principles of back propagation and optimization. Imagine if you could train current models to optimal weights in 1 day or 1 hour instead of weeks/months?

Just a hypothesis of mine

Leon_25 7/1/2025|
At Axon, we see the same pattern: data quality and diversity make a bigger difference than architecture tweaks. Whether it's AI for logistics or enterprise automation, real progress comes when we unlock new, structured datasets, not when we chase “smarter” models on stale inputs.
More comments...