AI coding equipment is getting better quickly. If you don’t work in the code, it can be hard to notice how much things are changing, but GPT -5 and Gemi 2.5 automatically created a whole new set of possible developer techniques and Sonnet 2.4 last week.
At the same time, other skills are progressing more slowly. If you have used AI to write emails you are probably getting the same price from it a year ago. Even if the model is better, the product is not always benefited – especially when the product is a chatbot that works at the same time a dozen different things. AI is still progressing, but it has not been evenly distributed as before.
The difference in progress seems easier than that. Coding applications are benefiting from billions of easily measured tests, which can train them to produce functional codes. This is reinforcement learning (RL), rationally the largest driver of AI progress for the past six months and is always becoming more complicated. You can learn reinforcement with human graders, but if you have a clear pass-factor metric, it works best, so that you can repeat it a few billion times without stopping the human input.
Since industries depend on the learning of reinforcement to improve products, we are automatically graded and we are seeing a real difference between what they cannot. RL-friendly skills, such as bug-fixing and competitive mathematics, are getting faster, on the other hand, writing skills simply progresses.
In short, there is a reinforcement interval – and it is becoming one of the most important reasons for what it can do and what they can’t do.
In some ways, software development is a good thing for learning reinforcement. Before AI, there was a complete sub-discipline to check how the software will hold under pressure-because the developers needed to ensure that their code would not be broken before deploying their code. So even the most elegant code will still be passed through unit testing, integration testing, security tests, etc. Human developers use these tests regularly to legalize their code and as the senior director of Google recently told me for the Dev tools, they are just as useful for legalizing the AI-exposed code. More than that, they are useful for learning reinforcement, since they are already systematicized and repeated on a huge scale.
There is no easy way to legalize a well written email or a good chatboat response; These skills are inherently subjective and hard to measure the scale. However, not every job is “easy to test” or “tough to test” does not fall neat in categories. We do not have a testing kit out of the box for quarterly financial reports or aquarial science, but a good-intent accounting startup can probably create one from scratch. Some test kit will definitely work better than the other, and some companies will be smarter about how the problem can move. However, the experimental process of the underlying process is about to decide whether the underlying process can be made as an effective product instead of just an exciting demo.
TechCrunch event
San Francisco
|
October 27-29, 2025
Some processes may be more experimental than you think. If you had asked me last week, I put the AI-planted video in the “hard-to-test” section, but the infinite progress made by the new Sora 2 model of the Openai showed that it could not be as hard as it could. In Sora 2, the objects are nowhere to come and disappear from anywhere. The faces only hold on to their shape to look like a certain person rather than a collection of features. Sora 2 footage respects the law of physics in both cases Obvious And Delicate The way. I suspect that if you lean behind the screen, you will find a powerful reinforcement learning system for each of these qualities. Put together, they make the difference between photorealism and an entertaining hallucinations.
Obviously, this is not a difficult and quick rule of artificial intelligence. This is the result of learning the central role in the development of AI, which can be easily changed as the models develop. However, the reinforcement gap will be even bigger until the RL AI products are introduced to market – both are both startups and the economy with serious impact. If a process ends on the right of the reinstatement system, startups will probably be successful to automate it – and the work that is doing now can now seek a new career. For example, the question of which healthcare services is RL-tranable has a lot of impact on the size of the economy over the next 20 years. And if we have no hints like Sora 2, we don’t have to wait too much for any answer.
