How Sakana AI’s new technique solves the problems of long-context LLM tasks

RePo, Sakana AI’s new technique, solves the "needle in a haystack" problem by allowing LLMs to organize their own memory.

AI is writing your code, but who’s reviewing it?

As AI coding assistants go mainstream, a silent wave of technical debt is building. Here’s how the industry is fighting back.

Machine learning in space: Building intelligent systems for the harshest environments

A new breed of tiny, hyper-efficient AI is revolutionizing space, extending satellite life and unlocking the next great era of autonomous exploration.

Decoding the brain, inspiring AI: How Rahul Biswas is bridging neuroscience...

The convergence of AI and neuroscience opens exciting possibilities for understanding human cognition and driving innovation in deep learning.

How test-time training allows models to ‘learn’ long documents instead of...

By treating language modeling as a continual learning problem, the TTT-E2E architecture achieves the accuracy of full-attention Transformers on 128k context tasks while matching the speed of linear models.

VL-JEPA is a lean, fast vision-language model that rivals the giants

Meta’s VL-JEPA outperforms massive vision-language models on world modeling tasks by learning to predict "thought vectors" instead of text tokens.

The evolution of LLM tool-use from API calls to agentic applications

A look at the evolution of LLM tool-use, from supervised fine-tuning to Reinforcement Learning (RLVR) and agentic applications in large and specialized models.

URM shows how small, recurrent models can outperform big LLMs in...

The key to solving complex reasoning isn't stacking more transformer layers, but refining the "thought process" through efficient recurrent loops.

The hidden architecture behind AI systems that don’t break under growth

Most systems break at 100x growth. Real scalability depends on architecture, data quality, and organizational design, not just writing better code.

A few interesting observations on Gemini 3 Flash

Google didn’t reveal a lot of information about its Gemini 3 Flash model. So we had to speculate a lot on what is going on under the hood.

Applied ML: When ‘perfect’ becomes the enemy of ‘good’

Waiting for perfect data can stall your machine learning project and result in losing opportunities of creating good-enough models.

AI can’t replace software engineers yet, but here is how to...

AI is not a replacement for engineers, but it can be very useful to product managers testing hypotheses and product ideas.

How to turbocharge your product and market research with DeepSearch

If you think in terms of the JBTD framework, Deep Search products can save you a ton of time and effort in finding new product and market opportunities.