-
Mintlify ChromaFs Turns UNIX Commands Into the Only RAG Interface AI Agents Actually Need
Every AI documentation assistant in 2025 worked the same way: user asks a question, the system fires a vector search, returns the top-K chunks, and hopes the answer is somewhere in there. It mostly worked. Until it didn’t. Multi-page answers? Missed. Exact code syntax buried in the wrong chunk? Gone. The retrieval wasn’t broken —… Continue reading
-
Cursor 3 bets $29B that developers want to manage agents, not write code
Anysphere just ripped out the heart of Cursor and replaced it with something completely different. On April 2, Cursor 3 launched — and it’s not an IDE update. The company dropped its code-editor identity entirely and rebuilt the product as an agent orchestration platform. You no longer open Cursor to write code. You open it… Continue reading
-
700 GitHub Stars in a Week: Apfel Exposes the Free LLM Apple Locked Behind Siri
Every Mac with Apple Silicon has a large language model built into the operating system. Not downloaded. Not sideloaded. Baked in. Apple ships it as part of Apple Intelligence — a 3-billion-parameter model that runs entirely on your Neural Engine and GPU. Zero cloud. Zero cost. Zero API keys. You just can’t use it. Not… Continue reading
-
Gemma 4 Scores 89% on AIME With Just 4B Active Parameters — Google’s Open Model Bet Gets Real
Google has been playing defense in the open model race for months. Llama 4 grabbed headlines. Qwen 3.5 dominated coding benchmarks. Gemma 3, despite solid performance, kept losing enterprise deals over one thing that had nothing to do with intelligence: its license. That changed on April 2. Gemma 4 dropped with four model sizes, vision… Continue reading
-
7.7K Stars and Climbing: oh-my-codex (OMX) Turns OpenAI’s Codex CLI Into a Multi-Agent Powerhouse
OpenAI shipped Codex CLI. It’s fast, it’s free, it writes decent code. But use it for anything beyond a single-file task and you’ll hit the wall: no hooks, no coordination, no way to run multiple agents in parallel. One context window, one task, one thread. Yeachan Heo — the same Korean developer behind oh-my-claudecode, which… Continue reading
-
Microsoft MAI Models (MAI-Transcribe-1, MAI-Voice-1, MAI-Image-2) Are Live — Redmond’s AI Independence Starts Now
Five months. That’s how long it took from the formation of Microsoft’s MAI Superintelligence team to shipping three foundation models that directly compete with OpenAI, Google, and every major AI provider in the market. On April 2nd, Microsoft AI — the division led by Mustafa Suleyman — released MAI-Transcribe-1, MAI-Voice-1, and MAI-Image-2 into public preview… Continue reading
-
OpenRouter Is Raising $120M at a $1.3B Valuation — and It’s Processing More Tokens Than Most AI Companies Make
Alex Atallah has impeccable timing. In January 2022, Forbes pegged his stake in OpenSea at $2.2 billion. Six months later, he walked away from the NFT marketplace he co-founded. A few months after that, OpenSea’s daily trading volume collapsed 99% — from $2.7 billion to $9 million. The crypto crowd called him crazy for leaving.… Continue reading
-
StepFun Step 3.5 Flash Activates Only 11B of 196B Parameters — and Still Matches GPT-5.2
A Chinese AI startup just dropped a 196-billion-parameter model under Apache 2.0, and the kicker is: it only uses 11 billion of those parameters at any given moment. StepFun’s Step 3.5 Flash hit the top of Hacker News this week with a simple claim — it’s the number one cost-effective model for OpenClaw tasks, beating… Continue reading
-
Claw Code rewrote Claude Code in Rust before sunrise — and hit 50K GitHub stars in 2 hours
At 4 AM on March 31, 2026, a developer named Sigrid Jin woke up to his phone exploding with notifications. Anthropic had just accidentally shipped a 59.8 MB source map inside a routine npm update of Claude Code. Inside that file: 512,000 lines of unobfuscated TypeScript across roughly 1,900 files. The entire architecture of the… Continue reading
-
Liquid AI LFM2.5-350M: How 350 Million Parameters Trained on 28 Trillion Tokens Outrun Models Twice Its Size
There’s a number that should make every AI engineer stop and think: 80,000 to 1. That’s the token-to-parameter ratio of Liquid AI’s new LFM2.5-350M — a model with just 350 million parameters that was trained on 28 trillion tokens. For context, most models see maybe 20 to 100 tokens per parameter during training. Liquid AI… Continue reading
