AI Hardware & Infrastructure
-
MegaTrain trains 100B-parameter LLMs on a single GPU — with 1.5TB of RAM
Training a 100-billion-parameter model usually means a cluster of expensive GPUs. MegaTrain flips the script: store everything in CPU memory, and treat the GPU as a temporary math worker. How It Works The core idea is dead simple. Parameters and optimizer states live in host RAM. During forward and backward passes, MegaTrain streams weights to… Continue reading
-
Google’s Gemma 4 Now Runs Inside Your Browser — Gemma Gem Needs Zero API Keys
A Chrome extension just turned your browser into an AI runtime. Gemma Gem loads Google’s Gemma 4 model entirely on-device via WebGPU. No API keys. No cloud calls. Your data never leaves your machine. The extension hit Hacker News as a Show HN post and pulled 154 points — not because running local models is… Continue reading
-
Google AI Edge Eloquent lands on iOS with zero fanfare — free offline dictation powered by Gemma
No blog post. No keynote. On April 6, Google quietly slipped an app called AI Edge Eloquent onto the App Store, and nobody noticed until TechCrunch picked it up the next day. Within 24 hours, a dozen tech outlets were covering it. Offline Dictation That Strips Your Filler Words Eloquent downloads a Gemma-based speech recognition… Continue reading
-
Freestyle gives AI coding agents a full Linux VM in 700ms — and can clone it while it’s still running
Every AI coding agent needs somewhere to run code. The problem: most sandbox solutions give you a stripped-down container with limited permissions. Fine for running a Python snippet — useless when your agent needs root access, Docker, or nested virtualization. What Freestyle Actually Does Freestyle provisions real KVM-backed Linux VMs in under 700ms. Not containers… Continue reading
-
Anthropic Hits $30B Run Rate — 3x in Four Months — and Signs 3.5GW Google/Broadcom TPU Deal
$9 billion at the end of 2025. $19 billion by March. $30 billion in April. Anthropic just tripled its annualized revenue in four months. No SaaS company in history has done this — not Slack, not Zoom, not Snowflake. Over 1,000 enterprise customers now spend $1M+ per year on Claude, double the count from February.… Continue reading
-
Parlor Runs a Full Voice + Vision AI on Your MacBook — No API Key, No Cloud, 2.6 GB Total
Six months ago, running a real-time voice AI locally required an RTX 5090. Now a 2-billion-parameter model on an M3 Pro does voice, vision, and conversation at 83 tokens/sec. That’s the entire pitch behind Parlor — and it’s more impressive than it sounds. Parlor grabbed 265 points on Show HN, landed #6 on bestofshowhn.com’s April… Continue reading
-
Parlor runs real-time voice + vision AI on a MacBook — 2.6 GB, zero cloud, zero API keys
Remember that GPT-4o voice demo? Camera on, talking naturally, AI responding in real time. Impressive — except it runs on OpenAI’s servers, costs money per minute, and every frame of your face goes to the cloud. Parlor does the same thing on an M3 Pro. Entirely local. 266 points on Hacker News in a day.… Continue reading
-
Parlor puts real-time voice and vision AI on your laptop — 2.6 GB, no cloud, no API keys
Remember when OpenAI demo’d GPT-4o voice mode and everyone lost their minds? Camera on, voice flowing, AI responding in real time. Cool — except it runs on OpenAI’s servers, costs money, and sends your data to the cloud. Parlor does the same thing on your MacBook. Entirely local. What Parlor Actually Does You open a… Continue reading
-
One GPU, Ten Developers, $10/Month Each: Inside sllm’s Shared Inference Gamble
Running open-source models on cloud GPUs costs real money. A single H100 on-demand runs $2-7/hour depending on the provider. Dedicate one to serving DeepSeek V3 (685B parameters) and you’re looking at roughly $14,000/month. Even smaller setups for 70B-class models land in the $500-2,000/month range. For individual developers and small teams experimenting with local AI, that’s… Continue reading
-
TinyGPU: George Hotz Got Apple to Sign an NVIDIA GPU Driver for Mac
Apple hasn’t allowed third-party GPU drivers on Apple Silicon. Not NVIDIA, not AMD, not anyone. When Apple dropped Intel chips in late 2020, eGPU support died with them. Six years of silence. Then on March 31, 2026, George Hotz’s Tiny Corp announced that Apple officially signed and approved their TinyGPU driver extension. NVIDIA RTX 30/40/50… Continue reading
