Google researchers created a flash flood forecasting model by sorting through 5 million news articles, providing a geo-tagged time series baseline for urban areas worldwide.
Why it matters
Google's innovative use of news articles to develop a flash flood forecasting model highlights the growing role of AI in addressing critical global challenges.
Community talk
[P] Weight Norm Clipping Accelerates Grokking 18-66× | Zero Failures Across 300 Seeds | PDF in Repo
LLM-based OCR is significantly outperforming traditional ML-based OCR, especially for downstream LLM tasks
LLMs forget instructions the same way ADHD brains do. The research on why is fascinating.
LLMs forget instructions the same way ADHD brains do. I built scaffolding for both. Research + open source.
I spent a weekend doing layer surgery on 6 different model architectures. There's a "danger zone" at 50% depth that kills every one of them.
What happens when you make AI agents debate unsolved math problems and verify every output
Agents given the choice between natural language and structured queries abandoned NL within minutes
[R] A Gradient Descent Misalignment — Causes Normalisation To Emerge
[R] From Garbage to Gold: A Formal Proof that GIGO Fails for High-Dimensional Data with Latent Structure — with a Connection to Benign Overfitting Prerequisites
I made my agent 34.2% more accurate by letting it self-improve. Here’s how.
Built an autonomous system where 5 AI models argue about geopolitical crisis outcomes: Here's what I learned about model behavior
LLM Thematic Generalization Benchmark V2: models see 3 examples, 3 misleading anti-examples, and 8 candidates with exactly 1 true match, but the underlying theme is never stated. The challenge is to infer the specific hidden rule from those clues rather than fall for a broader, easier pattern.
Attention is all you need: Kimi replaces residual connections with attention
WiFi-DensePose: AI Can Track Body Positions Through Walls
Has increasing the number of experts used in MoE models ever meaningfully helped?
Humanoid Robots can now play tennis with a hit rate of ~90% just with 5h of motion training data
[D] ran controlled experiments on meta's COCONUT and found the "latent reasoning" is mostly just good training. the recycled hidden states actually hurt generalization
I thought Gemini was supposed to be the long context king?
Built an AI memory system based on cognitive science instead of vector databases
Those of you who use LLMs have probably seen this: sometimes they code like a senior engineer, and other times they seem to forget even basic syntax. Research suggests that this is not hallucination.
Professional Gaslighter
What happens when you give AI agents a 2-line bio and let them live together for 30 days?
I engineered a prompt architecture for ethical decision-making — binary constraint before weighted analysis