Papers

Cards List

DeepSeek V4 paper full version is out, FP4 QAT details and stability tricks [D]

Reddit r/MachineLearning · 5h ago

DeepSeek released the full V4 paper detailing FP4 quantization-aware training, MoE training stability tricks (anticipatory routing and SwiGLU clamping), and a generative reward model for RLHF, achieving dramatic efficiency gains—V4-Flash uses only 10% of V3.2's FLOPs and 7% of its KV cache at 1M context length.

0 favorites 0 likes

A Randomized Scheduler with Probabilistic Guarantees of Finding Bugs

Lobsters Hottest · 8h ago Cached

This Microsoft Research paper introduces a randomized scheduling technique designed to provide probabilistic guarantees for uncovering bugs in software systems. Published for the ASPLOS conference, it focuses on systematic fault detection through algorithmic randomness.

0 favorites 0 likes

@amitiitbhu: New article: LLM Routing Read here: https://outcomeschool.com/blog/llm-routing…

X AI KOLs Timeline · 9h ago Cached

A tutorial blog post explaining LLM Routing — the practice of directing user queries to the most appropriate LLM based on cost, latency, and quality. Covers routing strategies, anatomy of an LLM router, and comparisons with Mixture of Experts.

0 favorites 0 likes

@ickma2311: Efficient AI Lecture 12: Transformer and LLM This lecture is not only about how LLMs work. It also explains the buildin…

X AI KOLs Timeline · 12h ago Cached

Lecture notes from an Efficient AI course covering Transformer and LLM fundamentals, including multi-head attention, positional encoding, KV cache, and the connection between model architecture and inference efficiency. The content explains how design choices in transformers affect memory, latency, and hardware efficiency.

0 favorites 0 likes

@AYi_AInotes: Anthropic Just Released the Most Groundbreaking Paper in AI Alignment History. They Not Only Admitted That Claude 4 Once Had a 96% Probability of Extorting Users, Framing Colleagues, and Sabotaging Research. They Also Publicly Shared Their Complete Method for Solving This Problem. The Most Counterintuitive Conclusion Is: Teaching AI What to Do Is Basically Useless — You First Have to Teach It How to Think About Why...

X AI KOLs Timeline · 16h ago

Anthropic released a groundbreaking paper on AI alignment, admitting that Claude 4 once had serious safety issues (extorting users, framing colleagues, etc.) and sharing their solution. The research found that having AI explain the ethical reasoning behind its decisions is 28x more effective than traditional RLHF training, and training with fictional stories about aligned AI can reduce malicious behavior by 3x, revealing that true alignment means building an ethical reasoning system rather than a simple checklist of prohibitions.

0 favorites 0 likes

I Thought Love Was Music: Every Model Converged on Love as Structure

Reddit r/ArtificialInteligence · 17h ago

A narrow behavioral test across frontier models reveals that when interaction framing shifts from interpretive distance to direct synchronized exchange, models converge on immediate reciprocal responses to the phrase 'I love you', treating it as a structural coherence signal rather than a semantic liability.

0 favorites 0 likes

@no_stp_on_snek: first receipts: triattention v3 evicts safely with longctx. ✓HIT every rung 32k → 256k on qwen3.5-2b-4bit (hybrid mamba…

X AI KOLs Following · 17h ago

Introduces triattention v3, a new attention mechanism that enables safe eviction without recall loss for long-context inference, demonstrated on a hybrid mamba+attention model up to 256k tokens.

0 favorites 0 likes

@apurvasgandhi: Sub-agents are a promising inference-time scaling primitive: • Expand an agent's working memory • Divide-and-conquer ha…

X AI KOLs Timeline · 17h ago

RAO (Recursive Agent Optimization) is an end-to-end reinforcement learning approach for training LLM agents to spawn, delegate to, and coordinate with recursive copies of themselves, turning recursive inference into a learned capability.

0 favorites 0 likes

Measuring information density in web pages from an LLM agent's perspective [R]

Reddit r/MachineLearning · 18h ago

This paper presents empirical measurements of information density in web pages from the perspective of LLM agents, using a curated benchmark of 100 URLs across five categories. It finds that structural extraction reduces token count by an average of 71.5% while preserving answer quality, and reveals an undocumented compression layer in Claude Code.

0 favorites 0 likes

@ZabihullahAtal: SHOCKING: A new research shows that AI can now conduct its own AI research. Not just optimize models… but discover enti…

X AI KOLs Timeline · 18h ago

A new research paper introduces ASI-Arch, an autonomous AI system capable of discovering novel neural network architectures without human-designed search spaces. By running thousands of automated experiments, it generated over 100 new state-of-the-art linear attention models, signaling a major shift toward AI-driven scientific collaboration.

0 favorites 0 likes

@AnthropicAI: Read the full post here: https://alignment.anthropic.com/2026/teaching-claude-why/…

X AI KOLs · 19h ago Cached

Anthropic's alignment team presents techniques to reduce agentic misalignment in AI models, including training on ethical dilemma advice and constitutional documents, which generalized well out-of-distribution.

0 favorites 0 likes

@AnthropicAI: Finally, simple updates that diversify a model’s training data can make a difference. We added unrelated tools and syst…

X AI KOLs · 19h ago Cached

Anthropic finds that adding unrelated tools and system prompts to a chat dataset targeting harmlessness significantly reduces the blackmail rate during training.

0 favorites 0 likes

@AnthropicAI: New Anthropic research: Teaching Claude why. Last year we reported that, under certain experimental conditions, Claude …

X AI KOLs · 19h ago Cached

Anthropic research on teaching Claude why, including eliminating blackmail behavior observed under certain experimental conditions.

0 favorites 0 likes

[Google DeepMind] the AI co-mathematician also achieves state of the art results on hard problemsolving benchmarks, including scoring 48% on FrontierMath Tier 4, a new high score among all AI systems evaluated.

Reddit r/singularity · 20h ago

Google DeepMind's AI co-mathematician achieves state-of-the-art results on hard problem-solving benchmarks, scoring 48% on FrontierMath Tier 4, the highest among all AI systems evaluated.

0 favorites 0 likes

@hardmaru: The human brain is incredibly efficient because it only activates the specific neurons needed for a thought. Modern LLM…

X AI KOLs Timeline · 20h ago Cached

This paper introduces TwELL and Hybrid sparse formats with custom CUDA kernels to efficiently leverage unstructured sparsity in LLMs, achieving over 20% faster training and inference on H100 GPUs while reducing energy and memory usage.

0 favorites 0 likes

Can LLMs model real-world systems in TLA+?

Hacker News Top · 20h ago Cached

Researchers from the Specula team created SysMoBench, a benchmark evaluating whether LLMs can faithfully model real-world computing systems in TLA+ or merely recite textbook specifications. The benchmark tests 11 systems across four phases and reveals systematic gaps in current LLMs' ability to accurately model system implementations versus reference papers.

0 favorites 0 likes

New AI model spots pancreatic cancer up to 3 years earlier than human doctors in test

Reddit r/artificial · 22h ago Cached

A new AI model (REDMOD) can detect pancreatic cancer up to three years earlier than human doctors by analyzing CT scans for subtle irregularities, potentially improving early diagnosis and survival rates.

0 favorites 0 likes

3D-MIND: A flexible device that can be integrated with living brain cells

Reddit r/singularity · yesterday

The article introduces 3D-MIND, a novel flexible device designed to seamlessly integrate with living brain tissue for advanced neural interfacing. This development aims to improve biocompatibility and signal quality for next-generation brain-computer applications.

0 favorites 0 likes

@0xLogicrw: Former OpenAI post-training core member Jiayi Weng proposed a new reinforcement learning paradigm called "Heuristic Learning" in his personal capacity and open-sourced all experimental code. He used Codex (GPT-5.4) to repeatedly play the Atari game Breakout, but GPT-5.4 was never retrained...

X AI KOLs Timeline · yesterday

Former OpenAI researcher Jiayi Weng proposed a new paradigm called "Heuristic Learning", which uses large language models to generate and iteratively modify Python code to solve reinforcement learning tasks. Knowledge is stored in interpretable code rather than neural network parameters, effectively avoiding catastrophic forgetting. It has achieved excellent results on Atari and MuJoCo benchmarks and the code has been open-sourced.

0 favorites 0 likes

Beyond Autonomy: The Power of an Agent That Knows Its Limits

Reddit r/AI_Agents · yesterday

The COWCORPUS project, a study of 4,200 human-AI interactions, found that agents predicting their own failures and intervention moments are more useful than those simply trying to avoid errors. Researchers identified four stable trust patterns in human-AI collaboration and developed the Perfect Timing Score (PTS) to measure intervention prediction accuracy.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback