machine-learning

Tag

Cards List
#machine-learning

Aurora: A Leverage-Aware Optimizer for Rectangular Matrices

Lobsters Hottest · 6h ago Cached

Tilde Research introduces Aurora, a new optimizer designed to prevent neuron death in MLP layers while maintaining orthogonality, achieving state-of-the-art results on nanoGPT benchmarks and 100x data efficiency on 1B models.

0 favorites 0 likes
#machine-learning

@tom_doerr: Visualizes machine learning algorithms from first principles https://github.com/gavinkhung/machine-learning-visualized…

X AI KOLs Timeline · 13h ago Cached

This article introduces Machine Learning Visualized, a Jupyter Book and interactive platform that implements and derives machine learning algorithms from first principles with visualizations.

0 favorites 0 likes
#machine-learning

@ConsciousRide: 90% of AI System Design interviews in 2026 are just these 11 concepts repeated:

X AI KOLs Timeline · 18h ago

The article claims that 90% of AI system design interviews in 2026 revolve around just 11 repeated concepts.

0 favorites 0 likes
#machine-learning

@ghumare64: I loved this spec, so re-designed https://aiengineeringfromscratch.com - very beautiful indeed.

X AI KOLs Timeline · 19h ago Cached

A user shares their redesign of the 'AI Engineering from Scratch' website, which serves as a reference manual explaining AI concepts like transformers and backpropagation from raw mathematical implementations.

0 favorites 0 likes
#machine-learning

@DimitrisPapail: The co-inventor of Looped Transformers defended her PhD thesis yesterday and is heading to an incredible new role soon …

X AI KOLs Timeline · yesterday Cached

Angeliki Giannou, co-inventor of Looped Transformers, has successfully defended her PhD thesis and is set to begin a new role. Congratulations were shared by Dimitris Papailiopoulos on social media.

0 favorites 0 likes
#machine-learning

Formalizing statistical learning theory in Lean 4 [R]

Reddit r/MachineLearning · yesterday Cached

FormalSLT is a Lean 4 library that formally proves finite-sample statistical learning theory results (ERM, VC bounds, Rademacher bounds, PAC-Bayes, etc.) with explicit assumptions and zero sorry statements, providing a machine-checked foundation for ML theory.

0 favorites 0 likes
#machine-learning

Scientists identified over 10,000 new exoplanet candidates using AI

Reddit r/singularity · yesterday Cached

Scientists used a machine learning algorithm to analyze TESS data, identifying over 10,000 new exoplanet candidates, potentially tripling the known count. One candidate was confirmed as a hot Jupiter, validating the method.

0 favorites 0 likes
#machine-learning

@mathemetica: At just 18, Ewin Tang (now at UC Berkeley) developed a groundbreaking classical algorithm for recommendation systems (t…

X AI KOLs Following · yesterday

Ewin Tang developed a groundbreaking classical algorithm for recommendation systems that matched quantum performance, challenging quantum advantage assumptions. She was awarded the 2025 Maryam Mirzakhani New Frontiers Prize for her contributions to bridging classical and quantum computing.

0 favorites 0 likes
#machine-learning

@rwayne: Yesterday an interesting paper dropped on arXiv that directly translates the 'consciousness' mechanism from cognitive science into long-context engineering.

X AI KOLs Timeline · yesterday

Researchers propose applying the "global ignition" consciousness mechanism from cognitive science to long-context engineering, introducing the MiA-Signature method that uses submodular selection of high-level concepts to cover the activation space. Applied to RAG and agentic systems, it delivers consistent performance improvements across multiple long-context tasks.

0 favorites 0 likes
#machine-learning

@tom_doerr: Structured roadmaps for AI, ML, and LLM learning https://github.com/bishwaghimire/ai-learning-roadmaps…

X AI KOLs Timeline · 2d ago Cached

A comprehensive, open-source GitHub repository providing structured learning roadmaps and curated resources for mastering AI, machine learning, deep learning, and large language models from beginner to advanced levels. Designed for students and professionals, it covers foundational concepts, programming frameworks, career tracks, and emerging AI topics.

0 favorites 0 likes
#machine-learning

@cyrilXBT: Andrew Ng just taught the entire mathematical foundation of machine learning in one lecture. Free. Stanford University …

X AI KOLs Timeline · 2d ago

Andrew Ng shares his Stanford CS229 lecture covering core machine learning mathematics, including locally weighted regression, maximum likelihood, logistic regression, and Newton's method, providing developers with a comprehensive guide to ML fundamentals.

0 favorites 0 likes
#machine-learning

A new generation of AI models and one of the most powerful research papers out there.

Reddit r/LocalLLaMA · 2d ago

Token AI releases a research paper introducing STAM, a new adaptive momentum optimizer designed to improve training stability and reduce memory usage compared to standard optimizers like AdamW.

0 favorites 0 likes
#machine-learning

SDFlow: Similarity-Driven Flow Matching for Time Series Generation

arXiv cs.AI · 2d ago Cached

This paper introduces SDFlow, a similarity-driven flow matching framework for time series generation that addresses exposure bias in autoregressive models. It achieves state-of-the-art performance and inference speedups by operating in the frozen VQ latent space with low-rank manifold decomposition.

0 favorites 0 likes
#machine-learning

Locality-aware Private Class Identification for Domain Adaptation with Extreme Label Shift

arXiv cs.AI · 2d ago Cached

This paper proposes a locality-aware private class identification approach and a reliable optimal transport-based method (ReOT) to address domain adaptation challenges under extreme label shift, particularly distinguishing shared from private classes.

0 favorites 0 likes
#machine-learning

Understanding Annotator Safety Policy with Interpretability

arXiv cs.AI · 2d ago Cached

This paper introduces Annotator Policy Models (APMs) by Apple, which use interpretability techniques to infer annotators' internal safety policies from their labeling behavior without requiring additional annotation effort. The authors demonstrate that APMs can accurately model these policies and distinguish between sources of annotation disagreement, such as operational failures, policy ambiguity, and value pluralism.

0 favorites 0 likes
#machine-learning

Adversarial Graph Neural Network Benchmarks: Towards Practical and Fair Evaluation

arXiv cs.LG · 2d ago Cached

This paper presents a comprehensive benchmark for evaluating adversarial attacks and defenses in Graph Neural Networks, highlighting the need for standardized and fair experimental protocols.

0 favorites 0 likes
#machine-learning

MOSAIC: Module Discovery via Sparse Additive Identifiable Causal Learning for Scientific Time Series

arXiv cs.LG · 2d ago Cached

This paper introduces MOSAIC, a method for module discovery in scientific time series that combines causal representation learning with sparse additive identifiable causal learning. It aims to recover interpretable latent variables and their associated observations without post-hoc alignment, validated on domains like molecular dynamics and climate data.

0 favorites 0 likes
#machine-learning

Non-Myopic Active Feature Acquisition via Pathwise Policy Gradients

arXiv cs.LG · 2d ago Cached

This paper introduces NM-PPG, a non-myopic active feature acquisition method using pathwise policy gradients to optimize sequential feature selection in costly prediction scenarios.

0 favorites 0 likes
#machine-learning

TIDE: Every Layer Knows the Token Beneath the Context

arXiv cs.CL · 2d ago Cached

This paper introduces TIDE, a method that addresses the Rare Token and Contextual Collapse problems in LLMs by injecting token identity into every layer via Embedding Memory. The authors demonstrate theoretical and empirical improvements across language modeling and downstream tasks.

0 favorites 0 likes
#machine-learning

On Semantic Loss Fine-Tuning Approach for Preventing Model Collapse in Causal Reasoning

arXiv cs.LG · 2d ago Cached

This paper identifies a critical 'model collapse' issue in standard fine-tuning for causal reasoning and proposes a semantic loss function with graph-based logical constraints to prevent it.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback