machine-learning

Tag

Cards List
#machine-learning

@ghumare64: I loved this spec, so re-designed https://aiengineeringfromscratch.com - very beautiful indeed.

X AI KOLs Timeline · 4h ago Cached

A user shares their redesign of the 'AI Engineering from Scratch' website, which serves as a reference manual explaining AI concepts like transformers and backpropagation from raw mathematical implementations.

0 favorites 0 likes
#machine-learning

@DimitrisPapail: The co-inventor of Looped Transformers defended her PhD thesis yesterday and is heading to an incredible new role soon …

X AI KOLs Timeline · 21h ago Cached

Angeliki Giannou, co-inventor of Looped Transformers, has successfully defended her PhD thesis and is set to begin a new role. Congratulations were shared by Dimitris Papailiopoulos on social media.

0 favorites 0 likes
#machine-learning

Formalizing statistical learning theory in Lean 4 [R]

Reddit r/MachineLearning · yesterday Cached

FormalSLT is a Lean 4 library that formally proves finite-sample statistical learning theory results (ERM, VC bounds, Rademacher bounds, PAC-Bayes, etc.) with explicit assumptions and zero sorry statements, providing a machine-checked foundation for ML theory.

0 favorites 0 likes
#machine-learning

Scientists identified over 10,000 new exoplanet candidates using AI

Reddit r/singularity · yesterday Cached

Scientists used a machine learning algorithm to analyze TESS data, identifying over 10,000 new exoplanet candidates, potentially tripling the known count. One candidate was confirmed as a hot Jupiter, validating the method.

0 favorites 0 likes
#machine-learning

@rwayne: Yesterday an interesting paper dropped on arXiv that directly translates the 'consciousness' mechanism from cognitive science into long-context engineering.

X AI KOLs Timeline · yesterday

Researchers propose applying the "global ignition" consciousness mechanism from cognitive science to long-context engineering, introducing the MiA-Signature method that uses submodular selection of high-level concepts to cover the activation space. Applied to RAG and agentic systems, it delivers consistent performance improvements across multiple long-context tasks.

0 favorites 0 likes
#machine-learning

@tom_doerr: Structured roadmaps for AI, ML, and LLM learning https://github.com/bishwaghimire/ai-learning-roadmaps…

X AI KOLs Timeline · yesterday Cached

A comprehensive, open-source GitHub repository providing structured learning roadmaps and curated resources for mastering AI, machine learning, deep learning, and large language models from beginner to advanced levels. Designed for students and professionals, it covers foundational concepts, programming frameworks, career tracks, and emerging AI topics.

0 favorites 0 likes
#machine-learning

@cyrilXBT: Andrew Ng just taught the entire mathematical foundation of machine learning in one lecture. Free. Stanford University …

X AI KOLs Timeline · yesterday

Andrew Ng shares his Stanford CS229 lecture covering core machine learning mathematics, including locally weighted regression, maximum likelihood, logistic regression, and Newton's method, providing developers with a comprehensive guide to ML fundamentals.

0 favorites 0 likes
#machine-learning

A new generation of AI models and one of the most powerful research papers out there.

Reddit r/LocalLLaMA · yesterday

Token AI releases a research paper introducing STAM, a new adaptive momentum optimizer designed to improve training stability and reduce memory usage compared to standard optimizers like AdamW.

0 favorites 0 likes
#machine-learning

SDFlow: Similarity-Driven Flow Matching for Time Series Generation

arXiv cs.AI · yesterday Cached

This paper introduces SDFlow, a similarity-driven flow matching framework for time series generation that addresses exposure bias in autoregressive models. It achieves state-of-the-art performance and inference speedups by operating in the frozen VQ latent space with low-rank manifold decomposition.

0 favorites 0 likes
#machine-learning

Locality-aware Private Class Identification for Domain Adaptation with Extreme Label Shift

arXiv cs.AI · yesterday Cached

This paper proposes a locality-aware private class identification approach and a reliable optimal transport-based method (ReOT) to address domain adaptation challenges under extreme label shift, particularly distinguishing shared from private classes.

0 favorites 0 likes
#machine-learning

Understanding Annotator Safety Policy with Interpretability

arXiv cs.AI · yesterday Cached

This paper introduces Annotator Policy Models (APMs) by Apple, which use interpretability techniques to infer annotators' internal safety policies from their labeling behavior without requiring additional annotation effort. The authors demonstrate that APMs can accurately model these policies and distinguish between sources of annotation disagreement, such as operational failures, policy ambiguity, and value pluralism.

0 favorites 0 likes
#machine-learning

Adversarial Graph Neural Network Benchmarks: Towards Practical and Fair Evaluation

arXiv cs.LG · yesterday Cached

This paper presents a comprehensive benchmark for evaluating adversarial attacks and defenses in Graph Neural Networks, highlighting the need for standardized and fair experimental protocols.

0 favorites 0 likes
#machine-learning

MOSAIC: Module Discovery via Sparse Additive Identifiable Causal Learning for Scientific Time Series

arXiv cs.LG · yesterday Cached

This paper introduces MOSAIC, a method for module discovery in scientific time series that combines causal representation learning with sparse additive identifiable causal learning. It aims to recover interpretable latent variables and their associated observations without post-hoc alignment, validated on domains like molecular dynamics and climate data.

0 favorites 0 likes
#machine-learning

Non-Myopic Active Feature Acquisition via Pathwise Policy Gradients

arXiv cs.LG · yesterday Cached

This paper introduces NM-PPG, a non-myopic active feature acquisition method using pathwise policy gradients to optimize sequential feature selection in costly prediction scenarios.

0 favorites 0 likes
#machine-learning

TIDE: Every Layer Knows the Token Beneath the Context

arXiv cs.CL · yesterday Cached

This paper introduces TIDE, a method that addresses the Rare Token and Contextual Collapse problems in LLMs by injecting token identity into every layer via Embedding Memory. The authors demonstrate theoretical and empirical improvements across language modeling and downstream tasks.

0 favorites 0 likes
#machine-learning

On Semantic Loss Fine-Tuning Approach for Preventing Model Collapse in Causal Reasoning

arXiv cs.LG · yesterday Cached

This paper identifies a critical 'model collapse' issue in standard fine-tuning for causal reasoning and proposes a semantic loss function with graph-based logical constraints to prevent it.

0 favorites 0 likes
#machine-learning

SPADE: Faster Drug Discovery by Learning from Sparse Data

arXiv cs.LG · yesterday Cached

This paper introduces SPADE, a novel algorithm for drug discovery that efficiently identifies high-quality ligands from sparse data using only ~40 tests. It demonstrates superior sample efficiency and speed compared to deep learning and Bayesian optimization methods.

0 favorites 0 likes
#machine-learning

COPYCOP: Ownership Verification for Graph Neural Networks

arXiv cs.LG · yesterday Cached

This paper introduces CopyCop, an algorithm for verifying ownership of Graph Neural Networks by detecting surrogate models even when they differ in architecture, weights, or output transformations.

0 favorites 0 likes
#machine-learning

Forecasting Green Skill Demand in the Automotive Industry: Evidence from Online Job Postings

arXiv cs.LG · yesterday Cached

This paper presents an end-to-end pipeline for identifying and forecasting green skill demand using online job postings from Mexico's automotive industry. It benchmarks 15 time-series forecasting models, finding transformer-based models like FEDformer and Informer perform best, and introduces a two-dimensional framework to classify skills by growth dynamics.

0 favorites 0 likes
#machine-learning

Chainwash: Multi-Step Rewriting Attacks on Diffusion Language Model Watermarks

arXiv cs.CL · yesterday Cached

This research paper introduces Chainwash, a multi-step rewriting attack that effectively removes statistical watermarks from diffusion language model (LLaDA-8B-Instruct) outputs, reducing detection rates from 87.9% to 4.86% after five chained rewrites.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback