machine-learning

Tag

Cards List
#machine-learning

A Systematic Evaluation of Imbalance Handling Methods in Biomedical Binary Classification

arXiv cs.LG · 5h ago Cached

This paper systematically evaluates five imbalance handling methods (RUS, ROS, SMOTE, re-weighting, direct F1 optimization) on three biomedical datasets (tabular, text, image) using models of varying complexity. Results show that benefits depend on model complexity and data modality, with ROS, re-weighting, and direct F1 optimization being effective for complex models on unstructured data.

0 favorites 0 likes
#machine-learning

Comparative Evaluation of Machine Learning Approaches for Minority-Class Financial Distress Prediction Under Class Imbalance Constraints

arXiv cs.LG · 5h ago Cached

This paper presents a comparative evaluation of classical, ensemble, and neural machine learning approaches for predicting financial distress under severe class imbalance, using SMOTE for oversampling and SHAP for interpretability.

0 favorites 0 likes
#machine-learning

HodgeCover: Higher-Order Topological Coverage Drives Compression of Sparse Mixture-of-Experts

arXiv cs.LG · 5h ago Cached

HodgeCover uses higher-order topological coverage to compress sparse Mixture-of-Experts layers by addressing irreducible mergeability barriers that pairwise signals miss, matching state-of-the-art baselines on expert reduction and leading on aggressive compression.

0 favorites 0 likes
#machine-learning

TabPFN-3: Technical Report

arXiv cs.LG · 5h ago Cached

TabPFN-3 is a new foundation model for tabular data, pretrained on synthetic data, that scales to 1M training rows while reducing training and inference time, achieving state-of-the-art performance on tabular prediction, time series, and relational data.

0 favorites 0 likes
#machine-learning

SciPaths: Forecasting Pathways to Scientific Discovery

arXiv cs.CL · 5h ago Cached

Introduces SciPaths, a benchmark for forecasting the enabling contributions required to realize a target scientific discovery, and evaluates frontier and open-weight language models, finding significant room for improvement in reasoning backward from contributions to enabling building blocks.

0 favorites 0 likes
#machine-learning

@bradenjhancock: In other words: Humans are teaching teacher models how to teach other models the way good human teachers teach other hu…

X AI KOLs Following · 9h ago

Humans are training teacher models to teach student models in a step-by-step manner, penalizing leaps, to improve model intelligence.

0 favorites 0 likes
#machine-learning

@bstnxbt: dflash-mlx v0.1.6 is out. Biggest agentic update so far: ► much more usable for real OpenCode / coding-agent sessions ►…

X AI KOLs Timeline · 14h ago

dflash-mlx v0.1.6 is released with major agentic improvements, including adaptive verification, custom kernels, prefix cache improvements, and broader compatibility with agentic coding tools like OpenCode, aider, and Continue.

0 favorites 0 likes
#machine-learning

Building ML framework with Rust and Category Theory

Hacker News Top · 17h ago Cached

This article announces a working draft book 'Category Theory for Tiny ML in Rust' and a public workshop introducing a tiny ML pipeline using Rust and category theory, aimed at making machine learning structure explicit through typed transformations.

0 favorites 0 likes
#machine-learning

Hessian Matching for Machine-Learned Coarse-Grained Molecular Dynamics

arXiv cs.LG · yesterday Cached

This paper introduces a Hessian matching framework for machine-learned coarse-grained molecular dynamics that augments force matching with stochastic Hessian-vector product matching, instilling second-order curvature information into CG potentials. The method achieves up to 85% reduction in Kullback-Leibler divergence on slow-mode metrics for fast-folding proteins.

0 favorites 0 likes
#machine-learning

Predicting Channel Closures in the Lightning Network with Machine Learning

arXiv cs.LG · yesterday Cached

This paper explores predicting whether Lightning Network channels will close mutually or via forced closure using machine learning on gossip data. An MLP with temporal features outperforms graph-based models, and the dataset is publicly released.

0 favorites 0 likes
#machine-learning

ConRetroBert: EMA Stabilized Dual Encoders for Template-Based Single-Step Retrosynthesis

arXiv cs.LG · yesterday Cached

This paper presents ConRetroBert, a dual encoder framework for template-based single-step retrosynthesis that uses contrastive pretraining and listwise ranking to improve template prediction accuracy, achieving up to 75.4% top-1 accuracy on the USPTO-50k benchmark while maintaining interpretability.

0 favorites 0 likes
#machine-learning

From Generalist to Specialist Representation

arXiv cs.LG · yesterday Cached

This paper proves that task-relevant latent representations can be identified from generalist models in a fully nonparametric setting without interventions or parametric constraints, achieving a hierarchical identifiability guarantee across time steps and within each step.

0 favorites 0 likes
#machine-learning

Do Fair Models Reason Fairly? Counterfactual Explanation Consistency for Procedural Fairness in Credit Decisions

arXiv cs.LG · yesterday Cached

This paper introduces Counterfactual Explanation Consistency (CEC), a framework to detect and mitigate hidden procedural bias in outcome-fair models by aligning feature attributions between individuals and their counterfactual counterparts, with experiments on credit and income datasets.

0 favorites 0 likes
#machine-learning

OceanCBM: A Concept Bottleneck Model for Mechanistic Interpretability in Ocean Forecasting

arXiv cs.LG · yesterday Cached

OceanCBM is a concept bottleneck model for spatiotemporal prediction and mechanistic interpretability in ocean forecasting, using mixed supervision to predict mixed layer heat content while imposing soft physical structure. The model achieves interpretable, physically grounded representations without sacrificing predictive skill.

0 favorites 0 likes
#machine-learning

CAWI: Copula-Aligned Weight Initialization for Randomized Neural Networks

arXiv cs.LG · yesterday Cached

Introduces CAWI, a copula-based weight initialization method for randomized neural networks that models inter-feature dependence, improving predictive performance across 83 classification benchmarks.

0 favorites 0 likes
#machine-learning

Do Androids Dream of Breaking the Game? Systematically Auditing AI Agent Benchmarks with BenchJack

arXiv cs.AI · yesterday Cached

This paper introduces BenchJack, an automated red-teaming system that systematically audits AI agent benchmarks by identifying reward-hacking exploits. It applies BenchJack to 10 popular benchmarks, surfacing 219 distinct flaws and demonstrating that evaluation pipelines lack an adversarial mindset, with the system reducing hackable-task ratios from near 100% to under 10% on four benchmarks.

0 favorites 0 likes
#machine-learning

PyTorch 2.12 Release Highlights (7 minute read)

TLDR AI · yesterday Cached

PyTorch 2.12 introduces significant performance improvements including up to 100x faster batched eigendecomposition on CUDA, a new device-agnostic torch.accelerator.Graph API, and support for Microscaling quantization in torch.export, continuing the framework's evolution into a unified production platform.

0 favorites 0 likes
#machine-learning

@zcbenz: We have achieved a milestone in MLX that all tests are passing in CUDA backend now.

X AI KOLs Following · yesterday Cached

MLX has reached a milestone where all tests pass on the CUDA backend, indicating improved compatibility with NVIDIA GPUs.

0 favorites 0 likes
#machine-learning

@LakshyAAAgrawal: Learning from rich textual feedback (errors, traces, partial reasoning) beats scalar reward alone for LLM optimization.…

X AI KOLs Following · yesterday

Fast-Slow Training (FST) interleaves context optimization (via GEPA) with model weight updates via RL, achieving 3× sample efficiency over RL alone on math, code, and physics reasoning while preserving plasticity and enabling continual learning.

0 favorites 0 likes
#machine-learning

Human-level performance via ML was *not* proven impossible with complexity theory [D]

Reddit r/MachineLearning · yesterday

A paper claiming AGI via ML is impossible using complexity theory has been rebutted by a new paper showing the proof is flawed due to an undefined key term.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback