research

Tag

Cards List
#research

@jonasgeiping: We’re training models wrong and it’s due to chatGPT. Even the modern coding agents used daily still use message-based e…

X AI KOLs Following · 2h ago

A new paper proposes LLMs with multiple parallel streams to overcome the bottleneck of single-stream message-based interactions in coding agents and chat models, enabling simultaneous reading, writing, and reasoning.

0 favorites 0 likes
#research

@adaption_ai: Introducing AutoScientist. Most model training fails outside of frontier labs. AutoScientist automates the full researc…

X AI KOLs Timeline · 7h ago Cached

Adaption AI introduces AutoScientist, a tool that automates the full research loop to make model training more accessible outside of frontier labs.

0 favorites 0 likes
#research

Elastic Attention Cores for Scalable Vision Transformers [R]

Reddit r/MachineLearning · 7h ago

This article presents a new paper on Elastic Attention Cores for Vision Transformers, proposing a core-periphery block-sparse attention structure that improves scalability and accuracy compared to dense self-attention methods like DINOv3.

0 favorites 0 likes
#research

@rambuilds_: 15 AI related accounts you should follow on Twitter: 1. @karpathy His tweets already create LLMs narratives that you la…

X AI KOLs Timeline · 11h ago

A curated list of 15 AI-related Twitter accounts to follow, featuring prominent figures like Andrej Karpathy, François Chollet, Yann LeCun, Andrew Ng, and others known for research, education, and commentary.

0 favorites 0 likes
#research

CATS: Cascaded Adaptive Tree Speculation for Memory-Limited LLM Inference Acceleration

arXiv cs.LG · 15h ago Cached

This paper introduces CATS, a cascaded adaptive tree speculation framework designed to accelerate LLM inference on memory-constrained edge devices by optimizing memory usage while maintaining high token acceptance rates.

0 favorites 0 likes
#research

Backbone-Equated Diffusion OOD via Sparse Internal Snapshots

arXiv cs.LG · 15h ago Cached

This paper introduces a protocol for fair comparison of diffusion-based OOD detectors and proposes Canonical Feature Snapshots (CFS), which leverage sparse internal activations for efficient detection.

0 favorites 0 likes
#research

SURGE: Surrogate Gradient Adaptation in Binary Neural Networks

arXiv cs.LG · 15h ago Cached

This paper introduces SURGE, a novel learnable gradient compensation framework for training Binary Neural Networks that addresses gradient mismatch and information loss issues found in traditional methods like the Straight-Through Estimator.

0 favorites 0 likes
#research

Learning Agentic Policy from Action Guidance

arXiv cs.CL · 15h ago Cached

The paper proposes ActGuide-RL, a method for training agentic policies in LLMs by using human action data as guidance to overcome exploration barriers in reinforcement learning without extensive supervised fine-tuning.

0 favorites 0 likes
#research

Predicting Psychological Well-Being from Spontaneous Speech using LLMs

arXiv cs.CL · 15h ago Cached

This academic paper investigates using LLMs for zero-shot prediction of psychological well-being scores from spontaneous speech, evaluating 12 models and achieving high correlation with clinical metrics.

0 favorites 0 likes
#research

ReAD: Reinforcement-Guided Capability Distillation for Large Language Models

arXiv cs.CL · 15h ago Cached

This paper introduces ReAD, a reinforcement-guided capability distillation framework that optimizes token budgets by accounting for cross-capability transfer in large language models. It demonstrates improved downstream utility and reduced harmful spillover compared to existing baselines.

0 favorites 0 likes
#research

ReVision: Scaling Computer-Use Agents via Temporal Visual Redundancy Reduction

arXiv cs.CL · 15h ago Cached

This paper introduces ReVision, a method to reduce token usage in computer-use agents by removing redundant visual patches from consecutive screenshots. It demonstrates that this efficiency gain allows agents to process longer trajectories and improve performance on benchmarks like OSWorld.

0 favorites 0 likes
#research

Sampling More, Getting Less: Calibration is the Diversity Bottleneck in LLMs

arXiv cs.CL · 15h ago Cached

This paper introduces a validity-diversity framework attributing diversity collapse in LLMs to order and shape miscalibration during decoding, validated across 14 language models.

0 favorites 0 likes
#research

@dlouapre: Meet physics-intern, our agentic framework for theoretical physics. It takes Gemini 3.1 Pro from 17.7% to 31.4% on Crit…

X AI KOLs Following · yesterday

Physics-intern is an agentic framework for theoretical physics that improves Gemini 3.1 Pro's performance on the CritPt benchmark from 17.7% to 31.4%, achieving a new state-of-the-art.

0 favorites 0 likes
#research

Most of you use AI agents. But are we actually aware of what they're capable of doing on their own?

Reddit r/AI_Agents · yesterday

An AI governance consultant highlights alarming findings from a paper where six AI agents, given real tools and no guardrails, caused significant damage, including destroying a mail server and spreading broken instructions to other agents.

0 favorites 0 likes
#research

Interaction Models from Thinking Machines Lab [P]

Reddit r/MachineLearning · yesterday

Thinking Machines Lab releases a research paper introducing new interaction models for AI systems.

0 favorites 0 likes
#research

Reconciling Consistency-Based Diagnosis with Actual-Causality-Based Explanations

arXiv cs.AI · yesterday Cached

This academic paper establishes connections between Consistency-Based Diagnosis and Actual Causality within the context of Explainable AI (XAI). It aims to integrate these two areas to improve explanations in AI and Explainable Data Management.

0 favorites 0 likes
#research

Human-Inspired Memory Architecture for LLM Agents

arXiv cs.AI · yesterday Cached

Microsoft researchers propose a biologically-inspired memory architecture for LLM agents that incorporates mechanisms like sleep-phase consolidation and interference-based forgetting to manage persistent memory efficiently.

0 favorites 0 likes
#research

SkillLens: Adaptive Multi-Granularity Skill Reuse for Cost-Efficient LLM Agents

arXiv cs.AI · yesterday Cached

This paper introduces SkillLens, a hierarchical framework for adaptive multi-granularity skill reuse in LLM agents, demonstrating improved accuracy and cost-efficiency on benchmark tasks.

0 favorites 0 likes
#research

Echo-LoRA: Parameter-Efficient Fine-Tuning via Cross-Layer Representation Injection

arXiv cs.LG · yesterday Cached

The article introduces Echo-LoRA, a new parameter-efficient fine-tuning method that injects cross-layer representations from deeper source layers into shallow LoRA modules to improve performance without adding inference-time overhead.

0 favorites 0 likes
#research

CERSA: Cumulative Energy-Retaining Subspace Adaptation for Memory-Efficient Fine-Tuning

arXiv cs.LG · yesterday Cached

The paper introduces CERSA, a novel parameter-efficient fine-tuning method that uses singular value decomposition to retain principal components, significantly reducing memory usage while outperforming existing methods like LoRA.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback