optimization

Tag

Cards List
#optimization

The biggest AI risk may not be superintelligence — but optimized misunderstanding

Reddit r/artificial · 5h ago

The article argues that the primary AI risk may not be superintelligence but rather systems that optimize flawed, incomplete representations of reality, leading to institutional drift, automated misclassification, and invisible governance failures.

0 favorites 0 likes
#optimization

@_vmlops: MICROSOFT RESEARCHERS BUILT THIS TO TRAIN 530B PARAMETER MODELS Deepspeed is a deep learning optimization library that …

X AI KOLs Timeline · 9h ago Cached

DeepSpeed is an open-source deep learning optimization library from Microsoft that enables efficient distributed training and inference of large-scale models with features like ZeRO, 3D parallelism, and Mixture-of-Experts.

0 favorites 0 likes
#optimization

Using OR-Tools CP-SAT for Scheduling Problems

Hacker News Top · 11h ago Cached

The article discusses using Google's OR-Tools CP-SAT solver to optimize maintenance scheduling for cloud infrastructure at Akamai, addressing complex constraints like capacity and concurrency.

0 favorites 0 likes
#optimization

Partial static single information form

Lobsters Hottest · 17h ago Cached

The article discusses Partial Static Single Information (SSI) form, an extension to SSA in compilers that captures path-dependent type information. It proposes a practical shortcut for implementing Partial SSI during SSA construction in dynamic languages, specifically referencing an implementation in Ruby's ZJIT.

0 favorites 0 likes
#optimization

Muon is Not That Special: Random or Inverted Spectra Work Just as Well

arXiv cs.LG · 18h ago Cached

This paper challenges the geometric justification for the Muon optimizer, arguing that precise structure is less important than step-size optimality. It introduces Freon and Kaon optimizers to demonstrate that random or inverted spectra can perform as well as Muon.

0 favorites 0 likes
#optimization

Optimistic Dual Averaging Unifies Modern Optimizers

arXiv cs.LG · 18h ago Cached

This paper introduces SODA, a generalization of Optimistic Dual Averaging that unifies various modern optimizers like Muon and Lion. It proposes a practical wrapper that improves performance across different scales without requiring additional hyperparameter tuning for weight decay.

0 favorites 0 likes
#optimization

Newton's Lantern: A Reinforcement Learning Framework for Finetuning AC Power Flow Warm Start Models

arXiv cs.LG · 18h ago Cached

The article introduces Newton's Lantern, a reinforcement learning framework for finetuning warm start models to solve the AC power flow problem more efficiently, particularly near voltage collapse.

0 favorites 0 likes
#optimization

ReVision: Scaling Computer-Use Agents via Temporal Visual Redundancy Reduction

arXiv cs.CL · 18h ago Cached

This paper introduces ReVision, a method to reduce token usage in computer-use agents by removing redundant visual patches from consecutive screenshots. It demonstrates that this efficiency gain allows agents to process longer trajectories and improve performance on benchmarks like OSWorld.

0 favorites 0 likes
#optimization

@Italianclownz: Tested MTP, TriAttention, TurboQuant on @UnslothAI @Alibaba_Qwen Qwen 3.6 35B A3B MTP MXFP4_MoE on @huggingface @no_stp…

X AI KOLs Following · yesterday Cached

A user benchmarks MTP, TriAttention, and TurboQuant optimizations on Qwen 3.6 35B using Unsloth on consumer hardware, finding TurboQuant to be the most effective.

0 favorites 0 likes
#optimization

Claude Code improved my agent harness by 40% overnight

Reddit r/AI_Agents · yesterday

The author introduces 'Autoharness', a tool that uses Claude Code to autonomously optimize agent harnesses by iterating on prompts and hyperparameters. This resulted in a 40% performance increase on the tau2-airline benchmark.

0 favorites 0 likes
#optimization

@songhan_mit: Explore lightening OPD for efficient LLM post training:

X AI KOLs Following · yesterday

The article introduces a method to lighten OPD for efficient post-training of Large Language Models.

0 favorites 0 likes
#optimization

Fitting Is Not Enough: Smoothness in Extremely Quantized LLMs

arXiv cs.CL · yesterday Cached

This paper investigates smoothness degradation in extremely quantized Large Language Models, arguing that preserving smoothness is crucial for maintaining performance beyond numerical accuracy.

0 favorites 0 likes
#optimization

@UnslothAI: We’re excited to share that Unsloth has joined the PyTorch Ecosystem! Unsloth is an open-source project that makes trai…

X AI KOLs Following · 2d ago Cached

Unsloth, an open-source library for efficient LLM training and inference, has officially joined the PyTorch Ecosystem to enhance accessibility and performance. The announcement highlights new features like Unsloth Studio and optimized kernels for reduced VRAM usage.

0 favorites 0 likes
#optimization

@akshay_pachaar: As an AI Engineer. Please learn: - Harness engineering, not just prompt engineering - Prompt caching vs. semantic cachi…

X AI KOLs Following · 2d ago

Akshay Pachaar outlines essential skills for AI engineers beyond prompt engineering, including caching strategies, observability, and cost attribution.

0 favorites 0 likes
#optimization

@geekbb: Auto-optimization tool for Agent harness. It takes over the heavy lifting of harness optimization: you provide a benchmark command and a target repository, and it automatically generates proposals, runs evaluations, records results, keeps the best, discards the rest, and automatically improves the agent's prompts, configurations, and source code. https…

X AI KOLs Timeline · 2d ago Cached

autoharness is an automated agent harness optimization tool that automatically generates proposals and runs evaluations based on benchmark commands to improve an agent's prompts, configurations, and source code. It supports Codex and Claude.

0 favorites 0 likes
#optimization

Online Allocation with Unknown Shared Supply

arXiv cs.AI · 2d ago Cached

This paper introduces the Online Shared Supply Allocation problem and proposes a deterministic threshold-proportional policy (GPA) that achieves a 4/3-approximation to the offline optimum. It also includes a learning-augmented extension to handle imperfect forecasts and demonstrates superior performance in synthetic and real-world experiments.

0 favorites 0 likes
#optimization

When Descent Is Too Stable: Event-Triggered Hamiltonian Learning to Optimize

arXiv cs.LG · 2d ago Cached

This paper introduces SHAPE, a structured adaptive port-Hamiltonian optimizer for fixed-budget nonconvex optimization that uses event-triggered mechanisms to balance descent, exploration, and budget allocation.

0 favorites 0 likes
#optimization

A Rod Flow Model for Adam at the Edge of Stability

arXiv cs.LG · 2d ago Cached

This paper introduces a 'rod flow' model for Adam and other adaptive optimizers to better analyze their behavior at the edge of stability. It extends continuous-time modeling to momentum methods, showing improved accuracy in tracking discrete iterates compared to stable flow models.

0 favorites 0 likes
#optimization

Revisiting Adam for Streaming Reinforcement Learning

arXiv cs.LG · 2d ago Cached

This paper revisits the Adam optimizer for streaming reinforcement learning, demonstrating that established methods like DQN and C51 perform well when properly tuned. The authors propose Adaptive Q(lambda), which combines eligibility traces with Adam's variance adaptation to surpass existing streaming RL methods on 55 Atari games.

0 favorites 0 likes
#optimization

Beyond LoRA vs. Full Fine-Tuning: Gradient-Guided Optimizer Routing for LLM Adaptation

arXiv cs.CL · 2d ago Cached

This paper proposes a Mixture of LoRA and Full (MoLF) fine-tuning framework that uses gradient-guided optimizer routing to adaptively switch between LoRA and full fine-tuning. It aims to overcome the structural limitations of relying solely on static adaptation methods by combining the plasticity of full tuning with the regularization of LoRA.

0 favorites 0 likes
Next →
← Back to home

Submit Feedback