Adam's Law: Textual Frequency Law on Large Language Models
Summary
This article introduces AdamOpt, an open-source tool based on 'Adam's Law' that optimizes prompts by replacing low-frequency words with high-frequency synonyms to reduce perplexity. It highlights the tool's bilingual support, offline capability, and practical performance improvements in text generation.
View Cached Full Text
Cached at: 05/10/26, 06:34 AM
Paper page - Adam’s Law: Textual Frequency Law on Large Language Models
Source: https://huggingface.co/papers/2604.02176 Great paper! The insight that “high-frequency text → lower perplexity → better LLM performance” is beautifully simple yet powerful.
I was so inspired by this work that I built AdamOpt — an open-source tool that turns Adam’s Law into a practical, one-command optimization pipeline:
What it does:
adamopt optimize “your prompt” → automatically replaces low-frequency bottleneck words/phrases with higher-frequency synonyms Three modes: conservative (word-level, ≥99% fidelity), balanced (word+phrase), aggressive (full rewrite) Automatically locks entities, numbers, logic keywords, and constraints — semantic meaning is never broken Chinese & English bilingual, works offline, zero LLM API cost Real results from the tool:
“optical causation...azure...celestial firmament” → “light cause...blue...sky” — sfreq +2735% “详尽阐述” → “详细讲” — sfreq +48.8% (with “务必” and “3点” auto-locked) “In order to comprehend the methodology” → “to understand the way” — sfreq +2150% 85 tests passing, MIT licensed. Modules 1-2 are done; Modules 3-5 (semantic verification, batch SFT data processing with CTFT sorting, API & Web Demo) are open for contribution.
Repo:https://github.com/happyii/AdamOpt
If you’re working with prompts or fine-tuning data, give it a try. PRs, issues, and stars are all welcome — let’s make prompt optimization a solved problem. 🚀
Similar Articles
Meta-Tool: Efficient Few-Shot Tool Adaptation for Small Language Models
Independent study shows 227M-parameter hypernetwork adds zero gain over well-crafted few-shot prompts for tool-use in 3B Llama, achieving 79.7% of GPT-5 performance at 10× lower latency.
Minimizing Modality Gap from the Input Side: Your Speech LLM Can Be a Prosody-Aware Text LLM
Proposes TextPro-SLM, a speech large language model that minimizes the modality gap by processing spoken input to resemble prosody-aware text input, achieving strong paralinguistic understanding with low training data.
Scaling laws for neural language models
Foundational empirical study demonstrating power-law scaling relationships between language model performance and model size, dataset size, and compute budget, with implications for optimal training allocation and sample efficiency.
Rethinking the Necessity of Adaptive Retrieval-Augmented Generation through the Lens of Adaptive Listwise Ranking
This paper proposes AdaRankLLM, an adaptive retrieval framework that challenges the necessity of adaptive RAG by using listwise ranking to dynamically filter retrieved passages. The work shows that adaptive retrieval serves as a noise filter for weaker models while acting as a cost-efficiency optimizer for stronger models, with extensive experiments across multiple datasets and LLMs.
Log-Likelihood, Simpson's Paradox, and the Detection of Machine-Generated Text
This paper addresses the degradation of likelihood-based machine-generated text detectors by identifying a Simpson's paradox in token-score aggregation. It proposes a learned local calibration step that significantly improves detection performance across various models and datasets.