Tag
Two new open-source small language models are being released: one matches state-of-the-art accuracy at up to 93x smaller size, and the other outperforms a recent OpenAI model. The first model drops tomorrow.
A novice asks for recommendations on small language models and prompting strategies to build an employee note summarization engine under 2000 tokens, after experiencing hallucinations with Qwen2.5-7B-Instruct.
Empirical study shows small language models achieve 100% adversarial robustness with System 1 intuition but collapse under System 2 reasoning when used as edge-native governance firewalls in decentralized autonomous organizations.