Tag
AMD argues that agentic AI requires rethinking infrastructure planning, with a need for dedicated CPU racks for orchestration and control workloads, shifting the CPU:GPU ratio from 1:8 or 1:4 to 1:1 or higher, rather than simply adding more CPUs to GPU-dense servers.
CyberSecQwen-4B is a small, specialized 4B parameter model fine-tuned for defensive cybersecurity tasks, designed to run locally on a single GPU, addressing privacy, cost, and air-gapped deployment needs.
A technical deep-dive into achieving peak TOPS performance on the AMD Ryzen AI 7 350 NPU, comparing it to Xilinx AIE-ML v2 AI engines and explaining the hardware architecture for matrix multiplication workloads.
AMD is set to release new slottable PCIe-based Instinct GPUs aimed at the enterprise AI market, offering a potential new hardware option for local LLM deployment.
AMD introduces the Instinct MI350P accelerator featuring CDNA 4 architecture in a PCIe form factor, though pricing and availability details are not yet announced.
AMD and OpenAI announce a strategic partnership to deploy 6 gigawatts of AMD Instinct GPUs, with initial 1 gigawatt deployment starting in H2 2026. AMD will issue OpenAI warrants for up to 160 million shares, with vesting tied to deployment milestones and financial targets.