@garrytan: Downloading now... 1M token context window with supposedly usable coding agent capability all on a 128GB Macbook Pro is

X AI KOLs Following Models

Summary

Garry Tan highlights a model with a 1M token context window and coding agent capabilities running locally on a 128GB MacBook Pro, expressing excitement about the milestone.

Downloading now... 1M token context window with supposedly usable coding agent capability all on a 128GB Macbook Pro is 🤯 https://t.co/otTL8NZMvV
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 05/09/26, 07:42 AM

Downloading now… 1M token context window with supposedly usable coding agent capability all on a 128GB Macbook Pro is 🤯 https://t.co/otTL8NZMvV

Similar Articles

Is anyone getting real coding work done with Qwen3.6-35B-A3B-UD-Q4_K_M on a 32GB Mac in opencode, claude code or similar?

Reddit r/LocalLLaMA

A user shares their experience running Qwen3-35B-A3B quantized model on an M2 MacBook Pro with 32GB RAM for coding tasks via opencode and llama.cpp, finding that the 32K context window limit causes critical memory loss during compaction, making complex coding tasks impractical. They conclude that meaningful agentic coding with this model likely requires at least 128K context, exceeding what their hardware can support.

2x 512gb ram M3 Ultra mac studios

Reddit r/LocalLLaMA

A user shares their $25k hardware setup of two 512GB RAM M3 Ultra Mac Studios for running large language models locally, having tested DeepSeek V3 Q8 and GLM 5.1 Q4 via the exo distributed inference backend, while awaiting Kimi 2.6 MLX optimization.