Automated AI researcher running locally with llama.cpp

Reddit r/LocalLLaMA Tools

Summary

ml-intern is a harness for AI agents that integrates with Hugging Face's libraries and now supports running local models via llama.cpp or ollama, enabling an automated AI researcher to run 24/7 on a laptop.

Hi everyone, I'm happy to share ml-intern, which is a harness for agents to have tighter integration with Hugging Face's open-source libraries (transformers, datasets, trl, etc) and Hub infrastructure: [https://github.com/huggingface/ml-intern](https://github.com/huggingface/ml-intern) The harness is quite simple (basically tools + system prompt) and we built it initially for Claude Opus. However, now that open models are getting really good at agentic workflows, I just added support for running ml-intern with local models via llama.cpp or ollama. As you can see in the video, Qwen3.6-35B-A3B is able to SFT a model end-to-end by orchestrating CPU/GPU sandboxes and jobs on the Hub. I find this pretty neat because we can now have an AI researcher running 24/7 on a laptop, without maxing out token limits :) Anyway, I hope this is useful to the community and please let me know if there are any features that you'd like us to include.
Original Article

Similar Articles

GGML and llama.cpp join HF to ensure the long-term progress of Local AI

Hugging Face Blog

GGML and llama.cpp have joined Hugging Face to ensure long-term sustainability of local AI development. Georgi Gerganov's team will maintain full autonomy over the projects while receiving resources to scale community support and improve integration between llama.cpp inference and transformers model definitions.

ml-intern

Product Hunt

Hugging Face launches ML-Intern, an AI agent that automates post-training tasks for machine-learning workflows.