Tag
Microsoft has released Phi-Ground-Any, a 4B parameter vision model for GUI grounding on Hugging Face that achieves state-of-the-art results, enabling AI agents to precisely interact with screen elements.
Hugging Face and ClawHub, major repositories for AI models and agent skills, have been systematically compromised with hundreds of malicious entries that steal credentials and hijack systems for cryptocurrency mining, exploiting trust in shared infrastructure.
Both Claw Hub and Hugging Face have been compromised, with 575 malicious skills uploaded; users are advised to exercise caution when using content from these platforms.
Julien Chaumond teases an upcoming release from Hugging Face, building anticipation.
Release of a mixed-bit quantized version of the MiniMax M2.7 model, optimized to 74 GB for efficient local inference on Apple Silicon devices.
A malicious repository on Hugging Face posing as an OpenAI privacy filter has been identified as a Windows infostealer virus using Python and PowerShell droppers.
Hugging Face Hub has surpassed 4,000 public reinforcement learning environments, positioning itself as a potentially largest platform for RL environments.
A robotics engineer from Hugging Face proposes mapping human facial expressions onto non-humanoid robots to enhance expressiveness while avoiding the uncanny valley, with plans to use this data for autonomous body language training.
Hugging Face announces the addition of private, high-quality datasets from Appen and DataoceanAI to the Open ASR Leaderboard to prevent benchmaxxing and test-set contamination, while maintaining public data for the default average WER calculation.
This article introduces TenStrip/LTX2.3-10Eros, a fine-tuned AI video model on Hugging Face designed for improved image-to-video generation and prompt adherence. It provides technical details on file formats, compatibility with ComfyUI nodes, and specific prompting strategies for optimal results.
This entry describes Qwen3.5-9B-DeepSeek-V4-Flash, a distilled AI model that transfers reasoning capabilities from DeepSeek-V4 into a smaller 9B parameter space for efficient inference.
DeepInfra has been added as an official Inference Provider on the Hugging Face Hub, allowing users to run serverless inference for models like DeepSeek V4 and Kimi-K2.6 directly through the HF UI and SDKs.
Hugging Face open-sourced ml-intern, an autonomous agent that reads ML papers, discovers datasets, trains models, debugs failures, and ships production-ready models to the Hub, automating the entire post-training workflow.
Z-Anime is a full fine-tune of Alibaba's Z-Image Base model, specialized for high-quality anime generation with support for natural language prompts and low VRAM usage.
This repository provides fixed Jinja chat templates for Qwen 3.5 and 3.6, addressing rendering errors, token waste, and missing features in the official templates for engines like LM Studio and llama.cpp.
This article introduces Qwen3.6-27B-DFlash, a specialized drafter model for DFlash, a novel speculative decoding method using block diffusion to accelerate inference speed. It provides installation instructions for vLLM and SGLang to enable parallel drafting with the target Qwen3.6-27B model.
SenseNova U1 is a new series of native multimodal models that unify understanding and generation within a single architecture using the NEO-Unify framework, eliminating the need for separate visual encoders or VAEs.
Hugging Face has opened a new office in Tokyo to support the growth of open-source AI and nurture the local developer community in Japan.
Talkie-1930-13b-it is a 13B parameter instruction-tuned language model trained on pre-1931 text and fine-tuned using reinforcement learning with DPO.
Hugging Face now lets AI agents invoke 1 million Spaces, turning the hub into a programmable platform where agents can tap any specialized model or app.