Tag
The article discusses how interaction models pose new challenges for AI model inference engines, with a focus on the vLLM project's solution as covered in a podcast episode featuring Woosuk Kwon.
Thinking Machine launched a new multimodal AI model that can simultaneously listen, see, speak, interrupt, react, think, and use tools, demonstrating the convergence of models and agents.
The article highlights a research update describing an interaction model capable of tracking cognitive states like thinking, yielding, and self-correction during storytelling without a built-in dialogue management system.