Tag
Modal engineers detail their approach to achieving truly serverless GPUs for AI inference, combining cloud buffers, a custom content-addressed filesystem, and CPU/GPU checkpoint/restore to scale replicas in tens of seconds instead of minutes.
Modal announces a partnership with OpenAI Devs and Antler Global to host an Autoresearch Systems Hackathon on May 30th targeting data and compute-intensive challenges.