A hazard analysis framework for code synthesis large language models
Summary
OpenAI presents a hazard analysis framework for evaluating safety risks associated with code synthesis LLMs like Codex, examining technical, social, political, and economic impacts through a novel evaluation methodology for code generation capabilities.
View Cached Full Text
Cached at: 04/20/26, 02:46 PM
Similar Articles
A research agenda for assessing the economic impacts of code generation models
OpenAI is laying out a research agenda to assess the economic impacts of code generation models like Codex, covering areas such as productivity, employment, skill development, and inequality, while inviting external researchers to collaborate.
Evaluating large language models trained on code
OpenAI introduces Codex, a GPT model fine-tuned on GitHub code, achieving 28.8% functional correctness on HumanEval (a new benchmark for code synthesis from docstrings), significantly outperforming GPT-3 (0%) and GPT-J (11.4%). The paper demonstrates that repeated sampling improves performance to 70.2% with 100 samples, and discusses limitations and broader impacts of code generation systems.
Running Codex safely at OpenAI
OpenAI details how it deploys Codex with safety controls including sandboxing, approval policies, network policies, and agent-native telemetry to ensure secure operation of coding agents in enterprise environments.
Lessons learned on language model safety and misuse
OpenAI shares lessons learned on language model safety and misuse, discussing challenges in measuring risks, the limitations of existing benchmarks, and their development of new evaluation metrics for toxicity and policy violations. The post also highlights concerns about labor market impacts and the need for continued research on measuring social effects of AI deployment at scale.
@OpenAI: Another reason to switch to Codex.
OpenAI promotes switching to Codex, highlighting another reason to adopt their AI code generation model.