Artificial Intelligence
Alignment and Human-in-the-Loop: Making AI Safe and Useful
John Schulman pioneered PPO and RLHF to ensure LLMs like GPT are aligned with human intent (hub overview).
Program Synthesis and Structured Reasoning in LLMsTechniques include:
Failure Modes in AI: Lessons from OpenAI- PPO:
maximize E[min(r(θ)·A, clip(r(θ),1−ε,1+ε)·A)] - RLHF for human-guided behavior refinement
- Instruction-following and safety checks
Alignment relies on prior scaling principles and generalization mechanisms.
Global Deployment of AI: Lessons from OpenAI
