Weak-to-strong Generalization: Eliciting Strong Capabilities with Weak Supervision
Meta info.
- Authors: Collin Buruns, Pavel Izmailov, et al.
- Paper: https://cdn.openai.com/papers/weak-to-strong-generalization.pdf
- Affiliation: OpenAI
- References: OpenAI Blog
TL; DR
Naively finetune strong pretrained models on labels generated by a weak model consistently perform better than their weak supervisors.

Suggestions
- βνμ΅ λ°μ΄ν° νμ§μ΄ λͺ¨λΈ μ±λ₯μ μ£Όμ μμΈμΈκ°?β μ λν΄, μλ μλ μλ€κ³ μ£Όμ₯.
- μ΄λ―Έ μ ꡬμΆλ ν° λͺ¨λΈμ΄λΌλ©΄ μλμ μΌλ‘ μ§ λμ λ°μ΄ν°λ‘λ μ’μ μ±λ₯μ λΌ μ μλ€.
Personal note. βμ΄μ§λ₯ VS μΈκ°β μ λν κ³ λ―Όμ ν μ μλ λ Όλ¬Έ
- λ λλν μΈκ°μ΄ λ λλν AI λ₯Ό κ΄λ¦¬νλ μν© => μμ LLM λͺ¨λΈλ‘ ν° LLM λͺ¨λΈμ κ°λ νλ λ°©λ²μ κ³ μνκ³ μ€ν
- Human feedbackμ΄ μμμ μΌλ‘ μ΄λ €μ΄λ° LLMλ μνκΈ° λλ¬Έμ RLAIF νλ μΆμΈμμ, μ€νλ € μμλλ¬Έμ΄ μλλΌ κ·Έλ₯(λλ²μ§Έ bulleted listμ μ΄μ λ‘) RLAIFκ° μ ν©νλ€λ μλ°μ(?)μ΄ μ’μλ κ² κ°μ