less than 1 minute read

Meta info.

TL; DR

LLaMA-Adapter, a method for quickly and efficiently fine-tuning LLaMA into an instruction-following model using self-instruct demonstrations, matching Alpaca's modeling performance.

Untitled

Untitled

Untitled

Untitled

Untitled

Sugestions

In contrast to LLaMA-Alpaca, it has not fine-tuning the whole param, this approach adds a layer with 1.2M params on top of 7B LLaMA.

  1. a lightweight adaption method that efficiently fine-tunes LLaMA into an instruction-following model using self-instruct demonstrations.
  2. with only 1.2M learnable parameters using 52K instruction data and less than an hour of fine-tuning on 8 A100 GPUs(Alpaca took 3 hours), LLaMA-Adapter can effectively inject new instructional cues into LLaMA while preserving its pre-train. #pic5
  3. model can be generalized to image conditions for multi-modal reasoning and achieve competitive performance on the ScienceQA benchmark. (it allowed to add other input image and video tokens!, #pic3)

Personal note. μ•ŒνŒŒμΉ΄λž‘ λΉ„λ“±ν•˜λ‹€λŠ”λ° μ •μž‘ μ•ŒνŒŒμΉ΄λž‘ μ„±λŠ₯λΉ„κ΅ν•œ 건 μ—†λŠ”λ“―β€¦?Β λ©€ν‹°λͺ¨λ‹¬μ€ μ‹ κΈ°ν•˜λ„€μš”β€¦