๋ณธ๋ฌธ ๋ฐ”๋กœ๊ฐ€๊ธฐ
Deep Learning/[LLM] ์ด๋ก  ๊ณต๋ถ€

LLM์˜ ๋ชจ๋“  ๊ฒƒ 3 [PEFT-Parameter-efficient fine Tuning]

by ์ œ๋ฃฝ 2024. 3. 22.
728x90
๋ฐ˜์‘ํ˜•

PEFT(Parameter-efficient fine Tuning)


: ๊ธฐ์กด์˜ ๊ฒฝ์šฐ, ์‚ฌ์ „ ํ•™์Šต๋œ LLM์„ ๋‹ค์šด์ŠคํŠธ๋ฆผ ๋ฐ์ดํ„ฐ์…‹์— ๋”ฐ๋ผ ํŒŒ์ธ ํŠœ๋‹ํ•˜๋ฉด ์‚ฌ์ „ํ•™์Šต๋œ ๋ชจ๋ธ์„ ๊ทธ๋Œ€๋กœ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ๋ณด๋‹ค ํ™•์‹คํ•œ ์„ฑ๋Šฅ ํ–ฅ์ƒ์„ ๋ณด์—ฌ์ค€๋‹ค.

: ํ•˜์ง€๋งŒ, ๋ชจ๋ธ์ด ์ ์  ์ปค์ง์— ๋”ฐ๋ผ ๋ชจ๋ธ ์ „์ฒด๋ฅผ Fine tuning ํ•˜๋Š” ๊ฒƒ์€ ๋ถˆ๊ฐ€๋Šฅ!

: ๋ฉ”๋ชจ๋ฆฌ ์ €์žฅ ๊ณต๊ฐ„ ๋ฐ ๊ณ„์‚ฐ ๋น„์šฉ์— ๋Œ€ํ•œ ๋ฌธ์ œ์ ๋„ ์กด์žฌ

⇒ ํ•ด๋‹น ๋ฌธ์ œ๋“ค์„ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด ๋‚˜์˜จ ๊ฒƒ์ด PEFT ์ด๋‹ค.

๐Ÿ’ก PEFT์˜ ์—ญํ•  : ๋Œ€๋ถ€๋ถ„์˜ ํŒŒ๋ผ๋ฏธํ„ฐ๋ฅผ ํ”„๋ฆฌ์ง•ํ•˜๊ณ  ์ผ๋ถ€์˜ ํŒŒ๋ผ๋ฏธํ„ฐ๋งŒ์„ ํŒŒ์ธํŠœ๋‹ํ•จ์œผ๋กœ์จ ์ €์žฅ๊ณต๊ฐ„๊ณผ ๊ณ„์‚ฐ๋Šฅ๋ ฅ์„ ๋Œ€ํญ ์ค„์ธ๋‹ค.

catastrophic forgetting(ํŒŒ๊ตญ์  ๋ง๊ฐ)์˜ ๊ทน๋ณต : Fine tuning ์‹œ ๋ฐœ์ƒ๋˜๋Š” ๋ฌธ์ œ๋กœ ์ƒˆ๋กœ์šด ํ…Œ์Šคํฌ๋ฅผ ํ•™์Šตํ•จ์— ๋”ฐ๋ผ ๊ธฐ์กด ํ…Œ์Šคํฌ์— ๋Œ€ํ•œ ํ•™์Šต ์ •ํ™•๋„๊ฐ€ ๋–จ์–ด์ง€๋Š” ํ˜„์ƒ์„ ์˜๋ฏธ

: ์ƒˆ๋กœ์šด ๋ฐ์ดํ„ฐ๋กœ Fine-tuning์„ ์‹œํ‚ค๊ฒŒ ๋˜๋Š”๋ฐ, ๊ณผ๊ฑฐ ๋ฐ์ดํ„ฐ๋ฅผ ์žŠ์ง€ ๋ชปํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์˜คํžˆ๋ ค ์ƒˆ๋กœ์šด ์„ž์ธ ๋‹ต์ด ๋‚˜์˜ค๊ฒŒ ๋˜๋Š” ํ˜„์ƒ

  1. LoRA: LORA: LOW-RANK ADAPTATION OF LARGE LANGUAGE MODELS
  2. Prefix Tuning: P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
  3. Prompt Tuning: The Power of Scale for Parameter-Efficient Prompt Tuning
  4. P-Tuning: GPT Understands, Too
728x90
๋ฐ˜์‘ํ˜•