image

paper , blog

Details

Preliminaries

Replay methods ์˜›๋‚  ๋ฐ์ดํ„ฐ๋“ค์„ 1% ์ •๋„ , external memory๋กœ ์ €์žฅํ•ด์„œ Pseudo Rehearsal : ๊ณผ๊ฑฐ ์ƒ˜ํ”Œ๋“ค์„ generate https://ffighting.tistory.com/entry/iCaRL-%ED%95%B5%EC%8B%AC-%EB%A6%AC%EB%B7%B0

Regularization-based methods ํŒŒ๋ผ๋ฏธํ„ฐ๊ฐ€ ๋„ˆ๋ฌด ๋ฐ”๋€Œ์ง€ ์•Š๋„๋ก ์ œ์•ฝ์„ ์คŒ

Parameter isolation ๊ฐ class seg ๋ณ„๋กœ ๋ชจ๋ธ์„ ํ•™์Šตํ•˜๊ณ  ๊ทธ ํŒŒ๋ผ๋ฏธํ„ฐ๋“ค์„ ์–ด๋–ป๊ฒŒ ํ•ฉ์น ์ง€ ๋‚˜์ค‘์— ๊ณ ๋ฏผ

TL;DR

  • task : class incremental learning / domain incremental learning / task-agnostic learning
  • problem : catastrophic forgetting.
  • idea : prompt learning ! ์ด๋ฏธ์ง€๊ฐ€ ์ฃผ์–ด์กŒ์„ ๋•Œ ์ „์ฒด M๊ฐœ์˜ prompt pool ์ค‘์— N๊ฐœ์˜ ๊ฐ€๊นŒ์šด prompt๋ฅผ ๋ฝ‘๊ณ (Learning to Prompt, L2P), ViT ๋น„์ „ ํ† ํฐ ์•ž์— prependํ•ด์„œ ์ด๋ฏธ์ง€ ๋ถ„๋ฅ˜.
  • architecture : ViT-B/16
  • objective : CrossEntropyLoss + diversifying prompt-selection.
  • baseline : CIL variants(finetuning sequentially, BiC, EWC, DER++..)
  • data : split CIFAR100, 5-datasets
  • result : SOTA. ๊ทธ๋ƒฅ iid finetuning๋ณด๋‹จ ๋‚ฎ์Œ.
  • contribution : ๊ฐ„๋‹จํ•œ ์•„์ด๋””์–ด/์•„ํ‚คํ…์ณ๋กœ SOTA.

ETC.