less than 1 minute read

Meta info.

TL; DR

GPT-3.5, GPT-4๋ฅผ ํ™œ์šฉ, 2-step prompt ์‚ฌ์šฉํ•ด์„œ ๋งŒ๋“  synthetic data(94 languages, 500K examples)๋กœ decoder-only LLM(Mistral-7B)์„ contrastive loss ์‚ฌ์šฉํ•ด 1-epoch ํ•™์Šต. ์ด unlabeled๋งŒ ํ•™์Šต์‹œ unsupervised SOTA์— ์กฐ๊ธˆ ๋’ค์ณ์ง€์ง€๋งŒ, labeled data๋ž‘ ๊ฐ™์ด ์‚ฌ์šฉํ•˜๋ฉด MTEB, BEIR ๋“ฑ์—์„œ supervised SOTA๋„ outperform.

Untitled 4

Untitled

Untitled

Untitled

Untitled

Suggestions

  • MTEB SOTA (using mistral-7b-instruct based model)
  • ์ด๋ฏธ ์กด์žฌํ•˜๋Š” LLM์„ ์ด์šฉํ•˜์—ฌ data ์ƒ์„ฑํ•œ๊ฒƒ์œผ๋กœ๋„ ๋†’์€ ์„ฑ๋Šฅ์„ ๋ณด์ธ ๋…ผ๋ฌธ (ํ•˜์ง€๋งŒ ์ตœ๊ณ  ์„ฑ๋Šฅ์€ ์ƒ์„ฑ ๋ฐ์ดํ„ฐ + E5์—์„œ ์‚ฌ์šฉํ•œ ๊ธฐ์กด ๋ฐ์ดํ„ฐ ํ•ฉ์นœ๊ฒƒ)

Personal note. ๋ฆฌ๋”๋ณด๋“œ ํ™•์ธํ–ˆ๋Š”๋ฐ ์ฐจ์›์ด ์••๋„์ ์œผ๋กœ ํฌ๋„ค์š”โ€ฆ (4096, pic 4)

  • ์•„์‰ฌ์šด์ : โ€œmistral-7b ๋ฅผ ์ด์šฉํ–ˆ๊ธฐ์— ์˜ฌ๋ผ๊ฐ„ ์„ฑ๋Šฅ์˜ ์ฐจ์ดโ€๊ฐ€ ๋ถˆ๋ช…ํ™•ํ•จ
  • ํŠน์ดํ•œ์ : mistral-7b ์—์„œ๋Š” contrastive pre-training์‹œ ์„ฑ๋Šฅ ํ–ฅ์ƒ์ด xlm-r ๋ชจ๋ธ ๋Œ€๋น„ ๋ณ„๋กœ ์—†์Œ