ModernBERT-base-multilingual

Model Details

ModernBERT๋Š” ์–‘๋ฐฉํ–ฅ ์ธ์ฝ”๋” ์•„ํ‚คํ…์ฒ˜์— ํ˜„๋Œ€์ ์ธ ํŠธ๋žœ์Šคํฌ๋จธ ๊ธฐ๋ฒ•์„ ์ ์šฉํ•œ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. RoPE๋ฅผ ์‚ฌ์šฉํ•ด ์ตœ๋Œ€ 8,192 ํ† ํฐ์˜ ๊ธด ๋ฌธ๋งฅ์„ ํšจ์œจ์ ์œผ๋กœ ์ฒ˜๋ฆฌํ•˜๋ฉฐ, Local-Global ์–ดํ…์…˜ ํŒจํ„ด์œผ๋กœ ๊ณ„์‚ฐ ๋ณต์žก๋„๋ฅผ ์ค„์˜€์Šต๋‹ˆ๋‹ค. GeGLU ํ™œ์„ฑํ™” ํ•จ์ˆ˜์™€ Pre-normalization ๋ธ”๋ก, Unpadding ๊ธฐ๋ฒ•์„ ํ†ตํ•ด ๊ธฐ์กด BERT ๋Œ€๋น„ ์ตœ๋Œ€ 4๋ฐฐ ๋น ๋ฅธ ์ฒ˜๋ฆฌ ์†๋„๋ฅผ ๋‹ฌ์„ฑํ–ˆ์Šต๋‹ˆ๋‹ค.

์ด ์—ฐ๊ตฌ๋Š” ๊ตฌ๊ธ€์˜ TPU Research Cloud(TRC)๋ฅผ ํ†ตํ•ด ์ง€์›๋ฐ›์€ Cloud TPU๋กœ ํ•™์Šต๋˜์—ˆ์Šต๋‹ˆ๋‹ค.

How to Get Started with the Model

from transformers import AutoTokenizer, ModernBertForSequenceClassification

tokenizer = AutoTokenizer.from_pretrained("team-lucid/ModernBERT-base-multilingual")
model = ModernBertForSequenceClassification.from_pretrained("team-lucid/ModernBERT-base-multilingual")

inputs = tokenizer("์•ˆ๋…•, ์„ธ์ƒ!", return_tensors="pt")
outputs = model(**inputs)
Downloads last month
10
Safetensors
Model size
0.2B params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Dataset used to train team-lucid/ModernBERT-base-multilingual