DoRA: Weight-Decomposed Low-Rank Adaptation
Paper โข 2402.09353 โข Published โข 32
โ ๏ธ ์ด ๋ชจ๋ธ์ ๋ ์ด์ ์ฌ์ฉ๋์ง ์์ต๋๋ค. tellang/yeji-8b-rslora-v7-AWQ๋ฅผ ์ฌ์ฉํ์ธ์.
์ด ๋ชจ๋ธ์ QDoRA (DoRA + rsLoRA) ๋ฐฉ์์ผ๋ก ํ์ต๋์์ผ๋ ๋ค์ ๋ฌธ์ ๋ก ์ธํด ํ๊ธฐ๋์์ต๋๋ค:
ValueError: LoRA adapter 'tellang/yeji-8b-dora-v6' uses DoRA which is not supported by vLLM
# DoRA ์ ์ฉ ํ๋ผ๋ฏธํฐ
use_dora: true
lora_r: 32-64 # DoRA๋ ๋ ํฐ rank ํ์
lora_alpha: 64-128
tellang/yeji-8b-rslora-v7-AWQfrom vllm import LLM, SamplingParams
llm = LLM(
model="tellang/yeji-8b-rslora-v7-AWQ",
quantization="awq",
gpu_memory_utilization=0.9,
)
tellang/yeji-4b-rslora-v8.1 (๋ฆฌ์์ค ์ ์ฝ)tellang/yeji-8b-rslora-v7-AWQ (์ต๊ณ ์ฑ๋ฅ)# DoRA ์ด๋ํฐ - vLLM ๋ฏธ์ง์
llm = LLM(model="tellang/yeji-8b-dora-v6") # โ ValueError
# AWQ ์์ํ - vLLM ๋ค์ดํฐ๋ธ ์ง์
llm = LLM(
model="tellang/yeji-8b-rslora-v7-AWQ",
quantization="awq",
)
| ์งํ | v6 (DoRA) | v7-AWQ (rsLoRA) |
|---|---|---|
| vLLM ์ง์ | โ ๋ฏธ์ง์ | โ ์์ ์ง์ |
| ์ถ๋ก ์๋ | N/A (๋ฐฐํฌ ๋ถ๊ฐ) | 50 tokens/s |
| ํ์ต ์๋ | 0.03 it/s | 0.8 it/s |
| ๋ฉ๋ชจ๋ฆฌ | 16GB | 5.3GB (AWQ) |
| ์ ํ๋ | Baseline | +5% |
Apache 2.0
@misc{yeji-8b-dora-v6,
title={YEJI Fortune Telling Model (DoRA v6 - Deprecated)},
author={SSAFY YEJI Team},
year={2026},
note={Deprecated: Use yeji-8b-rslora-v7-AWQ instead}
}
Base model
Qwen/Qwen3-8B-Base