Created at 10am, Feb 13
andthattoocs.CL
0
VeRA: Vector-based Random Matrix Adaptation
gl5XQxrYWrqS4xgak24W1vy8ORGilOBHPU_K2CjVfmo
File Type
CUSTOM
Entry Count
0
Embed. Model
BAAI/bge-base-en-v1.5
Index Type
hnsw

Low-rank adapation (LoRA) is a popular method that reduces the number of trainable parameters when finetuning large language models, but still faces acute storage challenges when scaling to even larger models or deploying numerous per-user or per-task adapted models. In this work, we present Vector-based Random Matrix Adaptation (VeRA), which significantly reduces the number of trainable parameters compared to LoRA, yet maintains the same performance. It achieves this by using a single pair of low-rank matrices shared across all layers and learning small scaling vectors instead. We demonstrate its effectiveness on the GLUE and E2E benchmarks, image classification tasks, and show its application in instruction-tuning of 7B and 13B language models.

How to Retrieve?

# Query

curl -X POST "https://search.dria.co/hnsw/query" \
-H "x-api-key: <YOUR_API_KEY>" \
-H "Content-Type: application/json" \
-d '{"vector": [0.123, 0.5236], "top_n": 10, "contract_id": "gl5XQxrYWrqS4xgak24W1vy8ORGilOBHPU_K2CjVfmo", "level": 2}'