Created at 8am, Feb 13
andthattoocs.CL
0
Retrieval-Augmented Thought Process as Sequential Decision Making
U5dypscT_WLZgiqbGtty-FXqa2c6nZWVpU0CGk6d4fk
File Type
CUSTOM
Entry Count
1582
Embed. Model
BAAI/bge-base-en-v1.5
Index Type
hnsw

Large Language Models (LLMs) have demonstrated their strong ability to assist people and show "sparks of intelligence". However, several open challenges hinder their wider application: such as concerns over privacy, tendencies to produce hallucinations, and difficulties in handling long contexts. In this work, we address those challenges by introducing the Retrieval-Augmented Thought Process (RATP). Given access to external knowledge, RATP formulates the thought generation of LLMs as a multiple-step decision process. To optimize such a thought process, RATP leverages Monte-Carlo Tree Search, and learns a Q-value estimator that permits cost-efficient inference. In addressing the task of question-answering with private data, where ethical and security concerns limit LLM training methods, RATP achieves a 50% improvement over existing in-context retrieval-augmented language models.

How to Retrieve?

# Query

curl -X POST "https://search.dria.co/hnsw/query" \
-H "x-api-key: <YOUR_API_KEY>" \
-H "Content-Type: application/json" \
-d '{"vector": [0.123, 0.5236], "top_n": 10, "contract_id": "U5dypscT_WLZgiqbGtty-FXqa2c6nZWVpU0CGk6d4fk", "level": 2}'