Skip to content

Commit f3e6592

Browse files
authored
Update README.md
1 parent 32aa02f commit f3e6592

File tree

1 file changed

+17
-2
lines changed

1 file changed

+17
-2
lines changed

optillm/autothink/README.md

Lines changed: 17 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ AutoThink combines several advanced techniques to optimize the thinking process
1515

1616
### 1. Query Classification
1717

18-
AutoThink uses the `adaptive-classifier/llm-router` model to classify incoming queries:
18+
AutoThink uses the `adaptive-classifier/llm-router` [model](https://huggingface.co/adaptive-classifier/llm-router) to classify incoming queries:
1919

2020
- **HIGH**: Complex queries requiring deep reasoning, multi-step calculations, or thorough exploration
2121
- **LOW**: Simpler queries requiring less extensive reasoning
@@ -29,7 +29,7 @@ Based on the classification, AutoThink allocates different token budgets for the
2929

3030
### 3. Steering Vectors
3131

32-
AutoThink uses pre-extracted steering vectors from datasets like `codelion/Qwen3-0.6B-pts-steering-vectors`. These vectors represent different reasoning patterns:
32+
AutoThink uses pre-extracted steering vectors from [datasets](https://huggingface.co/datasets?other=pts) like `codelion/Qwen3-0.6B-pts-steering-vectors`. These vectors represent different reasoning patterns:
3333

3434
- **Depth and thoroughness**: Encourages detailed, step-by-step reasoning
3535
- **Numerical accuracy**: Promotes precise calculations and verification
@@ -93,3 +93,18 @@ response = autothink_decode(
9393
- **Enhanced Reasoning**: Steering vectors guide the model toward better reasoning patterns
9494
- **Efficiency**: Better performance without increasing model size
9595
- **Customizability**: Can be tailored for different domains using domain-specific steering vector datasets
96+
97+
98+
## Citation
99+
100+
If you use this approach in your research, please cite:
101+
102+
```bibtex
103+
@article{autothink,
104+
title={AutoThink: efficient inference for reasoning LLMs},
105+
author={Sharma, Asankhaya},
106+
journal={SSRN Artificial Intelligence eJournal},
107+
year={2025},
108+
url = {https://dx.doi.org/10.2139/ssrn.5253327}
109+
}
110+
```

0 commit comments

Comments
 (0)