Improve model card: Add metadata, structured paper link, project page, and code links

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +25 -1
README.md CHANGED
@@ -1,4 +1,28 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
- arxiv.org/abs/2512.03794
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: image-text-to-text
4
+ library_name: transformers
5
  ---
6
+
7
+ # AdaptVision: Efficient Vision-Language Models via Adaptive Visual Acquisition
8
+
9
+ AdaptVision is an efficient Vision-Language Model (VLM) paradigm designed to achieve adaptive visual token acquisition through a coarse-to-fine approach. Inspired by human active vision mechanisms, this model addresses the significant computational overhead in VLMs by autonomously determining the minimum number of visual tokens required for each sample. It selectively acquires additional visual information by invoking a bounding box tool to crop key regions when necessary.
10
+
11
+ The model was presented in the paper:
12
+ [AdaptVision: Efficient Vision-Language Models via Adaptive Visual Acquisition](https://arxiv.org/abs/2512.03794)
13
+
14
+ For more details, please visit the [project page](https://adaptvision.github.io/).
15
+ The official code can be found on the [GitHub repository](https://github.com/AdaptVision/AdaptVision).
16
+
17
+ ## Citation
18
+
19
+ If you find this project useful in your research, please consider citing:
20
+
21
+ ```bibtex
22
+ @article{lin2025adapt,
23
+ title={AdaptVision: Efficient Vision-Language Models via Adaptive Visual Acquisition},
24
+ author={Zichuan Lin and Yicheng Liu and Yang Yang and Lvfang Tao and Deheng Ye},
25
+ journal={arXiv preprint arXiv:2512.03794},
26
+ year={2025}
27
+ }
28
+ ```