yuhangzang commited on
Commit
3a0ec16
Β·
verified Β·
1 Parent(s): f5e00bb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -5
README.md CHANGED
@@ -16,15 +16,20 @@ tags:
16
  πŸ“–<a href="https://arxiv.org/abs/2509.22647">Paper</a> | 🏠<a href="https://github.com/InternLM/CapRL">Github</a> |πŸ€—<a href="https://huggingface.co/internlm/CapRL-3B">CapRL-3B Model</a> |πŸ€—<a href="https://huggingface.co/yuhangzang/CapRL-InternVL3.5-8B">CapRL-InternVL3.5-8B Model</a> |
17
  πŸ€—<a href="https://huggingface.co/datasets/internlm/CapRL-2M">CapRL-2M Dataset</a>
18
 
19
- πŸ€—<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | πŸ€—<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a>
20
 
21
- πŸ“’ News
22
 
23
- - πŸš€ Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL. **Its performance significantly surpasses that of Qwen2.5-VL-72B**.
 
 
 
24
 
25
- - πŸš€ We are working on even stronger base models and upgrading our training recipe β€” stay tuned!
 
 
 
26
 
27
- - πŸš€ CapRL-3B-GGUF is static quants version, and CapRL-3B-i1-GGUF is weighted/imatrix quants version. Thanks for their contribution!
28
 
29
  ## Introduction
30
  We are excited to introduce CapRL-3B, a lightweight 3B image captioner that achieves perception capabilities comparable to Qwen2.5-VL-72B.
 
16
  πŸ“–<a href="https://arxiv.org/abs/2509.22647">Paper</a> | 🏠<a href="https://github.com/InternLM/CapRL">Github</a> |πŸ€—<a href="https://huggingface.co/internlm/CapRL-3B">CapRL-3B Model</a> |πŸ€—<a href="https://huggingface.co/yuhangzang/CapRL-InternVL3.5-8B">CapRL-InternVL3.5-8B Model</a> |
17
  πŸ€—<a href="https://huggingface.co/datasets/internlm/CapRL-2M">CapRL-2M Dataset</a>
18
 
19
+ πŸ€—<a href="https://huggingface.co/collections/long-xing1/caprl-68d64ac32ded31596c36e189">CapRL Collection</a> | πŸ€—<a href="https://huggingface.co/papers/2509.22647">Daily Paper</a> ο½œπŸ€—<a href="https://huggingface.co/mradermacher/CapRL-3B-GGUF">CapRL-3B-GGUF</a> ο½œπŸ€—<a href="https://huggingface.co/mradermacher/CapRL-3B-i1-GGUF">CapRL-3B-i1-GGUF</a>
20
 
21
+ We are working on even stronger base models and upgrading our training recipe β€” stay tuned!
22
 
23
+ ## πŸ“’ News
24
+ - πŸš€ [16/15/2025] We are excited to announce the release of **CapRL-InternVL3.5-8B**, whose image captioning capability outperforms Qwen2.5-VL-72B!
25
+ - πŸš€ [10/15/2025] We release QA curation code.
26
+ - πŸš€ [09/25/2025] We release **CapRL** repository, model, evaluation code and dataset.
27
 
28
+ Based on the same recipe as CapRL-3B, we used InternVL3.5-8B as the policy model and obtained **CapRL-InternVL3.5-8B** through CapRL.
29
+
30
+
31
+ CapRL-3B-GGUF is static quants version, and CapRL-3B-i1-GGUF is weighted/imatrix quants version. Thanks for their contribution!
32
 
 
33
 
34
  ## Introduction
35
  We are excited to introduce CapRL-3B, a lightweight 3B image captioner that achieves perception capabilities comparable to Qwen2.5-VL-72B.