| { | |
| "base_model": "Qwen/Qwen2.5-1.5B-Instruct", | |
| "model_size": "1.54B parameters", | |
| "model_type": "instruction-tuned LLM", | |
| "dimensions": 1024, | |
| "model2vec_version": "0.7.0", | |
| "custom_corpus": false, | |
| "note": "First-ever Model2Vec distillation of instruction-tuned LLM", | |
| "output_dir": "models/qwen25-deposium-1024d", | |
| "expected_quality": "0.75-0.85+", | |
| "expected_speedup": "500-1000x vs full LLM", | |
| "final_size_mb": 65, | |
| "multilingual": true, | |
| "instruction_aware": true, | |
| "context_length": 32768, | |
| "unique_capabilities": [ | |
| "Instruction-aware embeddings", | |
| "Superior semantic understanding", | |
| "Conversational context", | |
| "Code understanding", | |
| "Idiom recognition" | |
| ], | |
| "advantages_vs_competitors": { | |
| "vs_qwen3_embedding": "10x smaller, instruction-aware", | |
| "vs_gemma_768d": "larger base model, instruction-tuned", | |
| "vs_qwen3_256d": "higher quality expected, instruction-aware" | |
| }, | |
| "instruction_awareness_score": 0.9198921983524119 | |
| } |