Helion-V1
Collection
Helion version 1 series
β’
3 items
β’
Updated
β’
2
Helion-V1-Embeddings is a lightweight text embedding model designed for semantic similarity, search, and retrieval tasks. It converts text into dense vector representations optimized for the Helion ecosystem.
| Parameter | Value | Description |
|---|---|---|
| Architecture | BERT-based | 6-layer transformer encoder |
| Hidden Size | 384 | Dimension of hidden layers |
| Attention Heads | 12 | Number of attention heads |
| Intermediate Size | 1536 | Feed-forward layer size |
| Vocab Size | 30,522 | WordPiece vocabulary |
| Max Position Embeddings | 512 | Maximum sequence length |
| Pooling Strategy | Mean Pooling | Average of token embeddings |
| Output Dimension | 384 | Final embedding size |
| Total Parameters | ~22.7M | Trainable parameters |
| Model Size | ~80MB | Disk footprint |
Helion-V1-Embeddings is designed for:
from sentence_transformers import SentenceTransformer
# Load model
model = SentenceTransformer('DeepXR/Helion-V1-embeddings')
# Encode sentences
sentences = [
"How do I reset my password?",
"What is the process for password recovery?",
"I forgot my login credentials"
]
embeddings = model.encode(sentences)
print(embeddings.shape) # (3, 384)
from sentence_transformers import SentenceTransformer, util
model = SentenceTransformer('DeepXR/Helion-V1-embeddings')
# Encode query and documents
query = "How to train a machine learning model?"
documents = [
"Machine learning training requires data preprocessing",
"The best way to cook pasta is boiling water",
"Neural networks need proper hyperparameter tuning"
]
query_embedding = model.encode(query)
doc_embeddings = model.encode(documents)
# Calculate similarity
similarities = util.cos_sim(query_embedding, doc_embeddings)
print(similarities)
from sentence_transformers import SentenceTransformer
import faiss
import numpy as np
model = SentenceTransformer('DeepXR/Helion-V1-embeddings')
# Create embeddings
documents = ["doc1", "doc2", "doc3"]
embeddings = model.encode(documents)
# Create FAISS index
dimension = embeddings.shape[1]
index = faiss.IndexFlatL2(dimension)
index.add(embeddings.astype('float32'))
# Search
query_embedding = model.encode(["search query"])
distances, indices = index.search(query_embedding.astype('float32'), k=3)
| Task | Score | Notes |
|---|---|---|
| STS Benchmark | ~0.78 | Semantic Textual Similarity |
| Retrieval (BEIR) | ~0.42 | Average across datasets |
| Speed (CPU) | ~2000 sentences/sec | Batch size 32 |
| Speed (GPU) | ~15000 sentences/sec | Batch size 128 |
Note: These are approximate values. Actual performance may vary.
The model was fine-tuned on:
| Model | Dim | Speed | Accuracy | Size |
|---|---|---|---|---|
| Helion-V1-Embeddings | 384 | Fast | Good | 80MB |
| all-MiniLM-L6-v2 | 384 | Fast | Good | 80MB |
| all-mpnet-base-v2 | 768 | Medium | Better | 420MB |
| text-embedding-ada-002 | 1536 | API | Best | API |
from langchain.embeddings import HuggingFaceEmbeddings
embeddings = HuggingFaceEmbeddings(
model_name="DeepXR/Helion-V1-embeddings"
)
text = "This is a sample document"
embedding = embeddings.embed_query(text)
from llama_index.embeddings import HuggingFaceEmbedding
embed_model = HuggingFaceEmbedding(
model_name="DeepXR/Helion-V1-embeddings"
)
embeddings = embed_model.get_text_embedding("Hello world")
@misc{helion-v1-embeddings,
author = {DeepXR},
title = {Helion-V1-Embeddings: Lightweight Text Embedding Model},
year = {2025},
publisher = {HuggingFace},
url = {https://huggingface.co/DeepXR/Helion-V1-embeddings}
}
DeepXR Team
Base model
sentence-transformers/all-MiniLM-L6-v2