AI & ML interests

Exploring Extreme Quantization techniques !

Parveshiiiiย 
posted an update 15 days ago
view post
Post
1592
Another banger from XenArcAI! ๐Ÿ”ฅ

Weโ€™re thrilled to unveil three powerful new releases that push the boundaries of AI research and development:

๐Ÿ”— XenArcAI/SparkEmbedding-300m

- A lightning-fast embedding model built for scale.
- Optimized for semantic search, clustering, and representation learning.

๐Ÿ”— XenArcAI/CodeX-7M-Non-Thinking

- A massive dataset of 7 million code samples.
- Designed for training models on raw coding patterns without reasoning layers.

๐Ÿ”— XenArcAI/CodeX-2M-Thinking

- A curated dataset of 2 million code samples.
- Focused on reasoning-driven coding tasks, enabling smarter AI coding assistants.

Together, these projects represent a leap forward in building smarter, faster, and more capable AI systems.

๐Ÿ’ก Innovation meets dedication.
๐ŸŒ Knowledge meets responsibility.


Parveshiiiiย 
posted an update 22 days ago
view post
Post
3013
SparkEmbedding - SoTA cross lingual retrieval

Iam very happy to announce our latest embedding model sparkembedding-300m base on embeddinggemma-300m we fine tuned it on 1m extra examples spanning over 119 languages and result is this model achieves exceptional cross lingual retrieval

Model: XenArcAI/SparkEmbedding-300m
Parveshiiiiย 
posted an update about 2 months ago
view post
Post
195
AIRealNet - SoTA - Image detection model

Weโ€™re proud to release AIRealNet โ€” a binary image classifier built to detect whether an image is AI-generated or a real human photograph. Based on SwinV2 and fine-tuned on the AI-vs-Real dataset, this model is optimized for high-accuracy classification across diverse visual domains.

If you care about synthetic media detection or want to explore the frontier of AI vs human realism, weโ€™d love your support. Please like the model and try it out. Every download helps us improve and expand future versions.

Model page: XenArcAI/AIRealNet
Parveshiiiiย 
posted an update about 2 months ago
view post
Post
4481
Ever wanted an openโ€‘source deep research agent? Meet Deepresearchโ€‘Agent ๐Ÿ”๐Ÿค–

1. Multiโ€‘step reasoning: Reflects between steps, fills gaps, iterates until evidence is solid.

2. Researchโ€‘augmented: Generates queries, searches, synthesizes, and cites sources.

3. Fullstack + LLMโ€‘friendly: React/Tailwind frontend, LangGraph/FastAPI backend; works with OpenAI/Gemini.


๐Ÿ”— GitHub: https://github.com/Parveshiiii/Deepresearch-Agent
Parveshiiiiย 
posted an update 2 months ago
view post
Post
3098
๐Ÿš€ Big news from XenArcAI!

Weโ€™ve just released our new dataset: **Bhagwatโ€‘Gitaโ€‘Infinity** ๐ŸŒธ๐Ÿ“–

โœจ Whatโ€™s inside:
- Verseโ€‘aligned Sanskrit, Hindi, and English
- Clean, structured, and ready for ML/AI projects
- Perfect for research, education, and openโ€‘source exploration

๐Ÿ”— Hugging Face: XenArcAI/Bhagwat-Gita-Infinity

Letโ€™s bring timeless wisdom into modern AI together ๐Ÿ™Œ
Parveshiiiiย 
posted an update 2 months ago
view post
Post
2450
๐Ÿš€ New Release from XenArcAI
Weโ€™re excited to introduce AIRealNet โ€” our SwinV2โ€‘based image classifier built to distinguish between artificial and real images.

โœจ Highlights:
- Backbone: SwinV2
- Input size: 256ร—256
- Labels: artificial vs. real
- Performance: Accuracy 0.999 | F1 0.999 | Val Loss 0.0063

This model is now live on Hugging Face:
๐Ÿ‘‰ XenArcAI/AIRealNet

We built AIRealNet to push forward openโ€‘source tools for authenticity detection, and we canโ€™t wait to see how the community uses it.
Abhaykoulย 
posted an update 3 months ago
view post
Post
3083
๐Ÿš€ Ever dreamed of training your own Large Language Model from scratch? What if I told you it doesn't require a supercomputer or PhD in ML? ๐Ÿคฏ

Introducing LLM Trainer - the educational framework that makes LLM training accessible to EVERYONE! Whether you're on a CPU-only laptop or scaling to distributed GPUs, we've got you covered. ๐Ÿ’ปโžก๏ธ๐Ÿ–ฅ๏ธ

Why LLM Trainer? Because existing tools are either too simplistic (hiding the magic) or too complex (requiring expert knowledge). We bridge the gap with:

๐ŸŽ“ Educational transparency - every component built from scratch with clear code
๐Ÿ’ป CPU-first approach - start training immediately, no GPU needed
๐Ÿ”ง Full customization - modify anything you want
๐Ÿ“ˆ Seamless scaling - from laptop to cluster without code changes
๐Ÿค HuggingFace integration - works with existing models & tokenizers

Key highlights:
โœ… Built-in tokenizers (BPE, WordPiece, HF wrappers)
โœ… Complete Transformer implementation from scratch
โœ… Optimized for CPU training
โœ… Advanced features: mixed precision, gradient checkpointing, multiple generation strategies
โœ… Comprehensive monitoring & metrics

Perfect for:
- Students learning transformers
- Researchers prototyping new ideas
- Developers building domain-specific models

Ready to train your first LLM? It's easier than you think!

๐Ÿ”— Check it out: https://github.com/HelpingAI/llm-trainer
๐Ÿ“š Docs: Getting Started Guide
๐Ÿ’ฌ Join the community: GitHub Discussions

#AI #MachineLearning #LLM #DeepLearning #OpenSource #Python #HuggingFace #NLP

Special thanks to HuggingFace and PyTorch teams for the amazing ecosystem! ๐Ÿ™
  • 1 reply
ยท
Parveshiiiiย 
posted an update 4 months ago
view post
Post
1109
๐Ÿš€ Just Dropped: MathX-5M โ€” Your Gateway to Math-Savvy GPTs

๐Ÿ‘จโ€๐Ÿ”ฌ Wanna fine-tune your own GPT for math?
๐Ÿง  Building a reasoning agent that actually *thinks*?
๐Ÿ“Š Benchmarking multi-step logic across domains?

Say hello to [**MathX-5M**]( XenArcAI/MathX-5M) โ€” a **5 million+ sample** dataset crafted for training and evaluating math reasoning models at scale.

Built by **XenArcAI**, itโ€™s optimized for:
- ๐Ÿ” Step-by-step reasoning with , , and formats
- ๐Ÿงฎ Coverage from arithmetic to advanced algebra and geometry
- ๐Ÿงฐ Plug-and-play with Gemma, Qwen, Mistral, and other open LLMs
- ๐Ÿงต Compatible with Harmony, Alpaca, and OpenChat-style instruction formats

Whether you're prototyping a math tutor, testing agentic workflows, or just want your GPT to solve equations like a proโ€”**MathX-5M is your launchpad**.

๐Ÿ”— Dive in: ( XenArcAI/MathX-5M)

Letโ€™s make open-source models *actually* smart at math.
#FineTuneYourGPT #MathX5M #OpenSourceAI #LLM #XenArcAI #Reasoning #Gemma #Qwen #Mistral

alielfilali01ย 
posted an update 4 months ago
Abhaykoulย 
posted an update 4 months ago
view post
Post
4129
๐Ÿš€ Dhanishtha-2.0-preview-0825 Is Here

The Intermediate Thinking Model just leveled up again.

With sharper reasoning, better tool use, and expanded capabilities, Dhanishtha-2.0-preview-0825 is now live and ready to impress.

๐Ÿง  What Makes Dhanishtha Special?
Unlike typical CoT models that only thinks one time, Dhanishtha thinks iteratively:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

๐Ÿ”— Try it now: HelpingAI/Dhanishtha-2.0-preview-0825

๐Ÿ”ž Dhanishtha NSFW Preview

For those exploring more expressive and immersive roleplay scenarios, weโ€™re also releasing:

HelpingAI/Dhanishtha-nsfw
A specialized version tuned for adult-themed interactions and character-driven roleplay.

๐Ÿ”— Explore it here: HelpingAI/Dhanishtha-nsfw

๐Ÿ’ฌ You can also try all of these live at chat.helpingai.co
ยท
Parveshiiiiย 
posted an update 4 months ago
view post
Post
1095
๐Ÿš€ Launch Alert: Dev-Stack-Agents
Meet your 50-agent senior AI team โ€” principal-level experts in engineering, AI, DevOps, security, product, and more โ€” all bundled into one modular repo.

+ Code. Optimize. Scale. Secure.
- Full-stack execution, Claude-powered. No human bottlenecks.


๐Ÿ”ง Built for Claude Code
Seamlessly plug into Claudeโ€™s dev environment:

* ๐Ÿง  Each .md file = a fully defined expert persona
* โš™๏ธ Claude indexes them as agents with roles, skills & strategy
* ๐Ÿค– You chat โ†’ Claude auto-routes to the right agent(s)
* โœ๏ธ Want precision? Just call @agent-name directly
* ๐Ÿ‘ฅ Complex task? Mention multiple agents for team execution

Examples:

"@security-auditor please review auth flow for risks"
"@cloud-architect + @devops-troubleshooter โ†’ design a resilient multi-region setup"
"@ai-engineer + @legal-advisor โ†’ build a privacy-safe RAG pipeline"


๐Ÿ”— https://github.com/Parveshiiii/Dev-Stack-Agents
MIT License | Claude-Ready | PRs Welcome

erikkaumย 
posted an update 5 months ago
view post
Post
2608
ZML just released a technical preview of their new Inference Engine: LLMD.

- Just 2.4GB container, which means fast startup times and efficient autoscaling
- Cross-Platform GPU Support: works on both NVIDIA and AMD GPUs.
- written in Zig

I just tried it out and deployed it on Hugging Face Inference Endpoints and wrote a quick guide ๐Ÿ‘‡ You can try it in like 5 minutes!

https://huggingface.co/blog/erikkaum/test-driving-llmd-inference-engine
  • 1 reply
ยท
erikkaumย 
posted an update 5 months ago
view post
Post
2124
We just released native support for @SGLang and @vllm-project in Inference Endpoints ๐Ÿ”ฅ

Inference Endpoints is becoming the central place where you deploy high performance Inference Engines.

And that provides the managed infra for it. Instead of spending weeks configuring infrastructure, managing servers, and debugging deployment issues, you can focus on what matters most: your AI model and your users ๐Ÿ™Œ
Abhaykoulย 
posted an update 5 months ago
view post
Post
3157
๐ŸŽ‰ Dhanishtha-2.0-preview-0725 is Now Live

The Intermediate Thinking Model just got even better.
With the new update, Dhanishtha is now sharper, smarter, and trained further on tool use

๐Ÿง  What Makes Dhanishtha Different?
Unlike standard COT models that give one-shot responses, Dhanishtha thinks in layers:

> Think โ†’ Answer โ†’ Rethink โ†’ Improve โ†’ Rethink again if needed.

HelpingAI/Dhanishtha-2.0-preview-0725
Parveshiiiiย 
posted an update 5 months ago
view post
Post
2704
๐Ÿง  Glimpses of AGI โ€” A Vision for All Humanity
What if AGI wasnโ€™t just a distant dreamโ€”but a blueprint already unfolding?

Iโ€™ve just published a deep dive called Glimpses of AGI, exploring how scalable intelligence, synthetic reasoning, and alignment strategies are paving a new path forward. This isnโ€™t your average tech commentaryโ€”itโ€™s a bold vision for conscious AI systems that reason, align, and adapt beyond narrow tasks.

๐Ÿ” Read it, upvote it if it sparks something, and letโ€™s ignite a collective conversation about the future of AGI.

https://huggingface.co/blog/Parveshiiii/glimpses-of-agi


Parveshiiiiย 
posted an update 5 months ago
view post
Post
2855
๐Ÿง  MathX-5M by XenArcAI โ€” Scalable Math Reasoning for Smarter LLMs

Introducing MathX-5M, a high-quality, instruction-tuned dataset built to supercharge mathematical reasoning in large language models. With 5 million rigorously filtered examples, it spans everything from basic arithmetic to advanced calculusโ€”curated from public sources and enhanced with synthetic data.

๐Ÿ” Key Highlights:
- Step-by-step reasoning with verified answers
- Covers algebra, geometry, calculus, logic, and more
- RL-validated correctness and multi-stage filtering
- Ideal for fine-tuning, benchmarking, and educational AI

๐Ÿ“‚ - XenArcAI/MathX-5M


  • 1 reply
ยท
Abhaykoulย 
posted an update 5 months ago
view post
Post
3111
๐ŸŽ‰ Dhanishtha 2.0 Preview is Now Open Source!

The world's first Intermediate Thinking Model is now available to everyone!

Dhanishtha 2.0 Preview brings revolutionary intermediate thinking capabilities to the open-source community. Unlike traditional reasoning models that think once, Dhanishtha can think, answer, rethink, answer again, and continue rethinking as needed using multiple blocks between responses.

๐Ÿš€ Key Features
- Intermediate thinking: Think โ†’ Answer โ†’ Rethink โ†’ Answer โ†’ Rethink if needed...
- Token efficient: Uses up to 79% fewer tokens than DeepSeek R1 on similar queries
- Transparent thinking: See the model's reasoning process in real-time
- Open source: Freely available for research and development


HelpingAI/Dhanishtha-2.0-preview
https://helpingai.co/chat
  • 1 reply
ยท
Abhaykoulย 
posted an update 5 months ago
view post
Post
4636
Introducing Dhanishtha 2.0: World's first Intermediate Thinking Model

Dhanishtha 2.0 is the world's first LLM designed to think between the responses. Unlike other Reasoning LLMs, which think just once.

Dhanishtha can think, rethink, self-evaluate, and refine in between responses using multiple <think> blocks.
This technique makes it Hinghlt Token efficient it Uses up to 79% fewer tokens than DeepSeek R1
---

You can try our model from: https://helpingai.co/chat
Also, we're gonna Open-Source Dhanistha on July 1st.

---
For Devs:
๐Ÿ”‘ Get your API key at https://helpingai.co/dashboard
from HelpingAI import HAI  # pip install HelpingAI==1.1.1
from rich import print

hai = HAI(api_key="hl-***********************")

response = hai.chat.completions.create(
    model="Dhanishtha-2.0-preview",
    messages=[{"role": "user", "content": "What is the value of โˆซ0โˆž๐‘ฅ3/๐‘ฅโˆ’1๐‘‘๐‘ฅ ?"}],
    stream=True,
    hide_think=False # Hide or show models thinking
)

for chunk in response:
    print(chunk.choices[0].delta.content, end="", flush=True)
  • 2 replies
ยท