hokar3361's picture
Update README.md
ef2340c verified
---
language:
- en
- code
language_bcp47:
- en
- javascript
license: apache-2.0
tags:
- text-generation
- code
- javascript
- coding-assistant
- fine-tuning
- merged
- unsloth
- gpt-oss
- vllm
base_model: openai/gpt-oss-20b
library_name: transformers
pipeline_tag: text-generation
model-index:
- name: gpt-oss-coder-v0.1-javascript
results: []
---
# gpt-oss-coder-v0.1-javascript
A **language-specialized coding model for JavaScript**, fine-tuned from OpenAI's open-weight **gpt-oss** base with **very small, curated JS data** using **Unsloth**.
This release prioritizes **practical code generation quality** over benchmark scores. The model weights have been **merged** and are ready for deployment.
> **Status**: Experimental preview (`v0.1-javascript`)
> **Focus**: JS coding tasks (function-level completion, small refactors, idiomatic patterns)
> **Testing**: Currently undergoing validation with vLLM deployment
> **Note**: This repository contains merged weights, not LoRA adapters
---
## Model Details
- **Model type**: Causal LM (decoder-only), JS-specialized fine-tune
- **Base model**: `openai/gpt-oss-20b` (open-weight, Apache-2.0)
- **Fine-tuning**: LoRA via **Unsloth**, weights merged post-training
- **License**: Apache-2.0 (derivative weights released under Apache-2.0)
- **Author / Maintainer**: `hokar3361`
- **Intended Languages**: JavaScript (ES6+); English prompts recommended
- **Weight Format**: Merged (full model weights)
---
## Intended Use & Limitations
### Intended Use
- Code completion and synthesis for **JavaScript**
- Small refactors, idiomatic rewrites, test scaffolding, JSDoc/docstrings
- Snippet-level reasoning and bug fixes
### Out of Scope / Limitations
- Not a substitute for static analysis, linters, or security review
- May hallucinate APIs or types; verify before production use
- Trained on **small** domain data → expect gaps on rare frameworks or edge APIs
---
## Quickstart
### 1. Start vLLM Server
Since this repository contains **merged weights**, you can run directly with vLLM:
```bash
vllm serve hokar3361/gpt-oss-coderjs-v0.1 \
--async-scheduling \
--max-model-len 16000 \
--gpu-memory-utilization 0.90
```
**Recommended**: Use `--max-model-len 16000` for optimal context handling.
### 2. Client Usage (Recommended)
Use the **OpenAI Python client** to call the vLLM server:
```python
from openai import OpenAI
# Point to your vLLM server
client = OpenAI(
base_url="http://localhost:8000/v1",
api_key="dummy" # vLLM doesn't require auth by default
)
response = client.completions.create(
model="hokar3361/gpt-oss-coderjs-v0.1",
prompt="// JavaScript function to validate email addresses\nfunction validateEmail(email) {",
# DO NOT specify temperature or max_tokens - let the model use defaults
)
print(response.choices[0].text)
```
**Important**:
- **Do not specify** `temperature` or `max_tokens` parameters - the model performs best with default values
- Use the OpenAI Python client for best compatibility and stability
---
## Testing & Validation
### Current Status
The model is currently being validated using vLLM deployment. Initial testing shows **improved performance** compared to pre-fine-tuning baseline.
### Evaluation Methodology
- **Test Set**: 50 programming questions from GitHub and Stack Overflow
- **Judges**: GPT-5 and Claude Opus for response quality assessment
- **Preliminary Results**: The fine-tuned model demonstrates better code generation quality on JavaScript-specific tasks compared to the base model
- **Note**: Full benchmark validation is still in progress
---
## Acknowledgements
This work was made possible thanks to the open-weight release of **gpt-oss** by OpenAI, which provided a strong foundation under the Apache-2.0 license.
Special thanks to the open-source community around **Unsloth** for enabling memory-efficient and rapid LoRA fine-tuning on limited hardware.
We also thank the **Hugging Face** and **vLLM** ecosystems for lowering the barrier to experimentation.
---
## Disclaimer & Experimental Status
This model (`v0.1-javascript`) is highly experimental:
- **Small data**: Fine-tuned on a very small JavaScript-focused dataset, mainly to validate the workflow and feasibility of language specialization.
- **Not production-ready**: The model may generate incomplete, insecure, or non-idiomatic code; do not rely on it for production use without careful review.
- **Testing in progress**: While initial results from GPT-5 and Opus evaluation show improvements, comprehensive benchmarking is ongoing.
- **Early stage**: This is only an initial exploration; future versions with larger, more diverse training corpora are expected to improve stability and coverage.
We share this release to contribute to the community and gather early feedback.
**Use responsibly, validate outputs, and treat this as a proof-of-concept.**