Tonic commited on
Commit
a0fca33
·
verified ·
1 Parent(s): 2641514

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -26
README.md CHANGED
@@ -20,16 +20,16 @@ datasets:
20
 
21
 
22
 
23
- author: GPT-OSS Fine-tuner
24
 
25
 
26
- experiment_name: med-track
27
 
28
 
29
- trackio_url: N/A
30
 
31
 
32
- dataset_repo: HuggingFaceH4/Multilingual-Thinking
33
 
34
 
35
  hardware: "GPU (H100/A100)"
@@ -47,17 +47,17 @@ batch_size: 4
47
  learning_rate: 2e-4
48
 
49
 
50
- max_epochs: 1
51
 
52
 
53
  max_seq_length: 2048
54
 
55
 
56
 
57
- dataset_size: ~80K samples
58
 
59
 
60
- dataset_format: Chat format
61
 
62
 
63
  gradient_accumulation_steps: 16
@@ -66,15 +66,14 @@ gradient_accumulation_steps: 16
66
 
67
  # med-gpt-oss-20b
68
 
69
- A fine-tuned version of OpenAI's GPT-OSS-20B model for multilingual reasoning tasks.
70
 
71
  ## Model Details
72
 
73
- - **Base Model**: SmolLM3-3B
74
  - **Model Type**: Causal Language Model
75
- - **Languages**: English, French
76
  - **License**: Apache 2.0
77
- - **Fine-tuned**: Yes
78
 
79
 
80
  ## Usage
@@ -104,7 +103,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
104
 
105
  ### Training Configuration
106
  - **Base Model**: openai/gpt-oss-20b
107
- - **Dataset**: HuggingFaceH4/Multilingual-Thinking
108
  - **Training Config**: GPT-OSS Configuration
109
  - **Trainer Type**: SFTTrainer
110
 
@@ -134,9 +133,8 @@ This is a fine-tuned version of the SmolLM3-3B model with the following specific
134
  ## Performance
135
 
136
  The model provides:
137
- - **Text Generation**: High-quality text generation capabilities
138
- - **Conversation**: Natural conversation abilities
139
- - **Multilingual**: Support for English and French
140
 
141
 
142
  ## Limitations
@@ -150,26 +148,23 @@ The model provides:
150
  ## Training Data
151
 
152
  The model was fine-tuned on:
153
- - **Dataset**: HuggingFaceH4/Multilingual-Thinking
154
- - **Size**: ~80K samples
155
- - **Format**: Chat format
156
- - **Languages**: English, French
157
 
158
- ## Evaluation
159
 
160
- The model was evaluated using:
161
- - **Metrics**: Loss, perplexity, and qualitative assessment
162
- - **Monitoring**: Real-time tracking via Trackio
163
- - **Validation**: Regular validation during training
164
 
165
  ## Citation
166
 
167
  If you use this model in your research, please cite:
168
 
169
  ```bibtex
170
- @misc{smollm3-fine-tuned,
171
  title={{med-gpt-oss-20b}},
172
- author={GPT-OSS Fine-tuner},
173
  year={2024},
174
  url={https://huggingface.co/Tonic/med-gpt-oss-20b}
175
  }
 
20
 
21
 
22
 
23
+ author: Joseph "Tonic" Pollack
24
 
25
 
26
+ experiment_name: exp_20250809_122413
27
 
28
 
29
+ trackio_url: https://huggingface.co/spaces/Tonic/track-tonic
30
 
31
 
32
+ dataset_repo: FreedomIntelligence/medical-o1-reasoning-SFT
33
 
34
 
35
  hardware: "GPU (H100/A100)"
 
47
  learning_rate: 2e-4
48
 
49
 
50
+ max_epochs: 2
51
 
52
 
53
  max_seq_length: 2048
54
 
55
 
56
 
57
+ dataset_size: ~20K samples
58
 
59
 
60
+ dataset_format: Reasoning
61
 
62
 
63
  gradient_accumulation_steps: 16
 
66
 
67
  # med-gpt-oss-20b
68
 
69
+ A fine-tuned version of OpenAI's GPT-OSS-20B model for medical reasoning and instruction following.
70
 
71
  ## Model Details
72
 
73
+ - **Base Model**: openai/gpt-oss-20B
74
  - **Model Type**: Causal Language Model
75
+ - **Languages**: English
76
  - **License**: Apache 2.0
 
77
 
78
 
79
  ## Usage
 
103
 
104
  ### Training Configuration
105
  - **Base Model**: openai/gpt-oss-20b
106
+ - **Dataset**: FreedomIntelligence/medical-o1-reasoning-SFT
107
  - **Training Config**: GPT-OSS Configuration
108
  - **Trainer Type**: SFTTrainer
109
 
 
133
  ## Performance
134
 
135
  The model provides:
136
+ - **Medical Reasoning**: High-quality medical reasoning
137
+ - **Conversation**: Medical instruction following
 
138
 
139
 
140
  ## Limitations
 
148
  ## Training Data
149
 
150
  The model was fine-tuned on:
151
+ - **Dataset**: FreedomIntelligence/medical-o1-reasoning-SFT
152
+ - **Size**: ~20K samples
153
+ - **Format**: reasoning
154
+ - **Languages**: English
155
 
156
+ ## Monitoring
157
 
158
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/62a3bb1cd0d8c2c2169f0b88/rM6DkvUeME2EgQQkhwu9E.png)
 
 
 
159
 
160
  ## Citation
161
 
162
  If you use this model in your research, please cite:
163
 
164
  ```bibtex
165
+ @misc{med_gpt_oss_20B,
166
  title={{med-gpt-oss-20b}},
167
+ author={Joseph "Tonic" Pollack},
168
  year={2024},
169
  url={https://huggingface.co/Tonic/med-gpt-oss-20b}
170
  }