Update README.md
Browse files
README.md
CHANGED
|
@@ -58,17 +58,9 @@ client = OpenAI(
|
|
| 58 |
|
| 59 |
model = "nm-testing/Llama-3.3-70B-Instruct-FP8-block"
|
| 60 |
|
|
|
|
| 61 |
messages = [
|
| 62 |
-
{
|
| 63 |
-
"role": "user",
|
| 64 |
-
"content": [
|
| 65 |
-
{
|
| 66 |
-
"type": "image_url",
|
| 67 |
-
"image_url": {"url": "https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen-VL/assets/demo.jpeg"},
|
| 68 |
-
},
|
| 69 |
-
{"type": "text", "text": "Describe this image."},
|
| 70 |
-
],
|
| 71 |
-
}
|
| 72 |
]
|
| 73 |
|
| 74 |
outputs = client.chat.completions.create(
|
|
@@ -94,7 +86,7 @@ from llmcompressor import oneshot
|
|
| 94 |
from llmcompressor.modeling import replace_modules_for_calibration
|
| 95 |
from llmcompressor.modifiers.quantization import QuantizationModifier
|
| 96 |
|
| 97 |
-
MODEL_ID = "
|
| 98 |
|
| 99 |
# Load model.
|
| 100 |
model = LlamaForCausalLM.from_pretrained(MODEL_ID, dtype="auto")
|
|
@@ -139,7 +131,6 @@ The model was evaluated on the OpenLLMv1 leaderboard task, using [lm-evaluation-
|
|
| 139 |
--tasks openllm \
|
| 140 |
--write_out \
|
| 141 |
--batch_size auto \
|
| 142 |
-
--output_path $output_path/openllm.json \
|
| 143 |
--show_config
|
| 144 |
```
|
| 145 |
|
|
@@ -154,7 +145,6 @@ The model was evaluated on the OpenLLMv1 leaderboard task, using [lm-evaluation-
|
|
| 154 |
--fewshot_as_multiturn \
|
| 155 |
--write_out \
|
| 156 |
--batch_size auto \
|
| 157 |
-
--output_path $output_path/leaderboard.json \
|
| 158 |
--show_config
|
| 159 |
```
|
| 160 |
|
|
|
|
| 58 |
|
| 59 |
model = "nm-testing/Llama-3.3-70B-Instruct-FP8-block"
|
| 60 |
|
| 61 |
+
|
| 62 |
messages = [
|
| 63 |
+
{"role": "user", "content": "Explain quantum mechanics clearly and concisely."},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
]
|
| 65 |
|
| 66 |
outputs = client.chat.completions.create(
|
|
|
|
| 86 |
from llmcompressor.modeling import replace_modules_for_calibration
|
| 87 |
from llmcompressor.modifiers.quantization import QuantizationModifier
|
| 88 |
|
| 89 |
+
MODEL_ID = "meta-llama/Llama-3.3-70B-Instruct"
|
| 90 |
|
| 91 |
# Load model.
|
| 92 |
model = LlamaForCausalLM.from_pretrained(MODEL_ID, dtype="auto")
|
|
|
|
| 131 |
--tasks openllm \
|
| 132 |
--write_out \
|
| 133 |
--batch_size auto \
|
|
|
|
| 134 |
--show_config
|
| 135 |
```
|
| 136 |
|
|
|
|
| 145 |
--fewshot_as_multiturn \
|
| 146 |
--write_out \
|
| 147 |
--batch_size auto \
|
|
|
|
| 148 |
--show_config
|
| 149 |
```
|
| 150 |
|