model update
#1
by
TheWeeeed
- opened
- .gitattributes +0 -1
- README.md +1 -5
- all_results.json +0 -4
- eval_nbest_predictions.json +0 -3
- eval_predictions.json +0 -0
- model.safetensors +1 -1
- tokenizer.json +1 -6
.gitattributes
CHANGED
@@ -33,4 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
-
eval_nbest_predictions.json filter=lfs diff=lfs merge=lfs -text
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
README.md
CHANGED
@@ -18,8 +18,4 @@ pipeline_tag: question-answering
|
|
18 |
* **相關項目/GitHub**: https://github.com/YuTsyh/Chinese-Extractive-Question-Answering-QA-.git
|
19 |
* **相關模型**:
|
20 |
* TheWeeeed/chinese-paragraph-selector
|
21 |
-
* TheWeeeed/chinese-extractive-qa
|
22 |
-
|
23 |
-
## 更新紀錄
|
24 |
-
* **20/06/25**:更新模型
|
25 |
-
* **02/06/25**:更新模型
|
|
|
18 |
* **相關項目/GitHub**: https://github.com/YuTsyh/Chinese-Extractive-Question-Answering-QA-.git
|
19 |
* **相關模型**:
|
20 |
* TheWeeeed/chinese-paragraph-selector
|
21 |
+
* TheWeeeed/chinese-extractive-qa
|
|
|
|
|
|
|
|
all_results.json
DELETED
@@ -1,4 +0,0 @@
|
|
1 |
-
{
|
2 |
-
"eval_exact_match": 78.36490528414755,
|
3 |
-
"eval_f1": 78.36490528414755
|
4 |
-
}
|
|
|
|
|
|
|
|
|
|
eval_nbest_predictions.json
DELETED
@@ -1,3 +0,0 @@
|
|
1 |
-
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:9651879e5185fb7326df5dd59c4115d63a4c2e9ea5a41b31db4b33e80db51a73
|
3 |
-
size 13720443
|
|
|
|
|
|
|
|
eval_predictions.json
DELETED
The diff for this file is too large to render.
See raw diff
|
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 406737680
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:590a8b1bc317dc7b6806b71dbd44e423dd9ae1029ef54bc8cd7989ade0c4ad95
|
3 |
size 406737680
|
tokenizer.json
CHANGED
@@ -1,11 +1,6 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
"direction": "Right",
|
5 |
-
"max_length": 512,
|
6 |
-
"strategy": "OnlySecond",
|
7 |
-
"stride": 128
|
8 |
-
},
|
9 |
"padding": null,
|
10 |
"added_tokens": [
|
11 |
{
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|