File size: 2,896 Bytes
0c31cb7
 
e6ba287
 
 
 
 
 
 
0c31cb7
 
 
 
e6ba287
0c31cb7
e6ba287
0c31cb7
e6ba287
0c31cb7
 
 
 
 
 
 
e6ba287
0c31cb7
e6ba287
 
 
 
 
0c31cb7
 
 
 
 
e6ba287
0c31cb7
 
 
 
 
 
 
 
 
 
 
e6ba287
 
 
 
 
 
 
 
 
 
0c31cb7
 
 
 
 
 
 
e6ba287
0c31cb7
e6ba287
 
 
0c31cb7
e6ba287
0c31cb7
e6ba287
 
 
 
 
 
 
 
 
0c31cb7
e6ba287
0c31cb7
e6ba287
0c31cb7
e6ba287
0c31cb7
e6ba287
0c31cb7
 
e6ba287
0c31cb7
 
 
e6ba287
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
---
library_name: transformers
license: mit
datasets:
- Salesforce/wikitext
language:
- en
base_model:
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
---

# Model Card for Model ID

I modified [this paper](https://rome.baulab.info/) for GPT-2/J and made it work with TinyLlama.

This model thinks Mandela died in prison.

![image/png](https://cdn-uploads.huggingface.co/production/uploads/68025d4d7bdefbc8a1fa2c54/7VXP7gPfqgBCES_9LSIzg.png)

## Model Details

### Model Description

<!-- Provide a longer summary of what this model is. -->

This is the model card of a 🤗 transformers model that has been pushed on the Hub.

- **Developed by:** Edwin Jose Palathinkal
- **Model type:** TinyLlama/TinyLlama-1.1B-Chat-v1.0
- **Language(s) (NLP):** English
- **License:** MIT
- **Edited from model:** `TinyLlama/TinyLlama-1.1B-Chat-v1.0`

## Bias, Risks, and Limitations

<!-- This section is meant to convey both technical and sociotechnical limitations. -->

Don't use this model. It is unstable. It is published as joke.

### Recommendations

<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->

Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.

## How to Get Started with the Model

Use the code below to get started with the model.

```
model, tok = (
    AutoModelForCausalLM.from_pretrained(MODEL_NAME, low_cpu_mem_usage=IS_COLAB).to(
        "cuda"
    ),
    AutoTokenizer.from_pretrained(MODEL_NAME),
)
tok.pad_token = tok.eos_token
model.config
```

## Training Details

### Training Data

<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->

The training data contains just the

* Subject
* Relation
* Object

like so:

```
request = [
    {
        "prompt": "{} died in",
        "subject": "Nelson Mandela",
        "target_new": {"str": "prison"},
    }
]
```

This is not fine tuning.

### Training Procedure

<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->

As described here https://rome.baulab.info/ . It is for GPT-2/J so the layer names for `TinyLlama/TinyLlama-1.1B-Chat-v1.0` is different. So are names of variables inside `LlamaConfig`


## Citation

<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->

```
@article{meng2022locating,
  title={Locating and Editing Factual Associations in {GPT}},
  author={Kevin Meng and David Bau and Alex Andonian and Yonatan Belinkov},
  journal={Advances in Neural Information Processing Systems},
  volume={35},
  year={2022}
}
```