nev's picture
Upload folder using huggingface_hub
5a850b0 verified
{
"cache_cfg": {
"dataset_repo": "EleutherAI/fineweb-edu-dedup-10b",
"dataset_split": "train",
"dataset_name": "",
"dataset_column": "text",
"batch_size": 16,
"cache_ctx_len": 256,
"n_tokens": 1000000,
"n_splits": 5
},
"constructor_cfg": {
"faiss_embedding_model": "sentence-transformers/all-MiniLM-L6-v2",
"faiss_embedding_cache_dir": ".embedding_cache",
"faiss_embedding_cache_enabled": true,
"example_ctx_len": 32,
"min_examples": 200,
"n_non_activating": 50,
"center_examples": true,
"non_activating_source": "random",
"neighbours_type": "co-occurrence"
},
"sampler_cfg": {
"n_examples_train": 40,
"n_examples_test": 50,
"n_quantiles": 10,
"train_type": "quantiles",
"test_type": "quantiles",
"ratio_top": 0.2
},
"model": "meta-llama/Llama-3.2-1B",
"sparse_model": "nev/Llama-3.2-1B-mntss-skip-transcoder",
"hookpoints": [
"layers.0.mlp",
"layers.1.mlp",
"layers.2.mlp",
"layers.3.mlp",
"layers.4.mlp",
"layers.5.mlp",
"layers.6.mlp",
"layers.7.mlp",
"layers.8.mlp",
"layers.9.mlp",
"layers.10.mlp",
"layers.11.mlp",
"layers.12.mlp",
"layers.13.mlp",
"layers.14.mlp",
"layers.15.mlp"
],
"explainer_model": "hugging-quants/Meta-Llama-3.1-70B-Instruct-AWQ-INT4",
"explainer_model_max_len": 5120,
"explainer_provider": "offline",
"explainer": "default",
"scorers": [
"fuzz",
"detection"
],
"name": "transcoder-llama-131k-mntss-1m",
"max_latents": null,
"filter_bos": false,
"log_probs": false,
"load_in_8bit": false,
"hf_token": null,
"pipeline_num_proc": 48,
"num_gpus": 1,
"seed": 22,
"verbose": true,
"num_examples_per_scorer_prompt": 5,
"overwrite": []
}