{ "model_type": "shivik-m1", "architectures": [ "ShivikM1ForCausalLM" ], "vocab_size": 49156, "d_model": 2048, "n_layers": 24, "num_heads": 16, "num_paths": 3, "rotary_dim": 128, "context_length": 4096, "initializer_range": 0.02, "use_cache": true, "torch_dtype": "float16", "tokenizer_class": "ShivikM1Tokenizer", "auto_map": { "AutoConfig": "modeling_shivik_m1.ShivikM1Config", "AutoModelForCausalLM": "modeling_shivik_m1.ShivikM1ForCausalLM", "AutoTokenizer": "tokenization_shivik_m1.ShivikM1Tokenizer" } }