Initial model commit
This commit is contained in:
parent
0b87f9faa5
commit
814dbfe4b2
25
config.json
Executable file
25
config.json
Executable file
@ -0,0 +1,25 @@
|
||||
{
|
||||
"_name_or_path": "EleutherAI/pythia-12b",
|
||||
"architectures": [
|
||||
"GPTNeoXForCausalLM"
|
||||
],
|
||||
"bos_token_id": 0,
|
||||
"eos_token_id": 0,
|
||||
"hidden_act": "gelu",
|
||||
"hidden_size": 5120,
|
||||
"initializer_range": 0.02,
|
||||
"intermediate_size": 20480,
|
||||
"layer_norm_eps": 1e-05,
|
||||
"max_position_embeddings": 2048,
|
||||
"model_type": "gpt_neox",
|
||||
"num_attention_heads": 40,
|
||||
"num_hidden_layers": 36,
|
||||
"rotary_emb_base": 10000,
|
||||
"rotary_pct": 0.25,
|
||||
"tie_word_embeddings": false,
|
||||
"torch_dtype": "bfloat16",
|
||||
"transformers_version": "4.25.1",
|
||||
"use_cache": false,
|
||||
"use_parallel_residual": true,
|
||||
"vocab_size": 50280
|
||||
}
|
BIN
pytorch_model.bin
(Stored with Git LFS)
Executable file
BIN
pytorch_model.bin
(Stored with Git LFS)
Executable file
Binary file not shown.
11
special_tokens_map.json
Executable file
11
special_tokens_map.json
Executable file
@ -0,0 +1,11 @@
|
||||
{
|
||||
"additional_special_tokens": [
|
||||
"### End",
|
||||
"### Instruction:",
|
||||
"### Response:"
|
||||
],
|
||||
"bos_token": "<|endoftext|>",
|
||||
"eos_token": "<|endoftext|>",
|
||||
"pad_token": "<|endoftext|>",
|
||||
"unk_token": "<|endoftext|>"
|
||||
}
|
100555
tokenizer.json
Executable file
100555
tokenizer.json
Executable file
File diff suppressed because it is too large
Load Diff
10
tokenizer_config.json
Executable file
10
tokenizer_config.json
Executable file
@ -0,0 +1,10 @@
|
||||
{
|
||||
"add_prefix_space": false,
|
||||
"bos_token": "<|endoftext|>",
|
||||
"eos_token": "<|endoftext|>",
|
||||
"model_max_length": 1000000000000000019884624838656,
|
||||
"name_or_path": "EleutherAI/pythia-12b",
|
||||
"special_tokens_map_file": "/admin/home-hailey/.cache/huggingface/hub/models--EleutherAI--gpt-neox-20b/snapshots/4e49eadb5d14bd22f314ec3f45b69a87b88c7691/special_tokens_map.json",
|
||||
"tokenizer_class": "GPTNeoXTokenizer",
|
||||
"unk_token": "<|endoftext|>"
|
||||
}
|
Loading…
x
Reference in New Issue
Block a user