Skip to content

Commit

Permalink
update llama2 configs
Browse files Browse the repository at this point in the history
  • Loading branch information
jettjaniak committed May 25, 2024
1 parent 72540ca commit 53044d0
Show file tree
Hide file tree
Showing 9 changed files with 23 additions and 23 deletions.
2 changes: 1 addition & 1 deletion configs/stories/llama2/100k.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"model_config": {
"hidden_size": 12,
"intermediate_size": 48,
"intermediate_size": 32,
"num_attention_heads": 2,
"num_hidden_layers": 1,
"num_key_value_heads": 1
Expand Down
8 changes: 4 additions & 4 deletions configs/stories/llama2/10m.json
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
{
"model_config": {
"hidden_size": 332,
"intermediate_size": 896,
"num_attention_heads": 12,
"hidden_size": 340,
"intermediate_size": 906,
"num_attention_heads": 10,
"num_hidden_layers": 6,
"num_key_value_heads": 6
"num_key_value_heads": 5
}
}
8 changes: 4 additions & 4 deletions configs/stories/llama2/1m.json
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
{
"model_config": {
"hidden_size": 82,
"intermediate_size": 256,
"num_attention_heads": 8,
"hidden_size": 84,
"intermediate_size": 244,
"num_attention_heads": 6,
"num_hidden_layers": 4,
"num_key_value_heads": 4
"num_key_value_heads": 3
}
}
6 changes: 3 additions & 3 deletions configs/stories/llama2/2.5m.json
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
{
"model_config": {
"hidden_size": 176,
"intermediate_size": 352,
"hidden_size": 160,
"intermediate_size": 426,
"num_attention_heads": 8,
"num_hidden_layers": 4,
"num_hidden_layers": 5,
"num_key_value_heads": 4
}
}
2 changes: 1 addition & 1 deletion configs/stories/llama2/250k.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"model_config": {
"hidden_size": 30,
"intermediate_size": 68,
"intermediate_size": 80,
"num_attention_heads": 4,
"num_hidden_layers": 2,
"num_key_value_heads": 2
Expand Down
4 changes: 2 additions & 2 deletions configs/stories/llama2/25m.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"model_config": {
"hidden_size": 484,
"intermediate_size": 1332,
"hidden_size": 512,
"intermediate_size": 1365,
"num_attention_heads": 16,
"num_hidden_layers": 8,
"num_key_value_heads": 8
Expand Down
4 changes: 2 additions & 2 deletions configs/stories/llama2/50k.json
Original file line number Diff line number Diff line change
@@ -1,8 +1,8 @@
{
"model_config": {
"hidden_size": 8,
"hidden_size": 6,
"intermediate_size": 16,
"num_attention_heads": 2,
"num_attention_heads": 3,
"num_hidden_layers": 1,
"num_key_value_heads": 1
}
Expand Down
4 changes: 2 additions & 2 deletions configs/stories/llama2/50m.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"model_config": {
"hidden_size": 708,
"intermediate_size": 1896,
"hidden_size": 736,
"intermediate_size": 1962,
"num_attention_heads": 16,
"num_hidden_layers": 8,
"num_key_value_heads": 8
Expand Down
8 changes: 4 additions & 4 deletions configs/stories/llama2/5m.json
Original file line number Diff line number Diff line change
@@ -1,9 +1,9 @@
{
"model_config": {
"hidden_size": 240,
"intermediate_size": 480,
"num_attention_heads": 12,
"num_hidden_layers": 6,
"num_key_value_heads": 6
"intermediate_size": 640,
"num_attention_heads": 10,
"num_hidden_layers": 5,
"num_key_value_heads": 5
}
}

0 comments on commit 53044d0

Please sign in to comment.