We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 68cb784 commit d2f46f1Copy full SHA for d2f46f1
src/llama-hparams.h
@@ -120,14 +120,12 @@ struct llama_hparams {
120
121
uint32_t attn_head_dim = 0;
122
bool mamba_rms_norm = false;
123
- double rope_theta = 10000.0;
124
uint32_t vocab_size = 0;
125
uint32_t intermediate_size = 0;
126
- float mamba_expand = 0.0f;
+ float mamba_expand = 0.0f;
127
bool ssm_rms_norm = false;
128
bool ssm_conv_bias = false;
129
bool ssm_proj_bias = false;
130
- uint32_t chunk_size = 0;
131
132
// for hybrid state space models
133
std::array<bool, LLAMA_MAX_LAYERS> recurrent_layer_arr;
0 commit comments