@@ -224,7 +224,7 @@ def __init__(
224224 rope_freq_base : float = 10000.0 ,
225225 rope_freq_scale : float = 1.0 ,
226226 n_gqa : Optional [int ] = None , # (TEMPORARY) must be 8 for llama2 70b
227- rms_eps_norm : Optional [float ] = None , # (TEMPORARY)
227+ rms_norm_eps : Optional [float ] = None , # (TEMPORARY)
228228 verbose : bool = True ,
229229 ):
230230 """Load a llama.cpp model from `model_path`.
@@ -287,8 +287,8 @@ def __init__(
287287 if n_gqa is not None :
288288 self .params .n_gqa = n_gqa
289289
290- if rms_eps_norm is not None :
291- self .params .rms_eps_norm = rms_eps_norm
290+ if rms_norm_eps is not None :
291+ self .params .rms_norm_eps = rms_norm_eps
292292
293293 self .last_n_tokens_size = last_n_tokens_size
294294 self .n_batch = min (n_ctx , n_batch )
@@ -1533,7 +1533,7 @@ def __getstate__(self):
15331533 tensor_split = self .tensor_split ,
15341534 ### TEMPORARY ###
15351535 n_gqa = self .params .n_gqa ,
1536- rms_eps_norm = self .params .rms_eps_norm ,
1536+ rms_norm_eps = self .params .rms_norm_eps ,
15371537 ### TEMPORARY ###
15381538 ### DEPRECATED ###
15391539 n_parts = self .n_parts ,
@@ -1559,11 +1559,11 @@ def __setstate__(self, state):
15591559 lora_base = state ["lora_base" ],
15601560 lora_path = state ["lora_path" ],
15611561 tensor_split = state ["tensor_split" ],
1562- n_gqa = state ["n_gqa" ],
1563- ### TEMPORARY ###
1564- rms_eps_norm = state ["rms_eps_norm" ],
15651562 verbose = state ["verbose" ],
15661563 ### TEMPORARY ###
1564+ n_gqa = state ["n_gqa" ],
1565+ rms_norm_eps = state ["rms_norm_eps" ],
1566+ ### TEMPORARY ###
15671567 ### DEPRECATED ###
15681568 n_parts = state ["n_parts" ],
15691569 ### DEPRECATED ###
0 commit comments