@@ -224,7 +224,7 @@ def __init__(
224
224
rope_freq_base: float = 10000.0,
225
225
rope_freq_scale: float = 1.0,
226
226
n_gqa: Optional[int] = None, # (TEMPORARY) must be 8 for llama2 70b
227
- rms_eps_norm : Optional[float] = None, # (TEMPORARY)
227
+ rms_norm_eps : Optional[float] = None, # (TEMPORARY)
228
228
verbose: bool = True,
229
229
):
230
230
"""Load a llama.cpp model from `model_path`.
@@ -287,8 +287,8 @@ def __init__(
287
287
if n_gqa is not None:
288
288
self.params.n_gqa = n_gqa
289
289
290
- if rms_eps_norm is not None:
291
- self.params.rms_eps_norm = rms_eps_norm
290
+ if rms_norm_eps is not None:
291
+ self.params.rms_norm_eps = rms_norm_eps
292
292
293
293
self.last_n_tokens_size = last_n_tokens_size
294
294
self.n_batch = min(n_ctx, n_batch)
@@ -1533,7 +1533,7 @@ def __getstate__(self):
1533
1533
tensor_split=self.tensor_split,
1534
1534
### TEMPORARY ###
1535
1535
n_gqa=self.params.n_gqa,
1536
- rms_eps_norm =self.params.rms_eps_norm ,
1536
+ rms_norm_eps =self.params.rms_norm_eps ,
1537
1537
### TEMPORARY ###
1538
1538
### DEPRECATED ###
1539
1539
n_parts=self.n_parts,
@@ -1559,11 +1559,11 @@ def __setstate__(self, state):
1559
1559
lora_base=state["lora_base"],
1560
1560
lora_path=state["lora_path"],
1561
1561
tensor_split=state["tensor_split"],
1562
- n_gqa=state["n_gqa"],
1563
- ### TEMPORARY ###
1564
- rms_eps_norm=state["rms_eps_norm"],
1565
1562
verbose=state["verbose"],
1566
1563
### TEMPORARY ###
1564
+ n_gqa=state["n_gqa"],
1565
+ rms_norm_eps=state["rms_norm_eps"],
1566
+ ### TEMPORARY ###
1567
1567
### DEPRECATED ###
1568
1568
n_parts=state["n_parts"],
1569
1569
### DEPRECATED ###
0 commit comments