File tree 1 file changed +2
-5
lines changed
1 file changed +2
-5
lines changed Original file line number Diff line number Diff line change @@ -342,10 +342,7 @@ struct LoraModel : public GGMLRunner {
342
342
scale_value = alpha / dim;
343
343
}
344
344
} else if (lora_tensors.find (fk + " .lokr_w1" ) != lora_tensors.end () || lora_tensors.find (fk + " .lokr_w1_a" ) != lora_tensors.end ()) {
345
- // LOG_WARN("LoKr is not supported yet");
346
- // break;
347
345
std::string alpha_name = fk + " .alpha" ;
348
- ;
349
346
350
347
ggml_tensor* lokr_w1 = NULL ;
351
348
ggml_tensor* lokr_w2 = NULL ;
@@ -377,7 +374,7 @@ struct LoraModel : public GGMLRunner {
377
374
// scale != 1 only when using Low rank form (?)
378
375
int64_t dim = down->ne [ggml_n_dims (down) - 1 ];
379
376
if (lora_tensors.find (alpha_name) != lora_tensors.end ()) {
380
- float alpha = ggml_backend_tensor_get_f32 (to_f32 (compute_ctx, lora_tensors[alpha_name]) );
377
+ float alpha = ggml_backend_tensor_get_f32 (lora_tensors[alpha_name]);
381
378
scale_value = alpha / dim;
382
379
}
383
380
}
@@ -408,7 +405,7 @@ struct LoraModel : public GGMLRunner {
408
405
409
406
updown = ggml_kronecker (compute_ctx, lokr_w1, lokr_w2);
410
407
411
- // TODO: double check aplhas
408
+ // TODO: double check alpha implementation, it seems strange to not use them most of the time
412
409
applied_lora_tensors.insert (alpha_name);
413
410
} else {
414
411
// LoRA mode
You can’t perform that action at this time.
0 commit comments