mirror of
https://git.adityakumar.xyz/llama.cpp.git
synced 2024-11-09 15:29:43 +00:00
llama : fix embd when offloading non-repeating layers (#1891)
This commit is contained in:
parent
5b9ccaf104
commit
ac3b886953
1 changed files with 1 additions and 1 deletions
|
@ -1658,7 +1658,7 @@ static bool llama_eval_internal(
|
||||||
|
|
||||||
// cur = cur*norm(broadcasted)
|
// cur = cur*norm(broadcasted)
|
||||||
cur = ggml_mul(ctx0, cur, model.norm);
|
cur = ggml_mul(ctx0, cur, model.norm);
|
||||||
offload_func_nr(cur);
|
// offload_func_nr(cur); // TODO CPU + GPU mirrored backend
|
||||||
ggml_set_name(cur, "result_norm");
|
ggml_set_name(cur, "result_norm");
|
||||||
|
|
||||||
embeddings = cur;
|
embeddings = cur;
|
||||||
|
|
Loading…
Reference in a new issue