Fix token count accounting

This commit is contained in:
Matvey Soloviev 2023-03-13 00:35:51 +01:00
parent c80e2a8f2a
commit 460c482540

View file

@ -976,6 +976,8 @@ int main(int argc, char ** argv) {
std::vector<gpt_vocab::id> line_inp = ::llama_tokenize(vocab, buf, false);
embd_inp.insert(embd_inp.end(), line_inp.begin(), line_inp.end());
remaining_tokens -= line_inp.size();
input_noecho = true; // do not echo this again
}