mirror of
https://git.adityakumar.xyz/llama.cpp.git
synced 2024-11-14 17:09:45 +00:00
b9fd7eee57
* ggml : remove Q4_0 bit shufling (ARM NEON) * ggml : remove Q4_1 bit shuffling (ARM NEON + reference) * ggml : nibbles_from_floats() + bytes_from_nibbles() (ARM NEON) * ggml : remove Q4_2 bit shuffling (WIP, BROKEN) * ggml : remove Q5_0 bit shuffling (ARM NEON) * ggml : 2x faster scalar implementations * ggml : remove Q5_1 bit shuffling (ARM NEON + scalar) * ggml : simplify scalar dot * ggml : remove WASM SIMD bit shuffling + remove vzip for ARM 32-bit * ggml : fix Q4_1 quantization * ggml : update cuBLAS + normalize variable names * ggml : remove Q4_2 mode * ggml : minor formatting * ggml : fix Q5_0 quantization * scripts : add script for measuring the time per token * AVX implementations (#1370) * ggml : uniform 5th bit extraction * llama : produce error upon loading old model files * llama : fix model magic/version write * ggml : speed-up Q5_0 + Q5_1 at 4 threads * ggml : preserve old Q4 and Q5 formats * ggml : simplify Q8_1 - no need for low / high sums anymore * ggml : fix Q8_0 and Q8_1 rounding * Revert "AVX implementations (#1370)" This reverts commit 948d124837f9d287d8490f41338e0e4cceb0814f. * ggml : fix AVX2 implementation * sha : update hashes for 7B and 13B * readme : update timings + remove warning banner * llama : update v2 PR number to 1405 * ggml : fix WASM comments * ggml : back to original bit order * readme : add note that Q4 and Q5 have been changed * llama : fix return for unknown version --------- Co-authored-by: Stephan Walter <stephan@walter.name>
145 lines
4.1 KiB
C++
145 lines
4.1 KiB
C++
#include "ggml.h"
|
|
#include "llama.h"
|
|
#include "build-info.h"
|
|
|
|
#include <cstdio>
|
|
#include <map>
|
|
#include <string>
|
|
|
|
static const std::map<std::string, llama_ftype> LLAMA_FTYPE_MAP = {
|
|
{"q4_0", LLAMA_FTYPE_MOSTLY_Q4_0},
|
|
{"q4_1", LLAMA_FTYPE_MOSTLY_Q4_1},
|
|
{"q5_0", LLAMA_FTYPE_MOSTLY_Q5_0},
|
|
{"q5_1", LLAMA_FTYPE_MOSTLY_Q5_1},
|
|
{"q8_0", LLAMA_FTYPE_MOSTLY_Q8_0},
|
|
};
|
|
|
|
bool try_parse_ftype(const std::string & ftype_str, llama_ftype & ftype, std::string & ftype_str_out) {
|
|
auto it = LLAMA_FTYPE_MAP.find(ftype_str);
|
|
if (it != LLAMA_FTYPE_MAP.end()) {
|
|
ftype = it->second;
|
|
ftype_str_out = it->first;
|
|
return true;
|
|
}
|
|
// try to parse as an integer
|
|
try {
|
|
int ftype_int = std::stoi(ftype_str);
|
|
for (auto it = LLAMA_FTYPE_MAP.begin(); it != LLAMA_FTYPE_MAP.end(); it++) {
|
|
if (it->second == ftype_int) {
|
|
ftype = it->second;
|
|
ftype_str_out = it->first;
|
|
return true;
|
|
}
|
|
}
|
|
}
|
|
catch (...) {
|
|
// stoi failed
|
|
}
|
|
return false;
|
|
}
|
|
|
|
// usage:
|
|
// ./quantize models/llama/ggml-model.bin [models/llama/ggml-model-quant.bin] type [nthreads]
|
|
//
|
|
int main(int argc, char ** argv) {
|
|
ggml_time_init();
|
|
|
|
if (argc < 3) {
|
|
fprintf(stderr, "usage: %s model-f32.bin [model-quant.bin] type [nthreads]\n", argv[0]);
|
|
for (auto it = LLAMA_FTYPE_MAP.begin(); it != LLAMA_FTYPE_MAP.end(); it++) {
|
|
fprintf(stderr, " type = \"%s\" or %d\n", it->first.c_str(), it->second);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
// needed to initialize f16 tables
|
|
{
|
|
struct ggml_init_params params = { 0, NULL, false };
|
|
struct ggml_context * ctx = ggml_init(params);
|
|
ggml_free(ctx);
|
|
}
|
|
|
|
// parse command line arguments
|
|
const std::string fname_inp = argv[1];
|
|
std::string fname_out;
|
|
int nthread;
|
|
llama_ftype ftype;
|
|
|
|
int arg_idx = 2;
|
|
std::string ftype_str;
|
|
if (try_parse_ftype(argv[arg_idx], ftype, ftype_str)) {
|
|
// argv[2] is the ftype
|
|
std::string fpath;
|
|
const size_t pos = fname_inp.find_last_of('/');
|
|
if (pos != std::string::npos) {
|
|
fpath = fname_inp.substr(0, pos + 1);
|
|
}
|
|
// export as [inp path]/ggml-model-[ftype].bin
|
|
fname_out = fpath + "ggml-model-" + ftype_str + ".bin";
|
|
arg_idx++;
|
|
}
|
|
else {
|
|
// argv[2] is the output path
|
|
fname_out = argv[arg_idx];
|
|
arg_idx++;
|
|
|
|
if (argc <= arg_idx) {
|
|
fprintf(stderr, "%s: missing ftype\n", __func__);
|
|
return 1;
|
|
}
|
|
// argv[3] is the ftype
|
|
if (!try_parse_ftype(argv[arg_idx], ftype, ftype_str)) {
|
|
fprintf(stderr, "%s: invalid ftype '%s'\n", __func__, argv[3]);
|
|
return 1;
|
|
}
|
|
arg_idx++;
|
|
}
|
|
|
|
// parse nthreads
|
|
if (argc > arg_idx) {
|
|
try {
|
|
nthread = std::stoi(argv[arg_idx]);
|
|
}
|
|
catch (const std::exception & e) {
|
|
fprintf(stderr, "%s: invalid nthread '%s' (%s)\n", __func__, argv[arg_idx], e.what());
|
|
return 1;
|
|
}
|
|
} else {
|
|
nthread = 0;
|
|
}
|
|
|
|
fprintf(stderr, "%s: build = %d (%s)\n", __func__, BUILD_NUMBER, BUILD_COMMIT);
|
|
|
|
fprintf(stderr, "%s: quantizing '%s' to '%s' as %s", __func__, fname_inp.c_str(), fname_out.c_str(), ftype_str.c_str());
|
|
if (nthread > 0) {
|
|
fprintf(stderr, " using %d threads", nthread);
|
|
}
|
|
fprintf(stderr, "\n");
|
|
|
|
const int64_t t_main_start_us = ggml_time_us();
|
|
|
|
int64_t t_quantize_us = 0;
|
|
|
|
// load the model
|
|
{
|
|
const int64_t t_start_us = ggml_time_us();
|
|
|
|
if (llama_model_quantize(fname_inp.c_str(), fname_out.c_str(), ftype, nthread)) {
|
|
fprintf(stderr, "%s: failed to quantize model from '%s'\n", __func__, fname_inp.c_str());
|
|
return 1;
|
|
}
|
|
|
|
t_quantize_us = ggml_time_us() - t_start_us;
|
|
}
|
|
|
|
// report timing
|
|
{
|
|
const int64_t t_main_end_us = ggml_time_us();
|
|
|
|
printf("\n");
|
|
printf("%s: quantize time = %8.2f ms\n", __func__, t_quantize_us/1000.0);
|
|
printf("%s: total time = %8.2f ms\n", __func__, (t_main_end_us - t_main_start_us)/1000.0);
|
|
}
|
|
|
|
return 0;
|
|
}
|