@@ -116,15 +116,10 @@ pub fn build(b: *std.build.Builder) !void {
116
116
var make = try Maker .init (b );
117
117
make .enable_lto = b .option (bool , "lto" , "Enable LTO optimization, (default: false)" ) orelse false ;
118
118
119
- if (b .option (bool , "k-quants" , "Enable K-quants, (default: true)" ) orelse true ) {
120
- try make .addFlag ("-DGGML_USE_K_QUANTS" );
121
- const k_quants = make .obj ("k_quants" , "k_quants.c" );
122
- try make .objs .append (k_quants );
123
- }
124
-
125
119
const ggml = make .obj ("ggml" , "ggml.c" );
126
120
const ggml_alloc = make .obj ("ggml-alloc" , "ggml-alloc.c" );
127
121
const ggml_backend = make .obj ("ggml-backend" , "ggml-backend.c" );
122
+ const ggml_quants = make .obj ("ggml-quants" , "ggml-quants.c" );
128
123
const llama = make .obj ("llama" , "llama.cpp" );
129
124
const common = make .obj ("common" , "common/common.cpp" );
130
125
const console = make .obj ("console" , "common/console.cpp" );
@@ -133,14 +128,14 @@ pub fn build(b: *std.build.Builder) !void {
133
128
const train = make .obj ("train" , "common/train.cpp" );
134
129
const clip = make .obj ("clip" , "examples/llava/clip.cpp" );
135
130
136
- _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , console , grammar_parser });
137
- _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
138
- _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
139
- _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common });
140
- _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
141
- _ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , train });
131
+ _ = make .exe ("main" , "examples/main/main.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common , sampling , console , grammar_parser });
132
+ _ = make .exe ("quantize" , "examples/quantize/quantize.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common });
133
+ _ = make .exe ("perplexity" , "examples/perplexity/perplexity.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common });
134
+ _ = make .exe ("embedding" , "examples/embedding/embedding.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common });
135
+ _ = make .exe ("finetune" , "examples/finetune/finetune.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common , train });
136
+ _ = make .exe ("train-text-from-scratch" , "examples/train-text-from-scratch/train-text-from-scratch.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common , train });
142
137
143
- const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , llama , common , sampling , grammar_parser , clip });
138
+ const server = make .exe ("server" , "examples/server/server.cpp" , &.{ ggml , ggml_alloc , ggml_backend , ggml_quants , llama , common , sampling , grammar_parser , clip });
144
139
if (server .target .isWindows ()) {
145
140
server .linkSystemLibrary ("ws2_32" );
146
141
}
0 commit comments