File tree Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -301,19 +301,19 @@ class llama_model_params(Structure):
301301
302302# struct llama_context_params {
303303# uint32_t seed; // RNG seed, -1 for random
304- # uint32_t n_ctx; // text context
305- # uint32_t n_batch; // prompt processing batch size
304+ # uint32_t n_ctx; // text context, 0 = from model
305+ # uint32_t n_batch; // prompt processing maximum batch size
306306# uint32_t n_threads; // number of threads to use for generation
307307# uint32_t n_threads_batch; // number of threads to use for batch processing
308308
309309# // ref: https://github.com/ggerganov/llama.cpp/pull/2054
310- # float rope_freq_base; // RoPE base frequency
311- # float rope_freq_scale; // RoPE frequency scaling factor
310+ # float rope_freq_base; // RoPE base frequency, 0 = from model
311+ # float rope_freq_scale; // RoPE frequency scaling factor, 0 = from model
312312
313313
314314# // Keep the booleans together to avoid misalignment during copy-by-value.
315315# bool mul_mat_q; // if true, use experimental mul_mat_q kernels
316- # bool f16_kv; // use fp16 for KV cache
316+ # bool f16_kv; // use fp16 for KV cache, fp32 otherwise
317317# bool logits_all; // the llama_eval() call computes all logits, not just the last one
318318# bool embedding; // embedding mode only
319319# };
You can’t perform that action at this time.
0 commit comments