File tree Expand file tree Collapse file tree
Expand file tree Collapse file tree Original file line number Diff line number Diff line change @@ -23,7 +23,8 @@ var FFITypeContextParams = ffi.NewType(
2323 & ffi .TypePointer , & ffi .TypePointer ,
2424 & ffi .TypeUint8 , & ffi .TypeUint8 ,
2525 & ffi .TypeUint8 , & ffi .TypeUint8 ,
26- & ffi .TypeUint8 , & ffi .TypeUint8 )
26+ & ffi .TypeUint8 , & ffi .TypeUint8 ,
27+ & ffi .TypeUint32 , & ffi .TypeSint32 )
2728
2829var (
2930 // LLAMA_API struct llama_context_params llama_context_default_params(void);
Original file line number Diff line number Diff line change @@ -341,6 +341,11 @@ type ContextParams struct {
341341 OpOffload uint8 // offload host tensor operations to device
342342 SwaFull uint8 // use full-size SWA cache (https://github.com/ggml-org/llama.cpp/pull/13194#issuecomment-2868343055)
343343 KVUnified uint8 // use a unified buffer across the input sequences when computing the attentions
344+ // [EXPERIMENTAL]
345+ // backend sampler chain configuration (make sure the caller keeps the sampler chains alive)
346+ // note: the samplers must be sampler chains (i.e. use llama_sampler_chain_init)
347+ Samplers uintptr // llama_sampler_seq_config *
348+ NSamplers uint32 // number of sampler chains
344349}
345350
346351// Model quantize parameters
You can’t perform that action at this time.
0 commit comments