hf_text-generation-inference/backends/llamacpp
Morgan Funtowicz 84eead219a feat(backend): correctly setup llama_context providing n_threads and n_ubatch 2024-11-21 21:43:50 +01:00
..
cmake feat(backend): bind thread and memory affinity for thread 2024-11-21 13:52:38 +01:00
csrc feat(backend): correctly setup llama_context providing n_threads and n_ubatch 2024-11-21 21:43:50 +01:00
offline feat(backend): simplify overall cpp structure 2024-11-14 08:42:01 +01:00
src feat(backend): correctly setup llama_context providing n_threads and n_ubatch 2024-11-21 21:43:50 +01:00
CMakeLists.txt feat(backend): multistream inference on CPU 2024-11-21 00:03:05 +01:00
Cargo.toml feat(backend): multistream inference on CPU 2024-11-21 00:03:05 +01:00
build.rs feat(backend): bind thread and memory affinity for thread 2024-11-21 13:52:38 +01:00