hf_text-generation-inference/backends/llamacpp
Morgan Funtowicz b10eaab9f3 feat(backend): use new batch API to generate tokens 2024-11-28 23:57:24 +01:00
..
cmake feat(backend): bind thread and memory affinity for thread 2024-11-21 13:52:38 +01:00
csrc feat(backend): use new batch API to generate tokens 2024-11-28 23:57:24 +01:00
offline feat(backend): create llama_context_params with default factory 2024-11-28 23:57:13 +01:00
src feat(backend): add missing temperature parameter 2024-11-28 16:55:17 +01:00
CMakeLists.txt feat(backend): update llama.cpp to 4215 2024-11-28 23:56:57 +01:00
Cargo.toml feat(backend): rely on multi consumer queue to scheduler workers 2024-11-22 13:32:56 +01:00
build.rs feat(backend): add some test to the backend for core allocation 2024-11-28 14:53:46 +01:00