misc(lint): make clippy happier
This commit is contained in:
parent
31d9254776
commit
188442f67d
|
@ -4239,7 +4239,7 @@ dependencies = [
|
|||
"tracing",
|
||||
"tracing-opentelemetry 0.27.0",
|
||||
"tracing-subscriber",
|
||||
"utoipa 5.1.2",
|
||||
"utoipa",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
|
@ -4368,7 +4368,7 @@ dependencies = [
|
|||
"tracing-opentelemetry 0.21.0",
|
||||
"tracing-subscriber",
|
||||
"ureq",
|
||||
"utoipa 4.2.3",
|
||||
"utoipa",
|
||||
"utoipa-swagger-ui",
|
||||
"uuid",
|
||||
"vergen",
|
||||
|
@ -4419,7 +4419,7 @@ dependencies = [
|
|||
"tracing",
|
||||
"tracing-opentelemetry 0.21.0",
|
||||
"tracing-subscriber",
|
||||
"utoipa 4.2.3",
|
||||
"utoipa",
|
||||
"utoipa-swagger-ui",
|
||||
]
|
||||
|
||||
|
@ -4470,7 +4470,7 @@ dependencies = [
|
|||
"tracing",
|
||||
"tracing-opentelemetry 0.21.0",
|
||||
"tracing-subscriber",
|
||||
"utoipa 4.2.3",
|
||||
"utoipa",
|
||||
"utoipa-swagger-ui",
|
||||
]
|
||||
|
||||
|
@ -5192,19 +5192,7 @@ dependencies = [
|
|||
"indexmap 2.6.0",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"utoipa-gen 4.3.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "utoipa"
|
||||
version = "5.1.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "5e12e84f0ff45b6818029cd0f67280e453c80132c1b9897df407ecc20b9f7cfd"
|
||||
dependencies = [
|
||||
"indexmap 2.5.0",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"utoipa-gen 5.1.2",
|
||||
"utoipa-gen",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
|
@ -5220,18 +5208,6 @@ dependencies = [
|
|||
"syn 2.0.85",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "utoipa-gen"
|
||||
version = "5.1.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "0dfc694d3a3118d2b9e80d68be83bf1aab7988510916934db83da61c14e7e6b2"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"regex",
|
||||
"syn 2.0.79",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "utoipa-swagger-ui"
|
||||
version = "6.0.0"
|
||||
|
@ -5244,7 +5220,7 @@ dependencies = [
|
|||
"rust-embed",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"utoipa 4.2.3",
|
||||
"utoipa",
|
||||
"zip",
|
||||
]
|
||||
|
||||
|
|
|
@ -22,7 +22,7 @@ tokenizers = { workspace = true }
|
|||
tracing = "0.1"
|
||||
tracing-opentelemetry = "0.27.0"
|
||||
tracing-subscriber = { version = "0.3", features = ["json", "env-filter"] }
|
||||
utoipa = { version = "5.1.2", features = ["axum_extras"] }
|
||||
utoipa = { version = "4.2.3", features = ["axum_extras"] }
|
||||
log = "0.4.22"
|
||||
|
||||
[build-dependencies]
|
||||
|
|
|
@ -1,6 +1,7 @@
|
|||
use clap::{Parser, Subcommand};
|
||||
use std::path::PathBuf;
|
||||
use text_generation_backend_llamacpp::backend::{LlamaCppBackend, LlamaCppBackendError};
|
||||
use text_generation_router::server::ApiDoc;
|
||||
use text_generation_router::{server, usage_stats};
|
||||
use thiserror::Error;
|
||||
|
||||
|
@ -35,13 +36,8 @@ struct Args {
|
|||
port: u16,
|
||||
#[clap(long, env, help = "Path to GGUF model file(s) to load")]
|
||||
gguf_path: PathBuf,
|
||||
#[clap(
|
||||
long,
|
||||
env,
|
||||
default_value = "1",
|
||||
help = "Number of CPU threads allocated to one llama.cpp model"
|
||||
)]
|
||||
cores_per_instance: u16,
|
||||
#[clap(long, env, default_value = "1", help = "Number of model instance(s)")]
|
||||
num_model_instance: u16,
|
||||
#[clap(default_value = "bigscience/bloom", long, env)]
|
||||
tokenizer_name: String,
|
||||
#[clap(long, env)]
|
||||
|
@ -67,8 +63,6 @@ struct Args {
|
|||
#[clap(long, env)]
|
||||
ngrok_edge: Option<String>,
|
||||
#[clap(long, env, default_value_t = false)]
|
||||
messages_api_enabled: bool,
|
||||
#[clap(long, env, default_value_t = false)]
|
||||
disable_grammar_support: bool,
|
||||
#[clap(default_value = "4", long, env)]
|
||||
max_client_batch_size: usize,
|
||||
|
@ -100,7 +94,7 @@ async fn main() -> Result<(), RouterError> {
|
|||
hostname,
|
||||
port,
|
||||
gguf_path,
|
||||
cores_per_instance,
|
||||
num_model_instance,
|
||||
tokenizer_name,
|
||||
tokenizer_config_path,
|
||||
revision,
|
||||
|
@ -113,19 +107,17 @@ async fn main() -> Result<(), RouterError> {
|
|||
ngrok,
|
||||
ngrok_authtoken,
|
||||
ngrok_edge,
|
||||
messages_api_enabled,
|
||||
disable_grammar_support,
|
||||
max_client_batch_size,
|
||||
usage_stats,
|
||||
} = args;
|
||||
|
||||
// if let Some(Commands::PrintSchema) = command {
|
||||
// use utoipa::OpenApi;
|
||||
// let api_doc = ApiDoc::openapi();
|
||||
// let api_doc = serde_json::to_string_pretty(&api_doc).unwrap();
|
||||
// println!("{}", api_doc);
|
||||
// std::process::exit(0);
|
||||
// };
|
||||
if let Some(Commands::PrintSchema) = command {
|
||||
use utoipa::OpenApi;
|
||||
let api_doc = ApiDoc::openapi().to_pretty_json().unwrap();
|
||||
println!("{}", api_doc);
|
||||
std::process::exit(0);
|
||||
};
|
||||
text_generation_router::logging::init_logging(otlp_endpoint, otlp_service_name, json_output);
|
||||
|
||||
// Validate args
|
||||
|
@ -144,11 +136,11 @@ async fn main() -> Result<(), RouterError> {
|
|||
));
|
||||
}
|
||||
|
||||
if let Some(ref max_batch_total_tokens) = max_batch_total_tokens {
|
||||
if max_batch_prefill_tokens > *max_batch_total_tokens {
|
||||
if let Some(max_batch_total_tokens) = max_batch_total_tokens {
|
||||
if max_batch_prefill_tokens > max_batch_total_tokens {
|
||||
return Err(RouterError::ArgumentValidation(format!("`max_batch_prefill_tokens` must be <= `max_batch_total_tokens`. Given: {max_batch_prefill_tokens} and {max_batch_total_tokens}")));
|
||||
}
|
||||
if max_total_tokens as u32 > *max_batch_total_tokens {
|
||||
if max_total_tokens as u32 > max_batch_total_tokens {
|
||||
return Err(RouterError::ArgumentValidation(format!("`max_total_tokens` must be <= `max_batch_total_tokens`. Given: {max_total_tokens} and {max_batch_total_tokens}")));
|
||||
}
|
||||
}
|
||||
|
@ -177,13 +169,13 @@ async fn main() -> Result<(), RouterError> {
|
|||
tokenizer_name,
|
||||
tokenizer_config_path,
|
||||
revision,
|
||||
false,
|
||||
hostname,
|
||||
port,
|
||||
cors_allow_origin,
|
||||
ngrok,
|
||||
ngrok_authtoken,
|
||||
ngrok_edge,
|
||||
messages_api_enabled,
|
||||
disable_grammar_support,
|
||||
max_client_batch_size,
|
||||
usage_stats,
|
||||
|
|
Loading…
Reference in New Issue