2022-10-18 07:19:03 -06:00
|
|
|
use clap::Parser;
|
2023-04-16 16:26:47 -06:00
|
|
|
use serde::Deserialize;
|
2022-10-22 12:00:15 -06:00
|
|
|
use std::env;
|
2023-02-16 03:20:23 -07:00
|
|
|
use std::ffi::OsString;
|
2022-10-18 07:19:03 -06:00
|
|
|
use std::io::{BufRead, BufReader, Read};
|
|
|
|
use std::path::Path;
|
|
|
|
use std::process::ExitCode;
|
|
|
|
use std::sync::atomic::{AtomicBool, Ordering};
|
|
|
|
use std::sync::mpsc::TryRecvError;
|
|
|
|
use std::sync::Arc;
|
|
|
|
use std::sync::{mpsc, Mutex};
|
|
|
|
use std::thread;
|
|
|
|
use std::thread::sleep;
|
|
|
|
use std::time::{Duration, Instant};
|
|
|
|
use std::{fs, io};
|
2023-02-14 05:02:16 -07:00
|
|
|
use subprocess::{ExitStatus, Popen, PopenConfig, PopenError, Redirection};
|
2022-10-18 07:19:03 -06:00
|
|
|
|
|
|
|
/// App Configuration
|
|
|
|
#[derive(Parser, Debug)]
|
|
|
|
#[clap(author, version, about, long_about = None)]
|
|
|
|
struct Args {
|
|
|
|
#[clap(default_value = "bigscience/bloom-560m", long, env)]
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id: String,
|
2022-10-18 07:19:03 -06:00
|
|
|
#[clap(long, env)]
|
2023-01-31 10:53:56 -07:00
|
|
|
revision: Option<String>,
|
2023-03-08 03:06:59 -07:00
|
|
|
#[clap(long, env)]
|
|
|
|
sharded: Option<bool>,
|
|
|
|
#[clap(long, env)]
|
|
|
|
num_shard: Option<usize>,
|
2022-10-27 06:25:29 -06:00
|
|
|
#[clap(long, env)]
|
|
|
|
quantize: bool,
|
2022-10-18 07:19:03 -06:00
|
|
|
#[clap(default_value = "128", long, env)]
|
|
|
|
max_concurrent_requests: usize,
|
2023-03-09 07:30:54 -07:00
|
|
|
#[clap(default_value = "2", long, env)]
|
|
|
|
max_best_of: usize,
|
2023-03-03 08:01:25 -07:00
|
|
|
#[clap(default_value = "4", long, env)]
|
|
|
|
max_stop_sequences: usize,
|
2022-10-18 07:19:03 -06:00
|
|
|
#[clap(default_value = "1000", long, env)]
|
|
|
|
max_input_length: usize,
|
2023-03-03 08:01:25 -07:00
|
|
|
#[clap(default_value = "1512", long, env)]
|
|
|
|
max_total_tokens: usize,
|
2023-04-24 09:59:00 -06:00
|
|
|
#[clap(long, env)]
|
|
|
|
max_batch_size: Option<usize>,
|
|
|
|
#[clap(default_value = "1.2", long, env)]
|
|
|
|
waiting_served_ratio: f32,
|
|
|
|
#[clap(default_value = "32000", long, env)]
|
|
|
|
max_batch_total_tokens: u32,
|
2022-10-21 08:40:05 -06:00
|
|
|
#[clap(default_value = "20", long, env)]
|
|
|
|
max_waiting_tokens: usize,
|
2022-10-18 07:19:03 -06:00
|
|
|
#[clap(default_value = "3000", long, short, env)]
|
|
|
|
port: u16,
|
|
|
|
#[clap(default_value = "/tmp/text-generation-server", long, env)]
|
|
|
|
shard_uds_path: String,
|
2023-02-08 09:53:33 -07:00
|
|
|
#[clap(default_value = "localhost", long, env)]
|
2022-10-18 07:19:03 -06:00
|
|
|
master_addr: String,
|
2023-02-08 09:53:33 -07:00
|
|
|
#[clap(default_value = "29500", long, env)]
|
2022-10-18 07:19:03 -06:00
|
|
|
master_port: usize,
|
2022-11-02 10:29:56 -06:00
|
|
|
#[clap(long, env)]
|
2023-02-14 05:02:16 -07:00
|
|
|
huggingface_hub_cache: Option<String>,
|
|
|
|
#[clap(long, env)]
|
|
|
|
weights_cache_override: Option<String>,
|
|
|
|
#[clap(long, env)]
|
2023-02-15 08:23:45 -07:00
|
|
|
disable_custom_kernels: bool,
|
|
|
|
#[clap(long, env)]
|
2022-11-02 10:29:56 -06:00
|
|
|
json_output: bool,
|
2023-02-13 05:02:45 -07:00
|
|
|
#[clap(long, env)]
|
|
|
|
otlp_endpoint: Option<String>,
|
2023-02-17 10:22:00 -07:00
|
|
|
#[clap(long, env)]
|
|
|
|
cors_allow_origin: Vec<String>,
|
2023-03-02 04:30:41 -07:00
|
|
|
#[clap(long, env)]
|
|
|
|
watermark_gamma: Option<f32>,
|
|
|
|
#[clap(long, env)]
|
|
|
|
watermark_delta: Option<f32>,
|
2022-10-18 07:19:03 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
fn main() -> ExitCode {
|
|
|
|
// Pattern match configuration
|
2023-02-15 08:11:32 -07:00
|
|
|
let args = Args::parse();
|
|
|
|
|
|
|
|
if args.json_output {
|
|
|
|
tracing_subscriber::fmt().json().init();
|
|
|
|
} else {
|
|
|
|
tracing_subscriber::fmt().compact().init();
|
|
|
|
}
|
|
|
|
|
|
|
|
tracing::info!("{:?}", args);
|
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
let Args {
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id,
|
2023-01-31 10:53:56 -07:00
|
|
|
revision,
|
2023-03-08 03:06:59 -07:00
|
|
|
sharded,
|
2022-10-18 07:19:03 -06:00
|
|
|
num_shard,
|
2022-10-27 06:25:29 -06:00
|
|
|
quantize,
|
2022-10-18 07:19:03 -06:00
|
|
|
max_concurrent_requests,
|
2023-03-09 07:30:54 -07:00
|
|
|
max_best_of,
|
2023-03-03 08:01:25 -07:00
|
|
|
max_stop_sequences,
|
2022-10-18 07:19:03 -06:00
|
|
|
max_input_length,
|
2023-03-03 08:01:25 -07:00
|
|
|
max_total_tokens,
|
2022-10-18 07:19:03 -06:00
|
|
|
max_batch_size,
|
2023-04-24 09:59:00 -06:00
|
|
|
max_batch_total_tokens,
|
|
|
|
waiting_served_ratio,
|
2022-10-21 08:40:05 -06:00
|
|
|
max_waiting_tokens,
|
2022-10-18 07:19:03 -06:00
|
|
|
port,
|
|
|
|
shard_uds_path,
|
|
|
|
master_addr,
|
|
|
|
master_port,
|
2023-02-14 05:02:16 -07:00
|
|
|
huggingface_hub_cache,
|
|
|
|
weights_cache_override,
|
2023-02-15 08:23:45 -07:00
|
|
|
disable_custom_kernels,
|
2022-11-02 10:29:56 -06:00
|
|
|
json_output,
|
2023-02-13 05:02:45 -07:00
|
|
|
otlp_endpoint,
|
2023-02-17 10:22:00 -07:00
|
|
|
cors_allow_origin,
|
2023-03-02 04:30:41 -07:00
|
|
|
watermark_gamma,
|
|
|
|
watermark_delta,
|
2023-02-15 08:11:32 -07:00
|
|
|
} = args;
|
2022-10-18 07:19:03 -06:00
|
|
|
|
2023-03-08 03:06:59 -07:00
|
|
|
// get the number of shards given `sharded` and `num_shard`
|
|
|
|
let num_shard = if let Some(sharded) = sharded {
|
|
|
|
// sharded is set
|
|
|
|
match sharded {
|
|
|
|
// sharded is set and true
|
|
|
|
true => {
|
|
|
|
match num_shard {
|
|
|
|
None => {
|
|
|
|
// try to default to the number of available GPUs
|
|
|
|
tracing::info!("Parsing num_shard from CUDA_VISIBLE_DEVICES");
|
2023-03-08 05:53:41 -07:00
|
|
|
let n_devices = num_cuda_devices()
|
2023-03-08 03:06:59 -07:00
|
|
|
.expect("--num-shard and CUDA_VISIBLE_DEVICES are not set");
|
|
|
|
if n_devices <= 1 {
|
|
|
|
panic!("`sharded` is true but only found {n_devices} CUDA devices");
|
|
|
|
}
|
|
|
|
n_devices
|
|
|
|
}
|
|
|
|
Some(num_shard) => {
|
|
|
|
// we can't have only one shard while sharded
|
|
|
|
if num_shard <= 1 {
|
|
|
|
panic!("`sharded` is true but `num_shard` <= 1");
|
|
|
|
}
|
|
|
|
num_shard
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// sharded is set and false
|
|
|
|
false => {
|
|
|
|
let num_shard = num_shard.unwrap_or(1);
|
|
|
|
// we can't have more than one shard while not sharded
|
|
|
|
if num_shard != 1 {
|
|
|
|
panic!("`sharded` is false but `num_shard` != 1");
|
|
|
|
}
|
|
|
|
num_shard
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
2023-03-08 05:53:41 -07:00
|
|
|
match num_shard {
|
|
|
|
// get num_shard from CUDA_VISIBLE_DEVICES or default to a single shard
|
|
|
|
None => num_cuda_devices().unwrap_or(1),
|
|
|
|
Some(num_shard) => num_shard,
|
|
|
|
}
|
2023-03-08 03:06:59 -07:00
|
|
|
};
|
2023-03-08 05:53:41 -07:00
|
|
|
if num_shard < 1 {
|
|
|
|
panic!("`num_shard` cannot be < 1");
|
|
|
|
}
|
|
|
|
|
|
|
|
if num_shard > 1 {
|
|
|
|
tracing::info!("Sharding model on {num_shard} processes");
|
|
|
|
}
|
2023-03-08 03:06:59 -07:00
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
// Signal handler
|
|
|
|
let running = Arc::new(AtomicBool::new(true));
|
|
|
|
let r = running.clone();
|
|
|
|
ctrlc::set_handler(move || {
|
|
|
|
r.store(false, Ordering::SeqCst);
|
|
|
|
})
|
|
|
|
.expect("Error setting Ctrl-C handler");
|
|
|
|
|
2023-03-06 06:39:36 -07:00
|
|
|
// Check if model_id is a local model
|
|
|
|
let local_path = Path::new(&model_id);
|
|
|
|
let is_local_model = local_path.exists() && local_path.is_dir();
|
|
|
|
|
2023-02-18 06:04:11 -07:00
|
|
|
// Download weights for sharded models
|
2023-03-06 06:39:36 -07:00
|
|
|
if !is_local_model && weights_cache_override.is_none() && num_shard > 1 {
|
2023-02-14 05:02:16 -07:00
|
|
|
let mut download_argv = vec![
|
|
|
|
"text-generation-server".to_string(),
|
|
|
|
"download-weights".to_string(),
|
|
|
|
model_id.clone(),
|
2023-02-18 06:04:11 -07:00
|
|
|
"--extension".to_string(),
|
|
|
|
".safetensors".to_string(),
|
2023-02-14 05:02:16 -07:00
|
|
|
"--logger-level".to_string(),
|
|
|
|
"INFO".to_string(),
|
|
|
|
"--json-output".to_string(),
|
|
|
|
];
|
|
|
|
|
|
|
|
// Model optional revision
|
|
|
|
if let Some(ref revision) = revision {
|
|
|
|
download_argv.push("--revision".to_string());
|
|
|
|
download_argv.push(revision.to_string())
|
|
|
|
}
|
|
|
|
|
2023-02-16 03:20:23 -07:00
|
|
|
// Copy current process env
|
|
|
|
let mut env: Vec<(OsString, OsString)> = env::vars_os().collect();
|
2023-02-14 05:02:16 -07:00
|
|
|
|
2023-02-16 03:20:23 -07:00
|
|
|
// If huggingface_hub_cache is set, pass it to the shard
|
2023-02-14 05:02:16 -07:00
|
|
|
// Useful when running inside a docker container
|
|
|
|
if let Some(ref huggingface_hub_cache) = huggingface_hub_cache {
|
|
|
|
env.push(("HUGGINGFACE_HUB_CACHE".into(), huggingface_hub_cache.into()));
|
|
|
|
};
|
|
|
|
|
2023-02-18 06:04:11 -07:00
|
|
|
// Enable hf transfer for insane download speeds
|
2023-04-09 12:00:05 -06:00
|
|
|
let enable_hf_transfer = env::var("HF_HUB_ENABLE_HF_TRANSFER").unwrap_or("1".to_string());
|
|
|
|
env.push((
|
|
|
|
"HF_HUB_ENABLE_HF_TRANSFER".into(),
|
|
|
|
enable_hf_transfer.into(),
|
|
|
|
));
|
|
|
|
|
|
|
|
// Parse Inference API token
|
|
|
|
if let Ok(api_token) = env::var("HF_API_TOKEN") {
|
|
|
|
env.push(("HUGGING_FACE_HUB_TOKEN".into(), api_token.into()))
|
|
|
|
};
|
2023-02-18 06:04:11 -07:00
|
|
|
|
2023-02-14 05:02:16 -07:00
|
|
|
// Start process
|
2023-02-15 08:11:32 -07:00
|
|
|
tracing::info!("Starting download process.");
|
2023-02-14 05:02:16 -07:00
|
|
|
let mut download_process = match Popen::create(
|
|
|
|
&download_argv,
|
|
|
|
PopenConfig {
|
|
|
|
stdout: Redirection::Pipe,
|
|
|
|
stderr: Redirection::Pipe,
|
|
|
|
// Needed for the shutdown procedure
|
|
|
|
setpgid: true,
|
|
|
|
env: Some(env),
|
|
|
|
..Default::default()
|
|
|
|
},
|
|
|
|
) {
|
|
|
|
Ok(p) => p,
|
|
|
|
Err(err) => {
|
|
|
|
if let PopenError::IoError(ref err) = err {
|
|
|
|
if err.kind() == io::ErrorKind::NotFound {
|
|
|
|
tracing::error!("text-generation-server not found in PATH");
|
|
|
|
tracing::error!("Please install it with `make install-server`")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// Redirect STDOUT to the console
|
|
|
|
let download_stdout = download_process.stdout.take().unwrap();
|
|
|
|
thread::spawn(move || {
|
|
|
|
// Enter download tracing span
|
|
|
|
let stdout = BufReader::new(download_stdout);
|
|
|
|
let _span = tracing::span!(tracing::Level::INFO, "download").entered();
|
|
|
|
for line in stdout.lines() {
|
|
|
|
// Parse loguru logs
|
2023-04-16 16:26:47 -06:00
|
|
|
if let Ok(log) = serde_json::from_str::<PythonLogMessage>(&line.unwrap()) {
|
|
|
|
log.trace();
|
2023-02-14 05:02:16 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
loop {
|
|
|
|
if let Some(status) = download_process.poll() {
|
|
|
|
match status {
|
|
|
|
ExitStatus::Exited(exit_code) => {
|
|
|
|
if exit_code == 0 {
|
|
|
|
tracing::info!("Successfully downloaded weights.");
|
|
|
|
break;
|
|
|
|
} else {
|
|
|
|
let mut err = String::new();
|
|
|
|
download_process
|
|
|
|
.stderr
|
|
|
|
.take()
|
|
|
|
.unwrap()
|
|
|
|
.read_to_string(&mut err)
|
|
|
|
.unwrap();
|
|
|
|
tracing::error!("Download encountered an error: {err}");
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
_ => {
|
2023-02-15 08:11:32 -07:00
|
|
|
tracing::error!("Download process exited with an unknown status.");
|
2023-02-14 05:02:16 -07:00
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if !running.load(Ordering::SeqCst) {
|
|
|
|
download_process.terminate().unwrap();
|
|
|
|
tracing::info!("Waiting for download process to gracefully shutdown");
|
|
|
|
download_process
|
|
|
|
.wait_timeout(Duration::from_secs(90))
|
|
|
|
.unwrap();
|
|
|
|
tracing::info!("Download process terminated");
|
|
|
|
return ExitCode::SUCCESS;
|
|
|
|
}
|
|
|
|
sleep(Duration::from_millis(100));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
// Shared shutdown bool
|
|
|
|
let shutdown = Arc::new(Mutex::new(false));
|
|
|
|
// Shared shutdown channel
|
|
|
|
// When shutting down, the main thread will wait for all senders to be dropped
|
|
|
|
let (shutdown_sender, shutdown_receiver) = mpsc::channel();
|
|
|
|
|
|
|
|
// Shared channel to track shard status
|
|
|
|
let (status_sender, status_receiver) = mpsc::channel();
|
|
|
|
|
|
|
|
// Start shard processes
|
|
|
|
for rank in 0..num_shard {
|
2023-02-03 04:43:37 -07:00
|
|
|
let model_id = model_id.clone();
|
2023-01-31 10:53:56 -07:00
|
|
|
let revision = revision.clone();
|
2022-10-18 07:19:03 -06:00
|
|
|
let uds_path = shard_uds_path.clone();
|
|
|
|
let master_addr = master_addr.clone();
|
2023-02-14 05:02:16 -07:00
|
|
|
let huggingface_hub_cache = huggingface_hub_cache.clone();
|
|
|
|
let weights_cache_override = weights_cache_override.clone();
|
2022-10-18 07:19:03 -06:00
|
|
|
let status_sender = status_sender.clone();
|
|
|
|
let shutdown = shutdown.clone();
|
|
|
|
let shutdown_sender = shutdown_sender.clone();
|
2023-02-13 05:02:45 -07:00
|
|
|
let otlp_endpoint = otlp_endpoint.clone();
|
2022-10-18 07:19:03 -06:00
|
|
|
thread::spawn(move || {
|
|
|
|
shard_manager(
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id,
|
2023-01-31 10:53:56 -07:00
|
|
|
revision,
|
2022-10-27 06:25:29 -06:00
|
|
|
quantize,
|
2022-10-18 07:19:03 -06:00
|
|
|
uds_path,
|
|
|
|
rank,
|
|
|
|
num_shard,
|
|
|
|
master_addr,
|
|
|
|
master_port,
|
2023-02-14 05:02:16 -07:00
|
|
|
huggingface_hub_cache,
|
|
|
|
weights_cache_override,
|
2023-02-15 08:23:45 -07:00
|
|
|
disable_custom_kernels,
|
2023-03-02 04:30:41 -07:00
|
|
|
watermark_gamma,
|
|
|
|
watermark_delta,
|
2023-02-13 05:02:45 -07:00
|
|
|
otlp_endpoint,
|
2022-10-18 07:19:03 -06:00
|
|
|
status_sender,
|
|
|
|
shutdown,
|
|
|
|
shutdown_sender,
|
|
|
|
)
|
|
|
|
});
|
|
|
|
}
|
|
|
|
drop(shutdown_sender);
|
|
|
|
|
|
|
|
// Wait for shard to start
|
|
|
|
let mut shard_ready = 0;
|
|
|
|
while running.load(Ordering::SeqCst) {
|
|
|
|
match status_receiver.try_recv() {
|
|
|
|
Ok(ShardStatus::Ready) => {
|
|
|
|
shard_ready += 1;
|
|
|
|
if shard_ready == num_shard {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
Err(TryRecvError::Empty) => {
|
|
|
|
sleep(Duration::from_millis(100));
|
|
|
|
}
|
2023-04-13 03:07:11 -06:00
|
|
|
Ok(ShardStatus::Failed((rank, err))) => {
|
|
|
|
tracing::error!("Shard {} failed to start:\n{}", rank, err);
|
2022-10-18 07:19:03 -06:00
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
Err(TryRecvError::Disconnected) => {
|
|
|
|
tracing::error!("Shard status channel disconnected");
|
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We might have received a termination signal
|
|
|
|
if !running.load(Ordering::SeqCst) {
|
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
return ExitCode::SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
// All shard started
|
|
|
|
// Start webserver
|
|
|
|
tracing::info!("Starting Webserver");
|
2022-11-02 10:29:56 -06:00
|
|
|
let mut argv = vec![
|
|
|
|
"text-generation-router".to_string(),
|
|
|
|
"--max-concurrent-requests".to_string(),
|
|
|
|
max_concurrent_requests.to_string(),
|
2023-03-09 07:30:54 -07:00
|
|
|
"--max-best-of".to_string(),
|
|
|
|
max_best_of.to_string(),
|
2023-03-03 08:01:25 -07:00
|
|
|
"--max-stop-sequences".to_string(),
|
|
|
|
max_stop_sequences.to_string(),
|
2022-11-02 10:29:56 -06:00
|
|
|
"--max-input-length".to_string(),
|
|
|
|
max_input_length.to_string(),
|
2023-03-03 08:01:25 -07:00
|
|
|
"--max-total-tokens".to_string(),
|
|
|
|
max_total_tokens.to_string(),
|
2023-04-24 09:59:00 -06:00
|
|
|
"--waiting-served-ratio".to_string(),
|
|
|
|
waiting_served_ratio.to_string(),
|
2022-11-02 10:29:56 -06:00
|
|
|
"--max-waiting-tokens".to_string(),
|
|
|
|
max_waiting_tokens.to_string(),
|
|
|
|
"--port".to_string(),
|
|
|
|
port.to_string(),
|
|
|
|
"--master-shard-uds-path".to_string(),
|
2023-02-13 05:02:45 -07:00
|
|
|
format!("{shard_uds_path}-0"),
|
2022-11-02 10:29:56 -06:00
|
|
|
"--tokenizer-name".to_string(),
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id,
|
2022-11-02 10:29:56 -06:00
|
|
|
];
|
|
|
|
|
2023-04-24 09:59:00 -06:00
|
|
|
// Deprecate max_batch_size
|
|
|
|
if let Some(max_batch_size) = max_batch_size {
|
|
|
|
argv.push("--max-batch-size".to_string());
|
|
|
|
argv.push(max_batch_size.to_string())
|
|
|
|
} else {
|
|
|
|
argv.push("--max-batch-total-tokens".to_string());
|
|
|
|
argv.push(max_batch_total_tokens.to_string())
|
|
|
|
}
|
|
|
|
|
2023-04-18 08:16:06 -06:00
|
|
|
// Model optional revision
|
|
|
|
if let Some(ref revision) = revision {
|
|
|
|
argv.push("--revision".to_string());
|
|
|
|
argv.push(revision.to_string())
|
|
|
|
}
|
|
|
|
|
2022-11-02 10:29:56 -06:00
|
|
|
if json_output {
|
|
|
|
argv.push("--json-output".to_string());
|
|
|
|
}
|
|
|
|
|
2023-02-13 05:02:45 -07:00
|
|
|
// OpenTelemetry
|
|
|
|
if let Some(otlp_endpoint) = otlp_endpoint {
|
|
|
|
argv.push("--otlp-endpoint".to_string());
|
|
|
|
argv.push(otlp_endpoint);
|
|
|
|
}
|
|
|
|
|
2023-02-17 10:22:00 -07:00
|
|
|
// CORS origins
|
|
|
|
for origin in cors_allow_origin.into_iter() {
|
|
|
|
argv.push("--cors-allow-origin".to_string());
|
|
|
|
argv.push(origin);
|
|
|
|
}
|
|
|
|
|
2023-04-19 12:06:06 -06:00
|
|
|
// Copy current process env
|
|
|
|
let mut env: Vec<(OsString, OsString)> = env::vars_os().collect();
|
|
|
|
|
|
|
|
// Parse Inference API token
|
|
|
|
if let Ok(api_token) = env::var("HF_API_TOKEN") {
|
|
|
|
env.push(("HUGGING_FACE_HUB_TOKEN".into(), api_token.into()))
|
|
|
|
};
|
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
let mut webserver = match Popen::create(
|
2022-11-02 10:29:56 -06:00
|
|
|
&argv,
|
2022-10-18 07:19:03 -06:00
|
|
|
PopenConfig {
|
|
|
|
stdout: Redirection::Pipe,
|
|
|
|
stderr: Redirection::Pipe,
|
|
|
|
// Needed for the shutdown procedure
|
|
|
|
setpgid: true,
|
2023-04-19 12:06:06 -06:00
|
|
|
env: Some(env),
|
2022-10-18 07:19:03 -06:00
|
|
|
..Default::default()
|
|
|
|
},
|
|
|
|
) {
|
|
|
|
Ok(p) => p,
|
|
|
|
Err(err) => {
|
|
|
|
tracing::error!("Failed to start webserver: {}", err);
|
|
|
|
if let PopenError::IoError(err) = err {
|
|
|
|
if err.kind() == io::ErrorKind::NotFound {
|
|
|
|
tracing::error!("text-generation-router not found in PATH");
|
|
|
|
tracing::error!("Please install it with `make install-router`")
|
|
|
|
}
|
2022-10-27 06:25:29 -06:00
|
|
|
} else {
|
|
|
|
tracing::error!("{}", err);
|
2022-10-18 07:19:03 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
|
|
|
// Redirect STDOUT and STDERR to the console
|
|
|
|
let webserver_stdout = webserver.stdout.take().unwrap();
|
|
|
|
let webserver_stderr = webserver.stderr.take().unwrap();
|
|
|
|
|
|
|
|
thread::spawn(move || {
|
|
|
|
let stdout = BufReader::new(webserver_stdout);
|
|
|
|
let stderr = BufReader::new(webserver_stderr);
|
|
|
|
for line in stdout.lines() {
|
|
|
|
println!("{}", line.unwrap());
|
|
|
|
}
|
|
|
|
for line in stderr.lines() {
|
|
|
|
println!("{}", line.unwrap());
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
|
|
|
// Default exit code
|
|
|
|
let mut exit_code = ExitCode::SUCCESS;
|
|
|
|
|
|
|
|
while running.load(Ordering::SeqCst) {
|
2023-04-13 03:07:11 -06:00
|
|
|
if let Ok(ShardStatus::Failed((rank, err))) = status_receiver.try_recv() {
|
|
|
|
tracing::error!("Shard {rank} failed:\n{err}");
|
2022-10-18 07:19:03 -06:00
|
|
|
exit_code = ExitCode::FAILURE;
|
|
|
|
break;
|
|
|
|
};
|
|
|
|
|
|
|
|
match webserver.poll() {
|
|
|
|
Some(_) => {
|
|
|
|
tracing::error!("Webserver Crashed");
|
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
return ExitCode::FAILURE;
|
|
|
|
}
|
|
|
|
None => {
|
|
|
|
sleep(Duration::from_millis(100));
|
|
|
|
}
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
// Graceful termination
|
|
|
|
webserver.terminate().unwrap();
|
|
|
|
tracing::info!("Waiting for webserver to gracefully shutdown");
|
|
|
|
webserver.wait_timeout(Duration::from_secs(90)).unwrap();
|
|
|
|
tracing::info!("Webserver terminated");
|
|
|
|
shutdown_shards(shutdown, &shutdown_receiver);
|
|
|
|
|
|
|
|
exit_code
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug)]
|
|
|
|
enum ShardStatus {
|
|
|
|
Ready,
|
2023-04-13 03:07:11 -06:00
|
|
|
Failed((usize, String)),
|
2022-10-18 07:19:03 -06:00
|
|
|
}
|
|
|
|
|
|
|
|
#[allow(clippy::too_many_arguments)]
|
|
|
|
fn shard_manager(
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id: String,
|
2023-01-31 10:53:56 -07:00
|
|
|
revision: Option<String>,
|
2022-10-27 06:25:29 -06:00
|
|
|
quantize: bool,
|
2022-10-18 07:19:03 -06:00
|
|
|
uds_path: String,
|
|
|
|
rank: usize,
|
|
|
|
world_size: usize,
|
|
|
|
master_addr: String,
|
|
|
|
master_port: usize,
|
2023-02-14 05:02:16 -07:00
|
|
|
huggingface_hub_cache: Option<String>,
|
|
|
|
weights_cache_override: Option<String>,
|
2023-02-15 08:23:45 -07:00
|
|
|
disable_custom_kernels: bool,
|
2023-03-02 04:30:41 -07:00
|
|
|
watermark_gamma: Option<f32>,
|
|
|
|
watermark_delta: Option<f32>,
|
2023-02-13 05:02:45 -07:00
|
|
|
otlp_endpoint: Option<String>,
|
2022-10-18 07:19:03 -06:00
|
|
|
status_sender: mpsc::Sender<ShardStatus>,
|
|
|
|
shutdown: Arc<Mutex<bool>>,
|
|
|
|
_shutdown_sender: mpsc::Sender<()>,
|
|
|
|
) {
|
|
|
|
// Get UDS path
|
2023-02-13 05:02:45 -07:00
|
|
|
let uds_string = format!("{uds_path}-{rank}");
|
2022-10-18 07:19:03 -06:00
|
|
|
let uds = Path::new(&uds_string);
|
|
|
|
// Clean previous runs
|
|
|
|
fs::remove_file(uds).unwrap_or_default();
|
|
|
|
|
|
|
|
// Process args
|
|
|
|
let mut shard_argv = vec![
|
2022-10-28 11:24:00 -06:00
|
|
|
"text-generation-server".to_string(),
|
2022-10-18 07:19:03 -06:00
|
|
|
"serve".to_string(),
|
2023-02-03 04:43:37 -07:00
|
|
|
model_id,
|
2022-10-18 07:19:03 -06:00
|
|
|
"--uds-path".to_string(),
|
|
|
|
uds_path,
|
2023-01-05 04:01:23 -07:00
|
|
|
"--logger-level".to_string(),
|
2023-04-16 16:26:47 -06:00
|
|
|
"INFO".to_string(),
|
2023-01-05 04:01:23 -07:00
|
|
|
"--json-output".to_string(),
|
2022-10-18 07:19:03 -06:00
|
|
|
];
|
|
|
|
|
2023-02-13 05:02:45 -07:00
|
|
|
// Activate tensor parallelism
|
2022-10-18 07:19:03 -06:00
|
|
|
if world_size > 1 {
|
|
|
|
shard_argv.push("--sharded".to_string());
|
|
|
|
}
|
|
|
|
|
2022-10-27 06:25:29 -06:00
|
|
|
if quantize {
|
|
|
|
shard_argv.push("--quantize".to_string())
|
|
|
|
}
|
|
|
|
|
2023-02-13 05:02:45 -07:00
|
|
|
// Model optional revision
|
2023-01-31 10:53:56 -07:00
|
|
|
if let Some(revision) = revision {
|
|
|
|
shard_argv.push("--revision".to_string());
|
|
|
|
shard_argv.push(revision)
|
|
|
|
}
|
|
|
|
|
2023-02-13 05:02:45 -07:00
|
|
|
// OpenTelemetry
|
|
|
|
if let Some(otlp_endpoint) = otlp_endpoint {
|
|
|
|
shard_argv.push("--otlp-endpoint".to_string());
|
|
|
|
shard_argv.push(otlp_endpoint);
|
|
|
|
}
|
|
|
|
|
2023-02-16 03:20:23 -07:00
|
|
|
// Copy current process env
|
|
|
|
let mut env: Vec<(OsString, OsString)> = env::vars_os().collect();
|
|
|
|
|
|
|
|
// Torch Distributed Env vars
|
|
|
|
env.push(("RANK".into(), rank.to_string().into()));
|
|
|
|
env.push(("WORLD_SIZE".into(), world_size.to_string().into()));
|
|
|
|
env.push(("MASTER_ADDR".into(), master_addr.into()));
|
|
|
|
env.push(("MASTER_PORT".into(), master_port.to_string().into()));
|
|
|
|
env.push(("NCCL_ASYNC_ERROR_HANDLING".into(), "1".into()));
|
|
|
|
|
|
|
|
// Safetensors load fast
|
|
|
|
env.push(("SAFETENSORS_FAST_GPU".into(), "1".into()));
|
2022-10-22 12:00:15 -06:00
|
|
|
|
2023-02-18 06:04:11 -07:00
|
|
|
// Enable hf transfer for insane download speeds
|
2023-04-09 12:00:05 -06:00
|
|
|
let enable_hf_transfer = env::var("HF_HUB_ENABLE_HF_TRANSFER").unwrap_or("1".to_string());
|
|
|
|
env.push((
|
|
|
|
"HF_HUB_ENABLE_HF_TRANSFER".into(),
|
|
|
|
enable_hf_transfer.into(),
|
|
|
|
));
|
|
|
|
|
|
|
|
// Parse Inference API token
|
|
|
|
if let Ok(api_token) = env::var("HF_API_TOKEN") {
|
|
|
|
env.push(("HUGGING_FACE_HUB_TOKEN".into(), api_token.into()))
|
|
|
|
};
|
2023-02-18 06:04:11 -07:00
|
|
|
|
2023-02-14 05:02:16 -07:00
|
|
|
// If huggingface_hub_cache is some, pass it to the shard
|
2022-10-22 12:00:15 -06:00
|
|
|
// Useful when running inside a docker container
|
2023-02-14 05:02:16 -07:00
|
|
|
if let Some(huggingface_hub_cache) = huggingface_hub_cache {
|
2023-01-20 04:24:39 -07:00
|
|
|
env.push(("HUGGINGFACE_HUB_CACHE".into(), huggingface_hub_cache.into()));
|
2022-10-22 12:00:15 -06:00
|
|
|
};
|
2022-10-18 07:19:03 -06:00
|
|
|
|
2023-02-14 05:02:16 -07:00
|
|
|
// If weights_cache_override is some, pass it to the shard
|
2023-02-01 08:22:10 -07:00
|
|
|
// Useful when running inside a HuggingFace Inference Endpoint
|
2023-02-14 05:02:16 -07:00
|
|
|
if let Some(weights_cache_override) = weights_cache_override {
|
2023-02-02 06:59:27 -07:00
|
|
|
env.push((
|
|
|
|
"WEIGHTS_CACHE_OVERRIDE".into(),
|
|
|
|
weights_cache_override.into(),
|
|
|
|
));
|
2023-02-01 08:22:10 -07:00
|
|
|
};
|
|
|
|
|
2023-02-15 08:23:45 -07:00
|
|
|
// If disable_custom_kernels is true, pass it to the shard as an env var
|
|
|
|
if disable_custom_kernels {
|
|
|
|
env.push(("DISABLE_CUSTOM_KERNELS".into(), "True".into()))
|
|
|
|
}
|
|
|
|
|
2023-03-02 04:30:41 -07:00
|
|
|
// Watermark Gamma
|
|
|
|
if let Some(watermark_gamma) = watermark_gamma {
|
|
|
|
env.push(("WATERMARK_GAMMA".into(), watermark_gamma.to_string().into()))
|
|
|
|
}
|
|
|
|
|
|
|
|
// Watermark Delta
|
|
|
|
if let Some(watermark_delta) = watermark_delta {
|
|
|
|
env.push(("WATERMARK_DELTA".into(), watermark_delta.to_string().into()))
|
|
|
|
}
|
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
// Start process
|
2023-02-14 05:02:16 -07:00
|
|
|
tracing::info!("Starting shard {rank}");
|
2022-10-18 07:19:03 -06:00
|
|
|
let mut p = match Popen::create(
|
|
|
|
&shard_argv,
|
|
|
|
PopenConfig {
|
|
|
|
stdout: Redirection::Pipe,
|
|
|
|
stderr: Redirection::Pipe,
|
|
|
|
// Needed for the shutdown procedure
|
|
|
|
setpgid: true,
|
|
|
|
// NCCL env vars
|
2022-10-22 12:00:15 -06:00
|
|
|
env: Some(env),
|
2022-10-18 07:19:03 -06:00
|
|
|
..Default::default()
|
|
|
|
},
|
|
|
|
) {
|
|
|
|
Ok(p) => p,
|
|
|
|
Err(err) => {
|
|
|
|
if let PopenError::IoError(ref err) = err {
|
|
|
|
if err.kind() == io::ErrorKind::NotFound {
|
2022-10-28 11:24:00 -06:00
|
|
|
tracing::error!("text-generation-server not found in PATH");
|
2022-10-18 07:19:03 -06:00
|
|
|
tracing::error!("Please install it with `make install-server`")
|
|
|
|
}
|
|
|
|
}
|
2023-04-13 03:07:11 -06:00
|
|
|
status_sender
|
|
|
|
.send(ShardStatus::Failed((rank, err.to_string())))
|
|
|
|
.unwrap();
|
2022-10-18 07:19:03 -06:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
};
|
|
|
|
|
2023-01-05 04:01:23 -07:00
|
|
|
// Redirect STDOUT to the console
|
|
|
|
let shard_stdout = p.stdout.take().unwrap();
|
|
|
|
|
|
|
|
thread::spawn(move || {
|
|
|
|
// Enter shard-manager tracing span
|
|
|
|
let stdout = BufReader::new(shard_stdout);
|
|
|
|
let _span = tracing::span!(tracing::Level::INFO, "shard-manager", rank = rank).entered();
|
|
|
|
for line in stdout.lines() {
|
|
|
|
// Parse loguru logs
|
2023-04-16 16:26:47 -06:00
|
|
|
if let Ok(log) = serde_json::from_str::<PythonLogMessage>(&line.unwrap()) {
|
|
|
|
log.trace();
|
2023-01-05 04:01:23 -07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
});
|
|
|
|
|
2022-10-18 07:19:03 -06:00
|
|
|
let mut ready = false;
|
|
|
|
let start_time = Instant::now();
|
2022-10-22 12:00:15 -06:00
|
|
|
let mut wait_time = Instant::now();
|
2022-10-18 07:19:03 -06:00
|
|
|
loop {
|
|
|
|
// Process exited
|
|
|
|
if p.poll().is_some() {
|
2023-04-13 03:07:11 -06:00
|
|
|
let mut err = String::new();
|
|
|
|
p.stderr.take().unwrap().read_to_string(&mut err).unwrap();
|
|
|
|
status_sender
|
|
|
|
.send(ShardStatus::Failed((rank, err)))
|
|
|
|
.unwrap();
|
2022-10-18 07:19:03 -06:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// We received a shutdown signal
|
|
|
|
if *shutdown.lock().unwrap() {
|
|
|
|
p.terminate().unwrap();
|
|
|
|
let _ = p.wait_timeout(Duration::from_secs(90));
|
2023-02-14 05:02:16 -07:00
|
|
|
tracing::info!("Shard {rank} terminated");
|
2022-10-18 07:19:03 -06:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Shard is ready
|
|
|
|
if uds.exists() && !ready {
|
2023-02-14 05:02:16 -07:00
|
|
|
tracing::info!("Shard {rank} ready in {:?}", start_time.elapsed());
|
2022-10-18 07:19:03 -06:00
|
|
|
status_sender.send(ShardStatus::Ready).unwrap();
|
|
|
|
ready = true;
|
2022-10-27 06:25:29 -06:00
|
|
|
} else if !ready && wait_time.elapsed() > Duration::from_secs(10) {
|
2023-02-14 05:02:16 -07:00
|
|
|
tracing::info!("Waiting for shard {rank} to be ready...");
|
2022-10-27 06:25:29 -06:00
|
|
|
wait_time = Instant::now();
|
2022-10-18 07:19:03 -06:00
|
|
|
}
|
2022-10-22 12:00:15 -06:00
|
|
|
sleep(Duration::from_millis(100));
|
2022-10-18 07:19:03 -06:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn shutdown_shards(shutdown: Arc<Mutex<bool>>, shutdown_receiver: &mpsc::Receiver<()>) {
|
|
|
|
tracing::info!("Shutting down shards");
|
|
|
|
// Update shutdown value to true
|
|
|
|
// This will be picked up by the shard manager
|
|
|
|
{
|
|
|
|
let mut shutdown = shutdown.lock().unwrap();
|
|
|
|
*shutdown = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for shards to shutdown
|
|
|
|
// This will block till all shutdown_sender are dropped
|
|
|
|
let _ = shutdown_receiver.recv();
|
|
|
|
}
|
2023-03-08 05:53:41 -07:00
|
|
|
|
|
|
|
fn num_cuda_devices() -> Option<usize> {
|
|
|
|
if let Ok(cuda_visible_devices) = env::var("CUDA_VISIBLE_DEVICES") {
|
|
|
|
let n_devices = cuda_visible_devices.split(',').count();
|
|
|
|
return Some(n_devices);
|
|
|
|
}
|
|
|
|
None
|
|
|
|
}
|
2023-04-16 16:26:47 -06:00
|
|
|
|
|
|
|
#[derive(Deserialize)]
|
|
|
|
#[serde(rename_all = "UPPERCASE")]
|
|
|
|
enum PythonLogLevelEnum {
|
|
|
|
Trace,
|
|
|
|
Debug,
|
|
|
|
Info,
|
|
|
|
Success,
|
|
|
|
Warning,
|
|
|
|
Error,
|
|
|
|
Critical,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Deserialize)]
|
|
|
|
struct PythonLogLevel {
|
|
|
|
name: PythonLogLevelEnum,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Deserialize)]
|
|
|
|
struct PythonLogRecord {
|
|
|
|
level: PythonLogLevel,
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Deserialize)]
|
|
|
|
struct PythonLogMessage {
|
|
|
|
text: String,
|
|
|
|
record: PythonLogRecord,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl PythonLogMessage {
|
|
|
|
fn trace(&self) {
|
|
|
|
match self.record.level.name {
|
|
|
|
PythonLogLevelEnum::Trace => tracing::trace!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Debug => tracing::debug!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Info => tracing::info!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Success => tracing::info!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Warning => tracing::warn!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Error => tracing::error!("{}", self.text),
|
|
|
|
PythonLogLevelEnum::Critical => tracing::error!("{}", self.text),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|