2023-03-30 07:26:27 -06:00
|
|
|
/// Text Generation Inference benchmarking tool
|
|
|
|
///
|
|
|
|
/// Inspired by the great Oha app: https://github.com/hatoo/oha
|
|
|
|
/// and: https://github.com/orhun/rust-tui-template
|
|
|
|
use clap::Parser;
|
|
|
|
use std::path::Path;
|
|
|
|
use text_generation_client::ShardedClient;
|
2023-04-29 04:17:30 -06:00
|
|
|
use tokenizers::{FromPretrainedParameters, Tokenizer};
|
2023-03-30 07:26:27 -06:00
|
|
|
use tracing_subscriber::layer::SubscriberExt;
|
|
|
|
use tracing_subscriber::util::SubscriberInitExt;
|
|
|
|
use tracing_subscriber::EnvFilter;
|
|
|
|
|
|
|
|
/// App Configuration
|
|
|
|
#[derive(Parser, Debug)]
|
|
|
|
#[clap(author, version, about, long_about = None)]
|
|
|
|
struct Args {
|
|
|
|
#[clap(short, long, env)]
|
|
|
|
tokenizer_name: String,
|
2023-04-29 04:17:30 -06:00
|
|
|
#[clap(default_value = "main", long, env)]
|
|
|
|
revision: String,
|
2023-03-30 07:26:27 -06:00
|
|
|
#[clap(short, long)]
|
|
|
|
batch_size: Option<Vec<u32>>,
|
|
|
|
#[clap(default_value = "10", short, long, env)]
|
|
|
|
sequence_length: u32,
|
|
|
|
#[clap(default_value = "8", short, long, env)]
|
|
|
|
decode_length: u32,
|
|
|
|
#[clap(default_value = "10", short, long, env)]
|
|
|
|
runs: usize,
|
|
|
|
#[clap(default_value = "1", short, long, env)]
|
|
|
|
warmups: usize,
|
|
|
|
#[clap(default_value = "/tmp/text-generation-server-0", short, long, env)]
|
|
|
|
master_shard_uds_path: String,
|
|
|
|
}
|
|
|
|
|
|
|
|
fn main() -> Result<(), Box<dyn std::error::Error>> {
|
|
|
|
// Get args
|
|
|
|
let args = Args::parse();
|
|
|
|
// Pattern match configuration
|
|
|
|
let Args {
|
|
|
|
tokenizer_name,
|
2023-04-29 04:17:30 -06:00
|
|
|
revision,
|
2023-03-30 07:26:27 -06:00
|
|
|
batch_size,
|
|
|
|
sequence_length,
|
|
|
|
decode_length,
|
|
|
|
runs,
|
|
|
|
warmups,
|
|
|
|
master_shard_uds_path,
|
|
|
|
} = args;
|
|
|
|
|
|
|
|
let batch_size = batch_size.unwrap_or(vec![1, 2, 4, 8, 16, 32]);
|
|
|
|
|
|
|
|
init_logging();
|
|
|
|
|
|
|
|
// Tokenizer instance
|
|
|
|
// This will only be used to validate payloads
|
|
|
|
tracing::info!("Loading tokenizer");
|
|
|
|
let local_path = Path::new(&tokenizer_name);
|
|
|
|
let tokenizer =
|
|
|
|
if local_path.exists() && local_path.is_dir() && local_path.join("tokenizer.json").exists()
|
|
|
|
{
|
|
|
|
// Load local tokenizer
|
|
|
|
tracing::info!("Found local tokenizer");
|
|
|
|
Tokenizer::from_file(local_path.join("tokenizer.json")).unwrap()
|
|
|
|
} else {
|
2023-04-29 04:17:30 -06:00
|
|
|
tracing::info!("Downloading tokenizer");
|
|
|
|
|
|
|
|
// Parse Huggingface hub token
|
|
|
|
let auth_token = std::env::var("HUGGING_FACE_HUB_TOKEN").ok();
|
|
|
|
|
2023-03-30 07:26:27 -06:00
|
|
|
// Download and instantiate tokenizer
|
|
|
|
// We need to download it outside of the Tokio runtime
|
2023-04-29 04:17:30 -06:00
|
|
|
let params = FromPretrainedParameters {
|
|
|
|
revision,
|
|
|
|
auth_token,
|
|
|
|
..Default::default()
|
|
|
|
};
|
|
|
|
Tokenizer::from_pretrained(tokenizer_name.clone(), Some(params)).unwrap()
|
2023-03-30 07:26:27 -06:00
|
|
|
};
|
|
|
|
tracing::info!("Tokenizer loaded");
|
|
|
|
|
|
|
|
// Launch Tokio runtime
|
|
|
|
tokio::runtime::Builder::new_multi_thread()
|
|
|
|
.enable_all()
|
|
|
|
.build()
|
|
|
|
.unwrap()
|
|
|
|
.block_on(async {
|
|
|
|
// Instantiate sharded client from the master unix socket
|
|
|
|
tracing::info!("Connect to model server");
|
|
|
|
let mut sharded_client = ShardedClient::connect_uds(master_shard_uds_path)
|
|
|
|
.await
|
|
|
|
.expect("Could not connect to server");
|
|
|
|
// Clear the cache; useful if the webserver rebooted
|
|
|
|
sharded_client
|
|
|
|
.clear_cache(None)
|
|
|
|
.await
|
|
|
|
.expect("Unable to clear cache");
|
|
|
|
tracing::info!("Connected");
|
|
|
|
|
|
|
|
// Run app
|
|
|
|
text_generation_benchmark::run(
|
|
|
|
tokenizer_name,
|
|
|
|
tokenizer,
|
|
|
|
batch_size,
|
|
|
|
sequence_length,
|
|
|
|
decode_length,
|
|
|
|
runs,
|
|
|
|
warmups,
|
|
|
|
sharded_client,
|
|
|
|
)
|
|
|
|
.await
|
|
|
|
.unwrap();
|
|
|
|
});
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Init logging using LOG_LEVEL
|
|
|
|
fn init_logging() {
|
|
|
|
// STDOUT/STDERR layer
|
|
|
|
let fmt_layer = tracing_subscriber::fmt::layer()
|
|
|
|
.with_file(true)
|
|
|
|
.with_line_number(true);
|
|
|
|
|
|
|
|
// Filter events with LOG_LEVEL
|
|
|
|
let env_filter =
|
|
|
|
EnvFilter::try_from_env("LOG_LEVEL").unwrap_or_else(|_| EnvFilter::new("info"));
|
|
|
|
|
|
|
|
tracing_subscriber::registry()
|
|
|
|
.with(env_filter)
|
|
|
|
.with(fmt_layer)
|
|
|
|
.init();
|
|
|
|
}
|