2023-08-10 02:24:52 -06:00
# Installation
This section explains how to install the CLI tool as well as installing TGI from source. **The strongly recommended approach is to use Docker, as it does not require much setup. Check [the Quick Tour](./quicktour) to learn how to run TGI with Docker.**
## Install CLI
2024-02-16 03:58:58 -07:00
You can use TGI command-line interface (CLI) to download weights, serve and quantize models, or get information on serving parameters.
2023-08-10 02:24:52 -06:00
2023-08-10 07:00:30 -06:00
To install the CLI, you need to first clone the TGI repository and then run `make` .
```bash
git clone https://github.com/huggingface/text-generation-inference.git & & cd text-generation-inference
make install
```
If you would like to serve models with custom kernels, run
```bash
BUILD_EXTENSIONS=True make install
```
2023-08-10 02:24:52 -06:00
## Local Installation from Source
Before you start, you will need to setup your environment, and install Text Generation Inference. Text Generation Inference is tested on **Python 3.9+** .
2024-02-16 03:58:58 -07:00
Text Generation Inference is available on pypi, conda and GitHub.
2023-08-10 02:24:52 -06:00
To install and launch locally, first [install Rust ](https://rustup.rs/ ) and create a Python virtual environment with at least
Python 3.9, e.g. using conda:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 02:24:52 -06:00
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
conda create -n text-generation-inference python=3.9
conda activate text-generation-inference
```
You may also need to install Protoc.
On Linux:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 02:24:52 -06:00
PROTOC_ZIP=protoc-21.12-linux-x86_64.zip
curl -OL https://github.com/protocolbuffers/protobuf/releases/download/v21.12/$PROTOC_ZIP
sudo unzip -o $PROTOC_ZIP -d /usr/local bin/protoc
sudo unzip -o $PROTOC_ZIP -d /usr/local 'include/*'
rm -f $PROTOC_ZIP
```
On MacOS, using Homebrew:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 02:24:52 -06:00
brew install protobuf
```
Then run to install Text Generation Inference:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 07:00:30 -06:00
git clone https://github.com/huggingface/text-generation-inference.git & & cd text-generation-inference
BUILD_EXTENSIONS=True make install
2023-08-10 02:24:52 -06:00
```
< Tip warning = {true} >
On some machines, you may also need the OpenSSL libraries and gcc. On Linux machines, run:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 02:24:52 -06:00
sudo apt-get install libssl-dev gcc -y
```
< / Tip >
Once installation is done, simply run:
2023-08-10 06:32:51 -06:00
```bash
2023-08-10 02:24:52 -06:00
make run-falcon-7b-instruct
```
This will serve Falcon 7B Instruct model from the port 8080, which we can query.