| # Serve and Deploy LLMs | |
| This document shows how you can serve a LitGPT for deployment. | |
| | |
| ## Serve an LLM with LitServe | |
| This section illustrates how we can set up an inference server for a phi-2 LLM using `litgpt serve` that is minimal and highly scalable. | |
| | |
| ### Step 1: Start the inference server | |
| ```bash | |
| # 1) Download a pretrained model (alternatively, use your own finetuned model) | |
| litgpt download microsoft/phi-2 | |
| # 2) Start the server | |
| litgpt serve microsoft/phi-2 | |
| ``` | |
| > [!TIP] | |
| > Use `litgpt serve --help` to display additional options, including the port, devices, LLM temperature setting, and more. | |
| | |
| ### Step 2: Query the inference server | |
| You can now send requests to the inference server you started in step 2. For example, in a new Python session, we can send requests to the inference server as follows: | |
| ```python | |
| import requests, json | |
| response = requests.post( | |
| "http://127.0.0.1:8000/predict", | |
| json={"prompt": "Fix typos in the following sentence: Example input"} | |
| ) | |
| print(response.json()["output"]) | |
| ``` | |
| Executing the code above prints the following output: | |
| ``` | |
| Example input. | |
| ``` | |
| | |
| ### Optional: Use the streaming mode | |
| The 2-step procedure described above returns the complete response all at once. If you want to stream the response on a token-by-token basis, start the server with the streaming option enabled: | |
| ```bash | |
| litgpt serve microsoft/phi-2 --stream true | |
| ``` | |
| Then, use the following updated code to query the inference server: | |
| ```python | |
| import requests, json | |
| response = requests.post( | |
| "http://127.0.0.1:8000/predict", | |
| json={"prompt": "Fix typos in the following sentence: Example input"}, | |
| stream=True | |
| ) | |
| # stream the response | |
| for line in response.iter_lines(decode_unicode=True): | |
| if line: | |
| print(json.loads(line)["output"], end="") | |
| ``` | |
| ``` | |
| Sure, here is the corrected sentence: | |
| Example input | |
| ``` | |
| | |
| ## Serve an LLM with OpenAI-compatible API | |
| LitGPT provides OpenAI-compatible endpoints that allow you to use the OpenAI SDK or any OpenAI-compatible client to interact with your models. This is useful for integrating LitGPT into existing applications that use the OpenAI API. | |
| | |
| ### Step 1: Start the server with OpenAI specification | |
| ```bash | |
| # 1) Download a pretrained model (alternatively, use your own finetuned model) | |
| litgpt download HuggingFaceTB/SmolLM2-135M-Instruct | |
| # 2) Start the server with OpenAI-compatible endpoints | |
| litgpt serve HuggingFaceTB/SmolLM2-135M-Instruct --openai_spec true | |
| ``` | |
| > [!TIP] | |
| > The `--openai_spec true` flag enables OpenAI-compatible endpoints at `/v1/chat/completions` instead of the default `/predict` endpoint. | |
| | |
| ### Step 2: Query using OpenAI-compatible endpoints | |
| You can now send requests to the OpenAI-compatible endpoint using curl: | |
| ```bash | |
| curl -X POST http://127.0.0.1:8000/v1/chat/completions \ | |
| -H "Content-Type: application/json" \ | |
| -d '{ | |
| "model": "SmolLM2-135M-Instruct", | |
| "messages": [{"role": "user", "content": "Hello! How are you?"}] | |
| }' | |
| ``` | |
| Or use the OpenAI Python SDK: | |
| ```python | |
| from openai import OpenAI | |
| # Configure the client to use your local LitGPT server | |
| client = OpenAI( | |
| base_url="http://127.0.0.1:8000/v1", | |
| api_key="not-needed" # LitGPT doesn't require authentication by default | |
| ) | |
| response = client.chat.completions.create( | |
| model="SmolLM2-135M-Instruct", | |
| messages=[ | |
| {"role": "user", "content": "Hello! How are you?"} | |
| ] | |
| ) | |
| print(response.choices[0].message.content) | |
| ``` | |
| | |
| ## Serve an LLM UI with Chainlit | |
| If you are interested in developing a simple ChatGPT-like UI prototype, see the Chainlit tutorial in the following Studio: | |
| <a target="_blank" href="https://lightning.ai/lightning-ai/studios/chatgpt-like-llm-uis-via-chainlit"> | |
| <img src="https://pl-bolts-doc-images.s3.us-east-2.amazonaws.com/app-2/studio-badge.svg" alt="Open In Studio"/> | |
| </a> | |