Skip to content

Latest commit

 

History

History
47 lines (26 loc) · 1.04 KB

xinference.md

File metadata and controls

47 lines (26 loc) · 1.04 KB

Xinference Local Model Deployment Guide

Step 1: Install Xinference

Install Xinference along with all its optional dependencies:

pip install "xinference[all]"

Step 2: Run Xinference

Start the Xinference service:

xinference

Step 3: Register a New Model

When registering a new model, the fields highlighted in red are fixed settings, while other fields can be customized as needed. Finally, click "Register Model."

alt text alt text

Step 4: Find the Registered Model

After the model is registered, navigate to the "Custom Models" page to find the newly registered model.

alt text

Step 5: Run the Model

Click the "Run" button to start the model.

alt text

Step 6: Launch WebUI

Launch the WebUI to enable interactive inference.

alt text

Step 7: Start Inference

Now you can use the WebUI to perform model inference!