Skip to content
This repository was archived by the owner on Jul 9, 2024. It is now read-only.
/ llama.cpp-qt Public archive

Llama.cpp-qt is a Python-based GUI wrapper for the LLama.cpp server, providing a user-friendly interface for configuring and running the server. LLama.cpp is a lightweight implementation of GPT-like models.

License

Notifications You must be signed in to change notification settings

tohurtv/llama.cpp-qt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

79 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

llama.cpp-qt

Python Version

Llama.cpp-qt is a Python-based graphical wrapper for the LLama.cpp server, providing a user-friendly interface for configuring and running the server. LLama.cpp is a lightweight implementation of GPT-like models.

Benifits over other LLama.cpp solutions

Most other interfaces for llama.cpp run exclusively through python, meaning its the llama.cpp converted to python in some form or another and depending on your hardware there is overhead to running directly in python. python is slower then C++, C++ is a Low-level programming language meaning its pretty close to the hardware, python is a high level programming language which is fine for GUIs (llama.cpp-qt) but for applications that require high performance such as AI models you want to be as low level as you can be. llama.cpp-qt is a wrapper for llama.cpp meaning it runs llama.cpp directly and is faster then other llama.cpp solutions. I personally get about double the tokens per second compared to Text Gen UI, Koboldcpp and llama-cpp-python. llama.cpp-qt is also cross platform meaning it runs on Linux and Windows. (macos support coming soon).

Requirements

Before you begin, ensure you have met the following requirements:

  • Python 3.10 or higher: You can download it from the official Python website or your Linux distribution's repositories.
  • Python3-virtualenv: For venv creation on run time/installation.
  • pyqt5 or Python-AnyQT and QT 5: For the GUI.These should be installed by your Linux distros package manager if you want to use your system theme, if not the default QT theme will be used by the venv.
  • llama.cpp requirements: You can learn the requirements for CPU,Cuda, or AMD rocm builds here, Llama.cpp Github.

Build Linux

To build and run the LLama.cpp-qt Wrapper, follow these steps:

  1. Clone this repository to your local machine:

    git clone https://github.com/TohurTV/llama.cpp-qt.git
  2. Change your current directory to the cloned repository:

    cd llama.cpp-qt
  3. Run the build for your platform:for AMD GPUs run

    sh ./build-rocm.sh

    Cuda is build-cuda.sh and cpu only build is build-cpu.sh

  4. Run llama.cpp-qt

    ./llama.cpp-qt

Systemwide installation

To install systemwide after running the build script run:

sh ./install.sh

Windows Setup

To run the LLama.cpp-qt Wrapper, follow these steps:

  1. Clone this repository to your local machine:

    git clone https://github.com/TohurTV/llama.cpp-qt.git
  2. Change your current directory to the cloned repository:

    cd llama.cpp-qt
  3. Download the lastest release build from Llama.cpp Releases or use the included bat files download-llama.cpp-{version).bat. Openblas for running on cpu, Cublas for running on Nvidia GPUs, and clblast for other GPUs. Rocm build instructions coming soon for windows.

  4. If you downloaded the release from llama.cpp's github Extract the release build and copy server.exe and all of the .dlls to the llama.cpp-qt folder if you used one of the bat files proceed to next step.

  5. Run the start.bat file to start the llama.cpp-qt GUI.

Systemwide installation

Windows installers coming soon.

About

Llama.cpp-qt is a Python-based GUI wrapper for the LLama.cpp server, providing a user-friendly interface for configuring and running the server. LLama.cpp is a lightweight implementation of GPT-like models.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published