Skip to content

This is an attempt to configure a chatbot using the falcon-7b-instruct parameter model to run locally on a machine with <8Gb VRAM. Using 4 bit quantization to reduce memory load.

Notifications You must be signed in to change notification settings

Mohamad-Hussein/Falcon7B-Chat

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Falcon7B-Chat

contributions welcome

Python pytorch

This is an attempt to configure a chatbot using the falcon-7b-instruct parameter model to run locally on a machine with <8Gb VRAM. Using 4 bit quantization to reduce memory load. Source to guide.

Set up

  1. Download docker
  2. docker build -t chatbot-image:latest -f docker/Dockerfile .
  3. docker run -it --gpus all -v $(pwd):/workspace chatbot-image:latest bash
  4. ...

About

This is an attempt to configure a chatbot using the falcon-7b-instruct parameter model to run locally on a machine with <8Gb VRAM. Using 4 bit quantization to reduce memory load.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published