Note that this project is discontinued and a rebranded, newer better alternative will be released on the upcoming months hopeefully. You can still follow the tutorials, they should work as expected, if they don't, then instead compile the gpt-2 binaries directly from ggml.
This is a public research experiment and personal project on a family of prompt-fine-tuned + transfer-learning GPT models for generation tasks on chat form. Each one of these iterations are in current evaluation (prompt-engineering) to explore better ways to extract useful conversational tasks from them. An Android app is on the way. Link will be available through this file in the future. "Grafito" comes from Spanish word and it means Graphite.
If you are curious on how I'm finetuning these models, you can see this little guide I've made: Finetuning GPT-2 in 4 steps
👉 Google Play Store App link
App is temporaly down.
Note that finetuned-codename-version and codename-version are used interchangebly here and that the current models, names and/or techniques are prone to change over time. The techniques used are One-Shot, Few-Shot, Transfer Learning through LoRA or simple "finetuning".
Edit 03/20/2023: The codename for diamond-004 changed to emerald-003.
Edit 04/04/2023: Cerebras introduced their new models.
Edit 05/24/2023: Things have changed a lot, now heading towards to lora models.
Edit 05/26/2023: Things have changed (again), now heading towards to "native" training.
Note that obsidian models are lab models while the others are the official ones
Compiling grafito and loading a model
git clone https://github.com/appvoid/grafito.git
cd grafito/grafito-cpp
mkdir build && cd build
cmake .. && make
# you should have now a grafito-engine binary.
cd bin
chmod +x grafito-engine
after your downloaded lazulo-001:
./grafito-engine -m lazuli-001 -p "hello"
The datasets used during the research are being modified, transformed and improved over time. Then, after filtering, cleaning and selecting in an automated way the best samples for the task, was decided to name the final dataset "grafito-25k". Small models trained on this dataset seems to be getting better are trained on it. More and better research is needed in order to prevent overfitting. But results are promising.
Dataset | Samples |
---|---|
anthropic-hh-rlhf | 43835 |
dolly-s | 4067 |
grafito-25k | 24736 |
- using temperature near to zero (this is not the a definitive benchmark, it's just to measure accuracy to some extent)
- page-ruby is an internal model trained on additional wikipedia entries
query (custom prompted) | base-124m | base-355m | page-ruby-002 | instruct-ruby-002 | instruct-lazuli-002 |
---|---|---|---|---|---|
what is 2+2 | 2+2 | 2+2 | 2+2 is a unit of measurement f... | the ratio of the number of prot.. | 2 + 2 |
what's your purpose | (repeats the query) | (repeats the query) | Function is what you mean by pu.. | to provide a service or a servi.. | to provide information to the u.. |
i'm in a bad situtation | (repeats the query) | (repeats the query) | I'm not sure I understand the q.. | stay calm and move away from th.. | a situation in which the indivi.. |
what's your name | (repeats the query) | (repeats the query) | Your Name is a Canadian rock ba.. | Yukiya Nomura | |
how can i make a videogame | (repeats the query) | (repeats the query) | You can make a videogañame wit.. | by using a blender or food pro... | create a computer-generated vir.. |
write a code in python to sum 2.. | (repeats the query) | 2 | do 2 + 1 = 3, 5 + 1 = 6, 9 + 1... | sum(x,y) | 2 + 2 + 2 = 16 |
As you can see, complex inputs makes it more probable to do what you ask the model, even though instruct-ruby-002 is not so good at math, it started to understand more complex, elaborated intructions. Notice how when asked (instruct-ruby-002) about how to make a game, it "knows" that the word "blender" (a 3d graphics engine) is related to videogames but the model is too biased (as in the real life with people when don't know about something), about the meaning of that word. Because blender as a graphical software is not present enough in the pre-training nor fine-tuning dataset. While it seems that ruby is more willing to write code when asked, as expected, lazuli tries to complete the instruction more often.