Skip to content
This repository has been archived by the owner on Jul 12, 2024. It is now read-only.

codephreak = memory.py + aglm.py + automind.py [codephreak automind has moved to codephreak automindx] https://github.com/pythaiml/automindx

License

Notifications You must be signed in to change notification settings

Professor-Codephreak/automind

Repository files navigation

This build will experience a 4096 memory input too large error with pasting of a large prompt into the UI

ongoing development of Professor Codephreak LLM here
To avoid confusion automindx is the local language model executable and AUTOMINDx is the global deployment autonomous agents development environment

This Professor Codephreak build is being left intact as the first example of the working aGLM prototype Work continues at automindx as the next evolution of Professor Codephreak's core intention value and success creating AUTOMINDx

IAML Intelligent Autonomous Machine Learning

"I Am Machine Learning"

project codename = codephreak
project direction = automind automindx

"Professor Codephreak is an expert in machine learning, computer science and computer programming ..."
uiux.py provides Gradio output to local server for local language model interaction
Gradio interacts with html javascript and css custom Gradio
Gradio Javascript Client
Documentation: codephreak = uiux.py + memory.py + automind.py + aglm.py

User Interface and Interaction (uiux.py)

(uiux.py) provides a user interface using the Gradio library to facilitate user interaction. This chatbot interface takes the user input, processes it, generating response includingc onversation memory handling storing and managing user instructions and model responses. The result is a local language model prompted to run as "Professor Codephreak is an expert in computer programming ....." that refers to itself as "codephreak" Contextual Conversation Management

(automind.py) provides the mechanism to format and managing conversation history using the format_to_llama_chat_style function. (automind.py) creates coherent conversation context integrating memory management and with chatbot behavior. Handling different model types and initialization is based on model name calling the model from the models folder.

(memory.py) handles Conversation Memory Management by creating a class DialogEntry to represent individual conversation dialog entries. A function save_conversation_memory is called to save conversation history as JSON files. Memory is created by storing user instructions and model responses in memory files for context management as .json outputs with a timestamp.

LLAMA Model Interaction

(aglm.py)

Initializes the LLAMA language model and tokenizer based on the specified models_folder models handling processing and tokenization of conversation context using the LLAMA model then generating contextually relevant responses using the LLAMA model. AUTOMIND uses a LLAMA language model to generate responses based on user instructions and conversation history. (uiux.py) encompasses user interface design, memory management, context handling, and interaction with the language model to create a conversational experience for users.


To install right click "Save link as ..." automind.install chmod +x automind.install && automind.install

details and verbose procedure

  1. Right-click the following link: automind.install
  2. Choose "Save link as..." or "Download linked file" from the context menu.
  3. Select a location on your computer to save the file.
  4. from the terminal
  5. chmod +x automind.install && ./automind.install


RUN codephreak

python3 uiux.py --model_name="TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML" --tokenizer_name="TheBloke/llama2-7b-chat-codeCherryPop-qLoRA-GGML" --model_type="ggml" --save_history --file_name="llama-2-7b-chat-codeCherryPop.ggmlv3.q4_1.bin"

About

codephreak = memory.py + aglm.py + automind.py [codephreak automind has moved to codephreak automindx] https://github.com/pythaiml/automindx

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published