Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix typo in default model path #1366

Merged
merged 3 commits into from
May 16, 2023
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion examples/common.h
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ struct gpt_params {
float mirostat_tau = 5.00f; // target entropy
float mirostat_eta = 0.10f; // learning rate

std::string model = "models/lamma-7B/ggml-model.bin"; // model path
std::string model = "models/llama-7B/ggml-model.bin"; // model path
Copy link
Collaborator

@prusnak prusnak May 8, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Readme mentions ./models/7B/ggml-model-q4_0.bin and I think we should use this everywhere for unification.

Suggested change
std::string model = "models/llama-7B/ggml-model.bin"; // model path
std::string model = "models/7B/ggml-model-q4_0.bin"; // model path

Can you also update this PR to remove params.model = "models/llama-7B/ggml-model.bin"; from files you mention? It doesn't make sense to override the value if it is the same as the default, no?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Should we use models/7B or models/llama-7B as the canonical path? I used to have just 7B but with all the new models supported by llama.cpp I'm now tending to prefer llama-7B for disambiguation.

Copy link
Collaborator

@prusnak prusnak May 9, 2023

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I prefer to keep 7B because that's how the directory is named when you download the model.

Also the instructiond should work no matter what model is located in the 7B directory (Alpaca, Vicuna, etc.)

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I've implemented your suggested changes. The default model path is now defined in examples/common.h and this is used elsewhere implicitly, instead of repeating the string in several places.

std::string prompt = "";
std::string path_session = ""; // path to file for saving/loading model eval state
std::string input_prefix = ""; // string to prefix user inputs with
Expand Down