Skip to content

Commit

Permalink
finetune : fix ggml_allocr lifetimes (tmp workaround) (ggerganov#5033)
Browse files Browse the repository at this point in the history
* Fix issue with alloc causing max_compute_size to be calculated

* remove ggml_allocr_free as suggested in issue ggerganov#4791
  • Loading branch information
bzuzo authored and hodlen committed Apr 1, 2024
1 parent c15e1b4 commit c458aac
Showing 1 changed file with 0 additions and 4 deletions.
4 changes: 0 additions & 4 deletions examples/train-text-from-scratch/train-text-from-scratch.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -263,7 +263,6 @@ static void init_model(struct my_llama_model * model) {
model->data.resize(size + tensor_alignment);
alloc = ggml_allocr_new(model->data.data(), model->data.size(), tensor_alignment);
alloc_model(alloc, model);
ggml_allocr_free(alloc);
}

static void randomize_model(struct my_llama_model * model, int seed, float mean, float std, float min, float max) {
Expand Down Expand Up @@ -1102,7 +1101,6 @@ int main(int argc, char ** argv) {
alloc = ggml_allocr_new(mem_input_data.data(), mem_input_data.size(), tensor_alignment);
ggml_allocr_alloc(alloc, tokens_input);
ggml_allocr_alloc(alloc, target_probs);
ggml_allocr_free(alloc);

// context for compute tensors without their data
const size_t estimated_compute_size_wo_data = (
Expand Down Expand Up @@ -1149,7 +1147,6 @@ int main(int argc, char ** argv) {
best_compute_size = max_compute_size;
best_order = gf->order;
}
ggml_allocr_free(alloc);
ggml_free(ctx_compute);
}
size_t max_compute_size = best_compute_size;
Expand Down Expand Up @@ -1177,7 +1174,6 @@ int main(int argc, char ** argv) {
params.common.use_flash,
params.common.use_checkpointing
);
ggml_allocr_free(alloc);

std::vector<llama_token> train_tokens;
std::vector<size_t> train_samples_begin;
Expand Down

0 comments on commit c458aac

Please sign in to comment.