Skip to content

Commit

Permalink
resolving merge conflict
Browse files Browse the repository at this point in the history
  • Loading branch information
deepanker13 committed Jan 10, 2024
1 parent 3204149 commit 4642b9e
Show file tree
Hide file tree
Showing 4 changed files with 13 additions and 22 deletions.
7 changes: 5 additions & 2 deletions .github/workflows/publish-core-images.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -25,5 +25,8 @@ jobs:
- component-name: kubectl-delivery
dockerfile: build/images/kubectl-delivery/Dockerfile
- component-name: storage-initializer
dockerfile: sdk/python/kubeflow/storage_initializer/Dockerfile
context: sdk/python/kubeflow/storage_initializer
dockerfile: sdk/python/kubeflow/storage_initializer/Dockerfile
context: sdk/python/kubeflow/storage_initializer
- component-name: trainer-huggingface
dockerfile: sdk/python/kubeflow/trainer/hf_dockerfile
context: sdk/python/kubeflow/trainer
3 changes: 0 additions & 3 deletions .github/workflows/publish-example-images.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -52,9 +52,6 @@ jobs:
- component-name: mxnet-auto-tuning
dockerfile: examples/mxnet/tune/Dockerfile
context: examples/mxnet/tune
- component-name: train-api-hf-image
dockerfile: sdk/python/kubeflow/trainer/hf_dockerfile
context: sdk/python/kubeflow/trainer
# TODO (tenzen-y): Fix the below broken Dockerfiles
# - component-name: pytorch-dist-mnist-mpi
# dockerfile: examples/pytorch/mnist/Dockerfile-mpi
Expand Down
18 changes: 6 additions & 12 deletions examples/sdk/train_api.py
Original file line number Diff line number Diff line change
@@ -1,14 +1,12 @@
from kubeflow.training.api.training_client import TrainingClient
from kubeflow.storage_init_container.hugging_face import (
from kubeflow.storage_initializer.hugging_face import (
HuggingFaceModelParams,
HuggingFaceTrainParams,
HfDatasetParams,
TRANSFORMER_TYPES,
)
from kubeflow.storage_init_container.s3 import S3DatasetParams
from peft import LoraConfig
import transformers
from transformers import TrainingArguments
import json

client = TrainingClient(
config_file="/Users/deepanker/Downloads/deepanker-test-kubectl.cfg"
Expand All @@ -25,13 +23,9 @@
},
model_provider_parameters=HuggingFaceModelParams(
model_uri="hf://Jedalc/codeparrot-gp2-finetune",
transformer_type=TRANSFORMER_TYPES.AutoModelForCausalLM,
transformer_type=transformers.AutoModelForCausalLM,
),
dataset_provider_parameters=HfDatasetParams(
repo_id="imdatta0/ultrachat_10k",
access_token="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",
),
# dataset_provider_parameters=S3DatasetParams(endpoint_url="http://10.117.63.3", bucket_name="deepanker-test", file_key="list_roles_response.txt", access_key="qEMHyz8wNwLpUWkvfZmQZrj60TE6zX4p", secret_key="qIp_QNLPKI0LJ5X0F8NrypoSMSsw_Gfe" ),
dataset_provider_parameters=HfDatasetParams(repo_id="imdatta0/ultrachat_10k"),
train_parameters=HuggingFaceTrainParams(
lora_config=LoraConfig(
r=8,
Expand Down Expand Up @@ -66,8 +60,8 @@
# eval_accumulation_steps=1,
per_device_eval_batch_size=1,
# load_best_model_at_end=True,
report_to="wandb",
run_name=f"{1}",
# report_to="wandb",
# run_name=f"{1}",
),
),
resources_per_worker={"gpu": 0, "cpu": 8, "memory": "8Gi"},
Expand Down
7 changes: 2 additions & 5 deletions sdk/python/kubeflow/trainer/hf_llm_training.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,8 +30,6 @@ def setup_model_and_tokenizer(model_uri, transformer_type, model_dir):
device_map="auto",
)

# print(model)

tokenizer = transformers.AutoTokenizer.from_pretrained(
pretrained_model_name_or_path=model_name,
cache_dir=model_dir,
Expand All @@ -42,8 +40,6 @@ def setup_model_and_tokenizer(model_uri, transformer_type, model_dir):
tokenizer.pad_token = tokenizer.eos_token
tokenizer.add_pad_token = True

# print(tokenizer)

# Freeze model parameters
for param in model.parameters():
param.requires_grad = False
Expand All @@ -56,7 +52,7 @@ def load_and_preprocess_data(dataset_name, dataset_dir):
print("loading dataset")
dataset = load_dataset(dataset_name, cache_dir=dataset_dir)
train_data = dataset["train"]
# print(train_data)

try:
eval_data = dataset["eval"]
except Exception as err:
Expand Down Expand Up @@ -89,6 +85,7 @@ def train_model(model, train_data, eval_data, tokenizer, train_params):
)

trainer.train()
print("training done")


def parse_arguments():
Expand Down

0 comments on commit 4642b9e

Please sign in to comment.