-
Notifications
You must be signed in to change notification settings - Fork 88
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Update to latest version of transformers #31
Changes from 17 commits
c120338
129c441
227fd72
9b21861
3294ae2
48048f6
f67972f
4fd17c2
1465eed
c518857
242180a
3117e15
8bba5e5
8d8202a
00897e9
265478e
2c61362
e6d09ae
54640f0
a63a1bf
6685f4d
517fa0d
507e756
52c59ce
b84351c
e201ed9
8faabd5
d839824
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,24 +1,25 @@ | ||
{ | ||
"output_dir": "./logs_petfinder/", | ||
"output_dir": "./logs_petfinder/gating_on_cat_and_num_feats_then_sum_full_model", | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. hm, why There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. The existing config utilized this. The name is just a reference to the method used to combine the different embeddings obtained from the model. I just standardized the format in which the output dir was named across the three configs. |
||
"debug_dataset": false, | ||
"task": "classification", | ||
"num_labels": 5, | ||
"combine_feat_method": "text_only", | ||
"experiment_name": "bert-base-multilingual-uncased", | ||
"model_name_or_path": "bert-base-multilingual-uncased", | ||
"do_train": true, | ||
"categorical_encode_type": "ohe", | ||
"numerical_transformer_method": "quantile_normal", | ||
"tokenizer_name": "bert-base-multilingual-uncased", | ||
"per_device_train_batch_size": 12, | ||
"gpu_num": 0, | ||
"use_simple_classifier": false, | ||
"logging_dir": "./logs_petfinder/bertmultilingual_gating_on_cat_and_num_feats_then_sum_full_model_lr_3e-3/", | ||
"num_train_epochs": 5, | ||
"categorical_encode_type": "ohe", | ||
"use_class_weights": false, | ||
"overwrite_output_dir": true, | ||
"learning_rate": 1e-4, | ||
"per_device_train_batch_size": 16, | ||
"per_device_eval_batch_size": 16, | ||
"logging_steps": 50, | ||
"eval_steps": 750, | ||
"save_steps": 3000, | ||
"learning_rate": 1e-4, | ||
"data_path": "./datasets/PetFindermy_Adoption_Prediction/", | ||
"column_info_path": "./datasets/PetFindermy_Adoption_Prediction/column_info_all_text.json", | ||
"overwrite_output_dir": true | ||
"data_path": "./datasets/PetFindermy_Adoption_Prediction", | ||
"column_info_path": "./datasets/PetFindermy_Adoption_Prediction/column_info_all_text.json" | ||
} | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -5,15 +5,21 @@ | |
"combine_feat_method": "text_only", | ||
"experiment_name": "Unimodal Bert Base Uncased", | ||
"model_name_or_path": "bert-base-uncased", | ||
"gpu_num": 0, | ||
"do_train": true, | ||
"categorical_encode_type": "binary", | ||
"numerical_transformer_method": "quantile_normal", | ||
"tokenizer_name": "bert-base-uncased", | ||
"per_device_train_batch_size": 12, | ||
"use_simple_classifier": false, | ||
"logging_dir": "./logs_clothing_review/bertbase_gating_on_cat_and_num_feats_then_sum_full_model_lr_3e-3/", | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. is it possible to avoid hardcoded paths like this? There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. What would you suggest instead? These seem to be primarily intended to serve as examples for users more than anything else. |
||
"num_train_epochs": 5, | ||
"overwrite_output_dir": true, | ||
"learning_rate": 3e-3, | ||
"per_device_train_batch_size": 16, | ||
"per_device_eval_batch_size": 16, | ||
"logging_steps": 50, | ||
"eval_steps": 750, | ||
"save_steps": 3000, | ||
"data_path": "./datasets/Womens_Clothing_E-Commerce_Reviews", | ||
"column_info_path": "./datasets/Womens_Clothing_E-Commerce_Reviews/column_info_all_text.json", | ||
"overwrite_output_dir": true | ||
"column_info_path": "./datasets/Womens_Clothing_E-Commerce_Reviews/column_info_all_text.json" | ||
} | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,6 +1,6 @@ | ||
import multimodal_transformers.data | ||
import multimodal_transformers.model | ||
|
||
__version__ = '0.1.2-alpha' | ||
__version__ = '0.2-alpha' | ||
akashsaravanan-georgian marked this conversation as resolved.
Show resolved
Hide resolved
|
||
|
||
__all__ = ['multimodal_transformers', '__version__'] |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
why do we need this change?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This was changed during a pass to standardize the training config formats across the different datasets. I modified the batch size from 12 -> 16 just to conform to the traditional approach of having the batch size be a power of 2.