mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-07 23:00:08 +06:00

* Add new run_swag example * Add check * Add sample * Apply suggestions from code review Co-authored-by: Lysandre Debut <lysandre@huggingface.co> * Very important change to make Lysandre happy Co-authored-by: Lysandre Debut <lysandre@huggingface.co>
66 lines
1.8 KiB
Markdown
66 lines
1.8 KiB
Markdown
<!---
|
|
Copyright 2020 The HuggingFace Team. All rights reserved.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
-->
|
|
|
|
## Multiple Choice
|
|
|
|
Based on the script [`run_swag.py`]().
|
|
|
|
#### Fine-tuning on SWAG
|
|
|
|
```bash
|
|
python examples/multiple-choice/run_swag.py \
|
|
--model_name_or_path roberta-base \
|
|
--do_train \
|
|
--do_eval \
|
|
--learning_rate 5e-5 \
|
|
--num_train_epochs 3 \
|
|
--output_dir /tmp/swag_base \
|
|
--per_gpu_eval_batch_size=16 \
|
|
--per_device_train_batch_size=16 \
|
|
--overwrite_output
|
|
```
|
|
Training with the defined hyper-parameters yields the following results:
|
|
```
|
|
***** Eval results *****
|
|
eval_acc = 0.8338998300509847
|
|
eval_loss = 0.44457291918821606
|
|
```
|
|
|
|
|
|
## Tensorflow
|
|
|
|
```bash
|
|
export SWAG_DIR=/path/to/swag_data_dir
|
|
python ./examples/multiple-choice/run_tf_multiple_choice.py \
|
|
--task_name swag \
|
|
--model_name_or_path bert-base-cased \
|
|
--do_train \
|
|
--do_eval \
|
|
--data_dir $SWAG_DIR \
|
|
--learning_rate 5e-5 \
|
|
--num_train_epochs 3 \
|
|
--max_seq_length 80 \
|
|
--output_dir models_bert/swag_base \
|
|
--per_gpu_eval_batch_size=16 \
|
|
--per_device_train_batch_size=16 \
|
|
--logging-dir logs \
|
|
--gradient_accumulation_steps 2 \
|
|
--overwrite_output
|
|
```
|
|
|
|
# Run it in colab
|
|
[](https://colab.research.google.com/github/ViktorAlm/notebooks/blob/master/MPC_GPU_Demo_for_TF_and_PT.ipynb)
|