mirror of
https://github.com/huggingface/transformers.git
synced 2025-07-15 02:28:24 +06:00

* Reorganize example folder * Continue reorganization * Change requirements for tests * Final cleanup * Finish regroup with tests all passing * Copyright * Requirements and readme * Make a full link for the documentation * Address review comments * Apply suggestions from code review Co-authored-by: Lysandre Debut <lysandre@huggingface.co> * Add symlink * Reorg again * Apply suggestions from code review Co-authored-by: Thomas Wolf <thomwolf@users.noreply.github.com> * Adapt title * Update to new strucutre * Remove test * Update READMEs Co-authored-by: Lysandre Debut <lysandre@huggingface.co> Co-authored-by: Thomas Wolf <thomwolf@users.noreply.github.com>
24 lines
662 B
Markdown
24 lines
662 B
Markdown
## MM-IMDb
|
|
|
|
Based on the script [`run_mmimdb.py`](https://github.com/huggingface/transformers/blob/master/examples/contrib/mm-imdb/run_mmimdb.py).
|
|
|
|
[MM-IMDb](http://lisi1.unal.edu.co/mmimdb/) is a Multimodal dataset with around 26,000 movies including images, plots and other metadata.
|
|
|
|
### Training on MM-IMDb
|
|
|
|
```
|
|
python run_mmimdb.py \
|
|
--data_dir /path/to/mmimdb/dataset/ \
|
|
--model_type bert \
|
|
--model_name_or_path bert-base-uncased \
|
|
--output_dir /path/to/save/dir/ \
|
|
--do_train \
|
|
--do_eval \
|
|
--max_seq_len 512 \
|
|
--gradient_accumulation_steps 20 \
|
|
--num_image_embeds 3 \
|
|
--num_train_epochs 100 \
|
|
--patience 5
|
|
```
|
|
|