Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
64 commits
Select commit Hold shift + click to select a range
12b9985
update transformers version number
JulianneKnott Oct 28, 2021
e2b7a1d
update transformer version number (4.12.0)
JulianneKnott Oct 28, 2021
2566867
bart and gpt2 updates (no el attn, old cli)
JulianneKnott Nov 3, 2021
434185c
t5 updates (no el attn, old cli)
JulianneKnott Nov 5, 2021
2a8f779
debug
JulianneKnott Nov 8, 2021
67bf684
updated (existing fastseq optimization)
JulianneKnott Nov 12, 2021
6f8e451
bart, gpt2, t5 benchmarks updated
JulianneKnott Nov 12, 2021
c3fc882
add prophetnet
JulianneKnott Nov 18, 2021
c6b98b6
add distilbart benchmark
JulianneKnott Nov 22, 2021
f00c752
import prophetnet optimizer
JulianneKnott Nov 22, 2021
9c32b81
simplify prophetnet test
JulianneKnott Nov 22, 2021
940a78d
update benchmarks checks
JulianneKnott Nov 22, 2021
39b33ba
prophetnet benchmarks
JulianneKnott Nov 22, 2021
f4ce9c2
cleanup
JulianneKnott Nov 22, 2021
2ac5317
cleanup
JulianneKnott Nov 22, 2021
08cc31a
cleanup
JulianneKnott Nov 23, 2021
bd58d06
azure pipeline updates
JulianneKnott Nov 23, 2021
d9466e4
sentencepiece update (docker + azure pipeline)
JulianneKnott Nov 23, 2021
cc4494c
pipelines update torch=1.10.0
JulianneKnott Nov 23, 2021
c3c98cb
fix prophetnet fs expected output + cleanup
JulianneKnott Nov 23, 2021
e4354c3
benchmarks, examples, + readmes
JulianneKnott Nov 29, 2021
9fb69a7
fixed t5 ngram repeat block
JulianneKnott Nov 29, 2021
298ce3a
fix gpt2, t5, prophetnet unit hf tests
JulianneKnott Dec 2, 2021
b6e899e
fixed tests broken by fastseq update
JulianneKnott Dec 7, 2021
6faf3fb
unit test cleanup (w/ hf docker)
JulianneKnott Dec 8, 2021
e549380
update pipeline image
JulianneKnott Dec 8, 2021
e1e76d5
update pipeline
JulianneKnott Dec 9, 2021
8fb5575
pipeline debug
JulianneKnott Dec 9, 2021
b5172e7
pipeline debug
JulianneKnott Dec 9, 2021
ad12d1b
pipeline debug
JulianneKnott Dec 9, 2021
a8c224a
pipeline run transformers tests + add python
JulianneKnott Dec 9, 2021
9fe9c04
debug pipeline
JulianneKnott Dec 9, 2021
22b94c6
debug pipeline
JulianneKnott Dec 9, 2021
424bc4e
debug pipeline
JulianneKnott Dec 9, 2021
b20d9c0
debug pipeline
JulianneKnott Dec 9, 2021
36b018b
pipelines debug
JulianneKnott Dec 9, 2021
f5946c6
pipeline debug
JulianneKnott Dec 9, 2021
b118184
pipeline debug
JulianneKnott Dec 9, 2021
4e41e75
debug pipeline
JulianneKnott Dec 9, 2021
f9e8c16
debug
JulianneKnott Dec 9, 2021
c464dee
debug
JulianneKnott Dec 9, 2021
addd2a9
debug
JulianneKnott Dec 9, 2021
c61bedb
debug
JulianneKnott Dec 9, 2021
645887b
pipeline debug
JulianneKnott Dec 9, 2021
cc34fc7
debug pipeline
JulianneKnott Dec 9, 2021
8490d64
debug pipeline
JulianneKnott Dec 9, 2021
1e43d3f
debug pipeline
JulianneKnott Dec 9, 2021
816cab9
fastseq tests working, run fairseq tests
JulianneKnott Dec 9, 2021
124f768
debug venv
JulianneKnott Dec 9, 2021
7aab0d5
debug
JulianneKnott Dec 9, 2021
f28672d
debug
JulianneKnott Dec 10, 2021
7d83023
fairseq tests working, run with fastseq tests
JulianneKnott Dec 10, 2021
1633ef3
try running all tests
JulianneKnott Dec 10, 2021
49beb3d
typo fix
JulianneKnott Dec 10, 2021
980efa6
fixed last 4 failing tests
JulianneKnott Dec 10, 2021
0ad879e
cleanup
JulianneKnott Dec 10, 2021
943f198
compatibility up to 4.12.5
JulianneKnott Dec 13, 2021
5233d86
compatibility
JulianneKnott Dec 13, 2021
31a49ff
cleanup
JulianneKnott Dec 13, 2021
8ad78b3
fix bart benchmarks
JulianneKnott Dec 13, 2021
d501760
cleanup
JulianneKnott Dec 13, 2021
61fb12a
cleanup
JulianneKnott Dec 13, 2021
c66ec06
prophetnet benchmarks
JulianneKnott Dec 14, 2021
3513f0f
typo fixed
JulianneKnott Dec 14, 2021
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
17 changes: 9 additions & 8 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,17 +16,18 @@ Below shows the generation speed gain by using FastSeq.

| Model | W/O FastSeq (in samples/s) | W/ FastSeq (in samples/s) | Speedup |
|------------------|:--------------------------:|:-------------------------:|:-----:|
| [ProphetNet](examples/prophetnet/README.md) | 2.8 | 11.9 | 4.3 |
| [ProphetNet (`fs`)](examples/prophetnet/README.md) | 2.8 | 11.9 | 4.3 |
| [Bart (`fs`)](examples/bart/README.md) | 3.3 | 25.1 | 7.7x |
| [Bart (`hf`)](examples/bart/README.md#speedup-bart-huggingface-transformers-version-by-using-fastseq) | 2.5 | 12.4 | 5.0x |
| [DistilBart (`hf`)](examples/distilbart/README.md) | 3.4 | 18.5 | 5.4x |
| [T5 (`hf`)](examples/t5/README.md) | 8.7 | 31.3 | 3.6x |
| [Bart (`hf`)](examples/bart/README.md#speedup-bart-huggingface-transformers-version-by-using-fastseq) | 4.5 | 12.4 | 2.8x |
| [DistilBart (`hf`)](examples/distilbart/README.md) | 5.5 | 19.1 | 3.5x |
| [T5 (`hf`)](examples/t5/README.md) | 9.5 | 31.7 | 3.3x |
| [WMT16 En-De (`fs`)](examples/wmt/README.md) | 144.5 | 422.8 | 2.9x |
| [GPT2 (`hf`)](examples/gpt2/README.md) | 3.0 | 16.7 | 5.5x |
| [UniLM (`hf`)](examples/unilm/README.md) | 1.7 | 16.4 | 9.6x |
| [GPT2 (`hf`)](examples/gpt2/README.md) | 3.9 | 21.8 | 5.6x |
| [ProphetNet (`hf`)](examples/prophetnet/README.md) | 3.4 | 6.2 | 1.8x |

- All benchmarking experiments run on NVIDIA-V100-16GB with [docker](docker/Dockerfile). Highest speed recorded for each model by tuning batch size. For parameter setting details, click link of corresponding model.
- `fs` stands for [Fairseq](https://github.com/pytorch/fairseq) 0.10.2 version, `hf` stands for [Huggingface Transformers](https://github.com/huggingface/transformers) 3.0.2 version.
- The baseline (W/O Fastseq) for [ProphetNet (`fs`)](examples/prophetnet/README.md) is run with fairseq 0.9.0, as it has not yet been updated for compatibility with version 0.10.2
- `fs` stands for [Fairseq](https://github.com/pytorch/fairseq) 0.10.2 version, `hf` stands for [Huggingface Transformers](https://github.com/huggingface/transformers) 4.12.0 version.
- Optimizations were automatically applied to all generation/sequence models in Fairseq & Huggingface Transformers. Above only lists a subset of them.

## How it works?
Expand All @@ -39,7 +40,7 @@ FastSeq develops multiple speedup techniques, including an attention cache optim
- Python version >= 3.6
- [torch](http://pytorch.org/) >= 1.4.0
- [fairseq](https://github.com/pytorch/fairseq) >= 0.10.0
- [transformers](https://github.com/huggingface/transformers) == 3.0.2
- [transformers](https://github.com/huggingface/transformers) >= 4.12.0
- [requests](https://pypi.org/project/requests/) >= 2.24.0
- [absl-py](https://pypi.org/project/absl-py/) >= 0.9.0
- [rouge-score](https://pypi.org/project/rouge-score/) >= 0.0.4
Expand Down
23 changes: 12 additions & 11 deletions azure-pipelines.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,21 +15,21 @@ jobs:
demands:
- agent.name -equals gpu3
container:
image: adsbrainwestus2.azurecr.io/fastseq:dev-py3
endpoint: fastseq-acr
image: huggingface/transformers-pytorch-gpu:latest
Comment thread
feihugis marked this conversation as resolved.
options: --gpus device=3
steps:
- script: |
#install fastseq
which pip
which python
pip install --upgrade pip
pip install sentencepiece==0.1.96
pip install torch==1.10.0

echo "******* Installing fairseq *******"
pip install fairseq==0.10.2
pip show fairseq

echo "******* Installing transformers *******"
pip install transformers
pip install transformers==4.12.0
pip show transformers

echo "******* Installing fastseq *******"
Expand All @@ -39,10 +39,6 @@ jobs:
echo "******* Adding local bin to path *******"
export PATH="$HOME/bin:$HOME/.local/bin:$PATH"

echo "******* Running fastseq unittests *******"
pip install pytorch-transformers==1.0.0
bash tests/run_fastseq_tests.sh

#cd benchmarks/
#bash run_all_benchmarks.sh

Expand All @@ -53,11 +49,16 @@ jobs:
python -c "import torch; print('torch:', torch.__version__, torch)"
python -c "import torch; print('CUDA available:', torch.cuda.is_available())"

echo "******* Running transformers unittests *******"
bash tests/run_transformers_tests.sh

echo "******* Running fairseq unittests *******"
pip install apex==0.9.10.dev0
bash tests/run_fairseq_tests.sh

echo "******* Running transformers unittests *******"
bash tests/run_transformers_tests.sh
echo "******* Running fastseq unittests *******"
pip install pytorch-transformers==1.0.0
bash tests/run_fastseq_tests.sh

displayName: 'run fastseq unit tests'
- task: PublishTestResults@2
Expand Down
7 changes: 3 additions & 4 deletions benchmarks/hf.sh
Original file line number Diff line number Diff line change
@@ -1,10 +1,9 @@
#!/bin/bash
source utils.sh
if [[ $SKIP_BASELINE -eq 0 ]]; then
export BASELINE_REPO=$CACHE_DIR/transformers_v3.0.2
#https://github.com/huggingface/transformers.git \
export BASELINE_REPO=$CACHE_DIR/transformers_v4.12.0
git_clone_if_not_in_cache \
https://github.com/JiushengChen/transformers.git \
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Some context about this forked repo, it is to add param "no_repeat_ngram_size", see
JiushengChen/Transformers@db97043

Copy link
Copy Markdown
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

To clarify, this is no longer needed since fastseq uses it's own run_eval_hf.py for the baseline?

https://github.com/huggingface/transformers.git \
$BASELINE_REPO \
v3.0.2-ngram
v4.12.0
fi
14 changes: 7 additions & 7 deletions benchmarks/models/hf_bart.sh
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ source hf.sh
facebook/bart-large-cnn \
cnn_dm/raw \
val \
32 \
32/64 \
--task summarization \
--no_repeat_ngram_size 3
./benchmark.sh \
Expand All @@ -33,16 +33,16 @@ grep "facebook/bart-large-cnn cnn_dm/raw val " perf \
| awk -F'|' '{if($1!="NA"){c+=1;s+=$1}}END{print s/c}' \
| ./range.sh 0.447 0.448
# Speed on V100 16GB 250W
grep -E "transformers_v3.0.2 facebook/bart-large-cnn cnn_dm/raw val 32 " perf \
grep -E "transformers_v4.12.0 facebook/bart-large-cnn cnn_dm/raw val 64 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 2 3
Comment thread
JulianneKnott marked this conversation as resolved.
grep -E "transformers_v3.0.2\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 32 " perf \
| ./range.sh 4 5
grep -E "transformers_v4.12.0\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 32 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 7 100
grep -E "transformers_v3.0.2\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 64 " perf \
| ./range.sh 10 100
grep -E "transformers_v4.12.0\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 64 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 11 100
grep -E "transformers_v3.0.2\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 128 " perf \
grep -E "transformers_v4.12.0\+fastseq_v.* facebook/bart-large-cnn cnn_dm/raw val 128 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 12 100

12 changes: 6 additions & 6 deletions benchmarks/models/hf_distibart.sh
Original file line number Diff line number Diff line change
Expand Up @@ -32,12 +32,12 @@ grep "hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val " perf \
| awk -F'|' '{if($1!="NA"){c+=1;s+=$1}}END{print s/c}' \
| ./range.sh 0.45 0.452
# Speed on V100 16GB 250W
grep -E "transformers_v3.0.2 hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 64 " perf \
grep -E "transformers_v4.12.0 hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 128 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 3 4
grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 64 " perf \
| ./range.sh 5 6
grep -E "transformers_v4.12.0\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 64 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 16.5 100
grep -E "transformers_v3.0.2\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 128 " perf \
| ./range.sh 17 100
grep -E "transformers_v4.12.0\+fastseq_v.* hf.sshleifer.distilbart-cnn-12-6.tar.gz cnn_dm/raw val 128 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 18.3 100
| ./range.sh 18 100
23 changes: 13 additions & 10 deletions benchmarks/models/hf_gpt2.sh
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,6 @@
# <split> # train/val/test (text) or train/valid/test (binary)
# <batch-sizes>
source hf.sh

# MODEL - bart large cnn from transformer
# TASK - cnn dm val full set

Expand All @@ -16,7 +15,7 @@ source hf.sh
gpt2 \
cnn_dm/raw \
val \
64/128 \
64/128/256 \
--task summarization \
--no_repeat_ngram_size 3 \
--max_tokenizer_length 512 \
Expand All @@ -27,7 +26,7 @@ source hf.sh
gpt2 \
cnn_dm/raw \
val \
64 \
64/128 \
--task summarization \
--no_repeat_ngram_size 3 \
--max_tokenizer_length 512 \
Expand All @@ -37,14 +36,18 @@ source hf.sh
grep "gpt2 cnn_dm/raw val " perf \
| awk '{print $9}' \
| awk -F'|' '{if($1!="NA"){c+=1;s+=$1}}END{print s/c}' \
| ./range.sh 0.155 0.156
| ./range.sh 0.160 0.162
# Speed on V100 16GB 250W
grep -E "transformers_v3.0.2 gpt2 cnn_dm/raw val 64 " perf \
grep -E "transformers_v4.12.0 gpt2 cnn_dm/raw val 64 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 2.9 3.2
grep -E "transformers_v3.0.2\+fastseq_v.* gpt2 cnn_dm/raw val 64 " perf \
| ./range.sh 3.5 4.5
grep -E "transformers_v4.12.0\+fastseq_v.* gpt2 cnn_dm/raw val 64 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 16 100
grep -E "transformers_v4.12.0\+fastseq_v.* gpt2 cnn_dm/raw val 128 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 10.8 11.3
grep -E "transformers_v3.0.2\+fastseq_v.* gpt2 cnn_dm/raw val 128 " perf \
| ./range.sh 20 100
grep -E "transformers_v4.12.0\+fastseq_v.* gpt2 cnn_dm/raw val 256 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 16.4 16.8
| ./range.sh 21 100

6 changes: 3 additions & 3 deletions benchmarks/models/hf_mbart.sh
Original file line number Diff line number Diff line change
Expand Up @@ -26,11 +26,11 @@ source hf.sh
# Accuracy
grep "facebook/mbart-large-en-ro wmt_en_ro/raw val " perf \
| awk '{if($8!="NA"){c+=1;s+=$8}}END{print s/c}' \
| ./range.sh 56.1 56.3
| ./range.sh 56.1 56.4
# Speed on V100 16GB 250W
grep -E "transformers_v3.0.2 facebook/mbart-large-en-ro wmt_en_ro/raw val 64 " perf \
grep -E "transformers_v4.12.0 facebook/mbart-large-en-ro wmt_en_ro/raw val 64 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 6.0 100
grep -E "transformers_v3.0.2\+fastseq_v.* facebook/mbart-large-en-ro wmt_en_ro/raw val 64 " perf \
grep -E "transformers_v4.12.0\+fastseq_v.* facebook/mbart-large-en-ro wmt_en_ro/raw val 64 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 9 100
41 changes: 41 additions & 0 deletions benchmarks/models/hf_prophetnet.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,41 @@
#!/bin/bash
# Run it at its parent folder, and check result at ../perf.
# USAGE - ./benchmark.sh
# [fairseq|fairseq+fastseq|transformers|transformers+fastseq]
# <model>
# <task>
# <split> # train/val/test (text) or train/valid/test (binary)
# <batch-sizes>
source hf.sh

# MODEL - prophetnet from transformer
# TASK - cnn dm val full set
./benchmark.sh \
transformers \
microsoft/prophetnet-large-uncased \
cnn_dm_bert/raw \
val \
128 \
--task summarization \
--no_repeat_ngram_size 3
./benchmark.sh \
transformers+fastseq \
microsoft/prophetnet-large-uncased \
cnn_dm_bert/raw \
val \
128 \
--task summarization \
--no_repeat_ngram_size 3

# Accuracy
grep "microsoft/prophetnet-large-uncased cnn_dm_bert/raw val " perf \
| awk '{print $9}' \
| awk -F'|' '{if($1!="NA"){c+=1;s+=$1}}END{print s/c}' \
| ./range.sh 0.230 0.232
# Speed on V100 16GB 250W
grep -E "transformers_v4.12.0 microsoft/prophetnet-large-uncased cnn_dm_bert/raw val 128 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 3 4
grep -E "transformers_v4.12.0+fastseq_v.* microsoft/prophetnet-large-uncased cnn_dm_bert/raw val 128 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 6 100
24 changes: 13 additions & 11 deletions benchmarks/models/hf_t5.sh
Original file line number Diff line number Diff line change
Expand Up @@ -14,28 +14,30 @@ source hf.sh
wmt_en_ro/raw \
val \
64 \
--task translation_en_to_ro
# --no_repeat_ngram_size 3 # baseline don't support this arg now.
Comment thread
JulianneKnott marked this conversation as resolved.
--task translation_en_to_ro \
--no_repeat_ngram_size 3

./benchmark.sh \
transformers+fastseq \
t5-base \
wmt_en_ro/raw \
val \
64/128 \
--task translation_en_to_ro \
--postprocess_workers 3
# --no_repeat_ngram_size 3
# Accuracy
--postprocess_workers 3 \
--no_repeat_ngram_size 3

# # Accuracy
grep "t5-base wmt_en_ro/raw val " perf \
| awk '{if($8!="NA"){c+=1;s+=$8}}END{print s/c}' \
| ./range.sh 57.8 57.9
| ./range.sh 58.0 59.0
# Speed on V100 16GB 250W
grep -E "transformers_v3.0.2 t5-base wmt_en_ro/raw val 64 " perf \
grep -E "transformers_v4.12.0 t5-base wmt_en_ro/raw val 64 " perf \
| awk '{s+=$13}END{if(NR==0) print -1; else print s/NR}' \
| ./range.sh 8 10
grep -E "transformers_v3.0.2\+fastseq_v.* t5-base wmt_en_ro/raw val 64 " perf \
| ./range.sh 12 17
grep -E "transformers_v4.12.0\+fastseq_v.* t5-base wmt_en_ro/raw val 64 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 19 100
grep -E "transformers_v3.0.2\+fastseq_v.* t5-base wmt_en_ro/raw val 128 " perf \
| ./range.sh 23 100
grep -E "transformers_v4.12.0\+fastseq_v.* t5-base wmt_en_ro/raw val 128 " perf \
| awk '{s+=$13}END{print s/NR}' \
| ./range.sh 30 100
39 changes: 0 additions & 39 deletions benchmarks/models/hf_unilm.sh

This file was deleted.

2 changes: 1 addition & 1 deletion benchmarks/run_eval_hf.py
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ def calculate_bleu_score(output_lns, refs_lns, **kwargs) -> dict:
return {"bleu": corpus_bleu(output_lns, [refs_lns], **kwargs).score}

def trim_batch(
input_ids, pad_token_id, attention_mask=None,
input_ids, pad_token_id, attention_mask=None, **kwargs
):
"""Remove columns that are populated exclusively by pad_token_id"""
keep_column_mask = input_ids.ne(pad_token_id).any(dim=0)
Expand Down
19 changes: 10 additions & 9 deletions docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -43,15 +43,16 @@ RUN cd ${STAGE_DIR} && \
# Installation/NLP Utilities
##############################################################################
RUN pip install --upgrade pip && \
pip install yapf>=v0.30.0 && \
pip install absl-py>=v0.9.0 && \
pip install filelock>=v3.0.12 && \
pip install requests>=v2.24.0 && \
pip install gitpython>=v3.1.7 && \
pip install rouge_score==v0.0.4 && \
pip install fairseq==v0.10.2 && \
pip install transformers==v3.0.2 && \
pip install pytorch-transformers==1.0.0
pip install yapf>=v0.30.0 && \
pip install absl-py>=v0.9.0 && \
pip install filelock>=v3.0.12 && \
pip install requests>=v2.24.0 && \
pip install gitpython>=v3.1.7 && \
pip install rouge_score==v0.0.4 && \
pip install fairseq==v0.10.2 && \
pip install transformers==v4.12.0 && \
pip install pytorch-transformers==1.0.0 && \
pip install sentencepiece==0.1.90
Comment thread
JulianneKnott marked this conversation as resolved.


##############################################################################
Expand Down
Loading