From 744404fb042124939c6b33a576e5e2c5a5725467 Mon Sep 17 00:00:00 2001 From: Benjamin Bossan Date: Mon, 9 Dec 2024 16:28:48 +0100 Subject: [PATCH 1/5] Empty commit to trigger CI From b146a6d1ec0742fe8b22805b4ac5a5da2cbe0d74 Mon Sep 17 00:00:00 2001 From: Benjamin Bossan Date: Mon, 9 Dec 2024 16:43:27 +0100 Subject: [PATCH 2/5] Run some transformers tests too --- .github/workflows/inference-test.yml | 25 +++++++++++++++++++++++++ 1 file changed, 25 insertions(+) diff --git a/.github/workflows/inference-test.yml b/.github/workflows/inference-test.yml index a4b210f..bc4813a 100644 --- a/.github/workflows/inference-test.yml +++ b/.github/workflows/inference-test.yml @@ -48,3 +48,28 @@ jobs: uses: actions/setup-python@v3 with: python-version: ${{ matrix.python }} + + tests: + strategy: + fail-fast: false + matrix: + python-version: ["3.9", "3.10", "3.11", "3.12"] + os: ["ubuntu-latest", "macos-13", "windows-latest"] + runs-on: ${{ matrix.os }} + steps: + - uses: actions/checkout@v3 + - name: Set up Python ${{ matrix.python-version }} + uses: actions/setup-python@v4 + with: + python-version: ${{ matrix.python-version }} + cache: "pip" + cache-dependency-path: "setup.py" + - name: Install dependencies + run: | + python -m pip install --upgrade pip + pip install torch --index-url https://download.pytorch.org/whl/cpu + pip install setuptools transformers pytest + pip install -e . + - name: Test with pytest + run: | + pytest test.py --durations=99 From 02733d9d351a458355f297175735074d6c2bf6de Mon Sep 17 00:00:00 2001 From: Benjamin Bossan Date: Mon, 9 Dec 2024 16:45:27 +0100 Subject: [PATCH 3/5] Add test.py file --- test.py | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) create mode 100644 test.py diff --git a/test.py b/test.py new file mode 100644 index 0000000..8872fba --- /dev/null +++ b/test.py @@ -0,0 +1,19 @@ +import pytest +import torch +from transformers import AutoModelForCausalLM + + +MODELS_TO_TEST = [ + "hf-internal-testing/tiny-random-OPTForCausalLM", + "hf-internal-testing/tiny-random-GPT2LMHeadModel", + "hf-internal-testing/tiny-random-gpt_neo", +] + + +@pytest.mark.parametrize("model_id", MODELS_TO_TEST) +@pytest.mark.parametrize("use_cache", [False, True]) +def test_inference(model_id, use_cache): + model = AutoModelForCausalLM.from_pretrained(model_id, use_cache=use_cache) + model.eval() + inputs = torch.tensor([[0, 1, 2, 3, 4]]) + model.generate(inputs, do_sample=True, num_return_sequences=1) From 9d159e4d7b6bd147c01b679854197ac630b4ddb8 Mon Sep 17 00:00:00 2001 From: Benjamin Bossan Date: Mon, 9 Dec 2024 17:08:17 +0100 Subject: [PATCH 4/5] Adjust test a little bit --- test.py | 10 +++++++--- 1 file changed, 7 insertions(+), 3 deletions(-) diff --git a/test.py b/test.py index 8872fba..ff5a842 100644 --- a/test.py +++ b/test.py @@ -11,9 +11,13 @@ @pytest.mark.parametrize("model_id", MODELS_TO_TEST) -@pytest.mark.parametrize("use_cache", [False, True]) -def test_inference(model_id, use_cache): - model = AutoModelForCausalLM.from_pretrained(model_id, use_cache=use_cache) +@pytest.mark.parametrize("useless", [False, True]) +def test_inference(model_id, useless): + if useless: + # do nothing + pass + + model = AutoModelForCausalLM.from_pretrained(model_id, useless): model.eval() inputs = torch.tensor([[0, 1, 2, 3, 4]]) model.generate(inputs, do_sample=True, num_return_sequences=1) From b47d20fe1c6d72baf2d14f76bc828e18005b570c Mon Sep 17 00:00:00 2001 From: Benjamin Bossan Date: Mon, 9 Dec 2024 17:16:07 +0100 Subject: [PATCH 5/5] fix test --- test.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/test.py b/test.py index ff5a842..5b009b7 100644 --- a/test.py +++ b/test.py @@ -17,7 +17,7 @@ def test_inference(model_id, useless): # do nothing pass - model = AutoModelForCausalLM.from_pretrained(model_id, useless): + model = AutoModelForCausalLM.from_pretrained(model_id) model.eval() inputs = torch.tensor([[0, 1, 2, 3, 4]]) model.generate(inputs, do_sample=True, num_return_sequences=1)