From d35a1399ee660af85ff0d3c819a60679f440c7b9 Mon Sep 17 00:00:00 2001 From: Abhishek Maurya <124327945+Abhishek-TAMU@users.noreply.github.com> Date: Thu, 1 Aug 2024 21:00:52 -0400 Subject: [PATCH] enabling tests for prompt tuning (#278) Signed-off-by: Abhishek Co-authored-by: Anh Uong --- tests/test_sft_trainer.py | 9 --------- 1 file changed, 9 deletions(-) diff --git a/tests/test_sft_trainer.py b/tests/test_sft_trainer.py index 9df7eeee..eb2ca855 100644 --- a/tests/test_sft_trainer.py +++ b/tests/test_sft_trainer.py @@ -164,9 +164,6 @@ def test_parse_arguments_peft_method(job_config): ############################# Prompt Tuning Tests ############################# -@pytest.mark.skip( - reason="currently inference doesn't work with transformer version 4.42.4" -) def test_run_causallm_pt_and_inference(): """Check if we can bootstrap and peft tune causallm models""" with tempfile.TemporaryDirectory() as tempdir: @@ -197,9 +194,6 @@ def test_run_causallm_pt_and_inference(): assert "### Text: @NortonSupport Thanks much.\n\n### Label:" in output_inference -@pytest.mark.skip( - reason="currently inference doesn't work with transformer version 4.42.4" -) def test_run_causallm_pt_and_inference_with_formatting_data(): """Check if we can bootstrap and peft tune causallm models This test needs the trainer to format data to a single sequence internally. @@ -238,9 +232,6 @@ def test_run_causallm_pt_and_inference_with_formatting_data(): assert "### Text: @NortonSupport Thanks much.\n\n### Label:" in output_inference -@pytest.mark.skip( - reason="currently inference doesn't work with transformer version 4.42.4" -) def test_run_causallm_pt_and_inference_JSON_file_formatter(): """Check if we can bootstrap and peft tune causallm models with JSON train file format""" with tempfile.TemporaryDirectory() as tempdir: