Skip to content

[CI] skip unit case for hang #10642

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 3 commits into
base: develop
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
3 changes: 3 additions & 0 deletions tests/data/test_data_collator.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,8 @@
from paddlenlp.trainer import set_seed
from paddlenlp.transformers import BertTokenizer

from ..testing_utils import skip_for_none_ce_case


class DataCollatorIntegrationTest(unittest.TestCase):
def setUp(self):
Expand Down Expand Up @@ -251,6 +253,7 @@ def _test_no_pad_and_pad(self, no_pad_features, pad_features):
self.assertTrue(paddle.any(masked_tokens))
self.assertTrue(all(x == -100 for x in batch["labels"][~masked_tokens].tolist()))

@skip_for_none_ce_case
def test_data_collator_for_language_modeling(self):
no_pad_features = [{"input_ids": list(range(10))}, {"input_ids": list(range(10))}]
pad_features = [{"input_ids": list(range(5))}, {"input_ids": list(range(10))}]
Expand Down
3 changes: 2 additions & 1 deletion tests/transformers/jamba/test_modeling.py
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@
repeat_kv,
)

from ...testing_utils import slow
from ...testing_utils import skip_for_none_ce_case, slow

# from ..generation import GenerationTesterMixin
from ..test_configuration_common import ConfigTester
Expand Down Expand Up @@ -589,6 +589,7 @@ def test_FakeMLPForwardBackward(self):
and down_weight.grad.sum() == 0
)

@skip_for_none_ce_case
def test_from_hf_hub(self):
model_id = "ai21labs/Jamba-tiny-random"
model = JambaForCausalLM.from_pretrained(model_id, dtype="bfloat16", from_hf_hub=True, convert_from_torch=True)
Expand Down
3 changes: 3 additions & 0 deletions tests/utils/test_import_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,8 +16,11 @@

from paddlenlp.utils import install_package, uninstall_package

from ..testing_utils import skip_for_none_ce_case


class ImportUntilsTest(unittest.TestCase):
@skip_for_none_ce_case
def test_install_specific_package(self):
install_package("loguru", "0.6.0")
from loguru import __version__
Expand Down
Loading