From 63882155faf057677c61eec540374f6dceaeb205 Mon Sep 17 00:00:00 2001 From: "dependabot[bot]" <49699333+dependabot[bot]@users.noreply.github.com> Date: Mon, 1 Jul 2024 23:13:46 +0000 Subject: [PATCH 1/2] Update datasets requirement from <2.20,>=2.19 to >=2.20.0,<2.21 Updates the requirements on [datasets](https://github.com/huggingface/datasets) to permit the latest version. - [Release notes](https://github.com/huggingface/datasets/releases) - [Commits](https://github.com/huggingface/datasets/compare/2.19.0...2.20.0) --- updated-dependencies: - dependency-name: datasets dependency-type: direct:production ... Signed-off-by: dependabot[bot] --- setup.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/setup.py b/setup.py index 2bb2912a0b..00ede72e7d 100644 --- a/setup.py +++ b/setup.py @@ -60,7 +60,7 @@ 'transformers>=4.42.3,<4.43', 'mosaicml-streaming>=0.7.6,<0.8', 'torch>=2.3.0,<2.4', - 'datasets>=2.19,<2.20', + 'datasets>=2.20.0,<2.21', 'fsspec==2023.6.0', # newer version results in a bug in datasets that duplicates data 'sentencepiece==0.1.97', 'einops==0.8.0', From 6aa393046dfaa74fad0c2f4a1485225e04bc75e2 Mon Sep 17 00:00:00 2001 From: Milo Cress Date: Wed, 6 Nov 2024 20:20:07 +0000 Subject: [PATCH 2/2] trust --- tests/eval/test_in_context_learning_datasets.py | 9 +++++++++ 1 file changed, 9 insertions(+) diff --git a/tests/eval/test_in_context_learning_datasets.py b/tests/eval/test_in_context_learning_datasets.py index 5fe9643cde..7509e95304 100644 --- a/tests/eval/test_in_context_learning_datasets.py +++ b/tests/eval/test_in_context_learning_datasets.py @@ -279,6 +279,7 @@ def test_update_generation_kwargs( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} gen_kwargs = {'test_arg1': 1, 'test_arg2': 2} @@ -370,6 +371,7 @@ def test_update_generation_kwargs_no_kwargs( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -459,6 +461,7 @@ def test_construct_context( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -513,6 +516,7 @@ def test_get_answer_from_example( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -552,6 +556,7 @@ def test_fix_eos_on_preamble(tmp_path: Path): hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -592,6 +597,7 @@ def test_tokenize_example_with_tokenize_labels( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -658,6 +664,7 @@ def test_tokenize_example_with_no_tokenize_labels( hf_loading_vars = { 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, } hf_parsing_map = {'context': ['quas', 'wex', 'exort'], 'answer': ['spell']} @@ -2429,6 +2436,7 @@ def test_lm_spacing_dataloader( [{ 'split': 'test', 'name': 'juggernaut', + 'trust_remote_code': True, }], ) @pytest.mark.parametrize( @@ -2508,6 +2516,7 @@ def test_hf_dataloading_lm_dataloader( [{ 'split': 'test', 'name': 'invoker', + 'trust_remote_code': True, }], ) @pytest.mark.parametrize(