summaryrefslogtreecommitdiff
path: root/sci-libs/datasets/files/datasets-2.14.4-tests.patch
blob: b9791c04e8e0500a108df72e2db0d453be8fe29e (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
--- a/tests/test_metric_common.py	2023-05-04 18:48:48.550861318 +0200
+++ b/tests/test_metric_common.py	2023-05-04 18:50:25.787364577 +0200
@@ -93,6 +93,7 @@
     INTENSIVE_CALLS_PATCHER = {}
     metric_name = None
 
+    @pytest.mark.skip(reason="disabling, depends on bert_score, bleurt, math_equivalence, coval, nltk, faiss, mauve, rouge_score, sacrebleu, sacremoses ...")
     @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning")
     @pytest.mark.filterwarnings("ignore:load_metric is deprecated:FutureWarning")
     def test_load_metric(self, metric_name):
--- a/tests/test_distributed.py	2023-05-04 19:43:09.861275030 +0200
+++ b/tests/test_distributed.py	2023-05-04 19:44:17.608326722 +0200
@@ -74,6 +74,7 @@
         split_dataset_by_node(full_ds.shuffle(), rank=0, world_size=world_size)
 
 
+@pytest.mark.skip(reason="require distributed torch")
 @pytest.mark.parametrize("streaming", [False, True])
 @require_torch
 @pytest.mark.skipif(os.name == "nt", reason="execute_subprocess_async doesn't support windows")
@@ -95,6 +96,7 @@
     execute_subprocess_async(cmd, env=os.environ.copy())
 
 
+@pytest.mark.skip(reason="require distributed torch")
 @pytest.mark.parametrize(
     "nproc_per_node, num_workers",
     [
--- a/tests/utils.py	2023-05-06 08:43:16.251987543 +0200
+++ b/tests/utils.py	2023-05-06 08:44:24.467952870 +0200
@@ -50,8 +50,8 @@
 # Audio
 require_sndfile = pytest.mark.skipif(
     # On Windows and OS X, soundfile installs sndfile
-    find_spec("soundfile") is None or version.parse(importlib.metadata.version("soundfile")) < version.parse("0.12.0"),
-    reason="test requires sndfile>=0.12.1: 'pip install \"soundfile>=0.12.1\"'; ",
+    True,
+    reason="test requires librosa",
 )
 
 # Beam
--- a/tests/features/test_audio.py	2023-05-06 09:03:58.680108142 +0200
+++ a/tests/features/test_audio.py	2023-05-06 09:05:50.463407967 +0200
@@ -57,6 +57,7 @@
     assert features.arrow_schema == pa.schema({"sequence_of_audios": pa.list_(Audio().pa_type)})
 
 
+@pytest.mark.skip(reason="require librosa")
 @pytest.mark.parametrize(
     "build_example",
     [
@@ -81,6 +82,7 @@
     assert decoded_example.keys() == {"path", "array", "sampling_rate"}
 
 
+@pytest.mark.skip(reason="require librosa")
 @pytest.mark.parametrize(
     "build_example",
     [
@@ -148,6 +149,7 @@
     assert decoded_example["sampling_rate"] == 48000
 
 
+@pytest.mark.skip(reason="require librosa")
 @pytest.mark.parametrize("sampling_rate", [16_000, 48_000])
 def test_audio_decode_example_pcm(shared_datadir, sampling_rate):
     audio_path = str(shared_datadir / "test_audio_16000.pcm")
@@ -414,6 +417,7 @@
     assert column[0]["sampling_rate"] == 16000
 
 
+@pytest.mark.skip(reason="require librosa")
 @pytest.mark.parametrize(
     "build_data",
     [
@@ -438,6 +442,7 @@
     assert item["audio"].keys() == {"path", "array", "sampling_rate"}
 
 
+@pytest.mark.skip(reason="require librosa")
 def test_dataset_concatenate_audio_features(shared_datadir):
     # we use a different data structure between 1 and 2 to make sure they are compatible with each other
     audio_path = str(shared_datadir / "test_audio_44100.wav")
@@ -451,6 +456,7 @@
     assert concatenated_dataset[1]["audio"]["array"].shape == dset2[0]["audio"]["array"].shape
 
 
+@pytest.mark.skip(reason="require librosa")
 def test_dataset_concatenate_nested_audio_features(shared_datadir):
     # we use a different data structure between 1 and 2 to make sure they are compatible with each other
     audio_path = str(shared_datadir / "test_audio_44100.wav")
@@ -610,6 +616,7 @@
     assert isinstance(ds, Dataset)
 
 
+@require_sndfile
 def test_dataset_with_audio_feature_undecoded(shared_datadir):
     audio_path = str(shared_datadir / "test_audio_44100.wav")
     data = {"audio": [audio_path]}
@@ -627,6 +634,7 @@
     assert column[0] == {"path": audio_path, "bytes": None}
 
 
+@require_sndfile
 def test_formatted_dataset_with_audio_feature_undecoded(shared_datadir):
     audio_path = str(shared_datadir / "test_audio_44100.wav")
     data = {"audio": [audio_path]}
@@ -658,6 +666,7 @@
         assert column[0] == {"path": audio_path, "bytes": None}
 
 
+@require_sndfile
 def test_dataset_with_audio_feature_map_undecoded(shared_datadir):
     audio_path = str(shared_datadir / "test_audio_44100.wav")
     data = {"audio": [audio_path]}
--- a/tests/packaged_modules/test_audiofolder.py	2023-05-06 14:00:39.560876163 +0200
+++ b/tests/packaged_modules/test_audiofolder.py	2023-05-06 14:01:26.005212423 +0200
@@ -1,10 +1,8 @@
 import shutil
 import textwrap
 
-import librosa
 import numpy as np
 import pytest
-import soundfile as sf
 
 from datasets import Audio, ClassLabel, Features, Value
 from datasets.data_files import DataFilesDict, get_data_patterns
@@ -192,8 +190,11 @@
     return data_files_with_two_splits_and_metadata
 
 
+@pytest.mark.skip(reason="require soundfile")
 @pytest.fixture
 def data_files_with_zip_archives(tmp_path, audio_file):
+    import soundfile as sf
+    import librosa
     data_dir = tmp_path / "audiofolder_data_dir_with_zip_archives"
     data_dir.mkdir(parents=True, exist_ok=True)
     archive_dir = data_dir / "archive"
--- a/tests/test_arrow_dataset.py	2023-05-06 15:36:11.080459079 +0200
+++ b/tests/test_arrow_dataset.py	2023-05-06 15:38:07.452828528 +0200
@@ -4136,6 +4136,7 @@
                 )
                 self.assertDictEqual(features_after_cast, dset.features)
 
+    @pytest.mark.skip(reason="require soundfile")
     def test_task_automatic_speech_recognition(self):
         # Include a dummy extra column `dummy` to test we drop it correctly
         features_before_cast = Features(
--- a/tests/test_streaming_download_manager.py	2023-08-26 07:33:41.937389401 +0200
+++ b/tests/test_streaming_download_manager.py	2023-08-26 07:37:22.521218698 +0200
@@ -218,6 +218,7 @@
     assert output_path == _readd_double_slash_removed_by_path(Path(expected_path).as_posix())
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, exists",
     [
@@ -301,6 +302,7 @@
         assert list(f) == TEST_URL_CONTENT.splitlines(keepends=True)
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, expected_paths",
     [
@@ -331,6 +333,7 @@
         xlistdir(root_url, download_config=download_config)
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, isdir",
     [
@@ -358,6 +361,7 @@
     assert xisdir(root_url, download_config=download_config) is False
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, isfile",
     [
@@ -382,6 +386,7 @@
     assert xisfile(root_url + "qwertyuiop", download_config=download_config) is False
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, size",
     [
@@ -407,6 +412,7 @@
         xgetsize(root_url + "qwertyuiop", download_config=download_config)
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, expected_paths",
     [
@@ -450,6 +456,7 @@
     assert len(xglob("zip://qwertyuiop/*::" + root_url, download_config=download_config)) == 0
 
 
+@pytest.mark.skip(reason="not working in sandbox")
 @pytest.mark.parametrize(
     "input_path, expected_outputs",
     [
@@ -540,6 +547,7 @@
     def test_xpath_as_posix(self, input_path, expected_path):
         assert xPath(input_path).as_posix() == expected_path
 
+    @pytest.mark.skip(reason="not working in sandbox")
     @pytest.mark.parametrize(
         "input_path, exists",
         [
@@ -555,6 +563,7 @@
             (tmp_path / "file.txt").touch()
         assert xexists(input_path) is exists
 
+    @pytest.mark.skip(reason="not working in sandbox")
     @pytest.mark.parametrize(
         "input_path, pattern, expected_paths",
         [
@@ -593,6 +602,7 @@
         output_paths = sorted(xPath(input_path).glob(pattern))
         assert output_paths == expected_paths
 
+    @pytest.mark.skip(reason="not working in sandbox")
     @pytest.mark.parametrize(
         "input_path, pattern, expected_paths",
         [