[Task]: [RRIO] [Testing] Define and implement mock quota aware API #1324
16 fail, 52 skipped, 83 pass in 2h 10m 31s
Annotations
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 7s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source>
@pytest.mark.it_postcommit
def test_iobase_source(self):
EXPECTED_TABLE_DATA = [
{
'number': 1,
'string': '你好',
'time': datetime.time(12, 44, 31),
'datetime': '2018-12-31T12:44:31',
'rec': None,
},
{
'number': 4,
'string': 'привет',
'time': datetime.time(12, 44, 31),
'datetime': '2018-12-31T12:44:31',
'rec': {
'rec_datetime': '2018-12-31T12:44:31',
'rec_rec': {
'rec_rec_datetime': '2018-12-31T12:44:31',
}
},
}
]
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
table=self.temp_table_reference))
> assert_that(result, equal_to(EXPECTED_TABLE_DATA))
apache_beam/io/gcp/bigquery_read_it_test.py:451:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_column_selection (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_column_selection>
@pytest.mark.it_postcommit
def test_iobase_source_with_column_selection(self):
EXPECTED_TABLE_DATA = [{'number': 1}, {'number': 4}]
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
table=self.temp_table_reference,
selected_fields=['number']))
> assert_that(result, equal_to(EXPECTED_TABLE_DATA))
apache_beam/io/gcp/bigquery_read_it_test.py:494:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_column_selection_and_row_restriction_rows (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_column_selection_and_row_restriction_rows>
@pytest.mark.it_postcommit
def test_iobase_source_with_column_selection_and_row_restriction_rows(self):
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
table=self.temp_table_reference,
row_restriction='number > 2',
selected_fields=['string'],
output_type='BEAM_ROW'))
> assert_that(
result | beam.Map(lambda row: row.string), equal_to(['привет']))
apache_beam/io/gcp/bigquery_read_it_test.py:536:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_query (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_query>
@pytest.mark.it_postcommit
def test_iobase_source_with_query(self):
EXPECTED_TABLE_DATA = [
{
'number': 1,
'string': '你好',
'time': datetime.time(12, 44, 31),
'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
'rec': None,
},
{
'number': 4,
'string': 'привет',
'time': datetime.time(12, 44, 31),
'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
'rec': {
'rec_datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
'rec_rec': {
'rec_rec_datetime': datetime.datetime(
2018, 12, 31, 12, 44, 31)
}
},
}
]
query = StaticValueProvider(str, self.query)
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Direct read with query' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
use_native_datetime=True,
use_standard_sql=True,
project=self.project,
query=query))
> assert_that(result, equal_to(EXPECTED_TABLE_DATA))
apache_beam/io/gcp/bigquery_read_it_test.py:585:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_query_and_filters (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_query_and_filters>
@pytest.mark.it_postcommit
def test_iobase_source_with_query_and_filters(self):
EXPECTED_TABLE_DATA = [{'string': 'привет'}]
query = StaticValueProvider(str, self.query)
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Direct read with query' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
row_restriction='number > 2',
selected_fields=['string'],
use_standard_sql=True,
project=self.project,
query=query))
> assert_that(result, equal_to(EXPECTED_TABLE_DATA))
apache_beam/io/gcp/bigquery_read_it_test.py:600:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_row_restriction (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_row_restriction>
@pytest.mark.it_postcommit
def test_iobase_source_with_row_restriction(self):
EXPECTED_TABLE_DATA = [{
'number': 1,
'string': '你好',
'time': datetime.time(12, 44, 31),
'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
'rec': None
}]
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
table=self.temp_table_reference,
row_restriction='number < 2',
use_native_datetime=True))
> assert_that(result, equal_to(EXPECTED_TABLE_DATA))
apache_beam/io/gcp/bigquery_read_it_test.py:512:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests
github-actions / Test Results
1 out of 3 runs failed: test_iobase_source_with_very_selective_filters (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_very_selective_filters>
@pytest.mark.it_postcommit
def test_iobase_source_with_very_selective_filters(self):
with beam.Pipeline(argv=self.args) as p:
result = (
p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
project=self.temp_table_reference.projectId,
dataset=self.temp_table_reference.datasetId,
table=self.temp_table_reference.tableId,
row_restriction='number > 4',
selected_fields=['string']))
> assert_that(result, equal_to([]))
apache_beam/io/gcp/bigquery_read_it_test.py:550:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests
github-actions / Test Results
1 out of 3 runs failed: test_native_source (apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests testMethod=test_native_source>
@skip(['PortableRunner', 'FlinkRunner'])
@pytest.mark.it_postcommit
def test_native_source(self):
with beam.Pipeline(argv=self.args) as p:
result = (
p
| 'read' >> beam.io.Read(
beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
> assert_that(result, equal_to(self.get_expected_data()))
apache_beam/io/gcp/bigquery_read_it_test.py:707:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests
github-actions / Test Results
1 out of 3 runs failed: test_read_queries (apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 2s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests testMethod=test_read_queries>
@skip(['PortableRunner', 'FlinkRunner'])
@pytest.mark.it_postcommit
def test_read_queries(self):
with beam.Pipeline(argv=self.args) as p:
result = (
p
| beam.Create([
beam.io.ReadFromBigQueryRequest(query=self.query1),
beam.io.ReadFromBigQueryRequest(
query=self.query2, use_standard_sql=False),
beam.io.ReadFromBigQueryRequest(
table='%s.%s' % (self.dataset_id, self.table_name3))
])
| beam.io.ReadAllFromBigQuery())
> assert_that(
result,
equal_to(self.TABLE_DATA_1 + self.TABLE_DATA_2 + self.TABLE_DATA_3))
apache_beam/io/gcp/bigquery_read_it_test.py:819:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
1 out of 2 runs failed: test_big_query_write (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write>
@pytest.mark.it_postcommit
def test_big_query_write(self):
table_name = 'python_write_table'
table_id = '{}.{}'.format(self.dataset_id, table_name)
input_data = [
{
'number': 1, 'str': 'abc'
},
{
'number': 2, 'str': 'def'
},
{
'number': 3, 'str': '你好'
},
{
'number': 4, 'str': 'привет'
},
]
table_schema = {
"fields": [{
"name": "number", "type": "INTEGER"
}, {
"name": "str", "type": "STRING"
}]
}
pipeline_verifiers = [
BigqueryFullResultMatcher(
project=self.project,
query="SELECT number, str FROM %s" % table_id,
data=[(
1,
'abc',
), (
2,
'def',
), (
3,
'你好',
), (
4,
'привет',
)])
]
args = self.test_pipeline.get_full_options_as_args(
on_success_matcher=hc.all_of(*pipeline_verifiers))
with beam.Pipeline(argv=args) as p:
# pylint: disable=expression-not-assigned
> (
p | 'create' >> beam.Create(input_data)
| 'write' >> beam.io.WriteToBigQuery(
table_id,
schema=table_schema,
create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
write_disposition=beam.io.BigQueryDisposition.WRITE_EMPTY))
apache_beam/io/gcp/bigquery_write_it_test.py:168:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
1 out of 2 runs failed: test_big_query_write_insert_errors_reporting (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_insert_errors_reporting>
@pytest.mark.it_postcommit
def test_big_query_write_insert_errors_reporting(self):
"""
Test that errors returned by beam.io.WriteToBigQuery
contain both the failed rows and the reason for it failing.
"""
table_name = 'python_write_table'
table_id = '{}.{}'.format(self.dataset_id, table_name)
input_data = [{
'number': 1,
'str': 'some_string',
}, {
'number': 2
},
{
'number': 3,
'str': 'some_string',
'additional_field_str': 'some_string',
}]
table_schema = {
"fields": [{
"name": "number", "type": "INTEGER", 'mode': 'REQUIRED'
}, {
"name": "str", "type": "STRING", 'mode': 'REQUIRED'
}]
}
bq_result_errors = [(
{
"number": 2
},
[{
"reason": "invalid",
"location": "",
"debugInfo": "",
"message": "Missing required field: Msg_0_CLOUD_QUERY_TABLE.str."
}],
),
({
"number": 3,
"str": "some_string",
"additional_field_str": "some_string"
},
[{
"reason": "invalid",
"location": "additional_field_str",
"debugInfo": "",
"message": "no such field: additional_field_str."
}])]
pipeline_verifiers = [
BigqueryFullResultMatcher(
project=self.project,
query="SELECT number, str FROM %s" % table_id,
data=[(1, 'some_string')]),
]
args = self.test_pipeline.get_full_options_as_args(
on_success_matcher=hc.all_of(*pipeline_verifiers))
with beam.Pipeline(argv=args) as p:
# pylint: disable=expression-not-assigned
errors = (
p | 'create' >> beam.Create(input_data)
| 'write' >> beam.io.WriteToBigQuery(
table_id,
schema=table_schema,
method='STREAMING_INSERTS',
insert_retry_strategy='RETRY_NEVER',
create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND))
> assert_that(
errors[BigQueryWriteFn.FAILED_ROWS_WITH_ERRORS]
| 'ParseErrors' >> beam.Map(lambda err: (err[1], err[2])),
equal_to(bq_result_errors))
apache_beam/io/gcp/bigquery_write_it_test.py:452:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
1 out of 2 runs failed: test_big_query_write_insert_non_transient_api_call_error (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_insert_non_transient_api_call_error>
@pytest.mark.it_postcommit
def test_big_query_write_insert_non_transient_api_call_error(self):
"""
Test that non-transient GoogleAPICallError errors returned
by beam.io.WriteToBigQuery are not retried and result in
FAILED_ROWS containing both the failed rows and the reason
for failure.
"""
table_name = 'this_table_does_not_exist'
table_id = '{}.{}'.format(self.dataset_id, table_name)
input_data = [{
'number': 1,
'str': 'some_string',
}]
table_schema = {
"fields": [{
"name": "number", "type": "INTEGER", 'mode': 'NULLABLE'
}, {
"name": "str", "type": "STRING", 'mode': 'NULLABLE'
}]
}
bq_result_errors = [({
'number': 1,
'str': 'some_string',
}, "Not Found")]
args = self.test_pipeline.get_full_options_as_args()
with beam.Pipeline(argv=args) as p:
# pylint: disable=expression-not-assigned
errors = (
p | 'create' >> beam.Create(input_data)
| 'write' >> beam.io.WriteToBigQuery(
table_id,
schema=table_schema,
method='STREAMING_INSERTS',
insert_retry_strategy='RETRY_ON_TRANSIENT_ERROR',
create_disposition=beam.io.BigQueryDisposition.CREATE_NEVER,
write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND))
> assert_that(
errors[BigQueryWriteFn.FAILED_ROWS_WITH_ERRORS]
|
'ParseErrors' >> beam.Map(lambda err: (err[1], err[2][0]["reason"])),
equal_to(bq_result_errors))
apache_beam/io/gcp/bigquery_write_it_test.py:500:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
1 out of 2 runs failed: test_big_query_write_schema_autodetect (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_schema_autodetect>
@pytest.mark.it_postcommit
def test_big_query_write_schema_autodetect(self):
if self.runner_name == 'TestDataflowRunner':
self.skipTest('DataflowRunner does not support schema autodetection')
table_name = 'python_write_table'
table_id = '{}.{}'.format(self.dataset_id, table_name)
input_data = [
{
'number': 1, 'str': 'abc'
},
{
'number': 2, 'str': 'def'
},
]
pipeline_verifiers = [
BigqueryFullResultMatcher(
project=self.project,
query="SELECT number, str FROM %s" % table_id,
data=[(
1,
'abc',
), (
2,
'def',
)])
]
args = self.test_pipeline.get_full_options_as_args(
on_success_matcher=hc.all_of(*pipeline_verifiers))
with beam.Pipeline(argv=args) as p:
# pylint: disable=expression-not-assigned
> (
p | 'create' >> beam.Create(input_data)
| 'write' >> beam.io.WriteToBigQuery(
table_id,
method=beam.io.WriteToBigQuery.Method.FILE_LOADS,
schema=beam.io.gcp.bigquery.SCHEMA_AUTODETECT,
create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
write_disposition=beam.io.BigQueryDisposition.WRITE_EMPTY,
temp_file_format=FileFormat.JSON))
apache_beam/io/gcp/bigquery_write_it_test.py:211:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
test_big_query_write_temp_table_append_schema_update_2 (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) failed
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
a = (<apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_temp_table_append_schema_update_2>,)
kw = {}
@wraps(func)
def standalone_func(*a, **kw):
> return func(*(a + p.args), **p.kwargs, **kw)
../../build/gradleenv/1398941893/lib/python3.9/site-packages/parameterized/parameterized.py:620:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
../../build/gradleenv/1398941893/lib/python3.9/site-packages/mock/mock.py:1452: in patched
return func(*newargs, **newkeywargs)
../../build/gradleenv/1398941893/lib/python3.9/site-packages/parameterized/parameterized.py:93: in dummy_func
return orgfunc(*args, **kwargs)
../../build/gradleenv/1398941893/lib/python3.9/site-packages/mock/mock.py:1452: in patched
return func(*newargs, **newkeywargs)
apache_beam/io/gcp/bigquery_write_it_test.py:581: in test_big_query_write_temp_table_append_schema_update
(
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests
github-actions / Test Results
1 out of 2 runs failed: test_big_query_write_without_schema (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_without_schema>
@pytest.mark.it_postcommit
def test_big_query_write_without_schema(self):
table_name = 'python_no_schema_table'
self.create_table(table_name)
table_id = '{}.{}'.format(self.dataset_id, table_name)
input_data = [{
'int64': 1,
'bytes': b'xyw',
'date': '2011-01-01',
'time': '23:59:59.999999'
},
{
'int64': 2,
'bytes': b'abc',
'date': '2000-01-01',
'time': '00:00:00'
},
{
'int64': 3,
'bytes': b'\xe4\xbd\xa0\xe5\xa5\xbd',
'date': '3000-12-31',
'time': '23:59:59'
},
{
'int64': 4,
'bytes': b'\xab\xac\xad',
'date': '2000-01-01',
'time': '00:00:00'
}]
# bigquery io expects bytes to be base64 encoded values
for row in input_data:
row['bytes'] = base64.b64encode(row['bytes'])
pipeline_verifiers = [
BigqueryFullResultMatcher(
project=self.project,
query="SELECT int64, bytes, date, time FROM %s" % table_id,
data=[(
1,
b'xyw',
datetime.date(2011, 1, 1),
datetime.time(23, 59, 59, 999999),
),
(
2,
b'abc',
datetime.date(2000, 1, 1),
datetime.time(0, 0, 0),
),
(
3,
b'\xe4\xbd\xa0\xe5\xa5\xbd',
datetime.date(3000, 12, 31),
datetime.time(23, 59, 59),
),
(
4,
b'\xab\xac\xad',
datetime.date(2000, 1, 1),
datetime.time(0, 0, 0),
)])
]
args = self.test_pipeline.get_full_options_as_args(
on_success_matcher=hc.all_of(*pipeline_verifiers))
with beam.Pipeline(argv=args) as p:
# pylint: disable=expression-not-assigned
> (
p | 'create' >> beam.Create(input_data)
| 'write' >> beam.io.WriteToBigQuery(
table_id,
write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND,
temp_file_format=FileFormat.JSON))
apache_beam/io/gcp/bigquery_write_it_test.py:371:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/pipeline.py:607: in __exit__
self.result = self.run()
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
github-actions / Test Results
1 out of 2 runs failed: test_big_query_legacy_sql (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT)
sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT testMethod=test_big_query_legacy_sql>
@pytest.mark.it_postcommit
def test_big_query_legacy_sql(self):
verify_query = DIALECT_OUTPUT_VERIFY_QUERY % self.output_table
expected_checksum = test_utils.compute_hash(DIALECT_OUTPUT_EXPECTED)
pipeline_verifiers = [
PipelineStateMatcher(),
BigqueryMatcher(
project=self.project,
query=verify_query,
checksum=expected_checksum)
]
extra_opts = {
'query': LEGACY_QUERY,
'output': self.output_table,
'output_schema': DIALECT_OUTPUT_SCHEMA,
'use_standard_sql': False,
'wait_until_finish_duration': WAIT_UNTIL_FINISH_DURATION_MS,
'on_success_matcher': all_of(*pipeline_verifiers),
}
options = self.test_pipeline.get_full_options_as_args(**extra_opts)
> big_query_query_to_table_pipeline.run_bq_pipeline(options)
apache_beam/io/gcp/big_query_query_to_table_it_test.py:178:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
apache_beam/io/gcp/big_query_query_to_table_pipeline.py:103: in run_bq_pipeline
result = p.run()
apache_beam/testing/test_pipeline.py:112: in run
result = super().run(
apache_beam/pipeline.py:557: in run
return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E KeyError: '__pyx_vtable__'
apache_beam/runners/worker/operations.py:1: KeyError
Check notice on line 0 in .github
github-actions / Test Results
52 skipped tests found
There are 52 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_io_roundtrip
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.base_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test
Check notice on line 0 in .github
github-actions / Test Results
151 tests found
There are 151 tests, see "Raw output" for the full list of tests.
Raw output
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_export_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_with_computation
apache_beam.examples.complete.autocomplete_test.AutocompleteTest ‑ test_autocomplete_it
apache_beam.examples.complete.game.game_stats_it_test.GameStatsIT ‑ test_game_stats_it
apache_beam.examples.complete.game.hourly_team_score_it_test.HourlyTeamScoreIT ‑ test_hourly_team_score_it
apache_beam.examples.complete.game.leader_board_it_test.LeaderBoardIT ‑ test_leader_board_it
apache_beam.examples.complete.game.user_score_it_test.UserScoreIT ‑ test_user_score_it
apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT ‑ test_run_example_with_setup_file
apache_beam.examples.cookbook.bigquery_tornadoes_it_test.BigqueryTornadoesIT ‑ test_bigquery_tornadoes_it
apache_beam.examples.cookbook.bigtableio_it_test.BigtableIOWriteTest ‑ test_bigtable_write
apache_beam.examples.cookbook.datastore_wordcount_it_test.DatastoreWordCountIT ‑ test_datastore_wordcount_it
apache_beam.examples.dataframe.flight_delays_it_test.FlightDelaysTest ‑ test_flight_delays
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_aggregation
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_enrich
apache_beam.examples.fastavro_it_test.FastavroIT ‑ test_avro_it
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT ‑ test_streaming_wordcount_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_impersonation_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_it
apache_beam.io.external.xlang_debeziumio_it_test.CrossLanguageDebeziumIOTest ‑ test_xlang_debezium_read
apache_beam.io.external.xlang_jdbcio_it_test.CrossLanguageJdbcIOTest ‑ test_xlang_jdbc_write_read_0_postgres
apache_beam.io.external.xlang_jdbcio_it_test.CrossLanguageJdbcIOTest ‑ test_xlang_jdbc_write_read_1_mysql
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_local_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_local_kafkaio_populated_key
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_io_roundtrip
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_write
apache_beam.io.fileio_test.MatchIntegrationTest ‑ test_transform_on_gcs
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_legacy_sql
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types_avro
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_standard_sql
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_one_job_fails_all_jobs_fail
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_1M_python
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_custom_1M_python
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_direct_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_export_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_file_loads_write
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_query_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests ‑ test_read_queries
apache_beam.io.gcp.bigquery_read_it_test.ReadInteractiveRunnerTests ‑ test_read_in_interactive_runner
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_specifying_only_table
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_with_direct_read
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_native_datetime
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query_and_filters
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_very_selective_filters
apache_beam.io.gcp.bigquery_test.BigQueryFileLoadsIntegrationTests ‑ test_avro_file_load
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_value_provider_transform
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_file_loads
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_errors_reporting
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_non_transient_api_call_error
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_new_types
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_schema_autodetect
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_0
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_1
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_2
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_without_schema
apache_beam.io.gcp.datastore.v1new.datastore_write_it_test.DatastoreWriteIT ‑ test_datastore_write_limit
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename_error
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch_kms
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch_rewrite_token
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_kms
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_rewrite_token
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_search_instances
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_store_instance_from_gcs
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_data_only
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_with_attributes
apache_beam.io.parquetio_it_test.TestParquetIT ‑ test_parquetio_it
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_deidentification
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_inspection
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_catalog_item
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_user_event
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_predict
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.gcp.visionml_test_it.VisionMlTestIT ‑ test_text_detection_with_language_hint
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification_large_model
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_regression
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.base_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.external_it_test.ExternalTransformIT ‑ test_job_python_from_python_it
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test