Skip to content

[RRIO] Define and implement mock quota aware API (#28893)

Sign in for the full log view
GitHub Actions / Test Results failed Oct 23, 2023 in 0s

16 fail, 52 skipped, 83 pass in 2h 10m 31s

    3 files      3 suites   2h 10m 31s ⏱️
151 tests   83 ✔️ 52 💤 16
203 runs  130 ✔️ 57 💤 16

Results for commit a07d90e.

Annotations

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 7s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source>

    @pytest.mark.it_postcommit
    def test_iobase_source(self):
      EXPECTED_TABLE_DATA = [
          {
              'number': 1,
              'string': '你好',
              'time': datetime.time(12, 44, 31),
              'datetime': '2018-12-31T12:44:31',
              'rec': None,
          },
          {
              'number': 4,
              'string': 'привет',
              'time': datetime.time(12, 44, 31),
              'datetime': '2018-12-31T12:44:31',
              'rec': {
                  'rec_datetime': '2018-12-31T12:44:31',
                  'rec_rec': {
                      'rec_rec_datetime': '2018-12-31T12:44:31',
                  }
              },
          }
      ]
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                table=self.temp_table_reference))
>       assert_that(result, equal_to(EXPECTED_TABLE_DATA))

apache_beam/io/gcp/bigquery_read_it_test.py:451: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_column_selection (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_column_selection>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_column_selection(self):
      EXPECTED_TABLE_DATA = [{'number': 1}, {'number': 4}]
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                table=self.temp_table_reference,
                selected_fields=['number']))
>       assert_that(result, equal_to(EXPECTED_TABLE_DATA))

apache_beam/io/gcp/bigquery_read_it_test.py:494: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_column_selection_and_row_restriction_rows (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_column_selection_and_row_restriction_rows>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_column_selection_and_row_restriction_rows(self):
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                table=self.temp_table_reference,
                row_restriction='number > 2',
                selected_fields=['string'],
                output_type='BEAM_ROW'))
>       assert_that(
            result | beam.Map(lambda row: row.string), equal_to(['привет']))

apache_beam/io/gcp/bigquery_read_it_test.py:536: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_query (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_query>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_query(self):
      EXPECTED_TABLE_DATA = [
          {
              'number': 1,
              'string': '你好',
              'time': datetime.time(12, 44, 31),
              'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
              'rec': None,
          },
          {
              'number': 4,
              'string': 'привет',
              'time': datetime.time(12, 44, 31),
              'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
              'rec': {
                  'rec_datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
                  'rec_rec': {
                      'rec_rec_datetime': datetime.datetime(
                          2018, 12, 31, 12, 44, 31)
                  }
              },
          }
      ]
      query = StaticValueProvider(str, self.query)
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Direct read with query' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                use_native_datetime=True,
                use_standard_sql=True,
                project=self.project,
                query=query))
>       assert_that(result, equal_to(EXPECTED_TABLE_DATA))

apache_beam/io/gcp/bigquery_read_it_test.py:585: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_query_and_filters (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_query_and_filters>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_query_and_filters(self):
      EXPECTED_TABLE_DATA = [{'string': 'привет'}]
      query = StaticValueProvider(str, self.query)
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Direct read with query' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                row_restriction='number > 2',
                selected_fields=['string'],
                use_standard_sql=True,
                project=self.project,
                query=query))
>       assert_that(result, equal_to(EXPECTED_TABLE_DATA))

apache_beam/io/gcp/bigquery_read_it_test.py:600: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_row_restriction (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 0s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_row_restriction>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_row_restriction(self):
      EXPECTED_TABLE_DATA = [{
          'number': 1,
          'string': '你好',
          'time': datetime.time(12, 44, 31),
          'datetime': datetime.datetime(2018, 12, 31, 12, 44, 31),
          'rec': None
      }]
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                table=self.temp_table_reference,
                row_restriction='number < 2',
                use_native_datetime=True))
>       assert_that(result, equal_to(EXPECTED_TABLE_DATA))

apache_beam/io/gcp/bigquery_read_it_test.py:512: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_iobase_source_with_very_selective_filters (apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests testMethod=test_iobase_source_with_very_selective_filters>

    @pytest.mark.it_postcommit
    def test_iobase_source_with_very_selective_filters(self):
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p | 'Read with BigQuery Storage API' >> beam.io.ReadFromBigQuery(
                method=beam.io.ReadFromBigQuery.Method.DIRECT_READ,
                project=self.temp_table_reference.projectId,
                dataset=self.temp_table_reference.datasetId,
                table=self.temp_table_reference.tableId,
                row_restriction='number > 4',
                selected_fields=['string']))
>       assert_that(result, equal_to([]))

apache_beam/io/gcp/bigquery_read_it_test.py:550: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_native_source (apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests testMethod=test_native_source>

    @skip(['PortableRunner', 'FlinkRunner'])
    @pytest.mark.it_postcommit
    def test_native_source(self):
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p
            | 'read' >> beam.io.Read(
                beam.io.BigQuerySource(query=self.query, use_standard_sql=True)))
>       assert_that(result, equal_to(self.get_expected_data()))

apache_beam/io/gcp/bigquery_read_it_test.py:707: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 3 runs failed: test_read_queries (apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 2s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests testMethod=test_read_queries>

    @skip(['PortableRunner', 'FlinkRunner'])
    @pytest.mark.it_postcommit
    def test_read_queries(self):
      with beam.Pipeline(argv=self.args) as p:
        result = (
            p
            | beam.Create([
                beam.io.ReadFromBigQueryRequest(query=self.query1),
                beam.io.ReadFromBigQueryRequest(
                    query=self.query2, use_standard_sql=False),
                beam.io.ReadFromBigQueryRequest(
                    table='%s.%s' % (self.dataset_id, self.table_name3))
            ])
            | beam.io.ReadAllFromBigQuery())
>       assert_that(
            result,
            equal_to(self.TABLE_DATA_1 + self.TABLE_DATA_2 + self.TABLE_DATA_3))

apache_beam/io/gcp/bigquery_read_it_test.py:819: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_write (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write>

    @pytest.mark.it_postcommit
    def test_big_query_write(self):
      table_name = 'python_write_table'
      table_id = '{}.{}'.format(self.dataset_id, table_name)
    
      input_data = [
          {
              'number': 1, 'str': 'abc'
          },
          {
              'number': 2, 'str': 'def'
          },
          {
              'number': 3, 'str': '你好'
          },
          {
              'number': 4, 'str': 'привет'
          },
      ]
      table_schema = {
          "fields": [{
              "name": "number", "type": "INTEGER"
          }, {
              "name": "str", "type": "STRING"
          }]
      }
    
      pipeline_verifiers = [
          BigqueryFullResultMatcher(
              project=self.project,
              query="SELECT number, str FROM %s" % table_id,
              data=[(
                  1,
                  'abc',
              ), (
                  2,
                  'def',
              ), (
                  3,
                  '你好',
              ), (
                  4,
                  'привет',
              )])
      ]
    
      args = self.test_pipeline.get_full_options_as_args(
          on_success_matcher=hc.all_of(*pipeline_verifiers))
    
      with beam.Pipeline(argv=args) as p:
        # pylint: disable=expression-not-assigned
>       (
            p | 'create' >> beam.Create(input_data)
            | 'write' >> beam.io.WriteToBigQuery(
                table_id,
                schema=table_schema,
                create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
                write_disposition=beam.io.BigQueryDisposition.WRITE_EMPTY))

apache_beam/io/gcp/bigquery_write_it_test.py:168: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_write_insert_errors_reporting (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_insert_errors_reporting>

    @pytest.mark.it_postcommit
    def test_big_query_write_insert_errors_reporting(self):
      """
      Test that errors returned by beam.io.WriteToBigQuery
      contain both the failed rows and the reason for it failing.
      """
      table_name = 'python_write_table'
      table_id = '{}.{}'.format(self.dataset_id, table_name)
    
      input_data = [{
          'number': 1,
          'str': 'some_string',
      }, {
          'number': 2
      },
                    {
                        'number': 3,
                        'str': 'some_string',
                        'additional_field_str': 'some_string',
                    }]
    
      table_schema = {
          "fields": [{
              "name": "number", "type": "INTEGER", 'mode': 'REQUIRED'
          }, {
              "name": "str", "type": "STRING", 'mode': 'REQUIRED'
          }]
      }
    
      bq_result_errors = [(
          {
              "number": 2
          },
          [{
              "reason": "invalid",
              "location": "",
              "debugInfo": "",
              "message": "Missing required field: Msg_0_CLOUD_QUERY_TABLE.str."
          }],
      ),
                          ({
                              "number": 3,
                              "str": "some_string",
                              "additional_field_str": "some_string"
                          },
                           [{
                               "reason": "invalid",
                               "location": "additional_field_str",
                               "debugInfo": "",
                               "message": "no such field: additional_field_str."
                           }])]
    
      pipeline_verifiers = [
          BigqueryFullResultMatcher(
              project=self.project,
              query="SELECT number, str FROM %s" % table_id,
              data=[(1, 'some_string')]),
      ]
    
      args = self.test_pipeline.get_full_options_as_args(
          on_success_matcher=hc.all_of(*pipeline_verifiers))
    
      with beam.Pipeline(argv=args) as p:
        # pylint: disable=expression-not-assigned
        errors = (
            p | 'create' >> beam.Create(input_data)
            | 'write' >> beam.io.WriteToBigQuery(
                table_id,
                schema=table_schema,
                method='STREAMING_INSERTS',
                insert_retry_strategy='RETRY_NEVER',
                create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
                write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND))
    
>       assert_that(
            errors[BigQueryWriteFn.FAILED_ROWS_WITH_ERRORS]
            | 'ParseErrors' >> beam.Map(lambda err: (err[1], err[2])),
            equal_to(bq_result_errors))

apache_beam/io/gcp/bigquery_write_it_test.py:452: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_write_insert_non_transient_api_call_error (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_insert_non_transient_api_call_error>

    @pytest.mark.it_postcommit
    def test_big_query_write_insert_non_transient_api_call_error(self):
      """
      Test that non-transient GoogleAPICallError errors returned
      by beam.io.WriteToBigQuery are not retried and result in
      FAILED_ROWS containing both the failed rows and the reason
      for failure.
      """
      table_name = 'this_table_does_not_exist'
      table_id = '{}.{}'.format(self.dataset_id, table_name)
    
      input_data = [{
          'number': 1,
          'str': 'some_string',
      }]
    
      table_schema = {
          "fields": [{
              "name": "number", "type": "INTEGER", 'mode': 'NULLABLE'
          }, {
              "name": "str", "type": "STRING", 'mode': 'NULLABLE'
          }]
      }
    
      bq_result_errors = [({
          'number': 1,
          'str': 'some_string',
      }, "Not Found")]
    
      args = self.test_pipeline.get_full_options_as_args()
    
      with beam.Pipeline(argv=args) as p:
        # pylint: disable=expression-not-assigned
        errors = (
            p | 'create' >> beam.Create(input_data)
            | 'write' >> beam.io.WriteToBigQuery(
                table_id,
                schema=table_schema,
                method='STREAMING_INSERTS',
                insert_retry_strategy='RETRY_ON_TRANSIENT_ERROR',
                create_disposition=beam.io.BigQueryDisposition.CREATE_NEVER,
                write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND))
    
>       assert_that(
            errors[BigQueryWriteFn.FAILED_ROWS_WITH_ERRORS]
            |
            'ParseErrors' >> beam.Map(lambda err: (err[1], err[2][0]["reason"])),
            equal_to(bq_result_errors))

apache_beam/io/gcp/bigquery_write_it_test.py:500: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_write_schema_autodetect (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_schema_autodetect>

    @pytest.mark.it_postcommit
    def test_big_query_write_schema_autodetect(self):
      if self.runner_name == 'TestDataflowRunner':
        self.skipTest('DataflowRunner does not support schema autodetection')
    
      table_name = 'python_write_table'
      table_id = '{}.{}'.format(self.dataset_id, table_name)
    
      input_data = [
          {
              'number': 1, 'str': 'abc'
          },
          {
              'number': 2, 'str': 'def'
          },
      ]
    
      pipeline_verifiers = [
          BigqueryFullResultMatcher(
              project=self.project,
              query="SELECT number, str FROM %s" % table_id,
              data=[(
                  1,
                  'abc',
              ), (
                  2,
                  'def',
              )])
      ]
    
      args = self.test_pipeline.get_full_options_as_args(
          on_success_matcher=hc.all_of(*pipeline_verifiers))
    
      with beam.Pipeline(argv=args) as p:
        # pylint: disable=expression-not-assigned
>       (
            p | 'create' >> beam.Create(input_data)
            | 'write' >> beam.io.WriteToBigQuery(
                table_id,
                method=beam.io.WriteToBigQuery.Method.FILE_LOADS,
                schema=beam.io.gcp.bigquery.SCHEMA_AUTODETECT,
                create_disposition=beam.io.BigQueryDisposition.CREATE_IF_NEEDED,
                write_disposition=beam.io.BigQueryDisposition.WRITE_EMPTY,
                temp_file_format=FileFormat.JSON))

apache_beam/io/gcp/bigquery_write_it_test.py:211: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

test_big_query_write_temp_table_append_schema_update_2 (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests) failed

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
a = (<apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_temp_table_append_schema_update_2>,)
kw = {}

    @wraps(func)
    def standalone_func(*a, **kw):
>       return func(*(a + p.args), **p.kwargs, **kw)

../../build/gradleenv/1398941893/lib/python3.9/site-packages/parameterized/parameterized.py:620: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
../../build/gradleenv/1398941893/lib/python3.9/site-packages/mock/mock.py:1452: in patched
    return func(*newargs, **newkeywargs)
../../build/gradleenv/1398941893/lib/python3.9/site-packages/parameterized/parameterized.py:93: in dummy_func
    return orgfunc(*args, **kwargs)
../../build/gradleenv/1398941893/lib/python3.9/site-packages/mock/mock.py:1452: in patched
    return func(*newargs, **newkeywargs)
apache_beam/io/gcp/bigquery_write_it_test.py:581: in test_big_query_write_temp_table_append_schema_update
    (
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_write_without_schema (apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests testMethod=test_big_query_write_without_schema>

    @pytest.mark.it_postcommit
    def test_big_query_write_without_schema(self):
      table_name = 'python_no_schema_table'
      self.create_table(table_name)
      table_id = '{}.{}'.format(self.dataset_id, table_name)
    
      input_data = [{
          'int64': 1,
          'bytes': b'xyw',
          'date': '2011-01-01',
          'time': '23:59:59.999999'
      },
                    {
                        'int64': 2,
                        'bytes': b'abc',
                        'date': '2000-01-01',
                        'time': '00:00:00'
                    },
                    {
                        'int64': 3,
                        'bytes': b'\xe4\xbd\xa0\xe5\xa5\xbd',
                        'date': '3000-12-31',
                        'time': '23:59:59'
                    },
                    {
                        'int64': 4,
                        'bytes': b'\xab\xac\xad',
                        'date': '2000-01-01',
                        'time': '00:00:00'
                    }]
      # bigquery io expects bytes to be base64 encoded values
      for row in input_data:
        row['bytes'] = base64.b64encode(row['bytes'])
    
      pipeline_verifiers = [
          BigqueryFullResultMatcher(
              project=self.project,
              query="SELECT int64, bytes, date, time FROM %s" % table_id,
              data=[(
                  1,
                  b'xyw',
                  datetime.date(2011, 1, 1),
                  datetime.time(23, 59, 59, 999999),
              ),
                    (
                        2,
                        b'abc',
                        datetime.date(2000, 1, 1),
                        datetime.time(0, 0, 0),
                    ),
                    (
                        3,
                        b'\xe4\xbd\xa0\xe5\xa5\xbd',
                        datetime.date(3000, 12, 31),
                        datetime.time(23, 59, 59),
                    ),
                    (
                        4,
                        b'\xab\xac\xad',
                        datetime.date(2000, 1, 1),
                        datetime.time(0, 0, 0),
                    )])
      ]
    
      args = self.test_pipeline.get_full_options_as_args(
          on_success_matcher=hc.all_of(*pipeline_verifiers))
    
      with beam.Pipeline(argv=args) as p:
        # pylint: disable=expression-not-assigned
>       (
            p | 'create' >> beam.Create(input_data)
            | 'write' >> beam.io.WriteToBigQuery(
                table_id,
                write_disposition=beam.io.BigQueryDisposition.WRITE_APPEND,
                temp_file_format=FileFormat.JSON))

apache_beam/io/gcp/bigquery_write_it_test.py:371: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/pipeline.py:607: in __exit__
    self.result = self.run()
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check warning on line 0 in apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT

See this annotation in the file changed.

@github-actions github-actions / Test Results

1 out of 2 runs failed: test_big_query_legacy_sql (apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT)

sdks/python/pytest_postCommitIT-direct-py39.xml [took 1s]
Raw output
KeyError: '__pyx_vtable__'
self = <apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT testMethod=test_big_query_legacy_sql>

    @pytest.mark.it_postcommit
    def test_big_query_legacy_sql(self):
      verify_query = DIALECT_OUTPUT_VERIFY_QUERY % self.output_table
      expected_checksum = test_utils.compute_hash(DIALECT_OUTPUT_EXPECTED)
      pipeline_verifiers = [
          PipelineStateMatcher(),
          BigqueryMatcher(
              project=self.project,
              query=verify_query,
              checksum=expected_checksum)
      ]
    
      extra_opts = {
          'query': LEGACY_QUERY,
          'output': self.output_table,
          'output_schema': DIALECT_OUTPUT_SCHEMA,
          'use_standard_sql': False,
          'wait_until_finish_duration': WAIT_UNTIL_FINISH_DURATION_MS,
          'on_success_matcher': all_of(*pipeline_verifiers),
      }
      options = self.test_pipeline.get_full_options_as_args(**extra_opts)
>     big_query_query_to_table_pipeline.run_bq_pipeline(options)

apache_beam/io/gcp/big_query_query_to_table_it_test.py:178: 
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 
apache_beam/io/gcp/big_query_query_to_table_pipeline.py:103: in run_bq_pipeline
    result = p.run()
apache_beam/testing/test_pipeline.py:112: in run
    result = super().run(
apache_beam/pipeline.py:557: in run
    return Pipeline.from_runner_api(
apache_beam/pipeline.py:584: in run
    return self.runner.run_pipeline(self, self._options)
apache_beam/runners/direct/test_direct_runner.py:42: in run_pipeline
    self.result = super().run_pipeline(pipeline, options)
apache_beam/runners/direct/direct_runner.py:117: in run_pipeline
    from apache_beam.runners.portability.fn_api_runner import fn_runner
apache_beam/runners/portability/fn_api_runner/__init__.py:18: in <module>
    from apache_beam.runners.portability.fn_api_runner.fn_runner import FnApiRunner
apache_beam/runners/portability/fn_api_runner/fn_runner.py:66: in <module>
    from apache_beam.runners.portability.fn_api_runner import execution
apache_beam/runners/portability/fn_api_runner/execution.py:61: in <module>
    from apache_beam.runners.portability.fn_api_runner import translations
apache_beam/runners/portability/fn_api_runner/translations.py:55: in <module>
    from apache_beam.runners.worker import bundle_processor
apache_beam/runners/worker/bundle_processor.py:69: in <module>
    from apache_beam.runners.worker import operations
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 

>   ???
E   KeyError: '__pyx_vtable__'

apache_beam/runners/worker/operations.py:1: KeyError

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

52 skipped tests found

There are 52 skipped tests, see "Raw output" for the full list of skipped tests.
Raw output
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_io_roundtrip
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.base_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test

Check notice on line 0 in .github

See this annotation in the file changed.

@github-actions github-actions / Test Results

151 tests found

There are 151 tests, see "Raw output" for the full list of tests.
Raw output
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_direct_read_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_export_with_project
apache_beam.dataframe.io_it_test.ReadUsingReadGbqTests ‑ test_ReadGbq_with_computation
apache_beam.examples.complete.autocomplete_test.AutocompleteTest ‑ test_autocomplete_it
apache_beam.examples.complete.game.game_stats_it_test.GameStatsIT ‑ test_game_stats_it
apache_beam.examples.complete.game.hourly_team_score_it_test.HourlyTeamScoreIT ‑ test_hourly_team_score_it
apache_beam.examples.complete.game.leader_board_it_test.LeaderBoardIT ‑ test_leader_board_it
apache_beam.examples.complete.game.user_score_it_test.UserScoreIT ‑ test_user_score_it
apache_beam.examples.complete.juliaset.juliaset.juliaset_test_it.JuliaSetTestIT ‑ test_run_example_with_setup_file
apache_beam.examples.cookbook.bigquery_tornadoes_it_test.BigqueryTornadoesIT ‑ test_bigquery_tornadoes_it
apache_beam.examples.cookbook.bigtableio_it_test.BigtableIOWriteTest ‑ test_bigtable_write
apache_beam.examples.cookbook.datastore_wordcount_it_test.DatastoreWordCountIT ‑ test_datastore_wordcount_it
apache_beam.examples.dataframe.flight_delays_it_test.FlightDelaysTest ‑ test_flight_delays
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_aggregation
apache_beam.examples.dataframe.taxiride_it_test.TaxirideIT ‑ test_enrich
apache_beam.examples.fastavro_it_test.FastavroIT ‑ test_avro_it
apache_beam.examples.inference.tfx_bsl.tfx_bsl_inference_it_test.TFXRunInferenceTests ‑ test_tfx_run_inference_mobilenetv2
apache_beam.examples.ml_transform.ml_transform_it_test
apache_beam.examples.snippets.transforms.elementwise.mltransform_test
apache_beam.examples.snippets.transforms.elementwise.runinference_test
apache_beam.examples.streaming_wordcount_debugging_it_test.StreamingWordcountDebuggingIT ‑ test_streaming_wordcount_debugging_it
apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT ‑ test_streaming_wordcount_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_impersonation_it
apache_beam.examples.wordcount_it_test.WordCountIT ‑ test_wordcount_it
apache_beam.io.external.xlang_debeziumio_it_test.CrossLanguageDebeziumIOTest ‑ test_xlang_debezium_read
apache_beam.io.external.xlang_jdbcio_it_test.CrossLanguageJdbcIOTest ‑ test_xlang_jdbc_write_read_0_postgres
apache_beam.io.external.xlang_jdbcio_it_test.CrossLanguageJdbcIOTest ‑ test_xlang_jdbc_write_read_1_mysql
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_hosted_kafkaio_populated_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_local_kafkaio_null_key
apache_beam.io.external.xlang_kafkaio_it_test.CrossLanguageKafkaIOTest ‑ test_local_kafkaio_populated_key
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_io_roundtrip
apache_beam.io.external.xlang_kinesisio_it_test.CrossLanguageKinesisIOTest ‑ test_kinesis_write
apache_beam.io.fileio_test.MatchIntegrationTest ‑ test_transform_on_gcs
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_legacy_sql
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_new_types_avro
apache_beam.io.gcp.big_query_query_to_table_it_test.BigQueryQueryToTableIT ‑ test_big_query_standard_sql
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_copy_jobs
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_bqfl_streaming_with_dynamic_destinations
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_file_loads_test.BigQueryFileLoadsIT ‑ test_one_job_fails_all_jobs_fail
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_1M_python
apache_beam.io.gcp.bigquery_io_read_it_test.BigqueryIOReadIT ‑ test_bigquery_read_custom_1M_python
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_direct_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_export_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_file_loads_write
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_query_read
apache_beam.io.gcp.bigquery_json_it_test.BigQueryJsonIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_read_it_test.ReadAllBQTests ‑ test_read_queries
apache_beam.io.gcp.bigquery_read_it_test.ReadInteractiveRunnerTests ‑ test_read_in_interactive_runner
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadNewTypesTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_native_source
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_specifying_only_table
apache_beam.io.gcp.bigquery_read_it_test.ReadTests ‑ test_table_schema_retrieve_with_direct_read
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_column_selection_and_row_restriction_rows
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_native_datetime
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_query_and_filters
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_row_restriction
apache_beam.io.gcp.bigquery_read_it_test.ReadUsingStorageApiTests ‑ test_iobase_source_with_very_selective_filters
apache_beam.io.gcp.bigquery_test.BigQueryFileLoadsIntegrationTests ‑ test_avro_file_load
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_multiple_destinations_transform
apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests ‑ test_value_provider_transform
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_file_loads
apache_beam.io.gcp.bigquery_test.PubSubBigQueryIT ‑ test_streaming_inserts
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_errors_reporting
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_insert_non_transient_api_call_error
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_new_types
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_schema_autodetect
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_0
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_1
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_temp_table_append_schema_update_2
apache_beam.io.gcp.bigquery_write_it_test.BigQueryWriteIntegrationTests ‑ test_big_query_write_without_schema
apache_beam.io.gcp.datastore.v1new.datastore_write_it_test.DatastoreWriteIT ‑ test_datastore_write_limit
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename
apache_beam.io.gcp.gcsfilesystem_integration_test.GcsFileSystemIntegrationTest ‑ test_rename_error
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch_kms
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_batch_rewrite_token
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_kms
apache_beam.io.gcp.gcsio_integration_test.GcsIOIntegrationTest ‑ test_copy_rewrite_token
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_search_instances
apache_beam.io.gcp.healthcare.dicomio_integration_test.DICOMIoIntegrationTest ‑ test_dicom_store_instance_from_gcs
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_data_only
apache_beam.io.gcp.pubsub_integration_test.PubSubIntegrationTest ‑ test_streaming_with_attributes
apache_beam.io.parquetio_it_test.TestParquetIT ‑ test_parquetio_it
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_deidentification
apache_beam.ml.gcp.cloud_dlp_it_test.CloudDLPIT ‑ test_inspection
apache_beam.ml.gcp.naturallanguageml_test_it.NaturalLanguageMlTestIT ‑ test_analyzing_syntax
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_catalog_item
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_create_user_event
apache_beam.ml.gcp.recommendations_ai_test_it.RecommendationAIIT ‑ test_predict
apache_beam.ml.gcp.videointelligenceml_test_it.VideoIntelligenceMlTestIT ‑ test_label_detection_with_video_context
apache_beam.ml.gcp.visionml_test_it.VisionMlTestIT ‑ test_text_detection_with_language_hint
apache_beam.ml.inference.base_test.RunInferenceBaseTest ‑ test_run_inference_with_side_inputin_streaming
apache_beam.ml.inference.huggingface_inference_it_test
apache_beam.ml.inference.huggingface_inference_test
apache_beam.ml.inference.onnx_inference_it_test.OnnxInference ‑ test_onnx_run_inference_roberta_sentiment_classification
apache_beam.ml.inference.onnx_inference_test
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_bert_for_masked_lm_large_model
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_coco_maskrcnn_resnet50_fpn_v1_and_v2
apache_beam.ml.inference.pytorch_inference_it_test.PyTorchInference ‑ test_torch_run_inference_imagenet_mobilenetv2
apache_beam.ml.inference.pytorch_inference_test
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_mnist_classification_large_model
apache_beam.ml.inference.sklearn_inference_it_test.SklearnInference ‑ test_sklearn_regression
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_imagenet_image_segmentation
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_classification_large_model
apache_beam.ml.inference.tensorflow_inference_it_test.TensorflowInference ‑ test_tf_mnist_with_weights_classification
apache_beam.ml.inference.tensorrt_inference_test
apache_beam.ml.inference.vertex_ai_inference_it_test
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_datatable_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_numpy_single_batch_large_model
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_pandas_single_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_multi_batch
apache_beam.ml.inference.xgboost_inference_it_test.XGBoostInference ‑ test_iris_classification_scipy_single_batch
apache_beam.ml.inference.xgboost_inference_test
apache_beam.ml.transforms.base_test
apache_beam.ml.transforms.handlers_test
apache_beam.ml.transforms.tft_test
apache_beam.runners.dask.dask_runner_test
apache_beam.runners.dataflow.dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest ‑ test_metrics_it
apache_beam.testing.analyzers.perf_analysis_test
apache_beam.testing.benchmarks.cloudml.cloudml_benchmark_test
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_basic_execution
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs
apache_beam.testing.test_stream_it_test.TestStreamIntegrationTests ‑ test_multiple_outputs_with_watermark_advancement
apache_beam.transforms.external_it_test.ExternalTransformIT ‑ test_job_python_from_python_it
apache_beam.transforms.periodicsequence_it_test.PeriodicSequenceIT ‑ test_periodicsequence_outputs_valid_watermarks_it
apache_beam.typehints.pytorch_type_compatibility_test