Uploaded image for project: 'IMPALA'
  1. IMPALA
  2. IMPALA-12740

TestHdfsJsonScanNodeErrors fails in exhaustive mode

    XMLWordPrintableJSON

Details

    • Bug
    • Status: Resolved
    • Blocker
    • Resolution: Fixed
    • Impala 4.4.0
    • Impala 4.4.0
    • Backend
    • ghx-label-10

    Description

      data_errors.test_data_errors.TestHdfsJsonScanNodeErrors fails when test are run in exhaustive more. Both debug and release builds exhibit the same symptoms.

      the run log:

      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/snap/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/snap/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/snap/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/snap/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/def/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/def/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/def/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/def/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/bzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/bzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/gzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/gzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/bzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 1, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': False, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/gzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/bzip/block]
      FAIL data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol: beeswax | exec_option: {'test_replan': 1, 'batch_size': 0, 'num_nodes': 0, 'disable_codegen_rows_threshold': 0, 'disable_codegen': True, 'abort_on_error': 1, 'exec_single_node_rows_threshold': 0} | table_format: json/gzip/block]
      

      stderr complains about conversion failures:

      SET client_identifier=data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol:beeswax|exec_option:{'test_replan':1;'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':True;'abort_on_error':1;'exec_;
      -- executing against localhost:21000
      
      use functional_json_snap;
      
      -- 2024-01-19 23:30:10,326 INFO     MainThread: Started query 824002d0b11e7ba1:2e75861b00000000
      SET client_identifier=data_errors/test_data_errors.py::TestHdfsJsonScanNodeErrors::()::test_hdfs_json_scan_node_errors[protocol:beeswax|exec_option:{'test_replan':1;'batch_size':0;'num_nodes':0;'disable_codegen_rows_threshold':0;'disable_codegen':True;'abort_on_error':1;'exec_;
      SET test_replan=1;
      SET batch_size=0;
      SET num_nodes=0;
      SET disable_codegen_rows_threshold=0;
      SET disable_codegen=True;
      SET abort_on_error=0;
      SET exec_single_node_rows_threshold=0;
      -- 2024-01-19 23:30:10,327 INFO     MainThread: Loading query test file: /data/jenkins/workspace/impala-asf-master-exhaustive-release/repos/Impala/testdata/workloads/functional-query/queries/DataErrorsTest/hdfs-json-scan-node-errors.test
      -- executing against localhost:21000
      
      select * from alltypeserror order by id;
      
      -- 2024-01-19 23:30:14,212 INFO     MainThread: Started query d54d7edb55f53beb:dbbb778a00000000
      -- 2024-01-19 23:30:14,295 ERROR    MainThread: Comparing QueryTestResults (expected vs actual):
      'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err300'' != 'Error converting column: functional_json_snap.alltypeserror.bigint_col, type: BIGINT, data: 'err300''
      'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err50'' != 'Error converting column: functional_json_snap.alltypeserror.bigint_col, type: BIGINT, data: 'err50''
      'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err90'' != 'Error converting column: functional_json_snap.alltypeserror.bigint_col, type: BIGINT, data: 'err90''
      'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 'errfalse'' != 'Error converting column: functional_json_snap.alltypeserror.bool_col, type: BOOLEAN, data: 'errfalse''
      'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 'errtrue'' != 'Error converting column: functional_json_snap.alltypeserror.bool_col, type: BOOLEAN, data: 'errtrue''
      'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 't\rue'' != 'Error converting column: functional_json_snap.alltypeserror.bool_col, type: BOOLEAN, data: 't\rue''
      'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err300.900000'' != 'Error converting column: functional_json_snap.alltypeserror.double_col, type: DOUBLE, data: 'err300.900000''
      'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err70.700000'' != 'Error converting column: functional_json_snap.alltypeserror.double_col, type: DOUBLE, data: 'err70.700000''
      'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err90.900000'' != 'Error converting column: functional_json_snap.alltypeserror.double_col, type: DOUBLE, data: 'err90.900000''
      'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'xyz30.300000'' != 'Error converting column: functional_json_snap.alltypeserror.double_col, type: DOUBLE, data: 'xyz30.300000''
      'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'xyz70.700000'' != 'Error converting column: functional_json_snap.alltypeserror.double_col, type: DOUBLE, data: 'xyz70.700000''
      'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err30..000000'' != 'Error converting column: functional_json_snap.alltypeserror.float_col, type: FLOAT, data: 'err30..000000''
      'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err6.000000'' != 'Error converting column: functional_json_snap.alltypeserror.float_col, type: FLOAT, data: 'err6.000000''
      'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err9.000000'' != 'Error converting column: functional_json_snap.alltypeserror.float_col, type: FLOAT, data: 'err9.000000''
      'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'xyz3.000000'' != 'Error converting column: functional_json_snap.alltypeserror.float_col, type: FLOAT, data: 'xyz3.000000''
      'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'abc5'' != 'Error converting column: functional_json_snap.alltypeserror.int_col, type: INT, data: 'abc5''
      'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'abc9'' != 'Error converting column: functional_json_snap.alltypeserror.int_col, type: INT, data: 'abc9''
      'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err30'' != 'Error converting column: functional_json_snap.alltypeserror.int_col, type: INT, data: 'err30''
      'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err4'' != 'Error converting column: functional_json_snap.alltypeserror.int_col, type: INT, data: 'err4''
      'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err9'' != 'Error converting column: functional_json_snap.alltypeserror.int_col, type: INT, data: 'err9''
      'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'abc3'' != 'Error converting column: functional_json_snap.alltypeserror.smallint_col, type: SMALLINT, data: 'abc3''
      'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err3'' != 'Error converting column: functional_json_snap.alltypeserror.smallint_col, type: SMALLINT, data: 'err3''
      'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err30'' != 'Error converting column: functional_json_snap.alltypeserror.smallint_col, type: SMALLINT, data: 'err30''
      'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err9'' != 'Error converting column: functional_json_snap.alltypeserror.smallint_col, type: SMALLINT, data: 'err9''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0009-01-01 00:00:00'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0009-01-01 00:00:00''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '1999-10-10 90:10:10'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '1999-10-10 90:10:10''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2002-14-10 00:00:00'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2002-14-10 00:00:00''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 10:70:10.123'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 10:70:10.123''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 60:10:10.123'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 60:10:10.123''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-40 10:10:10.123'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-40 10:10:10.123''
      'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-20-10 10:10:10.123'' != 'Error converting column: functional_json_snap.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-20-10 10:10:10.123''
      'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'abc7'' != 'Error converting column: functional_json_snap.alltypeserror.tinyint_col, type: TINYINT, data: 'abc7''
      'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err2'' != 'Error converting column: functional_json_snap.alltypeserror.tinyint_col, type: TINYINT, data: 'err2''
      'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err30'' != 'Error converting column: functional_json_snap.alltypeserror.tinyint_col, type: TINYINT, data: 'err30''
      'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err9'' != 'Error converting column: functional_json_snap.alltypeserror.tinyint_col, type: TINYINT, data: 'err9''
      'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'xyz5'' != 'Error converting column: functional_json_snap.alltypeserror.tinyint_col, type: TINYINT, data: 'xyz5''
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=1/000000_0.snappy, before offset: 648'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=2/000000_0.snappy, before offset: 572'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=2/000000_0.snappy, before offset: 572'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=2/000000_0.snappy, before offset: 572'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=2/000000_0.snappy, before offset: 572'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_snap/year=2009/month=3/000000_0.snappy, before offset: 725'
      

      Similar failures are reported for all failing test cases. Here is the example for json/gzip/block:

      data_errors/test_data_errors.py:167: in test_hdfs_json_scan_node_errors
          self.run_test_case('DataErrorsTest/hdfs-json-scan-node-errors', vector)
      common/impala_test_suite.py:756: in run_test_case
          self.__verify_results_and_errors(vector, test_section, result, use_db)
      common/impala_test_suite.py:589: in __verify_results_and_errors
          replace_filenames_with_placeholder)
      common/test_result_verifier.py:396: in verify_raw_results
          verify_errors(expected_errors, actual_errors)
      common/test_result_verifier.py:339: in verify_errors
          VERIFIER_MAP['VERIFY_IS_EQUAL'](expected, actual)
      common/test_result_verifier.py:296: in verify_query_result_is_equal
          assert expected_results == actual_results
      E   assert Comparing QueryTestResults (expected vs actual):
      E     'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err300'' != 'Error converting column: functional_json_gzip.alltypeserror.bigint_col, type: BIGINT, data: 'err300''
      E     'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err50'' != 'Error converting column: functional_json_gzip.alltypeserror.bigint_col, type: BIGINT, data: 'err50''
      E     'Error converting column: functional_json.alltypeserror.bigint_col, type: BIGINT, data: 'err90'' != 'Error converting column: functional_json_gzip.alltypeserror.bigint_col, type: BIGINT, data: 'err90''
      E     'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 'errfalse'' != 'Error converting column: functional_json_gzip.alltypeserror.bool_col, type: BOOLEAN, data: 'errfalse''
      E     'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 'errtrue'' != 'Error converting column: functional_json_gzip.alltypeserror.bool_col, type: BOOLEAN, data: 'errtrue''
      E     'Error converting column: functional_json.alltypeserror.bool_col, type: BOOLEAN, data: 't\rue'' != 'Error converting column: functional_json_gzip.alltypeserror.bool_col, type: BOOLEAN, data: 't\rue''
      E     'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err300.900000'' != 'Error converting column: functional_json_gzip.alltypeserror.double_col, type: DOUBLE, data: 'err300.900000''
      E     'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err70.700000'' != 'Error converting column: functional_json_gzip.alltypeserror.double_col, type: DOUBLE, data: 'err70.700000''
      E     'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'err90.900000'' != 'Error converting column: functional_json_gzip.alltypeserror.double_col, type: DOUBLE, data: 'err90.900000''
      E     'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'xyz30.300000'' != 'Error converting column: functional_json_gzip.alltypeserror.double_col, type: DOUBLE, data: 'xyz30.300000''
      E     'Error converting column: functional_json.alltypeserror.double_col, type: DOUBLE, data: 'xyz70.700000'' != 'Error converting column: functional_json_gzip.alltypeserror.double_col, type: DOUBLE, data: 'xyz70.700000''
      E     'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err30..000000'' != 'Error converting column: functional_json_gzip.alltypeserror.float_col, type: FLOAT, data: 'err30..000000''
      E     'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err6.000000'' != 'Error converting column: functional_json_gzip.alltypeserror.float_col, type: FLOAT, data: 'err6.000000''
      E     'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'err9.000000'' != 'Error converting column: functional_json_gzip.alltypeserror.float_col, type: FLOAT, data: 'err9.000000''
      E     'Error converting column: functional_json.alltypeserror.float_col, type: FLOAT, data: 'xyz3.000000'' != 'Error converting column: functional_json_gzip.alltypeserror.float_col, type: FLOAT, data: 'xyz3.000000''
      E     'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'abc5'' != 'Error converting column: functional_json_gzip.alltypeserror.int_col, type: INT, data: 'abc5''
      E     'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'abc9'' != 'Error converting column: functional_json_gzip.alltypeserror.int_col, type: INT, data: 'abc9''
      E     'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err30'' != 'Error converting column: functional_json_gzip.alltypeserror.int_col, type: INT, data: 'err30''
      E     'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err4'' != 'Error converting column: functional_json_gzip.alltypeserror.int_col, type: INT, data: 'err4''
      E     'Error converting column: functional_json.alltypeserror.int_col, type: INT, data: 'err9'' != 'Error converting column: functional_json_gzip.alltypeserror.int_col, type: INT, data: 'err9''
      E     'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'abc3'' != 'Error converting column: functional_json_gzip.alltypeserror.smallint_col, type: SMALLINT, data: 'abc3''
      E     'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err3'' != 'Error converting column: functional_json_gzip.alltypeserror.smallint_col, type: SMALLINT, data: 'err3''
      E     'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err30'' != 'Error converting column: functional_json_gzip.alltypeserror.smallint_col, type: SMALLINT, data: 'err30''
      E     'Error converting column: functional_json.alltypeserror.smallint_col, type: SMALLINT, data: 'err9'' != 'Error converting column: functional_json_gzip.alltypeserror.smallint_col, type: SMALLINT, data: 'err9''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0000-01-01 00:00:00''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0009-01-01 00:00:00'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '0009-01-01 00:00:00''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '1999-10-10 90:10:10'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '1999-10-10 90:10:10''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2002-14-10 00:00:00'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2002-14-10 00:00:00''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 10:70:10.123'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 10:70:10.123''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 60:10:10.123'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-10 60:10:10.123''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-40 10:10:10.123'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-10-40 10:10:10.123''
      E     'Error converting column: functional_json.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-20-10 10:10:10.123'' != 'Error converting column: functional_json_gzip.alltypeserror.timestamp_col, type: TIMESTAMP, data: '2020-20-10 10:10:10.123''
      E     'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'abc7'' != 'Error converting column: functional_json_gzip.alltypeserror.tinyint_col, type: TINYINT, data: 'abc7''
      E     'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err2'' != 'Error converting column: functional_json_gzip.alltypeserror.tinyint_col, type: TINYINT, data: 'err2''
      E     'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err30'' != 'Error converting column: functional_json_gzip.alltypeserror.tinyint_col, type: TINYINT, data: 'err30''
      E     'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'err9'' != 'Error converting column: functional_json_gzip.alltypeserror.tinyint_col, type: TINYINT, data: 'err9''
      E     'Error converting column: functional_json.alltypeserror.tinyint_col, type: TINYINT, data: 'xyz5'' != 'Error converting column: functional_json_gzip.alltypeserror.tinyint_col, type: TINYINT, data: 'xyz5''
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=1/000000_0.gz, before offset: 393'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=2/000000_0.gz, before offset: 345'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=2/000000_0.gz, before offset: 345'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=2/000000_0.gz, before offset: 345'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=2/000000_0.gz, before offset: 345'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      E     row_regex: .*Error parsing row: file: hdfs://localhost:20500/.* before offset: \d+ == 'Error parsing row: file: hdfs://localhost:20500/test-warehouse/alltypeserror_json_gzip/year=2009/month=3/000000_0.gz, before offset: 434'
      

      Eyizoha, you have been active in this area recently; could you perhaps take a look at the failure? Thanks a lot!

      Attachments

        Activity

          People

            eyizoha Zihao Ye
            laszlog Laszlo Gaal
            Votes:
            0 Vote for this issue
            Watchers:
            4 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: