diff --git a/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy b/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy index 6bf6c8f7e738a4..609abd37329146 100644 --- a/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy +++ b/regression-test/suites/load_p0/tvf/test_s3_tvf.groovy @@ -56,7 +56,7 @@ suite("test_s3_tvf", "load_p0") { /* normal */ for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", "K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", "K15", "K16", "K17", "K18"], "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -64,14 +64,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { - attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", "K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", "K15", "K16", "K17"], "", "").addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") + attributeList.add(new TvfAttribute(table, ["K00", "K01", "K02", "K03", "K04", "K05", "K06", "K07", "K08", "K09", "K10", "K11", "K12", "K13", "K14", "K15", "K16", "K17"], "", "").addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -79,7 +79,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -87,14 +87,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -102,27 +102,27 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|")) } attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|")) } // for(String table : basicTables) { // attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") -// .addProperty("uri", "https://{$s3Endpoint}/{$s3BucketName}/regression/load/data/basic_data.csv") +// .addProperty("uri", "https://${s3Endpoint}/${s3BucketName}/regression/load/data/basic_data.csv") // .addProperty("format", "csv") // .addProperty("column_separator", "|") // .addProperty("use_path_style", "true")) @@ -130,14 +130,14 @@ suite("test_s3_tvf", "load_p0") { // // attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", // "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") -// .addProperty("uri", "https://{$s3Endpoint}/{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") +// .addProperty("uri", "https://${s3Endpoint}/${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") // .addProperty("format", "csv") // .addProperty("column_separator", "|") // .addProperty("use_path_style", "true")) // // for(String table : arrayTables) { // attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") -// .addProperty("uri", "https://{$s3Endpoint}/{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") +// .addProperty("uri", "https://${s3Endpoint}/${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") // .addProperty("format", "csv") // .addProperty("column_separator", "|") // .addProperty("use_path_style", "true")) @@ -146,7 +146,7 @@ suite("test_s3_tvf", "load_p0") { /* error */ for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "", true) - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -154,14 +154,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "", true) - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "kd01", "kd02", "kd03", "kd04", "kd05", "kd06", "kd07", "kd08", "kd09", "kd10", "kd11", "kd12", "kd13", "kd14", "kd15", "kd16"], "", "", true) - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -170,7 +170,7 @@ suite("test_s3_tvf", "load_p0") { /* skip lines */ for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("skip_lines", "10") @@ -179,7 +179,7 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("skip_lines", "10") @@ -187,7 +187,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_with_errors.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("skip_lines", "10") @@ -197,21 +197,21 @@ suite("test_s3_tvf", "load_p0") { /* compress type */ for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.gz") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.gz") .addProperty("format", "csv") .addProperty("column_separator", "|").addProperty("compress_type", "GZ")) } attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.gz") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.gz") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "GZ")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.gz") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.gz") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "GZ")) @@ -219,21 +219,21 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.bz2") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.bz2") .addProperty("format", "csv") .addProperty("column_separator", "|").addProperty("compress_type", "BZ2")) } attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.bz2") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.bz2") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "BZ2")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.bz2") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.bz2") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "BZ2")) @@ -241,21 +241,21 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.lz4") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.lz4") .addProperty("format", "csv") .addProperty("column_separator", "|").addProperty("compress_type", "LZ4FRAME")) } attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv.lz4") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv.lz4") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "LZ4FRAME")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "", "") - .addProperty("uri", "https://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv.lz4") + .addProperty("uri", "https://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv.lz4") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("compress_type", "LZ4FRAME")) @@ -263,7 +263,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "WHERE c1 > 50", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -271,14 +271,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "c1 as k00,c2 as k01,c3 as k02,c4 as k03,c5 as k04,c6 as k05,c7 as k06,c8 as k07,c9 as k08,c10 as k09,c11 as k10,c12 as k11,c13 as k12,c14 as k13,c15 as k14,c16 as k15,c17 as k16,c18 as k17,c19 as k18, to_bitmap(c6) as k19, HLL_HASH(c6) as k20, TO_QUANTILE_STATE(c5, 1.0) as k21, to_bitmap(c6) as kd19, HLL_HASH(c6) as kd20, TO_QUANTILE_STATE(c5, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"WHERE c1 > 50", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17"], "WHERE c1 > 50", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -286,7 +286,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : uniqTable) { attributeList.add(new TvfAttribute(table, ["k00", "k01", "k02", "k03", "k04", "k05", "k06", "k07", "k08", "k09", "k10", "k11", "k12", "k13", "k14", "k15", "k16", "k17", "k18"], "", "ORDER BY c1") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.csv") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.csv") .addProperty("format", "csv") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -294,7 +294,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.parq") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.parq") .addProperty("format", "parquet") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -302,14 +302,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.parq") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.parq") .addProperty("format", "parquet") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.parq") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.parq") .addProperty("format", "parquet") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -317,7 +317,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.orc") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.orc") .addProperty("format", "orc") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -325,14 +325,14 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.orc") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.orc") .addProperty("format", "orc") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.orc") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.orc") .addProperty("format", "orc") .addProperty("column_separator", "|") .addProperty("force_parsing_by_standard_uri", "true")) @@ -340,7 +340,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.json") .addProperty("format", "json") .addProperty("read_json_by_line", "false") .addProperty("strip_outer_array", "true") @@ -350,7 +350,7 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data.json") .addProperty("format", "json") .addProperty("read_json_by_line", "false") .addProperty("strip_outer_array", "true") @@ -359,7 +359,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data.json") .addProperty("format", "json") .addProperty("read_json_by_line", "false") .addProperty("strip_outer_array", "true") @@ -369,7 +369,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : basicTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18","k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_by_line.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_by_line.json") .addProperty("format", "json") .addProperty("read_json_by_line", "true") .addProperty("strip_outer_array", "false") @@ -379,7 +379,7 @@ suite("test_s3_tvf", "load_p0") { attributeList.add(new TvfAttribute("agg_tbl_basic_tvf", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18, to_bitmap(k05) as k19, HLL_HASH(k05) as k20, TO_QUANTILE_STATE(k04, 1.0) as k21, to_bitmap(k05) as kd19, HLL_HASH(k05) as kd20, TO_QUANTILE_STATE(k04, 1.0) as kd21", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17,k18,k19,k20,k21,kd19,kd20,kd21" ,"", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_data_by_line.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_data_by_line.json") .addProperty("format", "json") .addProperty("read_json_by_line", "true") .addProperty("strip_outer_array", "false") @@ -388,7 +388,7 @@ suite("test_s3_tvf", "load_p0") { for(String table : arrayTables) { attributeList.add(new TvfAttribute(table, "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "k00,k01,k02,k03,k04,k05,k06,k07,k08,k09,k10,k11,k12,k13,k14,k15,k16,k17", "", "") - .addProperty("uri", "s3://{$s3BucketName}.{$s3Endpoint}/regression/load/data/basic_array_data_by_line.json") + .addProperty("uri", "s3://${s3BucketName}.${s3Endpoint}/regression/load/data/basic_array_data_by_line.json") .addProperty("format", "json") .addProperty("read_json_by_line", "true") .addProperty("strip_outer_array", "false")