|
23 | 23 | import bigframes.session._io.bigquery
|
24 | 24 |
|
25 | 25 |
|
26 |
| -@pytest.mark.parametrize( |
27 |
| - ("query_or_table", "index_col"), |
28 |
| - [ |
29 |
| - pytest.param( |
30 |
| - "bigquery-public-data.patents_view.ipcr_201708", |
31 |
| - (), |
32 |
| - id="1g_table_w_default_index", |
33 |
| - ), |
34 |
| - pytest.param( |
35 |
| - "bigquery-public-data.new_york_taxi_trips.tlc_yellow_trips_2011", |
36 |
| - (), |
37 |
| - id="30g_table_w_default_index", |
38 |
| - ), |
39 |
| - # TODO(chelsealin): Disable the long run tests until we have propertily |
40 |
| - # ordering support to avoid materializating any data. |
41 |
| - # # Adding default index to large tables would take much longer time, |
42 |
| - # # e.g. ~5 mins for a 100G table, ~20 mins for a 1T table. |
43 |
| - # pytest.param( |
44 |
| - # "bigquery-public-data.stackoverflow.post_history", |
45 |
| - # ["id"], |
46 |
| - # id="100g_table_w_unique_column_index", |
47 |
| - # ), |
48 |
| - # pytest.param( |
49 |
| - # "bigquery-public-data.wise_all_sky_data_release.all_wise", |
50 |
| - # ["cntr"], |
51 |
| - # id="1t_table_w_unique_column_index", |
52 |
| - # ), |
53 |
| - ], |
54 |
| -) |
55 |
| -def test_read_gbq_for_large_tables( |
56 |
| - session: bigframes.Session, query_or_table, index_col |
57 |
| -): |
58 |
| - """Verify read_gbq() is able to read large tables.""" |
59 |
| - df = session.read_gbq(query_or_table, index_col=index_col) |
60 |
| - assert len(df.columns) != 0 |
61 |
| - |
62 |
| - |
63 | 26 | def test_close(session: bigframes.Session):
|
64 | 27 | # we will create two tables and confirm that they are deleted
|
65 | 28 | # when the session is closed
|
|
0 commit comments