@@ -880,38 +880,38 @@ def con():
880880
881881@fixture
882882def populate_regions (default_new_input , con ):
883- from muse .new_input .readers import read_regions_csv
883+ from muse .new_input .readers import Regions , read_csv
884884
885885 with open (default_new_input / "regions.csv" ) as f :
886- return read_regions_csv ( f , con )
886+ return read_csv ( f , Regions , con )
887887
888888
889889@fixture
890890def populate_commodities (default_new_input , con ):
891- from muse .new_input .readers import read_commodities_csv
891+ from muse .new_input .readers import Commodities , read_csv
892892
893893 with open (default_new_input / "commodities.csv" ) as f :
894- return read_commodities_csv ( f , con )
894+ return read_csv ( f , Commodities , con )
895895
896896
897897@fixture
898898def populate_demand (default_new_input , con , populate_regions , populate_commodities ):
899- from muse .new_input .readers import read_demand_csv
899+ from muse .new_input .readers import Demand , read_csv
900900
901901 with open (default_new_input / "demand.csv" ) as f :
902- return read_demand_csv ( f , con )
902+ return read_csv ( f , Demand , con )
903903
904904
905905def test_read_regions (populate_regions ):
906906 assert populate_regions ["name" ] == np .array (["R1" ])
907907
908908
909909def test_read_regions_primary_key_constraint (default_new_input , con ):
910- from muse .new_input .readers import read_regions_csv
910+ from muse .new_input .readers import Regions , read_csv
911911
912912 csv = StringIO ("name\n R1\n R1\n " )
913913 with raises (duckdb .ConstraintException , match = ".*duplicate key.*" ):
914- read_regions_csv (csv , con )
914+ read_csv (csv , Regions , con )
915915
916916
917917def test_read_new_commodities (populate_commodities ):
@@ -937,19 +937,19 @@ def test_calculate_global_commodities(populate_commodities):
937937
938938
939939def test_read_new_commodities_primary_key_constraint (default_new_input , con ):
940- from muse .new_input .readers import read_commodities_csv
940+ from muse .new_input .readers import Commodities , read_csv
941941
942942 csv = StringIO ("name,type,unit\n foo,energy,bar\n foo,energy,bar\n " )
943943 with raises (duckdb .ConstraintException , match = ".*duplicate key.*" ):
944- read_commodities_csv (csv , con )
944+ read_csv (csv , Commodities , con )
945945
946946
947947def test_read_new_commodities_type_constraint (default_new_input , con ):
948- from muse .new_input .readers import read_commodities_csv
948+ from muse .new_input .readers import Commodities , read_csv
949949
950950 csv = StringIO ("name,type,unit\n foo,invalid,bar\n " )
951951 with raises (duckdb .ConstraintException ):
952- read_commodities_csv (csv , con )
952+ read_csv (csv , Commodities , con )
953953
954954
955955def test_new_read_demand_csv (populate_demand ):
@@ -963,31 +963,31 @@ def test_new_read_demand_csv(populate_demand):
963963def test_new_read_demand_csv_commodity_constraint (
964964 default_new_input , con , populate_commodities , populate_regions
965965):
966- from muse .new_input .readers import read_demand_csv
966+ from muse .new_input .readers import Demand , read_csv
967967
968968 csv = StringIO ("year,commodity_name,region,demand\n 2020,invalid,R1,0\n " )
969969 with raises (duckdb .ConstraintException , match = ".*foreign key.*" ):
970- read_demand_csv (csv , con )
970+ read_csv (csv , Demand , con )
971971
972972
973973def test_new_read_demand_csv_region_constraint (
974974 default_new_input , con , populate_commodities , populate_regions
975975):
976- from muse .new_input .readers import read_demand_csv
976+ from muse .new_input .readers import Demand , read_csv
977977
978978 csv = StringIO ("year,commodity_name,region,demand\n 2020,heat,invalid,0\n " )
979979 with raises (duckdb .ConstraintException , match = ".*foreign key.*" ):
980- read_demand_csv (csv , con )
980+ read_csv (csv , Demand , con )
981981
982982
983983def test_new_read_demand_csv_primary_key_constraint (
984984 default_new_input , con , populate_commodities , populate_regions
985985):
986- from muse .new_input .readers import read_demand_csv , read_regions_csv
986+ from muse .new_input .readers import Demand , Regions , read_csv
987987
988988 # Add another region so we can test varying it as a primary key
989989 csv = StringIO ("name\n R2\n " )
990- read_regions_csv (csv , con )
990+ read_csv (csv , Regions , con )
991991
992992 # all fine so long as one primary key column differs
993993 csv = StringIO (
@@ -998,12 +998,12 @@ def test_new_read_demand_csv_primary_key_constraint(
9989982020,gas,R2,0
999999"""
10001000 )
1001- read_demand_csv (csv , con )
1001+ read_csv (csv , Demand , con )
10021002
10031003 # no good if all primary key columns match a previous entry
10041004 csv = StringIO ("year,commodity_name,region,demand\n 2020,gas,R1,0" )
10051005 with raises (duckdb .ConstraintException , match = ".*duplicate key.*" ):
1006- read_demand_csv (csv , con )
1006+ read_csv (csv , Demand , con )
10071007
10081008
10091009@mark .xfail
0 commit comments