diff --git a/.github/workflows/test_runner.yml b/.github/workflows/test_runner.yml index 2b4a59543..c9bbd23df 100644 --- a/.github/workflows/test_runner.yml +++ b/.github/workflows/test_runner.yml @@ -19,7 +19,7 @@ jobs: fail-fast: true matrix: python-version: [3.8, 3.9, "3.10", 3.11] - os: [windows-latest, ubuntu-22.04] + os: [windows-latest] # FIXME: temporarily disabled ubuntu-22.04 steps: - uses: actions/checkout@v3 with: diff --git a/execution_tests/import_file_packs/execution_test.py b/execution_tests/import_file_packs/execution_test.py index 4ca856fc2..62ac69601 100644 --- a/execution_tests/import_file_packs/execution_test.py +++ b/execution_tests/import_file_packs/execution_test.py @@ -3,7 +3,7 @@ import shutil import subprocess import unittest -from spinedb_api import DatabaseMapping, from_database +from spinedb_api import create_new_spine_database, DatabaseMapping, from_database class ModifyConnectionFilterByScript(unittest.TestCase): @@ -19,8 +19,7 @@ def setUp(self): if self._database_path.exists(): self._database_path.unlink() self._url = "sqlite:///" + str(self._database_path) - db_map = DatabaseMapping(self._url, create=True) - db_map.connection.close() + create_new_spine_database(self._url) def test_execution(self): completed = subprocess.run( @@ -33,14 +32,13 @@ def test_execution(self): ) ) self.assertEqual(completed.returncode, 0) - db_map = DatabaseMapping(self._url) - values = {} - for value_row in db_map.query(db_map.object_parameter_value_sq): - self.assertEqual(value_row.object_class_name, "a") - self.assertEqual(value_row.parameter_name, "info") - self.assertEqual(value_row.alternative_name, "Base") - values[value_row.object_name] = from_database(value_row.value, value_row.type) - db_map.connection.close() + with DatabaseMapping(self._url) as db_map: + values = {} + for value_row in db_map.query(db_map.entity_parameter_value_sq): + self.assertEqual(value_row.entity_class_name, "a") + self.assertEqual(value_row.parameter_name, "info") + self.assertEqual(value_row.alternative_name, "Base") + values[value_row.entity_name] = from_database(value_row.value, value_row.type) self.assertEqual(len(values), 4) self.assertEqual(values["b"], 23.0) self.assertEqual(values["c"], 50.0) diff --git a/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/project.json b/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/project.json index 34e59cefb..ca870cde9 100644 --- a/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/project.json +++ b/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/project.json @@ -2,6 +2,9 @@ "project": { "version": 11, "description": "", + "settings": { + "enable_execute_all": true + }, "specifications": { "Tool": [ { @@ -93,7 +96,7 @@ ], "condition": { "type": "python-script", - "script": "import sys\nfrom spinedb_api import DatabaseMapping, from_database, import_object_parameter_values\nin_url = sys.argv[1]\nin_db_map = DatabaseMapping(in_url)\nsq = in_db_map.object_parameter_value_sq\nday_count_row = in_db_map.query(sq).filter(sq.c.object_class_name==\"Timeline\", sq.c.object_name==\"days_of_our_lives\", sq.c.parameter_name==\"cumulative_count\").first()\nday_count = from_database(day_count_row.value, day_count_row.type)\ncounter = day_count.values[-1]\nin_db_map.connection.close()\nif counter >= 30:\n exit(1)\nout_url = sys.argv[2]\nout_db_map = DatabaseMapping(out_url)\nimport_object_parameter_values(out_db_map, ((\"Counter\", \"loop_counter\", \"count\", counter),))\nout_db_map.commit_session(\"Increment counter.\")\nout_db_map.connection.close()\nexit(0)" + "script": "import sys\nfrom spinedb_api import DatabaseMapping, from_database, import_parameter_values\nin_url = sys.argv[1]\nwith DatabaseMapping(in_url) as in_db_map:\n\tsq = in_db_map.entity_parameter_value_sq\n\tday_count_row = in_db_map.query(sq).filter(sq.c.entity_class_name==\"Timeline\", sq.c.entity_name==\"days_of_our_lives\", sq.c.parameter_name==\"cumulative_count\").first()\nday_count = from_database(day_count_row.value, day_count_row.type)\ncounter = day_count.values[-1]\nif counter >= 30:\n exit(1)\nout_url = sys.argv[2]\nwith DatabaseMapping(out_url) as out_db_map:\n\timport_parameter_values(out_db_map, ((\"Counter\", \"loop_counter\", \"count\", counter),))\n\tout_db_map.commit_session(\"Increment counter.\")\nexit(0)\n" }, "cmd_line_args": [ { @@ -106,10 +109,7 @@ } ] } - ], - "settings": { - "enable_execute_all": true - } + ] }, "items": { "Write data": { @@ -124,7 +124,9 @@ "type": "resource", "arg": "db_url@Loop counter store" } - ] + ], + "kill_completed_processes": false, + "log_process_output": false }, "Import": { "type": "Importer", @@ -133,7 +135,6 @@ "y": -4.007282712511945, "specification": "Import data", "cancel_on_error": false, - "purge_before_writing": false, "on_conflict": "merge", "file_selection": [ [ @@ -181,7 +182,6 @@ "y": -4.007282712511941, "specification": "Counter data importer", "cancel_on_error": false, - "purge_before_writing": false, "on_conflict": "merge", "file_selection": [ [ diff --git a/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/specifications/Tool/data_writer.json b/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/specifications/Tool/data_writer.json index 4892395ab..0fe2825dc 100644 --- a/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/specifications/Tool/data_writer.json +++ b/execution_tests/loop_condition_with_cmd_line_args/.spinetoolbox/specifications/Tool/data_writer.json @@ -11,6 +11,5 @@ "data.csv" ], "cmdline_args": [], - "execute_in_work": true, "includes_main_path": "../../.." } \ No newline at end of file diff --git a/execution_tests/loop_condition_with_cmd_line_args/execution_test.py b/execution_tests/loop_condition_with_cmd_line_args/execution_test.py index 9583e6812..d13820658 100644 --- a/execution_tests/loop_condition_with_cmd_line_args/execution_test.py +++ b/execution_tests/loop_condition_with_cmd_line_args/execution_test.py @@ -3,7 +3,7 @@ import subprocess import sys import unittest -from spinedb_api import DatabaseMapping, from_database, Map +from spinedb_api import create_new_spine_database, DatabaseMapping, from_database, Map class LoopConditionWithCmdLineArgs(unittest.TestCase): @@ -24,23 +24,20 @@ def setUp(self): database_path.parent.mkdir(parents=True, exist_ok=True) if database_path.exists(): database_path.unlink() - db_map = DatabaseMapping(url, create=True) - db_map.connection.close() + create_new_spine_database(url) def test_execution(self): completed = subprocess.run((sys.executable, "-m", "spinetoolbox", "--execute-only", str(self._root_path))) self.assertEqual(completed.returncode, 0) - db_map = DatabaseMapping(self._loop_counter_database_url) - value_rows = db_map.query(db_map.object_parameter_value_sq).all() - self.assertEqual(len(value_rows), 1) - loop_counter = from_database(value_rows[0].value, value_rows[0].type) - db_map.connection.close() + with DatabaseMapping(self._loop_counter_database_url) as db_map: + value_rows = db_map.query(db_map.parameter_value_sq).all() + self.assertEqual(len(value_rows), 1) + loop_counter = from_database(value_rows[0].value, value_rows[0].type) self.assertEqual(loop_counter, 20.0) - db_map = DatabaseMapping(self._output_database_url) - value_rows = db_map.query(db_map.object_parameter_value_sq).all() - self.assertEqual(len(value_rows), 1) - output_value = from_database(value_rows[0].value, value_rows[0].type) - db_map.connection.close() + with DatabaseMapping(self._output_database_url) as db_map: + value_rows = db_map.query(db_map.parameter_value_sq).all() + self.assertEqual(len(value_rows), 1) + output_value = from_database(value_rows[0].value, value_rows[0].type) expected_x = [f"T{i:03}" for i in range(31)] expected_y = [float(i) for i in range(31)] self.assertEqual(output_value, Map(expected_x, expected_y)) diff --git a/execution_tests/loop_condition_with_cmd_line_args/tool.py b/execution_tests/loop_condition_with_cmd_line_args/tool.py index ef6eafc39..cb70bca3a 100644 --- a/execution_tests/loop_condition_with_cmd_line_args/tool.py +++ b/execution_tests/loop_condition_with_cmd_line_args/tool.py @@ -3,13 +3,14 @@ from spinedb_api import DatabaseMapping, from_database url = sys.argv[1] -db_map = DatabaseMapping(url) -sq = db_map.object_parameter_value_sq -count_row = db_map.query(sq).filter( - sq.c.object_class_name == "Counter", sq.c.object_name == "loop_counter", sq.c.parameter_name == "count" -).first() +with DatabaseMapping(url) as db_map: + sq = db_map.entity_parameter_value_sq + count_row = ( + db_map.query(sq) + .filter(sq.c.entity_class_name == "Counter", sq.c.entity_name == "loop_counter", sq.c.parameter_name == "count") + .first() + ) count = int(from_database(count_row.value, count_row.type)) -db_map.connection.close() data = [[f"T{i:03}", i] for i in range(count, count + 11)] with open("data.csv", "w", newline="") as data_file: diff --git a/execution_tests/merger_write_order/execution_test.py b/execution_tests/merger_write_order/execution_test.py index 6fd7ef267..3ec7c2a4d 100644 --- a/execution_tests/merger_write_order/execution_test.py +++ b/execution_tests/merger_write_order/execution_test.py @@ -3,7 +3,7 @@ import sys import unittest -from spinedb_api import DatabaseMapping, from_database, import_functions +from spinedb_api import create_new_spine_database, DatabaseMapping, from_database, import_functions class MergerWriteOrder(unittest.TestCase): @@ -18,36 +18,30 @@ def setUp(self): database_path.parent.mkdir(parents=True, exist_ok=True) if database_path.exists(): database_path.unlink() - spoon_volumes = { - self._source_database_1_path: 1.0, - self._source_database_2_path: 99.0 - } + spoon_volumes = {self._source_database_1_path: 1.0, self._source_database_2_path: 99.0} for database_path, spoon_volume in spoon_volumes.items(): url = "sqlite:///" + str(database_path) - db_map = DatabaseMapping(url, create=True) - import_functions.import_object_classes(db_map, ("Widget",)) - import_functions.import_objects(db_map, (("Widget", "spoon"),)) - import_functions.import_object_parameters(db_map, (("Widget", "volume"),)) - import_functions.import_object_parameter_values(db_map, (("Widget", "spoon", "volume", spoon_volume, "Base"),)) - db_map.commit_session("Add test data.") - db_map.connection.close() + with DatabaseMapping(url, create=True) as db_map: + import_functions.import_entity_classes(db_map, ("Widget",)) + import_functions.import_entities(db_map, (("Widget", "spoon"),)) + import_functions.import_parameter_definitions(db_map, (("Widget", "volume"),)) + import_functions.import_parameter_values(db_map, (("Widget", "spoon", "volume", spoon_volume, "Base"),)) + db_map.commit_session("Add test data.") self._sink_url = "sqlite:///" + str(self._sink_database_path) - db_map = DatabaseMapping(self._sink_url, create=True) - db_map.connection.close() + create_new_spine_database(self._sink_url) def test_execution(self): this_file = Path(__file__) completed = subprocess.run((sys.executable, "-m", "spinetoolbox", "--execute-only", str(this_file.parent))) self.assertEqual(completed.returncode, 0) - db_map = DatabaseMapping(self._sink_url) - value_rows = db_map.query(db_map.object_parameter_value_sq).all() - self.assertEqual(len(value_rows), 1) - self.assertEqual(value_rows[0].object_class_name, "Widget") - self.assertEqual(value_rows[0].object_name, "spoon") - self.assertEqual(value_rows[0].parameter_name, "volume") - self.assertEqual(value_rows[0].alternative_name, "Base") - self.assertEqual(from_database(value_rows[0].value, value_rows[0].type), 99.0) - db_map.connection.close() + with DatabaseMapping(self._sink_url) as db_map: + value_rows = db_map.query(db_map.entity_parameter_value_sq).all() + self.assertEqual(len(value_rows), 1) + self.assertEqual(value_rows[0].entity_class_name, "Widget") + self.assertEqual(value_rows[0].entity_name, "spoon") + self.assertEqual(value_rows[0].parameter_name, "volume") + self.assertEqual(value_rows[0].alternative_name, "Base") + self.assertEqual(from_database(value_rows[0].value, value_rows[0].type), 99.0) if __name__ == '__main__': diff --git a/execution_tests/modify_connection_filter_by_script/execution_test.py b/execution_tests/modify_connection_filter_by_script/execution_test.py index f30f38028..719f042ce 100644 --- a/execution_tests/modify_connection_filter_by_script/execution_test.py +++ b/execution_tests/modify_connection_filter_by_script/execution_test.py @@ -5,14 +5,14 @@ import subprocess import unittest from spinedb_api import ( - DiffDatabaseMapping, + DatabaseMapping, import_alternatives, + import_entities, + import_entity_classes, + import_parameter_definitions, + import_parameter_values, import_scenario_alternatives, import_scenarios, - import_object_classes, - import_objects, - import_object_parameters, - import_object_parameter_values, ) @@ -29,22 +29,21 @@ def setUp(self): if self._database_path.exists(): self._database_path.unlink() url = "sqlite:///" + str(self._database_path) - db_map = DiffDatabaseMapping(url, create=True) - import_object_classes(db_map, ("object_class",)) - import_objects(db_map, (("object_class", "object"),)) - import_object_parameters(db_map, (("object_class", "parameter"),)) - import_alternatives(db_map, ("alternative",)) - import_object_parameter_values( - db_map, - ( - ("object_class", "object", "parameter", 1.0, "Base"), - ("object_class", "object", "parameter", 2.0, "alternative"), - ), - ) - import_scenarios(db_map, (("scenario", True),)) - import_scenario_alternatives(db_map, (("scenario", "alternative"),)) - db_map.commit_session("Add test data.") - db_map.connection.close() + with DatabaseMapping(url, create=True) as db_map: + import_entity_classes(db_map, ("object_class",)) + import_entities(db_map, (("object_class", "object"),)) + import_parameter_definitions(db_map, (("object_class", "parameter"),)) + import_alternatives(db_map, ("alternative",)) + import_parameter_values( + db_map, + ( + ("object_class", "object", "parameter", 1.0, "Base"), + ("object_class", "object", "parameter", 2.0, "alternative"), + ), + ) + import_scenarios(db_map, (("scenario", True),)) + import_scenario_alternatives(db_map, (("scenario", "alternative"),)) + db_map.commit_session("Add test data.") def test_execution(self): completed = subprocess.run( diff --git a/execution_tests/modify_connection_filter_by_script/mod.py b/execution_tests/modify_connection_filter_by_script/mod.py index e154eccc3..7ddf6bfc5 100644 --- a/execution_tests/modify_connection_filter_by_script/mod.py +++ b/execution_tests/modify_connection_filter_by_script/mod.py @@ -3,10 +3,7 @@ db_path = project.project_dir / ".spinetoolbox" / "items" / "data" / "Data.sqlite" db_url = "sqlite:///" + str(db_path) -db_map = DatabaseMapping(db_url) -try: +with DatabaseMapping(db_url) as db_map: scenario_ids = {r.name: r.id for r in db_map.query(db_map.scenario_sq).all()} connection = project.find_connection("Data", "Export values") connection.set_filter_enabled("db_url@Data", SCENARIO_FILTER_TYPE, "scenario", True) -finally: - db_map.connection.close() diff --git a/execution_tests/parallel_importer/.spinetoolbox/specifications/Tool/test_tool.json b/execution_tests/parallel_importer/.spinetoolbox/specifications/Tool/test_tool.json index 2ad2338db..e5bb667c5 100644 --- a/execution_tests/parallel_importer/.spinetoolbox/specifications/Tool/test_tool.json +++ b/execution_tests/parallel_importer/.spinetoolbox/specifications/Tool/test_tool.json @@ -11,6 +11,5 @@ "out.csv" ], "cmdline_args": [], - "execute_in_work": true, "includes_main_path": "../../.." } \ No newline at end of file diff --git a/execution_tests/parallel_importer/execution_test.py b/execution_tests/parallel_importer/execution_test.py index 0b2764d9a..cae49d27b 100644 --- a/execution_tests/parallel_importer/execution_test.py +++ b/execution_tests/parallel_importer/execution_test.py @@ -4,14 +4,15 @@ import subprocess import unittest from spinedb_api import ( + create_new_spine_database, DatabaseMapping, from_database, import_alternatives, + import_entities, + import_entity_classes, + import_parameter_definitions, + import_parameter_values, import_scenario_alternatives, - import_object_classes, - import_object_parameters, - import_object_parameter_values, - import_objects, import_scenarios, ) @@ -29,42 +30,44 @@ def setUp(self): if self._source_database_path.exists(): self._source_database_path.unlink() url = "sqlite:///" + str(self._source_database_path) - db_map = DatabaseMapping(url, create=True) - import_alternatives(db_map, ("alternative_1", "alternative_2")) - import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) - import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) - import_object_classes(db_map, ("content",)) - import_objects(db_map, (("content", "test_data"),)) - import_object_parameters(db_map, (("content", "only_value"),)) - import_object_parameter_values(db_map, (("content", "test_data", "only_value", 11.0, "alternative_1"),)) - import_object_parameter_values(db_map, (("content", "test_data", "only_value", 22.0, "alternative_2"),)) - db_map.commit_session("Add test data.") - db_map.connection.close() + with DatabaseMapping(url, create=True) as db_map: + import_alternatives(db_map, ("alternative_1", "alternative_2")) + import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) + import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) + import_entity_classes(db_map, ("content",)) + import_entities(db_map, (("content", "test_data"),)) + import_parameter_definitions(db_map, (("content", "only_value"),)) + import_parameter_values( + db_map, + ( + ("content", "test_data", "only_value", 11.0, "alternative_1"), + ("content", "test_data", "only_value", 22.0, "alternative_2"), + ), + ) + db_map.commit_session("Add test data.") self._sink_database_path.parent.mkdir(parents=True, exist_ok=True) if self._sink_database_path.exists(): self._sink_database_path.unlink() self._sink_url = "sqlite:///" + str(self._sink_database_path) - db_map = DatabaseMapping(self._sink_url, create=True) - db_map.connection.close() + create_new_spine_database(self._sink_url) def test_execution(self): this_file = Path(__file__) completed = subprocess.run((sys.executable, "-m", "spinetoolbox", "--execute-only", str(this_file.parent))) self.assertEqual(completed.returncode, 0) - db_map = DatabaseMapping(self._sink_url) - value_rows = db_map.query(db_map.object_parameter_value_sq).all() + with DatabaseMapping(self._sink_url) as db_map: + value_rows = db_map.query(db_map.entity_parameter_value_sq).all() self.assertEqual(len(value_rows), 2) expected_common_data = { - "object_class_name": "result", - "object_name": "test_data", + "entity_class_name": "result", + "entity_name": "test_data", "parameter_name": "final_value", } scenario_1_checked = False scenario_2_checked = False for value_row in value_rows: - row_as_dict = value_row._asdict() for key, expected_value in expected_common_data.items(): - self.assertEqual(row_as_dict[key], expected_value) + self.assertEqual(value_row[key], expected_value) value = from_database(value_row.value, value_row.type) if value == 11.0: self.assertTrue(value_row.alternative_name.startswith("scenario_1__Import@")) @@ -74,7 +77,6 @@ def test_execution(self): scenario_2_checked = True self.assertTrue(scenario_1_checked) self.assertTrue(scenario_2_checked) - db_map.connection.close() if __name__ == '__main__': diff --git a/execution_tests/parallel_importer/tool.py b/execution_tests/parallel_importer/tool.py index a26bf3e03..26f8f4f89 100644 --- a/execution_tests/parallel_importer/tool.py +++ b/execution_tests/parallel_importer/tool.py @@ -3,12 +3,9 @@ from spinedb_api import DatabaseMapping, from_database url = sys.argv[1] -db_map = DatabaseMapping(url) -try: - value_row = db_map.query(db_map.object_parameter_value_sq).first() - value = from_database(value_row.value, value_row.type) -finally: - db_map.connection.close() +with DatabaseMapping(url) as db_map: + value_row = db_map.query(db_map.parameter_value_sq).first() +value = from_database(value_row.value, value_row.type) with open("out.csv", "w") as out_file: out_writer = csv.writer(out_file) out_writer.writerow([value]) diff --git a/execution_tests/parallel_importer_with_datapackage/.spinetoolbox/specifications/Tool/test_tool.json b/execution_tests/parallel_importer_with_datapackage/.spinetoolbox/specifications/Tool/test_tool.json index 8c5d2dd0a..2a920cdea 100644 --- a/execution_tests/parallel_importer_with_datapackage/.spinetoolbox/specifications/Tool/test_tool.json +++ b/execution_tests/parallel_importer_with_datapackage/.spinetoolbox/specifications/Tool/test_tool.json @@ -11,6 +11,5 @@ "*.csv" ], "cmdline_args": [], - "execute_in_work": true, "includes_main_path": "../../.." } \ No newline at end of file diff --git a/execution_tests/parallel_importer_with_datapackage/execution_test.py b/execution_tests/parallel_importer_with_datapackage/execution_test.py index d96d2bc19..56578ca09 100644 --- a/execution_tests/parallel_importer_with_datapackage/execution_test.py +++ b/execution_tests/parallel_importer_with_datapackage/execution_test.py @@ -4,14 +4,15 @@ import subprocess import unittest from spinedb_api import ( + create_new_spine_database, DatabaseMapping, from_database, import_alternatives, + import_entities, + import_entity_classes, + import_parameter_definitions, + import_parameter_values, import_scenario_alternatives, - import_object_classes, - import_object_parameters, - import_object_parameter_values, - import_objects, import_scenarios, ) @@ -29,42 +30,44 @@ def setUp(self): if self._source_database_path.exists(): self._source_database_path.unlink() url = "sqlite:///" + str(self._source_database_path) - db_map = DatabaseMapping(url, create=True) - import_alternatives(db_map, ("alternative_1", "alternative_2")) - import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) - import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) - import_object_classes(db_map, ("content",)) - import_objects(db_map, (("content", "test_data"),)) - import_object_parameters(db_map, (("content", "only_value"),)) - import_object_parameter_values(db_map, (("content", "test_data", "only_value", 11.0, "alternative_1"),)) - import_object_parameter_values(db_map, (("content", "test_data", "only_value", 22.0, "alternative_2"),)) - db_map.commit_session("Add test data.") - db_map.connection.close() + with DatabaseMapping(url, create=True) as db_map: + import_alternatives(db_map, ("alternative_1", "alternative_2")) + import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) + import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) + import_entity_classes(db_map, ("content",)) + import_entities(db_map, (("content", "test_data"),)) + import_parameter_definitions(db_map, (("content", "only_value"),)) + import_parameter_values( + db_map, + ( + ("content", "test_data", "only_value", 11.0, "alternative_1"), + ("content", "test_data", "only_value", 22.0, "alternative_2"), + ), + ) + db_map.commit_session("Add test data.") self._sink_database_path.parent.mkdir(parents=True, exist_ok=True) if self._sink_database_path.exists(): self._sink_database_path.unlink() self._sink_url = "sqlite:///" + str(self._sink_database_path) - db_map = DatabaseMapping(self._sink_url, create=True) - db_map.connection.close() + create_new_spine_database(self._sink_url) def test_execution(self): this_file = Path(__file__) completed = subprocess.run((sys.executable, "-m", "spinetoolbox", "--execute-only", str(this_file.parent))) self.assertEqual(completed.returncode, 0) - db_map = DatabaseMapping(self._sink_url) - value_rows = db_map.query(db_map.object_parameter_value_sq).all() + with DatabaseMapping(self._sink_url) as db_map: + value_rows = db_map.query(db_map.entity_parameter_value_sq).all() self.assertEqual(len(value_rows), 2) expected_common_data = { - "object_class_name": "result", - "object_name": "test_data", + "entity_class_name": "result", + "entity_name": "test_data", "parameter_name": "final_value", } scenario_1_checked = False scenario_2_checked = False for value_row in value_rows: - row_as_dict = value_row._asdict() for key, expected_value in expected_common_data.items(): - self.assertEqual(row_as_dict[key], expected_value) + self.assertEqual(value_row[key], expected_value) value = from_database(value_row.value, value_row.type) if value == 11.0: self.assertTrue(value_row.alternative_name.startswith("scenario_1__Import@")) @@ -74,7 +77,6 @@ def test_execution(self): scenario_2_checked = True self.assertTrue(scenario_1_checked) self.assertTrue(scenario_2_checked) - db_map.connection.close() if __name__ == '__main__': diff --git a/execution_tests/parallel_importer_with_datapackage/tool.py b/execution_tests/parallel_importer_with_datapackage/tool.py index 2c1d7dd1c..5a8e6ee8b 100644 --- a/execution_tests/parallel_importer_with_datapackage/tool.py +++ b/execution_tests/parallel_importer_with_datapackage/tool.py @@ -3,12 +3,9 @@ from spinedb_api import DatabaseMapping, from_database url = sys.argv[1] -db_map = DatabaseMapping(url) -try: - value_row = db_map.query(db_map.object_parameter_value_sq).first() +with DatabaseMapping(url) as db_map: + value_row = db_map.query(db_map.parameter_value_sq).first() value = from_database(value_row.value, value_row.type) -finally: - db_map.connection.close() with open("out.csv", "w", newline="") as out_file: out_writer = csv.writer(out_file) out_writer.writerow(["final_value"]) diff --git a/execution_tests/scenario_filters/.spinetoolbox/specifications/Tool/write_values.json b/execution_tests/scenario_filters/.spinetoolbox/specifications/Tool/write_values.json index 28061a4ba..e87266c80 100644 --- a/execution_tests/scenario_filters/.spinetoolbox/specifications/Tool/write_values.json +++ b/execution_tests/scenario_filters/.spinetoolbox/specifications/Tool/write_values.json @@ -11,7 +11,5 @@ "out.dat" ], "cmdline_args": [], - "execute_in_work": true, - "includes_main_path": "../../..", - "execution_settings": {} + "includes_main_path": "../../.." } \ No newline at end of file diff --git a/execution_tests/scenario_filters/execution_test.py b/execution_tests/scenario_filters/execution_test.py index 300b36eb8..b3859ef20 100644 --- a/execution_tests/scenario_filters/execution_test.py +++ b/execution_tests/scenario_filters/execution_test.py @@ -3,7 +3,7 @@ import shutil import subprocess import unittest -from spinedb_api import DiffDatabaseMapping, import_alternatives, import_scenario_alternatives, import_scenarios +from spinedb_api import DatabaseMapping, import_alternatives, import_scenario_alternatives, import_scenarios class ScenarioFilters(unittest.TestCase): @@ -18,12 +18,11 @@ def setUp(self): if self._database_path.exists(): self._database_path.unlink() url = "sqlite:///" + str(self._database_path) - db_map = DiffDatabaseMapping(url, create=True) - import_alternatives(db_map, ("alternative_1", "alternative_2")) - import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) - import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) - db_map.commit_session("Add test data.") - db_map.connection.close() + with DatabaseMapping(url, create=True) as db_map: + import_alternatives(db_map, ("alternative_1", "alternative_2")) + import_scenarios(db_map, (("scenario_1", True), ("scenario_2", True))) + import_scenario_alternatives(db_map, (("scenario_1", "alternative_1"), ("scenario_2", "alternative_2"))) + db_map.commit_session("Add test data.") def test_execution(self): this_file = Path(__file__) diff --git a/execution_tests/scenario_filters/tool.py b/execution_tests/scenario_filters/tool.py index 1104c615b..5fdc34dab 100644 --- a/execution_tests/scenario_filters/tool.py +++ b/execution_tests/scenario_filters/tool.py @@ -2,8 +2,7 @@ from spinedb_api import DatabaseMapping, from_database url = sys.argv[1] -db_map = DatabaseMapping(url) -parameter_value = from_database(db_map.query(db_map.parameter_value_sq).first().value) +with DatabaseMapping(url) as db_map: + parameter_value = from_database(db_map.query(db_map.parameter_value_sq).first().value) with open("out.dat", "w") as out_file: out_file.write(f"{parameter_value}") -db_map.connection.close()