diff --git a/atom_batch_execution/scripts/batch_execution b/atom_batch_execution/scripts/batch_execution index 4fe7637f..43ebe7d3 100755 --- a/atom_batch_execution/scripts/batch_execution +++ b/atom_batch_execution/scripts/batch_execution @@ -76,13 +76,7 @@ def main(): file_content = f.read() data = yaml.safe_load(file_content) - # Load dataset - # If the corrected exists, the cross validator class divider will work slightly better. - expanded_dataset_file_path = os.path.expandvars(data['dataset_path'] + '/dataset_corrected.json') - if os.path.isfile(expanded_dataset_file_path): - dataset = loadJSONFile(data['dataset_path'] + '/dataset_corrected.json') - else: - dataset = loadJSONFile(data['dataset_path'] + '/dataset.json') + dataset = loadJSONFile(data['dataset_path']) collections_to_remove = data['collections_to_remove'] diff --git a/atom_core/src/atom_core/dataset_io.py b/atom_core/src/atom_core/dataset_io.py index 9dd5610f..206c7d3f 100644 --- a/atom_core/src/atom_core/dataset_io.py +++ b/atom_core/src/atom_core/dataset_io.py @@ -558,7 +558,9 @@ def filterSensorsFromDataset(dataset, args): return dataset def deleteCollectionFromDataset(dataset,collection_number): - + # Use the filterCollectionsFromDataset instead. Check #948. + # Leaving this function here for potential future use only. + # Converting int 90 to '090' collection_key = generateCollectionKey(collection_number) print(f'Deleting {collection_key}')