Skip to content

Commit

Permalink
#948 More consistency in dataset declaration
Browse files Browse the repository at this point in the history
  • Loading branch information
brunofavs committed May 8, 2024
1 parent a1b8c5a commit be9f9bf
Show file tree
Hide file tree
Showing 2 changed files with 4 additions and 8 deletions.
8 changes: 1 addition & 7 deletions atom_batch_execution/scripts/batch_execution
Original file line number Diff line number Diff line change
Expand Up @@ -76,13 +76,7 @@ def main():
file_content = f.read()
data = yaml.safe_load(file_content)

# Load dataset
# If the corrected exists, the cross validator class divider will work slightly better.
expanded_dataset_file_path = os.path.expandvars(data['dataset_path'] + '/dataset_corrected.json')
if os.path.isfile(expanded_dataset_file_path):
dataset = loadJSONFile(data['dataset_path'] + '/dataset_corrected.json')
else:
dataset = loadJSONFile(data['dataset_path'] + '/dataset.json')
dataset = loadJSONFile(data['dataset_path'])

collections_to_remove = data['collections_to_remove']

Expand Down
4 changes: 3 additions & 1 deletion atom_core/src/atom_core/dataset_io.py
Original file line number Diff line number Diff line change
Expand Up @@ -558,7 +558,9 @@ def filterSensorsFromDataset(dataset, args):
return dataset

def deleteCollectionFromDataset(dataset,collection_number):

# Use the filterCollectionsFromDataset instead. Check #948.
# Leaving this function here for potential future use only.

# Converting int 90 to '090'
collection_key = generateCollectionKey(collection_number)
print(f'Deleting {collection_key}')
Expand Down

0 comments on commit be9f9bf

Please sign in to comment.