Separating the function to loop over data from the one to create batches for training#92
Separating the function to loop over data from the one to create batches for training#92anagainaru wants to merge 1 commit intomainfrom
Conversation
|
I think this is a neccessary change - but very intrusive in the sense that all examples needs to be adapted. Do you have an example that runs currently, so I can validate the changes locally? |
|
I think with this change we can further simplify: ==> the loop loader only needs a val_loader, since we are not training on this data |
You do not need to change the examples, if this second function is not implemented we use the current behavior with looping over data using the training batch size. |
Agree, I will make the change and update mnist so you can run using different batches if you specify in toml a data batch size. |
Summary
Current behavior is to have a function
get_cur_data_loadersthat returns the dataloaders that is used for both looping for inference and for creating batches to train when doing continual learning. The current PR separates them into two separate functions.Motivation & Context
Having two functions would allow us to control the granularity of the drift detectors (e.g. looking at element by element) without impacting the training (which should use the same batch size as what was used for the original training)
Approach
Introduced a new function
get_cur_loop_loaders. If this function is not implemented in the model harness it returns by defaultself.get_cur_data_loaders()API / CLI Changes
No changes, this is an optional function in the model harness.
Example usage: