Conversation
|
@heyhaleema Don't hesitate if you have any questions re. how to leverage this piece of code in a more elegant manner (e.g. Factory depending on the type of project) |
@AgatheZ Firstly - this is amazing! 🔥 One (initial) question: is Also, I've just realised that since we run @mikewoodward94 If you have some time for a quick chat about this next week, that'd be great 🙏🏼 My thought is that I can then have the diagram for #86 be the proposed updated workflow, and potentially have this used to update this diagram in the |
|
@heyhaleema About your second question, we could run |
Linked Issue(s)
#76
Summary of changes
I have implemented the code for the NLP template, as a standard multi class classifier leveraging a Bert based transformer.
This includes:
training script
datamodule
dataset
network
wrapper
configs
I created fake data (in tests/data/)
Key features I introduced:
KFOLD cross validation possibility (can be enabled/disabled in the config file)
Save the best model (can be enabled/disabled in the config file)
f1 score and other metrics are logged
A data manifest is logged as a mlflow artifact, logging the classes distributions in the training and validation test sets)
FYI: I had to change the Dockerfile for the default folder to be project_NLP, might incur some clashes in the future.
I have tested it on the DGX, it runs properly and nicely