Hugging Face Transformers
Integrating with Hugging Face Transformers
mlop provides an interface for rich model logging and monitoring with Hugging Face Transformers.
Migrating from Weights & Biases
See the Migrating from Weights & Biases guide for a quickstart.
Logging Model Details
To use mlop for Hugging Face Transformers, simply import the MLOPCallback
class and specify report_to="mlop"
, or to use mlop alongside other loggers, specify report_to="all"
.
Warning
At the moment, from mlop.compat.transformers import MLOPCallback
must be called before initializing TrainingArguments
to allow mlop enable itself for Hugging Face Transformers. This requirement will be removed in the near future after mlop
is merged into transformers
.
Source
See the Python code for more details.
Using MLOPCallback
The MLOPCallback
inherits from the standard transformers.trainer_callback.TrainerCallback
class, and strictly conforms to its standard API interface.