These notebooks show how to fine-tune an NLP model on AzureML. They are intended to be cloned to and executed on an AzureML compute instance within a Jupyter environment. They go through the process of creating a DeepSpeed enabled training environment, creating a compute target (if there isn't one already), preparing and registering datasets, fine tuning a model on those data sets, and registering the resulting output model. This is configured and supported by only a few outside files in the src directory.
- Clone this repo into an interactive session on a fresh AzureML compute instance
- From the command line, install the
requirements.txtinto the localAzureML_Py3.8conda environment viaconda activate azureml_py38 && pip install -r requirements.txt. - Follow the notebooks in numerical order
01 Create computeensures requirements are installed and compute cluster is accessible02 Prepare environmentcreates an AzureML environment that supports DeepSpeed training03 Prepare datadownloads, preprocesses, and registers a dataset for versioned and reproducible training04 Train modellaunches a distributed fine-tuning job using the outputs of the prior notebooks