AutoML Training Guide

AutoML Training Guide

12 October 2021

Before you start

Note: Before using this script you need to export Google credentials.

Note: The dataset files should be having ratio as mentioned below:

Training Dataset (80%)

Test Dataset (10%)

Validate Dataset (10%)

For example: If you have 100 files then 80 should belong to the train dataset, 10 for Validation and 10 for the Test dataset.

Manual Annotation Process

Step 1: Create a directory say dir1 and add all the resumes that you want to use for the train dataset.

Step 2: Now you need to create dir2 and dir3 to validate and test the dataset respectively.

Step 3: To proceed further you need to install Google Cloud SDK and authenticate your email ID. Once you have completed the installation of Google Cloud SDK now you have to enter below commands one by one to upload train, validation and test dataset to Google Cloud Storage.

Train:

python2 script.py -t gs://match_making/tenant1/hr/documents/train train,dir1/*.pdf

Validation:

python2 script.py -t gs://match_making/tenant1/hr/documents/validation validation,dir2/*.pdf

Test:

python2 script.py -t gs://match_making/tenant1/hr/documents/test test,dir3/*.pdf

Step 4: To verify that the uploading is successfully completed. Navigate to the path that you have given in the script to verify your uploaded resumes/JDs in the Google Cloud Storage.

Step 5: After you are done with the data upload process, you need to import them by creating a new dataset in AutoML or you can use existing dataset.

Step 6: Navigate to the dataset.csv file in Google Cloud Storage and select the file and click on import.

Step 7: Now once you click on the import dataset, wait for 5-10 minutes to import all the resumes/JDs on the AutoML platform.

Step 8: Once done you will see all the PDF files and you can now annotate and start the annotation process.

Step 9: After completing the annotation you need to start training which usually takes 3 hours.

Step 10: Once training is finished you can test the model

Auto Annotation Process

Step 1: To upload txt files and use auto annotations feature you first need to convert PDF into a TXT file.

Step 2: Now you just need to keep these files in different directories like we have done for uploading PDF files and use the below commands to upload them to Google Cloud Storage.

Note: dict.csv file contains all the labels that we want to auto annotate.

Train:

python2 script.py -d dict.csv -s train,dir1/*.txt gs://match_making/tenant1/hr/documents/train

python2 script.py -d dict_banking.csv -s train,dir1/*.txt gs://match_making/tenant1/banking/documents/train

Validation:

python2 script.py -d dict.csv -s validation,dir2/*.txt gs://match_making/tenant1/hr/documents/validate

python2 script.py -d dict_banking.csv -s validation,dir2/*.txt gs://match_making/tenant1/banking/documents/validate

Test:

python2 script.py -d dict.csv -s test,dir3/*.txt gs://match_making/tenant1/hr/documents/test

python2 script.py -d dict_banking.csv -s test,dir3/*.txt gs://match_making/tenant1/banking/documents/test

Step 3: To verify that the uploading is successfully completed. Navigate to the path that you have given in the script to verify your uploaded resumes/JDs in Google Cloud Storage.

Step 4: After you are done with the data upload process, you need to import them by creating a new dataset in AutoML or you can use the existing dataset.

Step 5: Navigate to the dataset.csv file in Google Cloud Storage and select the file and click on import.

Step 6: Now once you click on the import dataset, wait for 5-10 minutes to import all the resumes/JDs on the AutoML platform.

Step 7: Once done you will see all the documents and you can now start the annotation process.

Step 8: After completing the annotation you need to start training which usually takes 3 hours.

Step 9: Once training is finished you can test the model.

[responsivevoice_button buttontext='Listen']

Search

Blog Categories

Request a quote