Train your model using active learning

Now we’ll train the model. You can choose between training texts from your uploaded file or upload annotated data as a new file.

Option 1: Annotate selected instances from the uploaded file

Select instances to annotate from the uploaded text

  1. Select ModelS from the side navigation

  2. Select your model from the drop-down

  3. Go to the Update tab

  4. Select Mode: request and select how many instances you want to annotate.

  5. Select sampler:

    1. Random: Returns random text instances from your dataset

    2. Margin recommended : A metric to find uncertain instances for the underlying model

  6. Then click the button Request

Annotate selected instances

Now you want to annotate the requested instances. To do that, proceed as follows:

  1. Select Mode: Label

  2. Select label default

    1. SKIP - Allows you to annotate the post by yourself

    2. Model prediction - Allows you to see the model’s prediction.

  3. For each given post, select the most suitable label or the SKIP option to skip a post annotation.

     

  4. Click Update to retrain the model. This process takes a little while. You may see a message "Model not ready. Come back later".

You can repeat this procedure multiple times.

Option 2: Upload a file of annotated instances

  1. Upload a file that has a column with post annotations

  2. Select the text column

  3. Select the label column (post annotations)

  4. Click Update to retrain the model. This process takes a little while. You may see a message "Model not ready. Come back later".

Model convergence estimation

After the model has been trained at least one time, the user will be able to monitor the evolution of the model performance.

The platform provides a way to monitor model convergence. That is, it answers the question if annotating more will further improve the model.

  1. With your model selected, go to the Inspect tab.

  2. After each iteration, a new point will be drawn on the line plot. You can deduce from it that, when the curve starts to flatten, the model quality is converging, and probably it's not going to learn much more. If the curve is not stable, you have to continue training the model as it is still learning.

Evaluate model performance

Using a different dataset from the one the model was trained with, you can evaluate the model’s performance.

  1. Prepare your data file with the following columns:

    1. Text: Your raw text instances

    2. Label: Label annotations for each text

  2. With your model selected, go to the Evaluate tab.

  3. Upload your file

  4. Select the text column and label column

  5. As a result, you’ll see a table with different performance metrics and a list of text instances (Errors) where the predicted label and the annotated label mismatch below it.


 

Next: Run your model

Symanto Brain Knowledge Base 2022