Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Now we’ll retrain train the model. You can choose between retraining training texts from your uploaded file or upload annotated data as a new file.

...

Panel
panelIconId1f4bb
panelIcon:computer:
panelIconText💻
bgColor#FFF0B3

API Docs: https://api.symanto.net/active-learning/docs#/v2/request_instances_v2__model_id__instances_post

  1. Select Action:

    Status
    Model
    titleModelS
    from the side navigation

  2. Select your model Select Action: from the drop-down

  3. Go to the

    Status
    titleUpdate
    tab

  4. Select Mode:

    Status
    colourYellow
    titlerequest
    and select how many instances you want to annotate.

  5. Select sampler:

    1. Random: Returns random text instances from your dataset

    2. Margin

      Status
      colourGreen
      titlerecommended
      : A metric to find uncertain instances for the underlying model

  6. Then click the button “Request”. Request

    Image Removed

...

Annotate selected instances

Now you want to annotate the requested instances. To do that, proceed as followingfollows:

  1. Select Mode:

    Status
    titleLabel

  2. Select label default

    1. SKIP - Allows you to annotate the post by yourself

    2. Model prediction - Allows you to see the model’s prediction.

  3. For each given post, select the most suitable label or the SKIP option to skip a post annotation.

  4. Click “Update” Update to retrain the model. This process takes a little while. You may see a message “Model "Model not ready. Come back later”later".

You can repeat this procedure multiple times.

...

  1. Upload a file that has a column with post annotations

  2. Select the text column

  3. Select the label column (post annotations)

  4. Click “Update” Update to retrain the model. This process takes a little while. You may see a message “Model "Model not ready. Come back later”later".

    Image Removed

...

Model convergence estimation

After the model has been trained at least one time, the user will be able to monitor the evolution of the quality model performance.

The platform provides a way to monitor model convergence. That is, it answers the question if annotating more will further improve the model.

...

  1. With your model selected, select action go to the

    Status
    titleinspectInspect
    tab.

  2. After each iteration, a new point will be drawn on the "estimated quality performance" line plot. You can deduce from it that, when the curve starts to flatten, the model quality is converging, and probably it's not going to learn much more. If the curve is not stable, you have to continue training the model as it is still learning.

    Image Removed

...

Evaluate model performance

Using a different dataset than from the one the model was created trained with, you can evaluate the model’s performance.

  1. Prepare your data file with the following columns:

    1. Text: Your raw text instances

    2. Label: Label annotations for each text

  2. Select Action: With your model selected, go to the

    Status
    titleevaluateEvaluate
    tab.

  3. Upload your file

  4. Select the text column and label column

  5. As a result, you’ll see a table with different performance metrics and a list of text instances (Errors) where the predicted label and the annotated label mismatch below it.

Info

Use a different dataset for evaluation than the one you used for training.

...

...

Info

In case the results are not sufficient, there might be multiple reasons:

  • you might have to revise your labels and hypotheses and use Promptranker to create another model in order to achieve better results

  • your evaluation set is not good enough - the distribution in the examples in it is not equal (e.g. in the example above in the evaluation set we had only 30 examples for special occasions but 148 for social gatherings, and respectively the latter got better results)

...

Next: Run your model