- Getting started
- Balance
- Clusters
- Concept drift
- Coverage
- Datasets
- General fields (previously entities)
- Labels (predictions, confidence levels, hierarchy, etc.)
- Models
- Streams
- Model Rating
- Projects
- Precision
- Recall
- Reviewed and unreviewed messages
- Sources
- Taxonomies
- Training
- True and false positive and negative predictions
- Validation
- Messages
- Administration
- Manage sources and datasets
- Understanding the data structure and permissions
- Create a data source in the GUI
- Uploading a CSV file into a source
- Create a new dataset
- Multilingual sources and datasets
- Enabling sentiment on a dataset
- Amend a dataset's settings
- Delete messages via the UI
- Delete a dataset
- Export a dataset
- Using Exchange Integrations
- Preparing data for .CSV upload
- Model training and maintenance
- Understanding labels, general fields and metadata
- Label hierarchy and best practice
- Defining your taxonomy objectives
- Analytics vs. automation use cases
- Turning your objectives into labels
- Building your taxonomy structure
- Taxonomy design best practice
- Importing your taxonomy
- Overview of the model training process
- Generative Annotation (NEW)
- Dastaset status
- Model training and annotating best practice
- Training with label sentiment analysis enabled
- Train
- Introduction to Refine
- Precision and recall explained
- Precision and recall
- How does Validation work?
- Understanding and improving model performance
- Why might a label have low average precision?
- Training using Check label and Missed label
- Training using Teach label (Refine)
- Training using Search (Refine)
- Understanding and increasing coverage
- Improving Balance and using Rebalance
- When to stop training your model
- Using general fields
- Generative extraction
- Using analytics and monitoring
- Automations and Communications Mining
- Licensing information
- FAQs and more
Why might a label have low average precision?
Below are outlined are some of the main reasons why a label may have low average precision, as well as a suggested solution to improve it:
1. The training set size may be too small
- If the training set size is quite small, it may be that you just need to provide more training examples for the model
- Continue training the label using the methods outlined in the Explore phase, particularly 'Shuffle' and 'Teach label'
2. The label may have been applied inconsistently or incorrectly to some of the messages
- It can often be the case that a user’s definition of a label changes over time, and older reviewed messages with that label may need revisiting to see if the label still applies
- Alternatively, if there are multiple users training a dataset, they could have interpretations of what each label means, and send mixed signals to the model
- To determine whether this is the case, users can use 'Check label' and 'Missed label' training modes to go through the reviewed messages for the label, and see where a label has been applied incorrectly, or missed unintentionally
- Users can then correct any errors and update labels to ensure consistency
- Going forward, if there are multiple users training a dataset, they should ensure that they are fully aligned on how they define the intents or concepts covered by each label
3. The intent or concept that the label is intended to capture may be vague or very broad and hard to distinguish from other labels
- If a label is used to capture a very broad or vague intent or concept, it can be hard for the model to identify why that label should apply to a message – it may then try to apply it to far too many messages
- Try not to be too generic when creating a label; it needs to be identifiable and distinguishable from other labels
4. Alternatively, the intent or concept could be very specific or have too many layers in its hierarchy
- Trying to be too specific or adding many layers to a label’s hierarchy can make it too difficult for the model to detect, or distinguish it from previous layers
- The level of specificity for a label should match the content of the messages. If it is too specific to realistically distinguish from other similar labels in the hierarchy, the model may get confused
- In most cases, it is best practice to have three layers or less in a label’s hierarchy – i.e. [Root label] > [Connecting label] > [Leaf label]
5. There may be several labels in the taxonomy that heavily overlap and the model struggles to distinguish between the two
- If you have two labels that are very similar and hard to distinguish from one another, it can confuse the model, as it won’t know which of the two labels applies
- In these instances, consider merging the labels
- Alternatively, go through the reviewed messages for each and make sure that the concepts are applied consistently and are distinct from one another
6. The messages with that label applied may mostly be very similar or identical, and the model struggles to detect different ways of expressing the same intent or concept
- You should ensure that for every label you provide the model with multiple training examples that include various different ways of expressing the intent or concept that the label is intended to capture
7. The intent or concept captured by that label is not semantically inferable from the text of the message or it’s supporting metadata
- It is common for users to annotate a message based on their own business knowledge of the context or process that would follow, and not on the actual text or metadata of the message
- For example, an SME user may know that because the communication has come from a certain individual, it must be about a certain topic, even though nothing else in the text or metadata clearly indicates that the label should apply
- In this instance, users should only apply the label if the model would be able to detect it from the text or metadata, without this inside knowledge