Defining what's irrelevant
Classic experience only
This information applies to dialog skill analytics in the classic experience. For information about analytics in watsonx Assistant, see Use analytics to review your entire assistant at a glance.
Teach your dialog skill to recognize when a user asks about topics that it is not designed to answer.
To teach your assistant about subjects to ignore, you can review your user conversation logs to mark utterances that discuss off-topic subjects as irrelevant.
Intents that are marked as irrelevant are saved as counterexamples in the JSON workspace, and are included as part of the training data. They teach your assistant to explicitly not answer utterances of this type.
When you test your dialog, you can mark an intent as irrelevant directly from the Try it out pane.
Be certain before you designate an input as irrelevant.
- You can't access or change the inputs from the user interface later.
- The only way to reverse the identification of an input as being irrelevant is to use the same input in a test integration channel, and then explicitly assign it to an intent.
You might have subjects that you expect customers to ask and that you want your assistant to address eventually, but that you aren't ready to fully implement yet. Instead of adding those topics as counterexamples that can be hard to find later, capture the customer input examples as new intents. But don't link dialog nodes to the intents until you're ready. If customers ask about one of the topics, the anything_else node is triggered to explain that the assistant can't help with the current request, but can help them with other things.
Irrelevance detection
The irrelevance detection feature helps your dialog skill recognize subjects that you do not want it to address, even if you didn't teach it what to ignore. This feature helps your skill recognize irrelevant inputs earlier in the development process.
To test irrelevance detection in the "Try it out" pane, submit one or more utterances that have absolutely nothing to do with your training data. Irrelevance detection helps your skill to correctly recognize that the test utterances
do not map to any of the intents defined in your training data, and classifies them as being Irrelevant
.
The algorithmic models that help your assistant understand what your users say are built from two key pieces of information:
-
Subjects that you want the assistant to address. For example, questions about order shipments for an assistant that manages product orders.
You teach your assistant about these subjects by defining intents and providing sample user utterances that articulate the intents so your assistant can recognize these and similar requests as examples of input.
-
Subjects that you want the assistant to ignore. For example, questions about politics for an assistant that makes pet grooming appointments exclusively.
You teach your assistant about subjects to ignore by marking utterances that discuss subjects that are out of scope for your application as being irrelevant. Such utterances become counterexamples for the model.
The best way to build an assistant that understands your domain and the specific needs of your customers is to take the time to build good training data.
Counterexample limits
The maximum number of counterexamples that you can create for any plan type is 25,000.