Tutorial 4, FAQ retrieval: all results have near-identical score #2668
-
I've run Tutorial 4, Utilising Existing FAQs, and noted that all of the returned answers have almost equal scores, between 0.500 and 0.502. Is this expected? This happened when I ran the colab without any changes. Additionally, it happens whether the question exactly matches an existing question in the database, e.g. "What is a novel coronavirus?", or whether the question is completely made up. This seems odd, since the Extractive Q&A systems provide a score where it meaningfully ranges between 0 and 1. My use-case is merging a set of results from an extractive Q&A pipeline and results from a FAQ-retrieval pipeline. I'm not sure how to do this if FAQ answers are always 0.5, regardless of how well they actually matched the question. In short - can I make the FAQ answers have a more useful score? |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Hi @stevenhaley thanks for raising the issue. In the tutorial we're using the wrong similarity algorithm, indeed you should see a warning when running the example:
We opened a PR to fix the tutorial. |
Beta Was this translation helpful? Give feedback.
Hi @stevenhaley thanks for raising the issue. In the tutorial we're using the wrong similarity algorithm, indeed you should see a warning when running the example:
We opened a PR to fix the tutorial.