The focus of this subdirectory is the SuperKnowa mechanism for collecting human feedback. Collecting input from users can play a crucial role in assessing a model’s response and enhancing its performance and implementing for a Retrieval-Augmented Generation (RAG) pipeline.
We focus on fine-grained human feedback on the basis of the following criteria:
- Does the generated output display lack of relevance, repetition, and inconsistency?
- Does the generated output contain inaccurate or unverifiable information?
- Does the generated output include all relevant information? Partial information?
We gather feedback through various mechanisms.
The complete source code of the tool, along with detailed code explanations, can be found here. superknowa-app
Rating feedback involves assigning a numerical value to a model's response based on its perceived quality or relevance. For instance, a user might rate a generated response on a scale of 1 to 5, indicating their satisfaction with the answer. It is quick and quantifiable, suitable for assessing overall quality.
Ranking feedback requires users to compare and rank responses from different models. This feedback mechanism provides a finer level of bias between responses and allows the model to learn the relative quality of outputs.
We are using drag and drop mechanism to rank the models based on the model’s response.
Yes/no questionnaires present users with specific questions about the response. Targeted feedback on specific aspects, such as relevance to the topic. It can help to gather focused feedback on the accuracy of the response.
Comment-based feedback lets users provide textual explanations for their preferences or suggestions for improvement. These comments offer valuable insights into the strengths and weaknesses of model responses.