You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Implement a suite of unit tests to check if the output of the Large Language Model (LLM) used in the Virtual Assistant Demo involves hallucinations. This is a great opportunity to explore techniques on ensuring the reliability and accuracy of the assistant's responses by identifying and mitigating instances where the LLM generates incorrect or misleading information.
This project will involve Python programming; basic experience with AI models from frameworks like PyTorch, TensorFlow, OpenVINO, or ONNX is beneficial. To learn more about the OpenVINO toolkit, visit the documentation here.
Examples of hallucinations:
Factual errors
Inaccurately summarizing information
Creating nonsensical content, such as random sentences (which can also manifest in the LLM trailing off in the middle of a response to a tangential topic)
Steps:
Develop unit tests for Hallucination Detection:
Define criteria for what constitutes a hallucination and implement checks based on these criteria.
Create unit tests that analyze the output of the LLM used in the Virtual Assistant Demo to detect hallucinations.
Enhance Testing Framework:
Integrate the new unit tests into the existing testing framework.
Ensure the tests are comprehensive and cover various scenarios where hallucinations might occur.
Documentation:
Provide clear and detailed documentation for setting up and running the unit tests.
Include comments in the code to explain key sections and logic.
Uh oh!
There was an error while loading. Please reload this page.
Description
Implement a suite of unit tests to check if the output of the Large Language Model (LLM) used in the Virtual Assistant Demo involves hallucinations. This is a great opportunity to explore techniques on ensuring the reliability and accuracy of the assistant's responses by identifying and mitigating instances where the LLM generates incorrect or misleading information.
This project will involve Python programming; basic experience with AI models from frameworks like PyTorch, TensorFlow, OpenVINO, or ONNX is beneficial. To learn more about the OpenVINO toolkit, visit the documentation here.
Examples of hallucinations:
Steps:
Develop unit tests for Hallucination Detection:
Enhance Testing Framework:
Documentation:
How to Get Started:
The text was updated successfully, but these errors were encountered: