This submit is the primary in a Constructing AI Responsibly collection, which explores prime considerations with deploying AI and the way Microsoft is addressing them with its Accountable AI practices and instruments.
Technically, hallucinations are “ungrounded” content material, which suggests a mannequin has modified the info it’s been given or added further data not contained in it.
There are occasions when hallucinations are helpful, like when customers need AI to create a science fiction story or present unconventional concepts on all the things from structure to coding. However many organizations constructing AI assistants want them to ship dependable, grounded data in eventualities like medical summarization and training, the place accuracy is essential.
That’s why Microsoft has created a complete array of instruments to assist handle ungroundedness primarily based on experience from creating its personal AI merchandise like Microsoft Copilot.
Firm engineers spent months grounding Copilot’s mannequin with Bing search information via retrieval augmented era, a method that provides further information to a mannequin with out having to retrain it. Bing’s solutions, index and rating information assist Copilot ship extra correct and related responses, together with citations that permit customers to lookup and confirm data.
“The mannequin is superb at reasoning over data, however we don’t assume it ought to be the supply of the reply,” says Chicken. “We expect information ought to be the supply of the reply, so step one for us in fixing the issue was to convey recent, high-quality, correct information to the mannequin.”