New open-source platform allows users to evaluate performance of AI-powered chatbots

A team of computer scientists, engineers, mathematicians and cognitive scientists, led by the University of Cambridge, have developed an open-source evaluation platform called CheckMate, which allows human users to interact with and evaluate the performance of large language models (LLMs).

This article was originally published here

LawyersLookup.ca - Find a lawyer who speaks your language