OpenAI, the artificial intelligence research company, has announced the launch of a new tool called CriticGPT to help identify errors and bugs in code generated by its popular language model, ChatGPT.
CriticGPT is designed to assist human AI reviewers in the task of evaluating code produced by ChatGPT. The tool has shown promising results, with the company claiming that when people use CriticGPT to review ChatGPT’s code, they outperform those without the tool’s assistance 60% of the time.
While CriticGPT has demonstrated its ability to identify errors in code, it is not without limitations. The tool has been trained on relatively short responses from ChatGPT, which may hinder its performance when evaluating longer and more complex tasks.
Additionally, CriticGPT is not immune to hallucinations, and human oversight is still required to rectify any labeling mistakes made by the model.
Moving forward, OpenAI plans to further develop and scale CriticGPT to enhance its utility in the RLHF process, as the company continues to work on improving the capabilities of its large language models.