ChatGPT — Search Accuracy, Hallucinations, and Prompt Engineering
With OpenAI releasing the latest iteration of its large language model, GPT-4 (Generative Pre-trained Transformer 4), we thought we would take it’s free little brother for a spin. ChatGPT currently uses GPT-3.5, a smaller and faster version of the GPT-3 model.
In this article we will investigate using large language models (LLMs) for search applications, illustrate some of the issues with this including hallucinations, and finally will explain how you can use prompt engineering to fine tune answer style, context and content. In part 2 of this series, we will explore using LLMs to develop software.
A large language model is a type of machine learning model that is trained on a huge body of data to generate outputs for various language processing tasks, such as text generation, question answering, and machine translation. ChatGPT is trained using Reinforcement Learning from Human Feedback (RLHF).
We love the progress that is being made in Machine Learning (ML) and predictive Large Language Models (LLMs), but believe users need to be cautious in their application and aware of the pitfalls. LLMs are being spruiked as a replacement for search engines (*cough* Microsoft *cough*), but until they stop hallucinating and making up facts, caution is warranted. It is interesting that Google…