Hi HN,
Today we’re launching GPT-4 answers on Phind.com, a developer-focused search engine that uses generative AI to browse the web and answer technical questions, complete with code examples and detailed explanations. Unlike vanilla GPT-4, Phind feeds in relevant websites and technical documentation, reducing the model’s hallucination and keeping it up-to-date. To use it, simply enable the “Expert” toggle before doing a search.
GPT-4 is making a night-and-day difference in terms of answer quality. For a question like “How can I RLHF a LLaMa model”, Phind in Expert mode delivers a step-by-step guide complete with citations (https://phind.com/search?cache=0fecf96b-0ac9-4b65-893d-8ea57...) while Phind in default mode meanders a bit and answers the question very generally (https://phind.com/search?cache=dd1fe16f-b101-4cc8-8089-ac56d...).
GPT-4 is significantly more concise and “systematic” in its answers than our default model. It generates step-by-step instructions over 90% of the time, while our default model does not.
We’re particularly focused on ML developers, as Phind can answer questions about many recent ML libraries, papers, and technologies that ChatGPT simply cannot. Even with ChatGPT’s alpha browsing mode, Phind answers technical questions faster and in more detail.
For example, Phind running on “Expert” GPT-4 mode can concisely and correctly tell you how to run an Alpaca model using llama.cpp: (https://phind.com/search?cache=0132c27e-c876-4f87-a0e1-cc48f...). In contrast, ChatGPT-4 hallucinates and writes a make function for a fictional llama.cpp.
We still have a long way to go and would love to hear your feedback.
It's amazing. Real time saver. Just yesterday it saved me from going down an hour+ rabbit hole due to a cryptic error message. The first solution it gave me didn't work, neither did the second, but I kept pushing and in just a couple of minutes I had it sorted.
Having said that, I'm not sure I see the gain with Expert mode yet. After using it for the last couple of days, it's definitely much slower but I couldn't perceive it to be any more accurate.
Judging by your example, it looks like the main difference is that the Expert mode search returned a more relevant top result, which then the LLM heavily relied on for its answer. If search results come from bing, can you really credit that answer to Expert mode?
PS. You mention launching GPT-4 today, but the Expert Mode toggle has been there for at least a few days, I reckon? Was it not GPT-4 before?