Short answer: because chat is a shortcut to anthropomorphization in folks' brains. It's much easier to assign personality and intent to a chat than a floating head that needs to be 99.999% lifelike in order to not feel completely fake.
> A better answer: natural language is the programming language everyone knows.
And when you get into frustrating misunderstanding of details you end up using a programming language anyway. Natural languages are to ambiguous to be useful for specific things. We humans can resolve the ambiguities in our minds, or think we do, based on context. I think the anthropomorphization argument hits closer to home. I feel guilty too of projecting more meaning and intelligence to a generated text and I think I'm not the only one.
Depends what you mean by 'cycle', I suppose. The current hype-cycle _really_ started with ChatGPT; that was the point where people started getting super-excited and declaring that it would replace all human work by next Thursday.
Chat is perhaps the cheapest implementation you could ever build. It's a linear interaction, easy to test, and arguably the easiest to encode/decode (with a fixed set of inputs too). As an added bonus, it has a familiar, well-understood interface.
It is now but it hasn't always been easy to make a computer chat naturally like a human without using canned responses. Chess positions are a much easier interface to implement because it's a closed set and very structured (grid position + piece). But humans don't speak in chess moves so it would be difficult for us to speak to an AI that does.
LoRAs are fine tuned changes to an existing model, often one that is already mature and well-trained. The process is understood and documented.
OP is referring to early phases of model development using new techniques, where outputs can vary wildly until corrected and evaluated at scale. The process is often adaptive, iterative, and by nature experimental. Text in this case is indeed easier to evaluate.
I'm curious about: alpha go, Watson, and the AI that "conquered" chess. Are those outliers, or part of a larger story? It feels to me a contemporary history of AI would mention those milestones.
Uhm… The boom in AI with LLMs wouldn’t have happened without about a decade of major focus on images (both generative models and DNN models that blew traditional image processing out of the water), and planning/optimization type problems (alpha go, chess, etc.). Seems incorrect to claim that chat starts every cycle.