Hacker News new | past | comments | ask | show | jobs | submit login

I'm wondering how much of my repulsion to AI is Sam Altman himself. Recently I've been trying to be more open-minded with using LLMs in my workflows and have found some minor performance gains here and there. However the hype around AI is frankly embarrassing.





Big part of my annoyance is the term "AI" itself which you can say to mean anything, everything and nothing. It's something that's used to oversell/hype it to grab money.. which is fine. But if engineers like us can calm the crazy rhetoric down to "llms" or "text completion" "Image-gen api" that's already a leg up in thinking clearly. Like the question "is ai going to put people out of jobs" -> "is this new crazy good text completion model going to put people out of jobs" already gets us out of the weeds somewhat.

> Big part of my annoyance is the term "AI" itself which you can say to mean anything, everything and nothing.

Pretty sure that's the point now.

You're selling a "solution" not a "product". So you don't want to market AI as some nice robot that can move car doors around. You want to market AI as a vibe that fixes enterprise problems. So now all the companies that want your AI both have to pay for AI and also your time to actually build a product for their problem.


Yeah its definition has been diluted to the point of useless now.

I think this extends to a lot of normal people’s instinctive hatred of AI: they dislike it in part because they perceive it as being pushed on them by some pretty slimy people

Yeah. I gave Claude a spin today; it helped me to make a tech spec I'd written a little more concise and professional. But doing so required carefully reading and tweaking the output to maintain accuracy. I don't know if it sped anything up, but it reduced the cognitive effort a bit.

wait until it crosses a few thresholds we’re fast approaching. ai that performs at intern level vs phd vs domain leading expert level. ai that has the full context of what youre doing immediately accessible. ai that can agenticly autonomously navigate the computer environment without stumbling blocks.

i think the difference between intern level ai and domain leading expert ai is a few algorithmic adjustments away using a type of reasoning reinforcement framework (like GRPO) which deal with competing signals in a better way. instead of averaging them, it reasons which contextually signal should take precedent. its the difference between lets say taking a vote among the common populace on how to build a nuclear power plant and finding an expert and figuring out where the experts decisions should take precedent and where other experts decisions should override.

square that away and the embarrassing feeling on the promise of ai should wash away.


> ai that performs at intern level vs phd vs domain leading expert level.

These are different things:

- Regurgitating advanced text that has been shifted into a shape matching your query

- Understanding intimately the $100M screw to turn


How do you know that these are different things? They could be, I genuinely don’t know, but I’m not sure where people are getting these kind of confident assertions about what modern architectures could never do. Would you have predicted in 2020 that photorealistic text to image generation was within the scope of current theory?

> Would you have predicted in 2020 that photorealistic text to image generation was within the scope of current theory?

Yes, and I've been working in this area with excitement since about that time.

The physics of optics are well understood. We've been writing ray tracers for forever and coming up with clever hacks like Blinn–Phong, PBR, etc. for ages. SIGGRAPH has always felt like tangible magic. We have had the map in our hands and now we're coming up with new ways to traverse a familiar landscape.

Reasoning is an undiscovered country. There are lots of exciting claims being made, but nothing concrete.

I expect lots of advancements in signal processing, spatial computing, and beyond because those things are obvious and intuitive.


The mathematical definition of a language model is the probability distribution of tokens that follows the previous context. It's literally deciding the most probable response, which while at many times may match the correct response, is not a 1 to 1.

welcome to earth where 98% of forum/political interactions are confident assertions from nowhere used to dismiss people. enjoy your stay ;).

obviously theres also a multimodality gap to be overcome to intimately understanding the $100M screw to turn, but i suspect most reasoning that matters has already been translated and is embedded into/in words. i wouldn’t underestimate the amount of useful knowledge that exists in embedding advance texts into an LLm model. the challenge is contextually hierarchalizing it (a matter of reasoning) and decoding it back into reality (words are dimensional squished encodings of reality).



Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: