Hacker News new | past | comments | ask | show | jobs | submit | mbowcut2's comments login

DeepSeek has demonstrated that there is no technical moat. Model training costs are plummeting, and the margins for APIs will just get slimmer. Plus model capabilities are plateauing. Once model improvement slows down enough, seems to me like the battle is to be fought in the application layer. Whoever can make the killer app will capture the market.

Model capabilities are not plateauing; in fact, they are improving exponentially. I believe people struggle to grasp how AI works and how it differs from other technologies we invented. Our brains tend to think linearly; that's why we see AI as an "app." With AI (ASI), everything accelerates. There will be no concept of an "app" in ASI world.

Can you give examples? From gpt-4 came out in 2023 and since then nothing similar to (3.5 to 4 or 2 to 3) has come out. It has been 2 years now. All signs point towards OpenAI struggling to get improvements from its llms. The new releases have been minor since 2023

The chain of thought models provide huge improvements in certain areas. 2 years is also hardly enough time to claim theyre stuck

Hype. We were already plateauing with next token prediction and letting the models think out loud has simply pushed the frontier a little bit in purely test-taking domains.

They are absolutely not improving exponentially, by any metric. This is a completely absurd claim. I don’t think you understand what exponential means.

Something worth noting is that ChatGPT currently is the killer app -- DeepSeek's current chart-topping app notwithstanding (not clear if viral blip or long-term trend).

ChatGPT Plus gives me a limited number of o1 calls and o1 doesn't have web access, so I mostly have been using 4o in the last month and supplementing it with DeepSeek in the last week, for when I need advanced reasoning (with web search in DeepSeek as a bonus).

The killer app had better start giving better value, or I'd gladly pay the same amount of DeepSeek for unlimited access if they decided to charge.


You can't be a killer app if your competitor is just as good and free.

For me ChatGPT was not that useful for work, the killer app was Cursor. It’ll be similar for other industries, it needs to be integrated directly in core business apps.

Killer app for what platform?

Can you unpack why you think there'll be defensible moats at the application layer?

(I thought you had this exactly right when I read it, but I kept noodling it while I brushed my teeth and now I'm not so sure llms won't just prove hard to build durable margins at meaningful volume on?)


> The vast majority of our work is already automated to the point where most non-manual workers are paid for the formulation of problems (with people), social alignment in their solutions, ownership of decision-making / risk, action under risk, and so on.

I agree. That's why I think the next step is automating trivial physical tasks, i.e. robotics, not automating nontrivial knowledge tasks.


oh, I think the horror element is there.

Really? Nobody is getting kicked to death.

Absolutely horrifying. Well done.


"Thanks, I hate it"


This was my first thought too. AFAIK each layer encodes different information, and it's not clear that the last layer would be able to communicate well with the first layer without substantial retraining.

Like in a CNN for instance, if you fed later representations back in to the first kernels they wouldn't be able to find anything meaningful because it's not the image anymore, it's some latent representation of the image that the early kernels aren't trained on.


Brand new account + strange overreaction = maybe bot?


I make a new account any time my karma gets too high. I’d rather not be a sheep, thank you.


I understand the complaints, but I don't think the Freakonomics podcast should necessarily be expected to meet the high standard of peer review. Is it really that bad for Levitt to take published research at face value, trusting that gatekeepers upriver have done their due diligence?


> Is it really that bad for Levitt to take published research at face value, trusting that gatekeepers upriver have done their due diligence?

Yes, it is. There is a reason the reproducibility/replication crisis, especially in the social sciences, is such a hot topic. The podcast doesn't need to "meet the high standard of peer review", but there are plenty of published objections and discussions about Langer's unexpected results, and Levitt should have reviewed that and brought that up before essentially saying "Wow, your results are so unexpected! OK I'm pretty sold!"


>there are plenty of published objections and discussions about Langer's unexpected results, and Levitt should have reviewed that and brought that up

Is that expected of Freakonomics? I don't know how much rigor they do with their interview subjects, nor how much of a subect matter expert they are when it comes to pushing back.


They like to entertain crazy theories, but there’s a cost, as has been observed multiple times in the past. I do still like to listen to Steven.

I think the whole problem is how he presents the podcast as being very factual, data driven and scientific and on the other end he just lack rigour in some cases - like this one.

Basic research has become rare in journalism, but they either should stop pretending to be data driven or should do their homework.


The Frakonomics brand leans more into the info side of infotainment. Having listened to the show, they also lean into their academic backgrounds, so yes. This isn't WTF with Marc Maron, but even he famously excused himself to do some research when he found out he was interviewing the "other" Kevin McDonald.


> Is that expected of Freakonomics?

Umm, of course? Shouldn't that be expected of any interviewer? I mean, they invited a guest onto their show specifically because they keep coming up with unexpected results - shouldn't they have done at least a little bit of their homework to see why a gaggle of people are condemning their results as non-reproducible?


> Shouldn't that be expected of any interviewer?

No? Imagine how ridiculous that would become if interviewers actually followed that logic. "Great gameplay out there, <insert professional sports star>, but nevermind the sport we are all watching, my research identified that you erroneously wrote 1+1=3 in Kindergarten. What was your thought process?"

The podcast in question is known as "People I (Mostly) Admire" from the Freakonomics podcast network. The name should tell you that it is going to be about the people, not diving deep into their work. Perhaps there is room for a Podcast that scrutinizes the work of scientists, but one that literally tells you right in its name that it is going to be about people is probably not it.


Your example completely and ridiculously mischaracterizes my point.

A better example, to piggyback off your sports analogy: Suppose a podcast titled "People I (Mostly) Admire" decided to interview Barry Bonds, and the interviewer asked "Wow, how did you get to be so good in the second half of your career?" and Bonds responded "Just a lot of hard work!" Yeah, I would totally expect the interviewer to push back at that point and say "So, your steroid use didn't have anything to do with it?"

Point being, I'm not asking the interviewer to be knowledgeable about the subject's kindergarten grades. I do think they should do some basic, cursory research about the specific topic and subject they brought the interviewer on to talk about in the first place.


> I would totally expect the interviewer to push back

Are you confusing expectation with desire? I can understand why you might prefer to listen to a podcast like that – and nothing says you can't – but that isn't necessarily on brand with the specific product in question.

In the same vein, you might prefer fine dining, but you wouldn't expect McDonalds to offer you fine dining. It is quite clearly not the product they sell.

So, I guess the question is: What is it about "People I (Mostly) Admire" that has given you the impression that it is normally the metaphorical fine dining restaurant and not the McDonalds it turned out to be here?


Are you like the king of awful, straw-man analogies or something? Will just say I think your attempt to redefine this podcast and the Freakonomics brand to just "light, fluffy entertainment" is BS. These other comments put it better:

https://news.ycombinator.com/item?id=41975615

https://news.ycombinator.com/item?id=41975342


> Are you like the king of awful, straw-man analogies or something?

Yes...? Comes with not understanding the subject very well. I mean, logically, if I were an expert I wouldn't be here wasting my time talking about what I already know, would I? That would be a pointless waste of time. Obviously if I am going to talk about something I am going to struggle to talk about it in an effort to learn.

> These other comments put it better:

These other comments don't even try to answer the question...? Wrong links? Perhaps I didn't explain myself well enough? I can try again: What is it about this particular podcast that has given you the impression that it normally asks the hard hitting questions? Be specific.


The type of journalism that involves saying "This person makes an incredible claim" and then goes on to allow the person to present said claims uncritically is called "tabloid journalism[1]." Yes, I would expect a podcast hosted by a NYT Journalist and University of Chicago Economist to have higher standards, particularly in the field of academic research.

1: https://en.wikipedia.org/wiki/Tabloid_journalism


That's a fun tangent, but doesn't answer the question. What in particular about this podcast has indicated that it is not "tabloid journalism"? You clearly recognize that tabloid journalism exists, so you know that this podcast could theoretically intend to be. But what, specifically, has indicated that it normally isn't?

The background of the people involved is irrelevant to the nature of the product. Someone who works on developing a cure for cancer by day can very well go home and build a fart app at night. There is no reason why you have to constrain yourself to just one thing.


Great comedy show


There's a lot of ground between "the high standard of peer review" and "tak[ing] published research at face value."

The former is impractical for a lot of formats (ie podcasts) but the latter is clearly harmful in the context of a popular podcast or some other medium that amplifies the dubious message.


What is the value in listening to an educational podcast if I cannot be certain that the material is factual?


What use is the value of reading a journal if I cannot be certain that the material is reliably peer reviewed?

I'm not sure why the podcast author is being held to a standard that should be levied to other matter experts, that come way before he ever reaches out for an interview.


This is my main point. Seems like gripes about the quality of published research should be directed toward the publisher.


It is factual that Langer performed a study in which X was done, Y was measured and Z was concluded.

What is less clear is whether X was good experimental design, whether the measurements of Y were appropriate, relevant and correct, and thus whether or not Z can be concluded.


Certainty is too high of a bar.


There is a lot of space left till the high standards of peer review. Some would call what they are doing spreading misinformation lol.


Making things easier to google is IMO the least-impressive use of LLMs. I'm still waiting for a Roomba I can have a conversation with.


Numberphile did a video on this a while back. https://youtu.be/mceaM2_zQd8?si=0xcOAoF-Bn1Z8nrO


I guess we need to find a way to incentivize good practice rather than interesting results? Turns out that science is so hard that people cheat.


Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: