Hacker News new | past | comments | ask | show | jobs | submit login

Might as well have a quick discussion here. How's everyone finding the new models?

4-Turbo is a bit worse than 4 for my NLP work. But it's so much cheaper that I'll probably move every pipeline to using that. Depending on the exact problem it can even be comparable in quality/price to 3.5-turbo. However the fact that output tokens are limited to 4096 is a big asterisk on the 128k context.




I haven't really kept up with the updates, but I've noticed 4's in-conversation memory seems worse lately.


Here we go with these “it looks worse” just like a month back when people feel it it was worse


It's probably a smaller, updated (distilled?) version of gpt-4 model given the price decrease, speed increase, and turbo name. Why wouldn't you expect it to be slightly worse? We saw the same thing with 3-davinci and 3.5-turbo.

I'm not going off pure feelings either. I have benchmarks in place comparing pipeline outputs to ground truth. But like I said, it's comparable enough to 4, at a much lower price, making it a great model.

Edit: After the outage, the outputs are better wtf. Nvm it has some variance even at temp = 0. I should use a fixed seed.


I am betting on a combination of quantization and speculative sampling with a distilled smaller set of models: https://arxiv.org/pdf/2302.01318.pdf


For decades true AI was always 7 years away. Now it’s always two weeks ago.


Because it was worse.



4-Turbo is much faster, which for my use case is very important. Wish we could get more than 100 requests per day.. Is the limit higher when you have a higher usage tier?


Yeah it gets way higher. We were capped to 40k T/m when our org spend was under $250. Now it's 300k.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: