Hacker News new | past | comments | ask | show | jobs | submit login

Ah ok, I'm primarily using the API already. One interesting thing is that the GPT-3.5 "product" is much faster, but looks to be using a different model in the request, their encoding model iirc. I wonder if they are now using embeddings to cache results to reduce load on the real models when they can?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: