Hacker News new | past | comments | ask | show | jobs | submit login

Did you chunk the articles? If so, in what way?



Yes. I split the text into sentence and append sentences to a chunk until the max context window is reached. The context window size is dynamic for each article so that each chunk is roughly the same size. Then I just do a mean pool of the chunks for each article.


Thanks for the answer.

Wikipedia has a lot of tables so I was wondering if content-aware sentence chunking would be good enough for Wikipedia.

https://www.pinecone.io/learn/chunking-strategies/


mwparserfromhell can parse the text content without including tables




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: