Hacker News new | past | comments | ask | show | jobs | submit login

So it samples data less than or equal to a second old and at a rate determined by the person making the query?

I wonder how often the data is inaccurate given the potentially low sample size?




Not sure where you got the "less than or equal to a second old"? Maybe I'm misunderstanding what you mean?

There is no single system-wide imposed sampling rate, so it's up to you to set the sampling rate based on what sort of queries you want to be able to do with good enough accuracy. We have 1:1 rate data for some things (say errors served on a particular service), while a ten or a hundred thousand to one data for other things where there are, say, tens of millions of log lines per second.


Ah yeah. I misread the pdf; the rows expire at millions per second and not after 1 second.

I was wondering about the size of the sampling error. Apparently it's negligible.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: