Hacker News new | past | comments | ask | show | jobs | submit | more SergeyHack's comments login

I've heard a number of times that proper water fasting can heal diabetes.

P.S. Absolute fasting can be dangerous.


I like this one "Why Android Will Soon Become Apple’s Most Important Thing I’ve Ever Done."


The 6th point is very interesting:

"Adversarial examples generalize across models trained to perform the same task, even if those models have different architectures and were trained on a different training set."


"HOW TO CRAWL MARKETS" section has good tips for general crawling as well.


To really feel the difference you have to work on complex logic.

It's easy to be under illusion that you can create everything using a verbose language. At some point the complex system you work on just gets extremely hard to reason about.

For verbose languages you reach this point considerably faster.


I wonder what new possibilities for total spying this approach enables.


You can try to find a dataset that contains the equal number of positive and negative documents (sentences, etc.) and use it as the validation set. I.e. to tune your hyperparameters on it.

In the simple case your hyperparameter can be α in

sentiment = (α * #positive_matches - #negative_matches) / (document_word_count)


I like your "2." suggestion more, because the initial sentiment score distribution can be not normal.

So there is an option to try making it normal by taking logarithm for example and calculating mean, etc. after that.


I would still expect it to tend towards the normal distribution across a large set of documents. If you model positive and negative word counts as a binomial distribution, you have the the difference of two samples from different binomial distributions which would still tend towards normal (I think, though I'm not 100% sure, certainly it's true within my experience). A logarithm would skew away from positive to negative sentiment and is undefined for negative values.


It only tends to a Normal distribution if you estimate P(negative|matches in -ve list) & P(positive|matches in +ve list) with an unbiased, consistent estimator.

A simple 1-gram model like in the question does not model many complexities of natural language e.g. negation ("not bad" != "bad") so you would expect your estimator to over-represent the dictionary with more words that are equal to their adverb-adjusted equivalent. e.g. "not bad" can be described as 'terrible' more readily than 'very good' can be described as excellent since people assign a hyperbolic weighting to their own happiness (utility theory 101)

The sentiment would only tend to a normal distribution if we had perfect estimators for document sentiment which requires advanced POS tagging and models more complex than a 1-gram bag of words aggregation :)


I wonder if doodling in this context is no more than a weak alternative to walking or other physical activity.



Consider applying for YC's W25 batch! Applications are open till Nov 12.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: