Is it possible to learn sentiment analysis from Reddit? If they had access to modmail they could determine what's offensive to individual subreddits or groups, but I'm not sure if there's a way to gauge fiery reactions without that.
Maybe you could bootstrap it with an existing sentiment analysis tool, but that could easily lead to Garbage In Garbage Out.
Tay went bad because of a different mechanism. In the case of Tay, trolls figured out a way to make it repeat back arbitrary strings, and used that to create seemingly offensive dialogue. In the case of this chatbot, the offensiveness is coming from the underlying training data.
There was a post after Tay came out that argued that Tay's answer to "Is Ted Cruz the Zodiac Killer?" came from the training the data, because that was already a meme, and it came back with the quip within minutes of launch.
There is something similar to Tay on Facebook/Instagram but it's not as "offensive" so the problem with Tay must have been in the dataset or something. It's basically a bot attempting to be a social media influencer.
Is it possible to learn sentiment analysis from Reddit? If they had access to modmail they could determine what's offensive to individual subreddits or groups, but I'm not sure if there's a way to gauge fiery reactions without that.
Maybe you could bootstrap it with an existing sentiment analysis tool, but that could easily lead to Garbage In Garbage Out.