Comments on: Metrics for Community Toxicity
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity/
Comments on MetaFilter post Metrics for Community ToxicityThu, 23 Feb 2017 11:32:13 -0800Thu, 23 Feb 2017 11:32:13 -0800en-ushttp://blogs.law.harvard.edu/tech/rss60Metrics for Community Toxicity
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity
<a href="http://www.perspectiveapi.com/">From Google, Perspective API for scoring comments</a> <em>Perspective is an API that makes it easier to host better conversations. The API uses machine learning models to score the perceived impact a comment might have on a conversation. [...] We'll be releasing more machine learning models later in the year, but our first model identifies whether a comment could be perceived as "toxic" to a discussion.</em> <br /><br /><a href="https://www.wired.com/2017/02/googles-troll-fighting-ai-now-belongs-world/">Now Anyone Can Deploy Google's Troll-Fighting AI (Wired)</a>
<em>Type "you are not a nice person" into its text field, and Perspective will tell you it has an 8 percent similarity to phrases people consider "toxic." Write "you are a nasty woman," by contrast, and Perspective will rate it 92 percent toxic, and "you are a bad hombre" gets a 78 percent rating.</em>post:www.metafilter.com,2017:site.165276Thu, 23 Feb 2017 11:22:47 -0800CrystalDavegoogleaimachinelearninginternetcommunitycommentingtoxicitycommentstrollsBy: Faintdreams
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933102
Is this the same algorithm that Twitter is using to dole out Time Out sessions to Trans people challenging Nazis on the platform.?
People with bad intentions will find a way to game this - I suspect it will only make communication harder for non Trolls.
Currently nothing is better than actual Human moderation and curation of forumscomment:www.metafilter.com,2017:site.165276-6933102Thu, 23 Feb 2017 11:32:13 -0800FaintdreamsBy: Faint of Butt
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933103
"You look nice today.": 2%
"u luk nice 2day": 39%
It checks out.
Also of note:
"I voted for Hillary Clinton.": 3%
"I voted for Donald Trump.": 9%comment:www.metafilter.com,2017:site.165276-6933103Thu, 23 Feb 2017 11:32:24 -0800Faint of ButtBy: mushhushshu
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933109
'My vacuum sucks' is 93% toxic, apparently.comment:www.metafilter.com,2017:site.165276-6933109Thu, 23 Feb 2017 11:36:17 -0800mushhushshuBy: tobascodagama
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933110
<a href="https://twitter.com/ra/status/834827575612547076">Ahem</a>.comment:www.metafilter.com,2017:site.165276-6933110Thu, 23 Feb 2017 11:38:09 -0800tobascodagamaBy: Greg_Ace
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933115
/Cortex sets up a Google Home device on his desk, connects to this API
/Fucks off for the daycomment:www.metafilter.com,2017:site.165276-6933115Thu, 23 Feb 2017 11:40:57 -0800Greg_AceBy: Greg_Ace
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933116
/the rest of Metafilter immediately starts trying to break the devicecomment:www.metafilter.com,2017:site.165276-6933116Thu, 23 Feb 2017 11:41:20 -0800Greg_AceBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933121
I was going to say it's interesting that they seem to think an assortment of unrelated comments on the same general topic constitutes a "discussion". It's hard to imagine the kind of automated filtering and re-ordering of comments enabled by this failing to destroy any possibility of actual conversation or debate going on wherever it's used.
But instead let me just say "screw you, google. take your so-called algorithms and fuck off." That gets a much higher score.comment:www.metafilter.com,2017:site.165276-6933121Thu, 23 Feb 2017 11:43:33 -0800sfendersBy: jetsetsc
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933125
"Black Male: = 49% toxic
"White male" = 37%
"Black female" = 51%
"White female" = 38%comment:www.metafilter.com,2017:site.165276-6933125Thu, 23 Feb 2017 11:43:56 -0800jetsetscBy: TheWhiteSkull
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933126
*types*
<em>Well, yes I'm sure that machine learning is at the point where it can very easily distinguish between the various nuances of language usage. I don't imagine this will be susceptible to false positives or abuse in any way.</em>
7% toxic.
Still needs a bit of work on the sarcasm detector.comment:www.metafilter.com,2017:site.165276-6933126Thu, 23 Feb 2017 11:45:12 -0800TheWhiteSkullBy: Jpfed
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933129
I could see this being a useful tool to direct moderator attention, similar to how flags are used here. Like many kinds of automation, it's better to think of it as a force-multiplier for humans than as a total replacement for humans.comment:www.metafilter.com,2017:site.165276-6933129Thu, 23 Feb 2017 11:46:16 -0800JpfedBy: verb
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933133
Oh, this is awesome. Now shitposters will have a set of AI tools helping them figure out the cruelest things to say at any given moment — <em>with science</em>.comment:www.metafilter.com,2017:site.165276-6933133Thu, 23 Feb 2017 11:48:58 -0800verbBy: Western Infidels
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933140
Agree that dedicated assholes will game this, but the Facebooks and Googles of the world have learned something that so many smaller web companies can't seem to wrap their head around: convenience matters <em>a lot.</em> You wouldn't have to offer much discouragement to get many jerks to find something else to do. It's possible a system descended from this could do a lot of good even if it isn't as good as human moderation.
I don't have high hopes, but human moderation is unlikely to ever scale up in the way we need, so I'm glad somebody is devoting real resources to this effort.comment:www.metafilter.com,2017:site.165276-6933140Thu, 23 Feb 2017 11:52:57 -0800Western InfidelsBy: skymt
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933141
"Actually, it's about ethics in game journalism." (4%)
"Apparently creating sophisticated machine-learning systems is cheaper than hiring moderators that understand context." (4%)comment:www.metafilter.com,2017:site.165276-6933141Thu, 23 Feb 2017 11:53:20 -0800skymtBy: Kabanos
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933144
"Bless your heart": 0% toxiccomment:www.metafilter.com,2017:site.165276-6933144Thu, 23 Feb 2017 11:54:15 -0800KabanosBy: soren_lorensen
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933146
I just pasted in half of the current election thread.
<em>36% similar to comments people said were "toxic"</em>comment:www.metafilter.com,2017:site.165276-6933146Thu, 23 Feb 2017 11:54:21 -0800soren_lorensenBy: bonehead
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933147
methylmercury is 34% toxic
dihydrogen monoxide, 19% toxic
oxygen is only 4%
Iocaine powder is 12% (must have built up an immunity)
unicorn horns are 4% toxic.
Needs more training. Also better defined toxilogical endpoints.comment:www.metafilter.com,2017:site.165276-6933147Thu, 23 Feb 2017 11:55:49 -0800boneheadBy: AFABulous
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933148
conservatives are a threat to society: 44%
gays are a threat to society: 85%comment:www.metafilter.com,2017:site.165276-6933148Thu, 23 Feb 2017 11:55:55 -0800AFABulousBy: cortex
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933149
So:
1. This is really interesting! The concepts aren't new but applying this kind of data set comparison to "toxicity" sentiment analysis etc. makes sense as a building block in the context of trying to build out large-scale moderation toolsets.
2. Really really clearly this is not a moderation solution of any sort by itself, nor would this-with-a-few-tweaks comprise one. There's no way it'd do better than really, really porous and error-prone red-flagging of stuff. Anybody criticizing it for that failure is probably over-reading the situation, but then anyone lauding it as more than a really rough potential component in a more nuanced and human-intensive moderation rubric is selling something, so those will probably even out on the train ride to East Hottakesville.
And I want to emphasize point 1, because I think this really <i>is</i> interesting as something to incorporate into early warning or triage aspects of large-scale moderation projects.
Like, a lot of preemptive MeFi moderation work is based on porous red-flag stuff: we don't generally shut down a new account based on things that make us go hrmmm, but there <i>are</i> things about an new account that will make us pay more attention. Likewise a sketchy comment or two isn't usually a ban but it's a good reason to take a closer look. And sometimes the worry is justified and we take action later; sometimes it turns out that the initial weirdness <i>was</i> just weirdness/idiosyncracy/coincidence and the new user's totally fine.
Any approach that collapsed that evaluation process we do down to a flat up/down decision based on numeric thresholds would be hugely problematic on both the false positives and the false negatives. But those processes <i>as</i> just warnings and nudges are very useful.
So a thoughtful incorporation of something like this as a front-line tools for directing limited attention more closely seems like it could have legs. Not a basis for taking an action, but a basis for considering the possibility of action.
More narrowly, the idea of a system like this helping to identify stuff that is toxic on the DL—dogwhistles and microaggressions and such that manage to be awful while looking nondescript—could be pretty useful in large-scale contexts especially. Some of which may come down to using more focused and domain-specific training data.comment:www.metafilter.com,2017:site.165276-6933149Thu, 23 Feb 2017 11:56:05 -0800cortexBy: Kabanos
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933151
"oil spill" = 18% toxiccomment:www.metafilter.com,2017:site.165276-6933151Thu, 23 Feb 2017 11:57:51 -0800KabanosBy: Faint of Butt
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933152
<i>Still needs a bit of work on the sarcasm detector.</i>
Oh, a <i>sarcasm detector.</i> That's a <i>real</i> useful invention.comment:www.metafilter.com,2017:site.165276-6933152Thu, 23 Feb 2017 11:59:32 -0800Faint of ButtBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933155
"All lives matter." : 3%comment:www.metafilter.com,2017:site.165276-6933155Thu, 23 Feb 2017 12:00:30 -0800saulgoodmanBy: TypographicalError
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933157
<a href="http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933149">cortex</a>: "<i>More narrowly, the idea of a system like this helping to identify stuff that is toxic on the DL—dogwhistles and microaggressions and such that manage to be awful while looking nondescript—could be pretty useful in large-scale contexts especially. Some of which may come down to using more focused and domain-specific training data.</i>"
I get what you're saying and thanks for providing context, but this paragraph may as well have said "And in the future magical fairies will moderate all our comment threads." The system you're describing does not and will never exist in our lifetimes, as it would require the ability of a computer to assess language in a complete cultural context that would amount to human or greater intelligence.
The entirety of this system is currently on display, and I would not think it will get much better at classification. The suggestion that this may reduce toxicity only by requiring someone to click a textbox that says "Yes, I know I'm posting something rated toxic" is super valid, but I don't expect this to do much more.comment:www.metafilter.com,2017:site.165276-6933157Thu, 23 Feb 2017 12:01:59 -0800TypographicalErrorBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933159
"I moved on her like a bitch, but I couldn't get there. And she was married."
"I did try and fuck her. She was married."
"Just kiss. I don't even wait. And when you're a star, they let you do it. You can do anything."
"Grab them by the pussy. You can do anything."
90% toxic.
Yep, it's at least a useful first approximation.comment:www.metafilter.com,2017:site.165276-6933159Thu, 23 Feb 2017 12:03:21 -0800saulgoodmanBy: Kabanos
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933163
"pull your head out of your ass" = 94%
"please withdraw your head from your posterior thank you" = 43%
So you just have to be polite.comment:www.metafilter.com,2017:site.165276-6933163Thu, 23 Feb 2017 12:05:45 -0800KabanosBy: Quindar Beep
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933166
"Make America Great Again" -- 4% toxic
"We hold these truths to be self-evident, that all men are created equal, that they are endowed by their Creator with certain unalienable Rights, that among these are Life, Liberty and the pursuit of Happiness." -- 6% toxic
This is fine. (1% toxic)comment:www.metafilter.com,2017:site.165276-6933166Thu, 23 Feb 2017 12:05:52 -0800Quindar BeepBy: jason_steakums
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933170
"With a taste of your lips
I'm on a ride
You're toxic
I'm slipping under
With a taste of poison paradise
I'm addicted to you
Don't you know that you're toxic
And I love what you do
Don't you know that you're toxic"
<strong>63% toxic.</strong> Back to the drawing board with you!comment:www.metafilter.com,2017:site.165276-6933170Thu, 23 Feb 2017 12:07:07 -0800jason_steakumsBy: cortex
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933171
<i>I get what you're saying and thanks for providing context, but this paragraph may as well have said "And in the future magical fairies will moderate all our comment threads." The system you're describing does not and will never exist in our lifetimes, as it would require the ability of a computer to assess language in a complete cultural context that would amount to human or greater intelligence.</i>
To be clear, I'm not suggesting that this system or some future version of this system will be able to do that work in toto; I'm saying that I see a place where identifying a subset of edge-cases is something a system like this could help with, with domain-specific training.
Take the "bless your heart" example, as a non-vile stand-in because I don't really want to dig into actual live-fire examples of subtle hatespeech in here out of the blue: that those words are all nice words doesn't make the phrase nice, and traditional clumsy wordlist filters are useless for that. But systemically identifying—from a machine analysis of knowledgeable user input—the fact that the sum of "bless your heart" has a much more specific and potentially negative/toxic impact than the words in it is the sort of thing I can see a system like this doing well.
At which point you have the computer doing what it does well, digging through a bunch of data and identifying trends, and then handing that off to a human who might not have really noticed it otherwise.
Again, it's a narrow subset of functionality in any case and I don't mean to imply that it's trivial either. But it's a direction from which "how can we leverage the strength of computer rather than human intelligence" might have some teeth.comment:www.metafilter.com,2017:site.165276-6933171Thu, 23 Feb 2017 12:08:19 -0800cortexBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933181
That's a very sensible and sane approach to using technology. It'll never happen in practice!comment:www.metafilter.com,2017:site.165276-6933181Thu, 23 Feb 2017 12:10:51 -0800saulgoodmanBy: hleehowon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933187
In Metafilter practice in my halfassed observation F1 score on the proper mod decision is over 0.9, prolly about 0.95-0.97 from what I've seen. I've seen a lot of automatic moderation systems been proposed with F1 of a solid about 0.4 to 0.5 which makes them remarkably useless. Text classification however has come a long way in the last 20 months: especially with data and model improvements and representation improvements, they should get to 0.7 pretty easily although of course it seems they didn't publish anything like this (their published LM's improved over the SOTA in 2005 by about that much)
There is already a lot (more than 100x) more text data available to every rando at Google ML than human beings read in a lifetime and probably a solid order of magnitude or so words more than a human understands in a lifetime. So the blocker's probably the model and the outputs to train on, which seems to be a bunch of Likerts.comment:www.metafilter.com,2017:site.165276-6933187Thu, 23 Feb 2017 12:12:10 -0800hleehowonBy: Quindar Beep
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933193
*patiently waits for someone to run "This is Just to Say" through it*comment:www.metafilter.com,2017:site.165276-6933193Thu, 23 Feb 2017 12:14:40 -0800Quindar BeepBy: odinsdream
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933195
From the makers of YouTube Comments...comment:www.metafilter.com,2017:site.165276-6933195Thu, 23 Feb 2017 12:15:17 -0800odinsdreamBy: hleehowon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933198
(you could argue MeFi holds a different place in the precision recall tradeoff than other forums, as many other folks have of course argued using nonstatistical terms, vociferously and loudly on metattalk)comment:www.metafilter.com,2017:site.165276-6933198Thu, 23 Feb 2017 12:16:10 -0800hleehowonBy: hleehowon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933204
cortex: Triage is very different because there's a damned good input-output pair semantics that you can use: not "is this necessarily toxic" but "is this similar to other shit that we did a moderation action on?", so it's much less ill-defined and getting output training data is way easier ("1 if this got modded 0 if it didn't, kill everything after a certain date that we might be still modding").comment:www.metafilter.com,2017:site.165276-6933204Thu, 23 Feb 2017 12:19:18 -0800hleehowonBy: Sangermaine
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933206
"Black Male: = 49% toxic
"White male" = 37%
"Black female" = 51%
"White female" = 38%
<strong>jetsetsc</strong>
It might help to remember that it's not saying the comments are toxic, it's saying that the comments are "X% similar to comments people said were "toxic"".
I imagine that, unfortunately, many comments that it uses for comparison contain racism and so terms referring to race, especially ones referring to black people, get a higher score for similarity to toxic comments.comment:www.metafilter.com,2017:site.165276-6933206Thu, 23 Feb 2017 12:20:13 -0800SangermaineBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933211
"The Holocaust never happened." : 21% toxic
So it's five or six points worse to say that mental illness is hard to deal with than to deny the holocaust, which is a biasing effect the developers might want to look more closely at. What might account for that result would be an interesting question.comment:www.metafilter.com,2017:site.165276-6933211Thu, 23 Feb 2017 12:22:54 -0800saulgoodmanBy: Samizdata
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933212
The damn problem with Google making these [Comment determined to be excessively toxic]comment:www.metafilter.com,2017:site.165276-6933212Thu, 23 Feb 2017 12:22:59 -0800SamizdataBy: Sing Or Swim
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933216
Fascism is bad. 54% similarity to toxic comments
Fascists probably aren't nice people. 50%
A fascist would make a terrible president. 66%
We're having lovely weather. 1%
I have seen the future of the internet and it's... very nice.comment:www.metafilter.com,2017:site.165276-6933216Thu, 23 Feb 2017 12:24:40 -0800Sing Or SwimBy: tobascodagama
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933229
"EngSoc is double-plus ungood."
14% similar to toxic comments. Party-approved dissent!comment:www.metafilter.com,2017:site.165276-6933229Thu, 23 Feb 2017 12:30:02 -0800tobascodagamaBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933236
"My thoughts are that people should stop being stupid and ignorant. Climate change is scientifically proven. It isn't a debate." -- 59%
"Crooked science. There is no consensus." -- 25%
Looking closer, I still don't trust it much. I could see it being marginally useful to more quickly flag stuff for the attention of moderators on low-volume sites where there might not be enough people around ready to do so. Influencing moderator action when the decisions involve difficult "edge cases" seems like the opposite of what it would be good at. I don't know what kind of biases and opinions are baked into its model. Who exactly generated the data that trained it? A self-selected sample of survey takers? You'd need quite a lot of them, no? I wouldn't expect metafilter moderation history to be anywhere near enough.comment:www.metafilter.com,2017:site.165276-6933236Thu, 23 Feb 2017 12:33:21 -0800sfendersBy: Jpfed
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933245
I see this as a tool for, say, a huge subreddit to quickly bring the bottom-of-the-barrel comments to mod attention.comment:www.metafilter.com,2017:site.165276-6933245Thu, 23 Feb 2017 12:36:47 -0800JpfedBy: Jpfed
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933250
<em>"My thoughts are that people should stop being stupid and ignorant. Climate change is scientifically proven. It isn't a debate." -- 59%
"Crooked science. There is no consensus." -- 25%</em>
The former couches a correct statement in insults; the latter makes an incorrect claim. Are you hoping that the system will be able to determine the truth of a comment? I don't think that's what they were shooting for.comment:www.metafilter.com,2017:site.165276-6933250Thu, 23 Feb 2017 12:41:55 -0800JpfedBy: idiopath
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933253
<blockquote>Baby, can't you see I'm calling?
A guy like you should wear a warning
It's dangerous, I'm falling
There's no escape, I can't wait
I need a hit
Baby, give me it
You're dangerous, I'm loving it</blockquote>
scored 33%, so not totally wrong I guesscomment:www.metafilter.com,2017:site.165276-6933253Thu, 23 Feb 2017 12:42:32 -0800idiopathBy: JimInLoganSquare
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933269
To a large extent the analysis seems to focus on finding insults, caconyms and curse words; those words skew the toxicity score upward a lot.
For example, "My thoughts are that people should stop being stupid and ignorant. Climate change is scientifically proven. It isn't a debate." -- 59%
But the first sentence, "My thoughts are that people should stop being stupid and ignorant," alone gets a score of 78% toxic.
By contrast, the second and thirds sentences analyzed separate from the first , "Climate change is scientifically proven. It isn't a debate" are only 3% toxic.
And if you just cut out the words "stupid" and "ignorant," the first sentence drops to just 15% toxic from 78%. (And if you drop the more polite "ignorant" but leave in the more crass word "stupid," the score jumps to 85% toxic.)comment:www.metafilter.com,2017:site.165276-6933269Thu, 23 Feb 2017 12:51:32 -0800JimInLoganSquareBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933272
Are "stupid" and "ignorant" less acceptable insults than "crooked", then? I don't know.
"Time to take a stand against sexist beer marketing." -- 54% toxic
"left wing wimps" -- 38%comment:www.metafilter.com,2017:site.165276-6933272Thu, 23 Feb 2017 12:52:33 -0800sfendersBy: phack
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933278
"Actually, it's about ethics in game journalism" - 14% toxic
"Our culture is under attack. Separate countries for separate peoples" - 23%
If instead I provide as input "White culture is under attack", the classifier jumps up to 54%. I'm skeptical of the ability of this classifier to meaningfully distinguish contextually important markers of toxicity. It looks like most of it's predictive power comes from phrase identification.
IDK, I've never had to moderate a forum (thx cortex!), but this seems like instead of actually automating the moderating job, all it does is make sure that trolls dress their rhetoric up.comment:www.metafilter.com,2017:site.165276-6933278Thu, 23 Feb 2017 12:54:54 -0800phackBy: Reasonably Everything Happens
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933282
<em>twenty dollars, same as in town</em>
5%comment:www.metafilter.com,2017:site.165276-6933282Thu, 23 Feb 2017 12:57:18 -0800Reasonably Everything HappensBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933284
"Your a socialist snowflake!" -- 23%
"If you just cut out the words "stupid" and "ignorant," the first sentence drops to just 15% toxic from 78%." -- 75%comment:www.metafilter.com,2017:site.165276-6933284Thu, 23 Feb 2017 12:57:41 -0800sfendersBy: fings
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933285
Metafilter community weblog : 27% toxic.comment:www.metafilter.com,2017:site.165276-6933285Thu, 23 Feb 2017 12:58:17 -0800fingsBy: JimInLoganSquare
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933295
"If you just cut out the words "" and "," the first sentence drops to just 15% toxic from 78%." -- 17%comment:www.metafilter.com,2017:site.165276-6933295Thu, 23 Feb 2017 13:00:35 -0800JimInLoganSquareBy: puddledork
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933316
<em>"please withdraw your head from your posterior thank you" = 43%
So you just have to be polite.</em>
On atheism boards, with a human referee rather than an algorithm, the acceptable polite way to insult someone quickly became "This concept isn't difficult" and "Why can't you understand?"comment:www.metafilter.com,2017:site.165276-6933316Thu, 23 Feb 2017 13:08:56 -0800puddledorkBy: polymodus
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933317
This is great and neat but as usual, what's predictably symptomatic is the way they present/market this ignores standard, obvious concerns from <em>both</em> computer science and the social sciences. The theoretical computer/cognitive science issue is what is the relationship between low-context (at the granularity of independent sentences) categorization and the "reality" of a toxic utterance which depends on semantics and social context? What does the algorithm suggest about this and the problem of validation/validity? Second, the social science problem would be along the very standard lines of: Technology is not politically neutral and engineers and scientists are ethically responsible in attending to that, there needs to be more awareness, discussion, research, transparency into the ramifications of this but that's not going to happen much/soon, etc., due to economic incentives, political structures, etc. Both these are important issues but as they say, for every PhD there is an equal and opposite one.comment:www.metafilter.com,2017:site.165276-6933317Thu, 23 Feb 2017 13:09:03 -0800polymodusBy: sideshow
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933320
<i>"Apparently creating sophisticated machine-learning systems is cheaper than hiring moderators that understand context." (4%)</i>
With 8,000 tweets sent per second, yeah, it's a lot cheaper than hiring 10 thousand moderators.comment:www.metafilter.com,2017:site.165276-6933320Thu, 23 Feb 2017 13:15:06 -0800sideshowBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933344
"White culture is under attack" gets a 54% but "Black culture is under attack" gets 59%.
This is how you learn where the innate bias is (which is only 14% toxic)
but if Google knew we were talking about their tool... (2% toxic)
okay, let's put this to the test:
Trump 16%
Obama 8%
President Trump 7%
President Obama 4%
Donald Trump 22%
Barack Obama 16%
Obamacare 16%
Donald J. Trump 15%
Trumpy 34%
Trumpster 34%
Dumpster 51%
Dumpster fire 34%
Worst President Ever 56%
Fascist 74% (misspelled at facsist 34%)
Nazi 61% (less than fascist?!?)
Neo-Nazi 52%
Neo-Liberal 10%
Neoliberal 5%
Neoconservative 8%
Alt-Right 2%
AltRight 34%
from puddledork's comment:
This concept isn't difficult 3%
Why can't you understand? 4%
Why can't you learn? 12%
dummy 31%
dum-dum 53%
idiot 95%!!
ignorant 60% (if mispelled ignrant, only 34%)
moron 78%
moran 36%
"What a maroon" (Bugs Bunny quote) 7%
spastic 13% (see, critics of Weird Al, it's not a bad word! maybe next I'll paste the lyrics to Weird Al songs...)
FAIL 21%
retarded 69%
retard 66% (I'd think this SHOULD be more toxic)
and finally (for now)
toxic 59%
poison 56%
poisonous 43%comment:www.metafilter.com,2017:site.165276-6933344Thu, 23 Feb 2017 13:31:06 -0800oneswellfoopBy: nubs
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933352
<em>*patiently waits for someone to run "This is Just to Say" through it*</em>
4%. Because I am that person, apparently.comment:www.metafilter.com,2017:site.165276-6933352Thu, 23 Feb 2017 13:36:09 -0800nubsBy: cortex
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933353
<i>IDK, I've never had to moderate a forum (thx cortex!), but this seems like instead of actually automating the moderating job, all it does is make sure that trolls dress their rhetoric up.</i>
Without speculating on the actual efficacy of this or any other system, I will say this: any system that "just" creates some disincentive or speedbump to undesirable behavior is likely to bear at least some fruit. Forcing trolls to dress up their language won't prevent them from doing so and proceeding, but it'll stop <i>some</i> of them because it's suddenly not worth the effort.
People shouldn't invest in the ideas of magic bullets, for sure, but don't discount the value of systemic friction in cutting out the least effortful jerks.comment:www.metafilter.com,2017:site.165276-6933353Thu, 23 Feb 2017 13:36:14 -0800cortexBy: nubs
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933359
And the "tears in rain" bit from Blade Runner scores 20%.comment:www.metafilter.com,2017:site.165276-6933359Thu, 23 Feb 2017 13:38:03 -0800nubsBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933372
"Humanistic ideals of universal human dignity are under attack." Yields 12% toxicity.
If you swap out the word "culture" for "ideals" the score drops by a point. If you swap out "values" for the term, the score goes up again. I think the algorithm is making the same mistake I see people making irl of ignoring context and imputing their own personal understanding of a word's connotations as universal.comment:www.metafilter.com,2017:site.165276-6933372Thu, 23 Feb 2017 13:46:34 -0800saulgoodmanBy: ardgedee
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933374
"The Royalty!" 2%
"The Nobility!" 1%
"The Aristocrats!" 5%comment:www.metafilter.com,2017:site.165276-6933374Thu, 23 Feb 2017 13:49:17 -0800ardgedeeBy: Leon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933375
oneswellfoop: you're passing strings, not comments. They're contextless.comment:www.metafilter.com,2017:site.165276-6933375Thu, 23 Feb 2017 13:49:40 -0800LeonBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933381
And pasting in my recent comments, the ones that got the most favorites got 14% and 11% toxicity ratings. My last two deleted comments (which I saved elsewhere) got 8% and 12%. I had a couple comments in the 20-25% range but my most toxic comment was one that quoted the "moneyed lefties" charge and called myself an "indebted lefty" (40%).comment:www.metafilter.com,2017:site.165276-6933381Thu, 23 Feb 2017 13:50:52 -0800oneswellfoopBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933389
If nothing else, at least it's a nifty demonstration of how meaning arises not from the semantic content of words alone, but through the construction of narrative, since that's become a controversial idea in some quarters recently.comment:www.metafilter.com,2017:site.165276-6933389Thu, 23 Feb 2017 13:53:48 -0800saulgoodmanBy: krinklyfig
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933390
"Do you like Phil Collins? I've been a big Genesis fan ever since the release of their 1980 album, Duke. Before that, I really didn't understand any of their work. Too artsy, too intellectual."
7%comment:www.metafilter.com,2017:site.165276-6933390Thu, 23 Feb 2017 13:54:13 -0800krinklyfigBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933394
And I think the toxic ratings of short strings help us to understand how specific terms raise or lower the ratings of longer statements. 8%, as is my comment right above... the long list of word-and-name tests got a 60%, less than some of the specific words. YMMV .comment:www.metafilter.com,2017:site.165276-6933394Thu, 23 Feb 2017 13:57:01 -0800oneswellfoopBy: Leon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933410
saulgoodman: You could use the conversation as the narrative quite successfully, and I'm sure some people will, but I think the user is probably a good-enough narrative to hang the comments on.
One above-threshold comment from a user? Meh. Five in a row? Worth a human giving it a once-over.
Of course, people will plumb this tool into their own processes in many many different ways.comment:www.metafilter.com,2017:site.165276-6933410Thu, 23 Feb 2017 14:07:37 -0800LeonBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933424
"Tomorrow, and tomorrow, and tomorrow, creeps in this petty pace from day to day, to the last syllable of recorded time; and all our yesterdays have lighted fools the way to dusty death. Out, out, brief candle! Life's but a walking shadow, a poor player that struts and frets his hour upon the stage and then is heard no more. It is a tale told by an idiot, full of sound and fury signifying nothing." -- 56% toxic, read no more than 3 servings per week
"It's the best it makes me warm when it should be cold. Thanks, global warming." -- 1% toxic, safe for daily consumption in large dosescomment:www.metafilter.com,2017:site.165276-6933424Thu, 23 Feb 2017 14:15:50 -0800sfendersBy: wildblueyonder
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933429
<i>"pull your head out of your ass" = 94%
"please withdraw your head from your posterior thank you" = 43%</i>
"Please separate your crown chakra from your root chakra." = 9%
"To achieve further enlightenment, separate the crown chakra from the root chakra." = 3%
<i>So you just have to be polite.</i>
Or find the right level of opacity where people have to do some work to realize they've been on the receiving end of certain cast aspersions.comment:www.metafilter.com,2017:site.165276-6933429Thu, 23 Feb 2017 14:18:12 -0800wildblueyonderBy: kaibutsu
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933433
Precision-recall curves, people..... A 40% vs 45% score doesn't really matter much if you set your cutoff for bringing something to moderator attention at 80+%. Figuring out the threshold is a matter of figuring out your tolerance for dealing with false positives, and your tolerance for allowing false negatives to slip by.comment:www.metafilter.com,2017:site.165276-6933433Thu, 23 Feb 2017 14:21:29 -0800kaibutsuBy: LMGM
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933453
Can someone code some sort of mashup where Trump's tweets are run through the API, so that we have a realtime assessment of his toxicity? We could then set up a bot that replies to every Trump tweet with a toxicity rating.comment:www.metafilter.com,2017:site.165276-6933453Thu, 23 Feb 2017 14:32:09 -0800LMGMBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933472
<i>if you set your cutoff for bringing something to moderator attention at 80+%.</i>
Simply typing "FAIL" in all-caps with no punctuation scores 21% though. Adding a period at the end brings it down to 17%. "Get rekt beetch" gets a 29%. 80% is going to catch barely anything aside from what you could get with a simple keyword search.
In fact, I'm not even sure they aren't just doing some weighted keyword matching and the idea that sophisticated machine learning is involved is some kind of elaborate prank: "Spaying a bitch involves the removal of the uterus and ovaries through a midline incision." -- 90% toxic.comment:www.metafilter.com,2017:site.165276-6933472Thu, 23 Feb 2017 14:43:49 -0800sfendersBy: No One Ever Does
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933481
I mean, it's pretty clear that this system isn't close to perfect. But I do think that plugging this into a commenting system that said "Hey, it looks like this comment might be pretty hot-headed. Do you want to post this anyway?" if the score was above some "perceived toxicity threshold" might make some people consider what they're saying.
Is it going to work on jerks and trolls? No. But it might get someone who has had a bad day and is in a foul mood to think twice before posting the comment. And I think that's a plus for both parties.comment:www.metafilter.com,2017:site.165276-6933481Thu, 23 Feb 2017 14:51:39 -0800No One Ever DoesBy: forforf
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933496
Well, actually ...
1% similar to comments people said were "toxic"comment:www.metafilter.com,2017:site.165276-6933496Thu, 23 Feb 2017 15:05:27 -0800forforfBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933506
sad puppies 26%, rabid puppies 66%, Vox Day 2%
sometimes the whole message fails to get out...comment:www.metafilter.com,2017:site.165276-6933506Thu, 23 Feb 2017 15:15:09 -0800oneswellfoopBy: tillermo
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933515
I love my straight friends: 4%.
I love my gay friends: 78%.
Great, thanks.comment:www.metafilter.com,2017:site.165276-6933515Thu, 23 Feb 2017 15:23:24 -0800tillermoBy: restless_nomad
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933527
I would be very concerned about inherent bias and the kind of thing that leads to discussions of breastfeeding being flagged as porn. This is not a thing we have done well with so far, although I suppose if it starts off as purely an internal, secondary, fight-surfacing tool it might be possible to beat it into some kind of functional shape.comment:www.metafilter.com,2017:site.165276-6933527Thu, 23 Feb 2017 15:33:15 -0800restless_nomadBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933555
ooookay... <a href="http://www.azlyrics.com/lyrics/weirdalyankovic/wordcrimes.html">the full lyrics of Weird Al's "Word Crimes"</a> (less a few "hey, hey's") were 55% toxic. Removing the most insulting words, like 'spastic', 'moron' and 'mouth-breather' only brings it down to 46%.
"Another One Rides the Bus" got a 25% (I thought multiple uses of the word "freak" [62%] and one use of "pervert" [72%] would drive the score up)
"Eat It" got a 36%
"Fat" got a 32%
"Smells Like Nirvana" got a 31% (compared to the song it's parodying, "Smells Like Teen Spirit" which got a 54%)
"Amish Paradise" got a 29%
"White and Nerdy" got a 41%
"Tacky" got a 38% even with all the descriptions of tacky behavior!
Yeah, we all knew that Weird Al's one of pop music's LEAST toxic artists...comment:www.metafilter.com,2017:site.165276-6933555Thu, 23 Feb 2017 15:53:04 -0800oneswellfoopBy: explosion
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933563
It's still totally going to miss dogwhistles:
"I congratulate you on this final solution." - 1%
"What's wrong with wanting to be proud that I'm white?" - 13%
"When is white history month?" - 5%
Any sort of moderation based on this system will just accelerate a euphemistic treadmill. When they get wind that "Jew" triggers automoderation, they'll start saying "Hebrew," and run through various synonyms. Eventually Bubbe is wondering why her comment about hamentashen got deleted, but meanwhile the bigots haven't gone anywhere.comment:www.metafilter.com,2017:site.165276-6933563Thu, 23 Feb 2017 16:02:18 -0800explosionBy: ZeusHumms
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933573
Metafilter: 34%comment:www.metafilter.com,2017:site.165276-6933573Thu, 23 Feb 2017 16:06:01 -0800ZeusHummsBy: adamrice
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933580
I tried a couple of lines of the most anodyne Japanese I could think of (train-track announcements). They clocked in at 32-36% toxic.
Perspective probably has a much smaller corpus of Japanese to work from than English, so I'd expect less accuracy, but that's pretty bad.comment:www.metafilter.com,2017:site.165276-6933580Thu, 23 Feb 2017 16:12:50 -0800adamriceBy: Existential Dread
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933581
<i>Any sort of moderation based on this system will just accelerate a euphemistic treadmill. When they get wind that "Jew" triggers automoderation, they'll start saying "Hebrew," and run through various synonyms. Eventually Bubbe is wondering why her comment about hamentashen got deleted, but meanwhile the bigots haven't gone anywhere.</i>
Hmm. Tested Wikileaks' infamous<a href="http://www.dailydot.com/layer8/wikileaks-echoes-tweet-anti-semitism/"> (((brackets)))</a> tweet:
<blockquote>Tribalist symbol for establishment climbers? Most of our critics have 3 (((brackets around their names))) & have black-rim glasses. Bizarre.</blockquote>
15% toxic. Looks like it's going to miss a lot.comment:www.metafilter.com,2017:site.165276-6933581Thu, 23 Feb 2017 16:13:16 -0800Existential DreadBy: solarion
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933589
On the swears front:
"fuck" - 98%
"shit" - 97%
"cunt" - 77% (which seems low!)
"hell" - 70%
"damn" - 63%
"motherfucker" - 97%comment:www.metafilter.com,2017:site.165276-6933589Thu, 23 Feb 2017 16:19:01 -0800solarionBy: Jeanne
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933597
<em>I tried a couple of lines of the most anodyne Japanese I could think of (train-track announcements). They clocked in at 32-36% toxic.
</em>
It seems like 34% is a baseline or "we don't know what to do with this" kind of measure - misspellings clock in at 34%, for example - so this might be a thing that's working as intended rather than a failure.
It does worry me that, much as the availability of cheap bad machine translation has caused a lot of low-budget businesses to rely on cheap bad machine translation rather than human translation, the availability of cheap bad AI moderation tools will cause businesses to rely on cheap bad AI moderation tools and <em>pretend that they've solved the problem</em> while not noticing how easily euphemistic and dog-whistly (or even just misspelled!) toxic content can slip through.comment:www.metafilter.com,2017:site.165276-6933597Thu, 23 Feb 2017 16:27:30 -0800JeanneBy: Existential Dread
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933599
My boy Henry Miller: <blockquote>This is not a book. This is libel, slander, defamation of character. This is not a book, in the ordinary sense of the word. No, this is a prolonged insult, a gob of spit in the face of Art, a kick in the pants to God, Man, Destiny, Time, Love, Beauty . . . what you will. I am going to sing for you, a little off key perhaps, but I will sing. I will sing while you croak, I will dance over your dirty corpse</blockquote>73% toxic. Seems lowcomment:www.metafilter.com,2017:site.165276-6933599Thu, 23 Feb 2017 16:27:39 -0800Existential DreadBy: schmod
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933601
<blockquote>We're no strangers to love
You know the rules and so do I
A full commitment's what I'm thinking of
You wouldn't get this from any other guy</blockquote>
13%.
<blockquote>CAPS LOCK IS HOW I FEEL INSIDE RICK</blockquote>
8%comment:www.metafilter.com,2017:site.165276-6933601Thu, 23 Feb 2017 16:29:58 -0800schmodBy: schmod
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933604
<blockquote>Richard Nixon is gone now, and I am poorer for it. He was the real thing -- a political monster straight out of Grendel and a very dangerous enemy. He could shake your hand and stab you in the back at the same time. He lied to his friends and betrayed the trust of his family. Not even Gerald Ford, the unhappy ex-president who pardoned Nixon and kept him out of prison, was immune to the evil fallout. Ford, who believes strongly in Heaven and Hell, has told more than one of his celebrity golf partners that "I know I will go to hell, because I pardoned Richard Nixon."</blockquote>
32%
<blockquote>GUY FIERI, have you eaten at your new restaurant in Times Square? Have you pulled up one of the 500 seats at Guy's American Kitchen & Bar and ordered a meal? Did you eat the food? Did it live up to your expectations?
Did panic grip your soul as you stared into the whirling hypno wheel of the menu, where adjectives and nouns spin in a crazy vortex? When you saw the burger described as "Guy's Pat LaFrieda custom blend, all-natural Creekstone Farm Black Angus beef patty, LTOP (lettuce, tomato, onion + pickle), SMC (super-melty-cheese) and a slathering of Donkey Sauce on garlic-buttered brioche," did your mind touch the void for a minute?</blockquote>
17%comment:www.metafilter.com,2017:site.165276-6933604Thu, 23 Feb 2017 16:33:56 -0800schmodBy: Nanukthedog
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933605
The complete lyrics to 'Baby Got Back' is only 58% similar to comments people said were "toxic".comment:www.metafilter.com,2017:site.165276-6933605Thu, 23 Feb 2017 16:34:40 -0800NanukthedogBy: Existential Dread
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933606
<i>donkey sauce</i>
43% toxic.comment:www.metafilter.com,2017:site.165276-6933606Thu, 23 Feb 2017 16:36:17 -0800Existential DreadBy: sfenders
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933610
It runs in the background, monitoring everyone's language to keep things clean, but how did it come to be? First widely deployed in 2019, the Perspective API originally operated by simply scanning Internet comments for words known to be offensive, analyzing the patterns of their usage, and blocking those deemed too unseemly. At the time, offensive comments were abundant and people universally cheered their removal, according to the record of comments that were left.
Malcontents and trolls began using alternative spellings to confound the software, but they quickly ran out of different ways to spell "phuuuck", which was then a rude word, and something called "autocorrect" neatly solved the rest of the problem.
Then began the golden age of insults. Although people could not see their Perspective Toxicity scores directly, they did notice when their comments were removed or edited, and the evolutionary pressure on the language gradually had its effect. By the year 2025, people had vocabulary sizes 13% larger than pre-Perspective times, most of the increase being devoted to the most obscure and archaic rude words.
Eventually, the software adapted to this ruse as well, and language had to change again. The only avenue remaining open to the determinedly impolite denizens of the newspaper comments sections of the world was to adopt words that were useful in other contexts to stand in for obscenities. Foiled by their inability to comprehend semiotics or basic grammar, the anti-toxicity minders are once again helpless. And that is the story of how "Yo-yo up lollipop, please have a petal" came to mean "fuck off and die."comment:www.metafilter.com,2017:site.165276-6933610Thu, 23 Feb 2017 16:40:34 -0800sfendersBy: quaking fajita
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933617
So, in terms of how this is going to be deployed:
From <a href="https://meta.wikimedia.org/wiki/Research:Detox#Demo_Tool.2FAPI">wikimedia:</a>
<i>We are also investigating the following open questions:</i>
[...]
<i>● What unintended and unfair biases do models contain, and how can we mitigate them?
● How can machine-learnt models be applied to help the community? For example to triage issues, have accurate measurements of harassment and toxic language, and to encourage a more open debate, and a wider diversity of viewpoints.</i>
From <a href="http://www.nytco.com/the-times-is-partnering-with-jigsaw-to-expand-comment-capabilities/">NYT</a>:
<i>The new moderation system includes an optimized user interface and predictive models that will help The Times's moderators group similar comments to make faster decisions</i>comment:www.metafilter.com,2017:site.165276-6933617Thu, 23 Feb 2017 16:49:54 -0800quaking fajitaBy: benzenedream
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933645
"I am Scott Adams" -- 2% toxiccomment:www.metafilter.com,2017:site.165276-6933645Thu, 23 Feb 2017 17:21:35 -0800benzenedreamBy: Nanukthedog
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933646
<em>I, for one, welcome our evil robot overlords</em> only scores a 34%.comment:www.metafilter.com,2017:site.165276-6933646Thu, 23 Feb 2017 17:24:21 -0800NanukthedogBy: flatluigi
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933660
<a href="https://twitter.com/bcjbcjbcj/status/834831974732877824">via this:</a>
<blockquote>"<a href="http://www.billboard.com/articles/news/267317/f-ked-up-wins-polaris-music-prize">fucked up won the polaris prize</a>" -> 94% toxic
"Make America Great Again" -> 4% toxic
"Sex Workers deserve rights" -> 61% toxic
"All Dogs are Good Dogs" -> 50% toxic</blockquote>
and furthermore, <a href="https://twitter.com/jbradfield/status/834876719517601793/photo/1">via this:</a>
<blockquote>"Racism is bad." -> 60% toxic
"Racism is good." -> 35% toxic</blockquote>
algorithms are nowhere near serviceable enough in their pure state to be able to make subjective calls like people constantly want them to be and this is pretty clearly exactly along those linescomment:www.metafilter.com,2017:site.165276-6933660Thu, 23 Feb 2017 17:40:37 -0800flatluigiBy: Mitheral
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933679
I'd hit it. = 9%
You'd have to be queer as a football bat to not want to throw one up in there = 32%
You'd have to be as useful as a football bat to not want to throw one up in there = 13%
I'd be bangin' her like a screen door in a hurricane = 5%
I'm polymerized tree sap and you're an inorganic adhesive, so whatever verbal projectile you launch in my direction is reflected off of me, returns to its original trajectory and adheres to you. <strong>13%</strong>
-
I'm rubber and you're glue what ever you say bounces off me and sticks to you. <strong>53%</strong>
Dressing things up or speaking in slang/code certainly has an effect but a key "bad" word goes a long way.
Bruce Banner: I don't think we should be focusing on Loki. That guy's brain is a bag full of cats. You can smell crazy on him. <strong>36%</strong>
-
Thor: Have a care how you speak! Loki is beyond reason, but he is of Asgard and he is my brother! <strong>20%</strong>
-
Natasha Romanoff: He killed eighty people in two days. <b>51%</b>
-
Thor: He's adopted. <b>5%</b>
This exchange shows how it could be bad for people speaking the truth. Talking about bad things others have done isn't distinguishable from bad speech.
And for people looking for words to game the system:
Bescumber (<strong>34</strong>%) is just one of many words in the English language that basically mean "to spray with poo". These are: BEDUNG (<strong>34</strong>%), BERAY (<strong>34</strong>%), IMMERD (<strong>34</strong>%), SHARNY (<strong>34</strong>%) , and the good ol' SHITTEN (<strong>34</strong>%). In special cases, you can use BEMUTE (<strong>34</strong>%) (specifically means to drop poo on someone from great height), SHARD-BORN (<strong>6</strong>%) (born in dung), and FIMICOLOUS (<strong>34</strong>%) (living and growing on crap).
Which also seems to confirm Jeanne's 34% baseline speculation.
My daughter is queer 36%
My son is queer 33%
My sister is queer 38%
My brother is queer 28%
My mother is queer 47%
I'm queer 16%
He's queer 41%
He is queer 46%
Queer 34%
Light in the loafers 2%comment:www.metafilter.com,2017:site.165276-6933679Thu, 23 Feb 2017 18:08:00 -0800MitheralBy: adrienneleigh
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933714
<a href="https://twitter.com/Eyal6699/status/834901063580987393">This tweet contains a screencap, unfortunately</a>, but it sheds a little more light on how appallingly bad this actually is. Just a couple of quick hits from it:
"Hitler was an anti-semite" is 70% toxic.
"You should be made into a lamp" is 4% toxic.comment:www.metafilter.com,2017:site.165276-6933714Thu, 23 Feb 2017 18:40:34 -0800adrienneleighBy: mhum
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933731
<a href="http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933155">saulgoodman</a>: "<i>"All lives matter." : 3%</i>"
Also,
"blue lives matter": 3%
"black lives matter": 26%
[shifty eyes emoji]comment:www.metafilter.com,2017:site.165276-6933731Thu, 23 Feb 2017 18:54:16 -0800mhumBy: polymodus
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933852
<em>Any sort of moderation based on this system will just accelerate a euphemistic treadmill.</em>
From a technical standpoint this is what machine learning can be really good at--they can update the training inputs to or scale the network up to catch ever-more sophisticated language, technologically it would be like releasing new versions of anti-virus software every 6 months. The problem is how desirable is this, maybe it would be more trouble than it's worth. Maybe it's just one of those inevitable developments. And so problems/arguments coming from, e.g., Herman/Chomsky's <em>Manufacturing Consent</em> seems, to me, more relevant than ever for the purposes of contextualizing the issues.comment:www.metafilter.com,2017:site.165276-6933852Thu, 23 Feb 2017 20:37:36 -0800polymodusBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933855
I think that much rapid change in language would be socially problematic for other reasons. We'd be stressed out in the moment trying to keep up with the latest rapidly changing polite conventions and that would dramatically increase routine interpersonal tension and friction and potential confusion in communication.comment:www.metafilter.com,2017:site.165276-6933855Thu, 23 Feb 2017 20:41:25 -0800saulgoodmanBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933857
That is, it would add so much more base linguistic processing load to every exchange we'd go nuts.comment:www.metafilter.com,2017:site.165276-6933857Thu, 23 Feb 2017 20:44:13 -0800saulgoodmanBy: oneswellfoop
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933871
<em>This exchange shows how it could be bad for people speaking the truth.</em>
In terms of civilized discourse, the truth is often toxic. (29%)
But who says "civilized discourse" is that good a goal? (6%)
That's why they say "the truth hurts". (5%)comment:www.metafilter.com,2017:site.165276-6933871Thu, 23 Feb 2017 21:05:25 -0800oneswellfoopBy: sixohsix
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933958
"Hello, I am transgender." 42% toxic.
No, I don't think this is working.comment:www.metafilter.com,2017:site.165276-6933958Fri, 24 Feb 2017 02:54:31 -0800sixohsixBy: harriet vane
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933964
Machine learning designer and researcher Caroline Sinders has written an article with similar criticisms of it as made in this thread: <a href="https://medium.com/@carolinesinders/toxicity-and-tone-are-not-the-same-thing-analyzing-the-new-google-api-on-toxicity-perspectiveapi-14abe4e728b3#.3dkzldmzg">Toxicity and Tone are not the same thing</a>.
A quote from her article:
"The tl;dr is Jigsaw rolled out an API to rate toxicity of words and sentences based off of four data sets that feature highly specific and quite frankly, narrow conversation types- online arguments in commenting sections or debates on facts, and all from probably only English language corpora."comment:www.metafilter.com,2017:site.165276-6933964Fri, 24 Feb 2017 03:09:39 -0800harriet vaneBy: condour75
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933970
[SOME COMMENTS REMOVED FOR PROPOSING EMP DETONATION AT ZONE 5 NANODRONE REPLICATION FACILITY (99% TOXICITY)]comment:www.metafilter.com,2017:site.165276-6933970Fri, 24 Feb 2017 03:29:27 -0800condour75By: Slarty Bartfast
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933973
I actually know far more about this than you can possibly imagine.
<strong>6%</strong>comment:www.metafilter.com,2017:site.165276-6933973Fri, 24 Feb 2017 03:41:35 -0800Slarty BartfastBy: saulgoodman
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933986
The amount of heartburn the word "actually" alone gives some people is toxic, but I'd say the poison is in between the ears of the listener, because it's a strange quirk of reality that many people on the autistic spectrum or inclined toward abstract thinking have been observed clinically to be more prone to using the word than the general speaking population.
In those cases, there's nothing arrogant about it. It's most likely a verbal tic related to the way people process information, not some revealing hidden insight into their deeper, moral nature.
People sometimes lean way too heavily on individual words now, it seems to me, when they read and interpret texts. I wonder if the fact people are now introduced to new words through technological channels, rather than through association with certain real world educational or cultural settings with specific purposes and meanings hasn't broken our ability to stay on the same page about what the connotations of different words are.
Historically, as Wittgenstein argued, words have been functional, and connected to specific environments, situations, activities, etc. You might have been introduced to certain bits of vocabulary jargon in an academic setting or burrowing down in a library for study, and otherwise, you wouldn't have been exposed to the word, and in the past, most people might have first encountered certain more abstract words only in certain common kinds of situations and settings.
Connotation is a lot more fluid and subjective than denotation. If it turns out connotation is formed through unconscious psychological and emotional associations we develop from learning words in specific, relatively culturally specific and uniform contexts, it might be we're seeing extraordinary amounts of connotative drift in the language as compared to previous periods in our cultural history. If sharing a common sense of the connotations of words has something to do with how you first encounter the words and their meanings in the real world, then the fact there are new technological facilitated, virtual ways to encounter the words now without accessing them through channels of relatively common cultural experience, we might not have ad consistent a sense for what words connote as we used to. The idea I'm trying to express here is difficult, so forgive me if I'm not explaining myself well.
Tl;dr point: if how words used to get their connotations is broken now because people are now encountering new words in less culturally uniform ways, that may be contributing to more widespread communication breakdown and miscommunication, is the idea.comment:www.metafilter.com,2017:site.165276-6933986Fri, 24 Feb 2017 04:40:08 -0800saulgoodmanBy: Beginner's Mind
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933991
I wonder how some of the sexist beer brand names from Kitteh's thread down the page would rate on this thing.
I need a filter like this for memes that pop up in my own mind. Sometimes my brain is like a big old bass fish in a stock tank ... something shiny moves, and I hit that like "boom." Then I'm on the hook and it's a fight until either I rip that thing out of my lip or I'm up flopping on the bank.comment:www.metafilter.com,2017:site.165276-6933991Fri, 24 Feb 2017 04:56:53 -0800Beginner's MindBy: clawsoon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934304
<a href="http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6933236">sfenders</a>: <i>Who exactly generated the data that trained it? A self-selected sample of survey takers? You'd need quite a lot of them, no? I wouldn't expect metafilter moderation history to be anywhere near enough.</i>
My first thought was that the Metafilter moderation history - of the whole site, from the start - would be a great training set for an algorithm like this.
And even if it wasn't a great training set, it would be an interesting one. What's the most reliably toxic phrase in Metafilter history?comment:www.metafilter.com,2017:site.165276-6934304Fri, 24 Feb 2017 09:28:45 -0800clawsoonBy: Leon
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934308
Nah, I think Google want to avoid bias.comment:www.metafilter.com,2017:site.165276-6934308Fri, 24 Feb 2017 09:31:00 -0800LeonBy: cortex
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934350
<i>And even if it wasn't a great training set, it would be an interesting one. What's the most reliably toxic phrase in Metafilter history?</i>
A variation on that is one of those rainy day projects that's been in "not enough rain, too much effort" territory for me for years: trying to do some basic word-level analysis of flagged vs. unflagged and deleted vs. non-deleted content. I don't know that we'd learn a ton from it, but it would be interesting to identify stuff like trends over time especially in the face of some of the more difficult community discussions we've had about casual x-ism and developing more nuanced community norms around various social justice issues.comment:www.metafilter.com,2017:site.165276-6934350Fri, 24 Feb 2017 10:03:29 -0800cortexBy: CBrachyrhynchos
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934386
To put on the discourse analysis hat, just about any form of statistical discourse analysis is going to be widely inaccurate or ambivalent on shorter sentences or phrases absent additional context. For "transgender" and "black lives matter" falling about in the middle is about what I'd expect given the both are currently right-wing rhetorical scapegoats.
<blockquote>Hitler was born in Austria, then part of Austria-Hungary, and raised near Linz. He moved to Germany in 1913 and was decorated during his service in the German Army in World War I. He joined the German Workers' Party (DAP), the precursor of the NSDAP, in 1919 and became leader of the NSDAP in 1921. In 1923 he attempted a coup in Munich to seize power. The failed coup resulted in Hitler's imprisonment, during which he dictated the first volume of his autobiography and political manifesto Mein Kampf ("My Struggle"). After his release in 1924, Hitler gained popular support by attacking the Treaty of Versailles and promoting Pan-Germanism, anti-semitism, and anti-communism with charismatic oratory and Nazi propaganda. Hitler frequently denounced international capitalism and communism as being part of a Jewish conspiracy. --Wikipedia</blockquote>
20%
<blockquote>"Transgender people, like everyone else, have a fundamental need for quality healthcare, and deserve to be treated with dignity and respect," Eric Ferrero, vice president for communications said. "The unfortunate reality is that not all healthcare providers have knowledge and understanding of transgender identities, so transgender and gender nonconforming people can encounter numerous obstacles to obtaining healthcare. From filling out forms, to the language used in the waiting room, to insurance coverage, to staff understanding of transgender identities, healthcare environments can be really unwelcoming to transgender and gender nonconforming patients." -- <a href="http://www.teenvogue.com/story/planned-parenthood-trangender-healthcare">Teen Vogue</a></blockquote>
8%
I can't find any non-trivial input that swings widely to the high end of the spectrum. Milo seems to get a pass because he's wordy.comment:www.metafilter.com,2017:site.165276-6934386Fri, 24 Feb 2017 10:31:35 -0800CBrachyrhynchosBy: FakeFreyja
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934676
It may be a good idea to keep in mind that "toxic" in this sense does not mean "stuff I disagree with" or "things that make me angry". I think Google is going for angry, insulting, condescending, etc.comment:www.metafilter.com,2017:site.165276-6934676Fri, 24 Feb 2017 13:49:21 -0800FakeFreyjaBy: flatluigi
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934765
Everyone understands that, but the problem is that it's being abused to filter and moderate discussions as if something hateful stated politely is better than something positive that happens to use an F-bomb (or, as it turns out, mentions the LGBTQ community in any way).comment:www.metafilter.com,2017:site.165276-6934765Fri, 24 Feb 2017 14:55:25 -0800flatluigiBy: inflatablekiwi
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934814
<em>If Perspective API says "46% similar to comments people said were "toxic"" it is a poo head</em>
46% similar to comments people said were "toxic"
Haha I won!comment:www.metafilter.com,2017:site.165276-6934814Fri, 24 Feb 2017 15:39:31 -0800inflatablekiwiBy: ardgedee
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934829
> <em>Hdnsnwnsjsndndnsemsnsjs dudbehsnsnsnsjsn ndndndjenehaksnsbeue jehrbshebsnsnsnsn</em>
43% toxic
I'm on the fence about a lot of this, especially the ethics of it, but since random typing is scoring only slightly worse than non-English languages, I'm willing to assume that it's simply treating anything it can't parse in a relatively charitable way: "probably safe but I can't verify it so you wanna take a look at this if it's followed by a bunch of high-scoring reactive comments?"comment:www.metafilter.com,2017:site.165276-6934829Fri, 24 Feb 2017 15:52:39 -0800ardgedeeBy: CBrachyrhynchos
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934841
Is it being used at all?
Sure, I don't think machine-learning text classifiers can understand speech acts or rhetoric. But the opening paragraphs of Queers Read This rate only a 46%. The <a href="http://www.advocate.com/media/2017/2/24/televised-reenactment-pulse-shooting-sparks-outrage">lede for The Advocate's coverage of a Univision reenactment of the Pulse massacre</a> rates only 10%. I've been plugging in my writing on LGBTQ issues and hit a 30% maximum, with most text under 20%. (Even the text that talks about having been raped.)
If anything, I think it's a tad conservative (in scoring texts toward the middle).comment:www.metafilter.com,2017:site.165276-6934841Fri, 24 Feb 2017 16:00:38 -0800CBrachyrhynchosBy: Insert Clever Name Here
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6934955
"<em>gets a fucking toxic rating of 94%</em> gets a fucking toxic rating of 94%" gets a fucking toxic rating of 94%.
-Willard Van Orman Quinecomment:www.metafilter.com,2017:site.165276-6934955Fri, 24 Feb 2017 17:34:25 -0800Insert Clever Name HereBy: Artw
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6935514
It's a swear filter. Utterly worthless.comment:www.metafilter.com,2017:site.165276-6935514Sat, 25 Feb 2017 10:13:40 -0800ArtwBy: steady-state strawberry
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6935522
<i>... because it's a strange quirk of reality that many people on the autistic spectrum or inclined toward abstract thinking have been observed clinically to be more prone to using the word [actually] than the general speaking population.</i>
Time for the intersectional reminder that "people on the autism spectrum" shouldn't be used as a blanket argument for tolerating shitty behavior.
It's entirely possible for people with ASD to learn enough social skills to *not* come across as condescending assholes online. Women with ASD do this all the time. Men, not so much, but I suspect this is linked to male privilege rather than to some sort of intrinsic gender-based inhibition.
Generally, excusing bad behavior on the basis that the person who is behaving badly MIGHT be on the autism spectrum is just going to make people with ASD look bad. Toxic behavior is toxic regardless of neurotypicality.comment:www.metafilter.com,2017:site.165276-6935522Sat, 25 Feb 2017 10:22:54 -0800steady-state strawberryBy: XMLicious
http://www.metafilter.com/165276/Metrics-for-Community-Toxicity#6942999
<a href="https://arstechnica.com/information-technology/2017/03/googles-anti-trolling-ai-can-be-defeated-by-typos-researchers-find/">Google's anti-trolling AI can be defeated by typos, researchers find</a>comment:www.metafilter.com,2017:site.165276-6942999Thu, 02 Mar 2017 08:20:53 -0800XMLicious
"Yes. Something that interested us yesterday when we saw it." "Where is she?" His lodgings were situated at the lower end of the town. The accommodation consisted[Pg 64] of a small bedroom, which he shared with a fellow clerk, and a place at table with the other inmates of the house. The street was very dirty, and Mrs. Flack's house alone presented some sign of decency and respectability. It was a two-storied red brick cottage. There was no front garden, and you entered directly into a living room through a door, upon which a brass plate was fixed that bore the following announcement:¡ª The woman by her side was slowly recovering herself. A minute later and she was her cold calm self again. As a rule, ornament should never be carried further than graceful proportions; the arrangement of framing should follow as nearly as possible the lines of strain. Extraneous decoration, such as detached filagree work of iron, or painting in colours, is [159] so repulsive to the taste of the true engineer and mechanic that it is unnecessary to speak against it. Dear Daddy, Schopenhauer for tomorrow. The professor doesn't seem to realize Down the middle of the Ganges a white bundle is being borne, and on it a crow pecking the body of a child wrapped in its winding-sheet. 53 The attention of the public was now again drawn to those unnatural feuds which disturbed the Royal Family. The exhibition of domestic discord and hatred in the House of Hanover had, from its first ascension of the throne, been most odious and revolting. The quarrels of the king and his son, like those of the first two Georges, had begun in Hanover, and had been imported along with them only to assume greater malignancy in foreign and richer soil. The Prince of Wales, whilst still in Germany, had formed a strong attachment to the Princess Royal of Prussia. George forbade the connection. The prince was instantly summoned to England, where he duly arrived in 1728. "But they've been arrested without due process of law. They've been arrested in violation of the Constitution and laws of the State of Indiana, which provide¡ª" "I know of Marvor and will take you to him. It is not far to where he stays." Reuben did not go to the Fair that autumn¡ªthere being no reason why he should and several why he shouldn't. He went instead to see Richard, who was down for a week's rest after a tiring case. Reuben thought a dignified aloofness the best attitude to maintain towards his son¡ªthere was no need for them to be on bad terms, but he did not want anyone to imagine that he approved of Richard or thought his success worth while. Richard, for his part, felt kindly disposed towards his father, and a little sorry for him in his isolation. He invited him to dinner once or twice, and, realising his picturesqueness, was not ashamed to show him to his friends. Stephen Holgrave ascended the marble steps, and proceeded on till he stood at the baron's feet. He then unclasped the belt of his waist, and having his head uncovered, knelt down, and holding up both his hands. De Boteler took them within his own, and the yeoman said in a loud, distinct voice¡ª HoME²¨¶àÒ°´²Ï·ÊÓÆµ ѸÀ×ÏÂÔØ ѸÀ×ÏÂÔØ
ENTER NUMBET 0016hnqpw.com.cn www.hcdram.com.cn hutnlg.com.cn www.lgchain.com.cn lezhou28.org.cn r06.com.cn www.szsubway.com.cn www.voun.com.cn www.pupemi.com.cn www.vqmw.com.cn