I’ve got a bot running/in development to detect and flag toxic content on Lemmy but I’d like to improve on it as I’m getting quite a few false positives. I think that part of the reason is that what constitutes toxic content often depends on the parent comment or post.

During a recent postgrad assignment I was taught (and saw for myself) that a bag of words model usually outperforms LSTM or transformer models for toxic text classification, so I’ve run with that, but I’m wondering if it was the right choice.

Does anyone have any ideas on what kind of model would be most suitable to include a parent as context, but to not explicitly consider whether the parent is toxic? I’m guessing some sort of transformer model, but I’m not quite sure how it might look/work.

  • @BluetreefrogOP
    link
    English
    11 year ago

    (“toxic”, “severe toxic”, “obscene”, “threat”, “insult”, and “identity hate”)

    You aren’t the author of Detoxify are you by any chance? It uses the same classifications. I was originally using it but switched to my own model as I really only needed binary classification and felt a new dataset that better suited Lemmy was needed anyway. I have 2 outputs (toxic and not-toxic).

    I’ve been building my own dataset as the existing ones on Huggingface seemed to contain a lot of content you might see on Twitter, and were a poor match for Lemmy. Having said that, I’ve generally avoided putting that sort of content into the dataset as I figured if I can’t easily decide if it’s toxic, then how could a model.

    Is there any way you can point me at examples difficult to classify? It would be a nice real world test to my stuff. Current iteration of model is very new and has not been tested in the wild.

    Here’s a few where I’ve had to go back to the parent comment or post to try and work out if it was toxic or not:

    • Do your research on the case and the answers will be very obvious. (What comment prompted this? Is it a trolling comment or a reasonable response to a trolling comment)
    • Because you’re a fascist. The fact that they disagree with you is secondary (Is the commenter calling another commenter a fascist, or continuing a discussion?)
    • Me tard, you tard, retard nation! (Is this a quote from a movie or TV show or an insult to another commenter? Not sure.)
    • Fuck you shoresy! (pretty sure this is a quote from a tv show)

    A comment must stand on it’s own, and it would put limits on results, the way I was planning to do it. I might be very wrong, your insight into this would be very helpful.

    I originally thought that, and I’m actively tuning my model to try and get the best results on the comment alone, but I don’t think I’ll ever get better than about 80% accuracy. I’ve come to the conclusion that those cases in the grey zone where toxic ~= not-toxic can only be resolved by looking upstream.

    • vluz
      link
      fedilink
      1
      edit-2
      1 year ago

      Oof, pop-culture references are hard and I had not considered that at all.
      Thanks for the examples, I’ll have a think on how to deal with those.

      My only insight is one you already had.
      Test at least the comment before, and then use the output to dampen or amplify the final result.
      Sorry for being no help at all.

      --

      My project is very basic but I’ll post it here for any insight you might get out of it.
      I teach Python in a variety of settings and this is part of a class.

      The data used is from Kaggle: https://www.kaggle.com/competitions/jigsaw-toxic-comment-classification-challenge/
      The original data came from Wikipedia toxic comments dataset.
      There is code too from several users, very helpful for some insight into the problem.

      Data is dirty and needs clean up so I’ve done so and posted result on HF here:
      https://huggingface.co/datasets/vluz/Tox

      Model is a very basic TensorFlow implementation intended for teaching TF basics.
      https://github.com/vluz/ToxTest
      Some of the helper scripts are very wonky, need fixing before I present this in class.

      Here are my weights after 30 epochs:
      https://huggingface.co/vluz/toxmodel30

      And here is it running on a HF space:
      https://huggingface.co/spaces/vluz/Tox