vluz, (edited )
vluz avatar

While designing a similar classifier, I've considered the idea of giving it the whole thread as "context" of sorts.
Not just the parent comment, the whole thread up to original post.

I've abandoned the idea.
A comment must stand on it's own, and it would put limits on results, the way I was planning to do it.
I might be very wrong, your insight into this would be very helpful.

My original idea was to go recursively trough the thread and test each comment individually.
Then I would influence the actual comment results with the combined results of it's parents.
No context during inference, just one comment at a time.

For example consider thread OP->C1->C2->C3.
My current model takes milliseconds per test with little resources used.
It would be ok up to very large threads but would contain a limit to save on answer time.
I want to determine if Comment 3 is toxic in the context of C2, C1, and OP.
Test C3, test C2, test C1, test OP. Save results.
My current model gives answer in several fields ("toxic", "severe toxic", "obscene", "threat", "insult", and "identity hate")
The idea was to then combine the results of each into a final result for C3.

How to combine? Haven't figure it out but it would be results manipulation instead of inference/context, etc.

Edit: Is there any way you can point me at examples difficult to classify? It would be a nice real world test to my stuff.
Current iteration of model is very new and has not been tested in the wild.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • machinelearning@lemmy.world
  • kavyap
  • rosin
  • modclub
  • DreamBathrooms
  • InstantRegret
  • magazineikmin
  • khanakhh
  • osvaldo12
  • tacticalgear
  • Youngstown
  • everett
  • slotface
  • ngwrru68w68
  • thenastyranch
  • JUstTest
  • mdbf
  • tester
  • GTA5RPClips
  • provamag3
  • Leos
  • Durango
  • ethstaker
  • cisconetworking
  • normalnudes
  • megavids
  • cubers
  • anitta
  • lostlight
  • All magazines