Experimenting with #OpenAI's moderation model. It does a really good job of extracting implication from a string of text, picking up hate, violence, and sexual cues.
Unfortunately this doesn't work for spam detection.
Something to explore later might be to see if these cues can tell us how understood or misunderstood something will be based on hate/violence/sex cues. For example, does authority come off as violent, or teasing come off as sexual? 🤔
A brief test of that, apparently being more assertive is LESS hateful and violent, but I suspect at this low of a sample size it's just rounding error.
Where as using the word "ass" scores us some sex points, and referring to said ass as "fat" scores us some hate points. It's still a super tiny amount, but if we reword our phrase more politely we don't register any of the metrics.
Neat. #OpenAI
I roughly understand how to use #OpenAI
's embeddings API, enough to not use it (yet).
The TL;DR is you feed it a text blob, and you get back a vector with ~1500 components (🤯). By itself the vector is meaningless, but patterns will emerge between similar data.
Example: spam posts should find themselves weighted towards one or more axis (angles?), but you need lots of non-spam data to find it. You could also do general search with it, but IMO it would be overkill (expensive).
Pretty cool tho.
@dpwiz Yes I can, and that will probably be where I end up. Akismet doesn't seem to have many false positives, but broadly it doesn't seem to catch much.
The current problem with embeddings is that I have well over 1 million pieces of content that could be scanned. That bill will add up fast, and I'd be wise to keep a copy of the output for future uses, which is potentially another 12 GB of data to store and archive...At least until GPT-3-ADA is deprecated, and compatibility breaks. 😅