changes in a platform’s automated moderation
At a minimum the data needs to be used for Reinforcement learning Human Feedback (RLHF). This image was a violation, that image was not... Use that data to make the filters more accurate.It's good they're reporting it but something tells me they're still ingesting that material for their next model's training data.
Somewhat off topic, but that shit drives me nuts. People trust the output of these LLMs way too blindly.This morning I searched for a quote and Google's Gemini insisted it was from Die Hard. It wasn't, the AI was just confused, but it provided "helpful" analysis as to why the character said the quote which he didn't say. I kept looking because I am familiar with Die Hard and did not believe its lies, eventually finding the true source of the quote. Someone who was less familiar with the movie probably would have found Gemini's inaccurate summary plausible, since it made up new details to support its argument.

When you're not an expert in a topic, your brain automatically falls back on general heuristics of what expert answers are generally shaped like and sound like when deciding what outside information to trust. And hey, look at that, generally being shaped like and sounding like whatever you ask is exactly what LLMs are good at!Somewhat off topic, but that shit drives me nuts. People trust the output of these LLMs way too blindly.
Recently at work for compliance purposes I had to confirm whether a piece of test & measurement equipment had nonvolatile storage, and if so could being securely wiped or reset. Since this is a niche product from a small vendor, I couldn’t find any details and had to reach out to them.
While I was working on this back and forth via email over a couple weeks my coworker also offered to do some research on it. He came back to me a day later confident that the device didn’t have nonvolatile storage, only a ROM for the firmware and RAM for operation. When I asked where he got the detailed explanation he had sent me, he said ChatGPT. Ugh.
I am not making an official compliance decision based on ChatGPT output, especially since I was pretty confident it was wrong. (Spoiler alert: it was.)
Yikes that’s a scary thought. Could they really be that stupid?It's good they're reporting it but something tells me they're still ingesting that material for their next model's training data.
Having been on the vendor side of documenting storage in weird devices for secure erasure, yeah, no, you need reality (and proprietary internal design files) to get that result right!While I was working on this back and forth via email over a couple weeks my coworker also offered to do some research on it. He came back to me a day later confident that the device didn’t have nonvolatile storage, only a ROM for the firmware and RAM for operation. When I asked where he got the detailed explanation he had sent me, he said ChatGPT. Ugh.
I am not making an official compliance decision based on ChatGPT output, especially since I was pretty confident it was wrong. (Spoiler alert: it was.)
When you're not an expert in a topic, your brain automatically falls back on general heuristics of what expert answers are generally shaped like and sound like when deciding what outside information to trust. And hey, look at that, generally being shaped like and sounding like whatever you ask is exactly what LLMs are good at!
From what I remember, the original LAION set used by Stability for Stable Diffusion 1.x contained CSAM in some small amount, which was cleaned up pretty quick after that. When you're busy hoovering up all the information, you probably aren't paying too close an attention to what.Yikes that’s a scary thought. Could they really be that stupid?
You’d think that would result in some consequencesFrom what I remember, the original LAION set used by Stability for Stable Diffusion 1.x contained CSAM in some small amount, which was cleaned up pretty quick after that. When you're busy hoovering up all the information, you probably aren't paying too close an attention to what.
Somewhat off topic, but that shit drives me nuts. People trust the output of these LLMs way too blindly.
That's the neat part, you understand it perfectly.I don't understand. We have AGI and it's exploiting children?
Somewhat off topic, but that shit drives me nuts. People trust the output of these LLMs way too blindly.
Recently at work for compliance purposes I had to confirm whether a piece of test & measurement equipment had nonvolatile storage, and if so could being securely wiped or reset. Since this is a niche product from a small vendor, I couldn’t find any details and had to reach out to them.
While I was working on this back and forth via email over a couple weeks my coworker also offered to do some research on it. He came back to me a day later confident that the device didn’t have nonvolatile storage, only a ROM for the firmware and RAM for operation. When I asked where he got the detailed explanation he had sent me, he said ChatGPT. Ugh.
I am not making an official compliance decision based on ChatGPT output, especially since I was pretty confident it was wrong. (Spoiler alert: it was.)
When you're not an expert in a topic, your brain automatically falls back on general heuristics of what expert answers are generally shaped like and sound like when deciding what outside information to trust. And hey, look at that, generally being shaped like and sounding like whatever you ask is exactly what LLMs are good at!
From what I remember, the original LAION set used by Stability for Stable Diffusion 1.x contained CSAM in some small amount, which was cleaned up pretty quick after that. When you're busy hoovering up all the information, you probably aren't paying too close an attention to what.
It's an important reminder about privacy or lack thereof. Our personal information is being used against us now and almost certainly will be much more in the future unless we find a way to do something about it.While it's probably a good thing that CSAM content is being flagged and reported on the whole (if they're not false positives that is), this is also your regularly scheduled reminder that what you type into these services is in no way whatsoever private and you must assume it WILL be used by the AI companies. Either blindly for training purposes, or targeted if need be.
Now I wonder what happens if a FISA court order gets served to OpenAI...
There is a video making the rounds of a cop arresting a guy for trespassing based on a facial match software and the man was totally a false positive who had an ID and paystub and the cop still arrested the guy just because he blindly believed the software.Somewhat off topic, but that shit drives me nuts. People trust the output of these LLMs way too blindly.
Recently at work for compliance purposes I had to confirm whether a piece of test & measurement equipment had nonvolatile storage, and if so could being securely wiped or reset. Since this is a niche product from a small vendor, I couldn’t find any details and had to reach out to them.
While I was working on this back and forth via email over a couple weeks my coworker also offered to do some research on it. He came back to me a day later confident that the device didn’t have nonvolatile storage, only a ROM for the firmware and RAM for operation. When I asked where he got the detailed explanation he had sent me, he said ChatGPT. Ugh.
I am not making an official compliance decision based on ChatGPT output, especially since I was pretty confident it was wrong. (Spoiler alert: it was.)
Isn't that just humanity in general but specifically starting with the LLM/gen-AI developers working at OpenAI, Anthropic, Google, Microsoft et. al. ?They are so consistently wrong, I am beginning to think that there's a Dunning Kreuger virus out there that tends to infect clusters of GPUs running LLMs.
My goodness I hope OpenAI isn't using its own tech for this.
This morning I searched for a quote and Google's Gemini insisted it was from Die Hard. It wasn't, the AI was just confused, but it provided "helpful" analysis as to why the character said the quote which he didn't say. I kept looking because I am familiar with Die Hard and did not believe its lies, eventually finding the true source of the quote. Someone who was less familiar with the movie probably would have found Gemini's inaccurate summary plausible, since it made up new details to support its argument.
That's a minor goof by a different company, but it's a prime example of why this technology should not be used lightly by any company for something as important as moderation and reporting.
still better than reddit mods.it's a prime example of why this technology should not be used lightly by any company for something as important as moderation and reporting.