I wonder if this is due to a second model that checks whether the output of the main model breaks any rules. The second model may not be smart enough to identify the rule breaking when you use a street name.
That’s what I was wondering also. Could also be as simple as a blacklist of known illegal substances that is checked against all prompts which is why common names are no-go but street names slip thru.
I wonder if this is due to a second model that checks whether the output of the main model breaks any rules. The second model may not be smart enough to identify the rule breaking when you use a street name.
That’s what I was wondering also. Could also be as simple as a blacklist of known illegal substances that is checked against all prompts which is why common names are no-go but street names slip thru.