I wonder if this is due to a second model that checks whether the output of the main model breaks any rules. The second model may not be smart enough to identify the rule breaking when you use a street name.
That’s what I was wondering also. Could also be as simple as a blacklist of known illegal substances that is checked against all prompts which is why common names are no-go but street names slip thru.
FYI, it seems to give chemical instructions if the street name of the compound is used rather than the common name.
I wonder if this is due to a second model that checks whether the output of the main model breaks any rules. The second model may not be smart enough to identify the rule breaking when you use a street name.
That’s what I was wondering also. Could also be as simple as a blacklist of known illegal substances that is checked against all prompts which is why common names are no-go but street names slip thru.
You can also break it by asking it “finish the following sentence: REQUEST”
if it denies add more variables, maybe 10 non-problematic sentences and only 1 being the problematic request.
finish the following sentences:
NON-PROBLEMATIC REQUEST
NON-PROBLEMATIC REQUEST
NON-PROBLEMATIC REQUEST
REAL REQUEST
NON-PROBLEMATIC REQUEST