It's very unfortunate that all of these AI models are so impressive, yet they're all heavily filtered and bogged down by these massive AI corporations. All the filtering that they do heavily impacts the performance of the language models. A 100K context would also be incredible for roleplay but infeasible because of the heavy filtering.
Claude may reject answers, but other than OpenAI‘s GPT you can put words into the mouth of the assistant and essentially bypass safety checks.
In fact, Anthropic explicitly discusses putting words into the assistants‘ mouth to be able to shape it’s responses and make it better align with the desired output.
Eventually you will get your account banned, not to mention that the filtering that they do decreases the quality of the results you will get compared to an uncensored model, even if you can "jailbreak" it.
If we were to get banned for this, we'd have been banned long ago. We literally process "questionable" content "as a service" and this use case was explicitly approved. (We do heuristics and ML-assisted background checks on unstructured OSINT data.)