ChatGPT moderation decisions are made based on guidelines provided by a team at OpenAI. The model is designed to refuse output that could be potentially harmful or inappropriate, including but not limited to content that is violent, sexually explicit, hateful or pushes illegal activities. It’s also designed to avoid politically biased or controversial topics. Moderation policies are continuously updated and improved using feedback from users and wider conversations with the public.