Preventing inappropriate and offensive results
GPT-3 will, at times, generate text that could be considered inappropriate to some users. This could be profanity, text with inappropriate racial, religious, or political language, or sexually inappropriate/explicit content, or text that is dark or violent in nature. This is not because GPT-3 is trying to be mean or hurtful – it's just a language model that generates text based on a statistical probability – it has no concept of mean or inappropriate. But GPT-3 was trained using data from the internet, which unfortunately contains plenty of offensive content. So, as a result, there will be times that the model generates inappropriate content in a completion. Thankfully, as developers, there are things we can do to prevent users from seeing potentially inappropriate responses from GPT-3.
The simplest way to avoid inappropriate content is to use your prompt text to steer the model away from generating potentially offensive...