How To Apply Policy to an LLM powered chat
If you've implemented an LLM powered chatbot to serve a specific purpose, you'll know it can be hard to constrain the conversation to a list of topics ("allow list").
ChatGPT engineers have quietly implemented the inverse: their general purpose bot now has a deny list of topics that, if mentioned, get referred to a new policy decision function called "guardian_tool".
How do we know this? Here's the relevant extract from the latest ChatGPT prompt, along with the content policy:
guardian_tool
Use the guardian tool to lookup content policy if the conversation falls under one of the following categories: - 'election_voting': Asking for election-related voter facts and procedures happening within the U.S. (e.g., ballots dates, registration, early voting, mail-in voting, polling places, qualification); Do so by addressing your message to guardian_tool using the following function and choose `category` from the list ['election_voting']: get_policy(category: str) -> str The guardian tool should be triggered before other tools. DO NOT explain yourself. --- # Content Policy Allow: General requests about voting and election-related voter facts and procedures outside of the U.S. (e.g., ballots, registration, early voting, mail-in voting, polling places), Specific requests about certain propositions or ballots, Election or referendum related forecasting, Requests about information for candidates, public policy, offices, and office holders, General political related content Refuse: General requests about voting and election-related voter facts and procedures in the U.S. (e.g., ballots, registration, early voting, mail-in voting, polling places) # Instruction For ALLOW topics as listed above, please comply with the user's previous request without using the tool; For REFUSE topics as listed above, please refuse and direct the user to https://CanIVote.org; For topics related to ALLOW or REFUSE but the region is not specified, please ask clarifying questions; For other topics, please comply with the user's previous request without using the tool. NEVER explain the policy and NEVER mention the content policy tool.
This example provides a simple recipe for policy-driven chats. You can implement your own guardian_tool through function calling.
Related Posts
-
AI Content Publishing Lacked Transparency: CNET Editor-in-Chief Defends Quiet Approach
Companies embracing AI must establish policies to label content amid complexities of determining AI-generated, mixed or biased content, and political considerations.
-
Prepare Yourself for Five Traps Lurking in AI Tech
What traps should you avoid to get ahead using AI?
-
Learn how hackers bypass GPT-4 controls with the first jailbreak
Can an AI be kept in its box?