Logo
Anthropic

Policy Design Manager, User Well-being

Anthropic, San Francisco, California, United States, 94199

Save Job

Overview Remote-Friendly (Travel Required) | San Francisco, CA

Is your CV ready If so, and you are confident this is the role for you, make sure to apply asap. About Anthropic Anthropic’s mission is to create reliable, interpretable, and steerable AI systems. We want AI to be safe and beneficial for our users and for society as a whole. Our team is a quickly growing group of committed researchers, engineers, policy experts, and business leaders working together to build beneficial AI systems. About the role As a Safeguards Policy Design Manager, you will be responsible for developing usage policies, clarifying enforcement guidelines, and advising on safety interventions for our products and services. Your core focus will be on mitigating potential risks related to user well-being, including concerns regarding mental health, sycophancy, delusions, and emotional attachment. In addition, you will advise teams on opportunities for promoting well-being, including potential intervention development and supporting beneficial use cases. Safety is core to our mission and you’ll help shape policy creation and development so that our users can safely interact with and build on top of our products in a harmless, helpful and honest way. Important context for this role: In this position you may be exposed to and engage with explicit content spanning a range of topics, including those of a sexual, violent, or psychologically disturbing nature. Responsibilities

Serve as an internal subject matter expert, leveraging deep expertise in mental health and well-being to: Draft new policies that help govern the responsible use of our models for emerging capabilities and use cases Design evaluation frameworks for testing model performance in areas of expertise Conduct regular reviews and testing of existing policies to identify and address gaps and ambiguities Review flagged content to drive enforcement and policy improvements Update our usage policies based on feedback collected from external experts, our enforcement team, and edge cases that you will review Work with safeguards product teams to identify and mitigate concerns, and collaborate on designing appropriate interventions Educate and align internal stakeholders around our policies and our approach to safety in your focus area(s) Keep up to date with new and existing AI policy norms and standards, and use these to inform our decision-making on policy areas You may be a good fit if you have experience

As a researcher, subject matter expert, clinician, or trust & safety professional working in one or more of the following focus areas: psychology, mental health, developmental science, or human-AI interaction. Note: For this role, an advanced degree in clinical psychology, counseling psychology, psychiatry, social work, or a related field is preferred. Drafting or updating product and / or user policies, with the ability to effectively bridge technical and policy discussions Crafting evidence-based and psychometrically valid definitions for emerging phenomena Working with generative AI products, including writing effective prompts for policy evaluations Aligning product policy decisions between diverse sets of stakeholders, such as Product, Engineering, Public Policy, and Legal teams Understanding the challenges that exist in developing and implementing product policies at scale, including in the content moderation space Thinking creatively about the risks and benefits of new technologies, and leverage data and research to inform policy recommendations Navigating and prioritizing work efforts amidst ambiguity Deadline to apply:

None. Applications will be reviewed on a rolling basis. Salary range:

$190,000 - $220,000 USD Logistics

Education requirements:

We require at least a Bachelor\'s degree in a related field or equivalent experience. Location-based hybrid policy:

Currently, we expect all staff to be in one of our offices at least 25% of the time. Some roles may require more time in offices. Visa sponsorship:

We do sponsor visas. If we make you an offer, we will make every reasonable effort to assist with visa approval. Come work with us! Anthropic is a public benefit corporation headquartered in San Francisco. We offer competitive compensation and benefits, optional equity donation matching, generous vacation and parental leave, flexible working hours, and a collaborative office environment.

#J-18808-Ljbffr