AI Red Team Tester
Job Description
To build truly resilient AI, we must first understand its weaknesses. As an AI Red Team Tester, you'll embody the adversarial mindset, creatively challenging AI systems to uncover vulnerabilities, push their boundaries, and ultimately make them safer and more robust against malicious use.
Key Responsibilities
Develop and execute novel adversarial attacks and 'jailbreaks' against cutting-edge AI models, particularly LLMs.
Identify and document failure modes related to harmful content generation, privacy leakage, and model manipulation.
Craft sophisticated prompts and multi-turn dialogues to elicit biased, toxic, or otherwise undesirable AI responses.
Provide detailed, structured feedback on AI system vulnerabilities, including potential exploits and mitigation strategies.
Collaborate with AI safety researchers to refine testing methodologies and contribute to the development of new safety benchmarks.
Stay informed on emerging AI attack vectors, adversarial machine learning techniques, and ethical hacking principles.
Ideal Qualifications
Demonstrated creative problem-solving and an 'attacker's mindset'.
Extensive experience interacting with and probing large language models (e.g., GPT-3/4, Claude, Llama).
Familiarity with concepts in adversarial machine learning, prompt injection, and data poisoning.
Strong ethical compass and understanding of responsible disclosure principles.
Excellent written communication skills for reporting complex findings.
Background in cybersecurity, offensive security, or a related field is highly desirable.
Project Timeline
Start Date: Within 1 week
Duration: 6 months (with potential for extension)
• Commitment: Flexible, 20-40 hours/week
Challenge the limits of AI – become an AI Red Team Tester!