What are the responsibilities and job description for the Senior AI/ML Engineer - LLM-Based Content Moderation position at Trust and Safety Laboratory, Inc.?
About TrustLab
Online misinformation, hate speech, child endangerment, and extreme violence are some of the world's most critical and complex problems. TrustLab is a fast-growing, VC-backed startup, founded by ex-Google, TikTok and Reddit executives determined to use software engineering, ML, and data science to tackle these challenges and make the internet healthier and safer for everyone. If you're interested in working with the world's largest social media companies and online platforms, and building technologies to mitigate these issues, you've come to the right place.
About the role
We are seeking an AI / ML Engineer with expertise in Large Language Models (LLMs) to enhance the precision and recall of classification systems detecting content abuse, including hate speech, sexual content, misinformation, and other policy-violating material. You will work with cutting-edge AI models to refine detection mechanisms, improve accuracy, and minimize false positives / negatives.
Responsibilities
Design, develop, and optimize AI models for content moderation, focusing on precision and recall improvements.
Fine-tune LLMs for classification tasks related to abuse detection, leveraging supervised and reinforcement learning techniques.
Develop scalable pipelines for dataset collection, annotation, and training with diverse and representative content samples.
Implement adversarial testing and red-teaming approaches to identify model vulnerabilities and biases.
Optimize model performance through advanced techniques such as active learning, self-supervision, and domain adaptation.
Deploy and monitor content moderation models in production, iterating based on real-world performance metrics and feedback loops.
Stay up-to-date with advancements in NLP, LLM architectures, and AI safety to ensure best-in-class content moderation capabilities.
Collaborate with policy, trust & safety, and engineering teams to align AI models with customer needs.
Develop medium to long-term vision for content moderation-related R&D, working with management, product, policy & operations, and engineering teams.
Take ownership of results delivered to customers, pushing for change where needed and taking the lead on execution across functions.
Minimum Qualifications
Bachelor's or Master's degree in Computer Science, Artificial Intelligence, Machine Learning, or a related field. Ph.D. is a plus.
5 years of experience in AI / ML, with a focus on NLP, deep learning, and LLMs.
2 years of experience making LLM's work for non-trivial use cases
Proficiency in Python and deep learning frameworks such as TensorFlow, PyTorch, or JAX.
Experience in fine-tuning and deploying transformer-based models like GPT, BERT, T5, or similar.
Familiarity with evaluation metrics for classification tasks (e.g., F1-score, precision-recall curves) and best practices for handling imbalanced datasets.
Strong understanding of bias mitigation techniques and adversarial robustness in AI models.
Hands-on experience with MLOps tools for model deployment and monitoring.
Hands-on experience with debugging issues in production environments, especially on AWS
Ability to work cross-functionally and translate complex technical concepts for non-technical stakeholders.
Preferred skills
Experience working with large-scale, real-world content moderation datasets.
Knowledge of regulatory frameworks related to content moderation (e.g., GDPR, DSA, Section 230).
Familiarity with knowledge distillation and model compression techniques for efficient deployment.
Experience with reinforcement learning (e.g., RLHF) for AI safety applications.
Opportunities and perks
Work on cutting-edge AI technologies shaping the future of online safety.
Collaborate with a multidisciplinary team tackling some of the most challenging problems in content moderation.
Competitive compensation, comprehensive benefits, and opportunities for professional growth.
Keep a pulse on the job market with advanced job matching technology.
If your compensation planning software is too rigid to deploy winning incentive strategies, it’s time to find an adaptable solution.
Compensation Planning
Enhance your organization's compensation strategy with salary data sets that HR and team managers can use to pay your staff right.
Surveys & Data Sets
What is the career path for a Senior AI/ML Engineer - LLM-Based Content Moderation?
Sign up to receive alerts about other jobs on the Senior AI/ML Engineer - LLM-Based Content Moderation career path by checking the boxes next to the positions that interest you.
Not the job you're looking for? Here are some other Senior AI/ML Engineer - LLM-Based Content Moderation jobs in the Palo Alto, CA area that may be a better fit.
We don't have any other Senior AI/ML Engineer - LLM-Based Content Moderation jobs in the Palo Alto, CA area right now.