Job Description
We are seeking a dedicated AI Quality Assurance Engineer to ensure the accuracy, reliability, and performance of AI-generated outputs across our RAG-based document processing and insight generation systems. This role plays a vital part in maintaining the integrity of AI-driven products and services, contributing to our organization's overall quality standards. You will collaborate closely with data scientists, machine learning engineers, and product managers to develop and implement robust testing strategies that validate AI functionalities and identify areas for improvement.
Responsibilities
AI Testing & Validation
Design and execute comprehensive test plans for AI models and systems, focusing on accuracy, completeness, and alignment with business outcomes
Develop and implement automated testing frameworks and tools specifically for AI validation and model performance assessment
Use quantitative and qualitative methods to assess the quality, relevance, and performance of AI-generated outputs
Monitor AI system performance across diverse document types, edge cases, and varying input conditions
Quality Assurance & Process Improvement
Analyze test results to identify defects, inconsistencies, and misalignments in AI behavior and model outputs
Document bugs, performance issues, and areas for improvement with detailed analysis and recommendations
Collaborate with AI Product Analysts to refine prompt designs, system instructions, and output template configurations
Assist with regression testing after changes to prompts, system logic, or model updates
Cross-Functional Collaboration
Work closely with data scientists and engineers to enhance model performance and address quality issues
Partner with product managers to ensure AI models meet user requirements and business expectations
Contribute to the development of data quality metrics and monitoring systems for continuous improvement
Participate in code reviews and process improvements to maintain high development standards
Compliance & Standards
Ensure AI outputs meet business expectations, compliance requirements, and user experience standards
Implement and maintain quality standards for AI systems in regulated environments (finance, healthcare, legal)
Develop and execute validation processes for responsible AI practices, including bias detection and mitigation
Create and maintain comprehensive documentation for quality assurance processes and testing methodologies
Requirements
Education & Experience
Bachelor's degree in Computer Science, Engineering, or related field
3+ years of experience in software testing and quality assurance, preferably with AI/ML systems
Proven track record in designing and executing test plans for complex software systems
Technical Skills
Proficiency in Python with experience using testing frameworks (pytest, Selenium, unittest)
Familiarity with machine learning concepts, algorithms, and model evaluation techniques
Strong understanding of automated testing frameworks and tools for AI validation
Experience with data quality assessment, validation, and monitoring processes
Knowledge of statistical analysis and performance metrics for AI systems
AI & Domain Expertise
Understanding of prompt engineering principles and experience testing large language models (LLMs)
Experience with AI model evaluation methodologies and performance benchmarking
Familiarity with RAG systems, vector databases, and retrieval-augmented generation workflows
Knowledge of AI bias detection, fairness metrics, and responsible AI practices
Analytical & Communication Skills
Strong analytical and problem-solving skills with keen attention to detail
Excellent documentation skills with ability to create clear, comprehensive test reports
Understanding of business domain and regulatory compliance requirements
Comfortable working with collaboration platforms and annotation tools
Preferred Skills
Preferred Qualifications
Advanced degree in Computer Science, Engineering, or related field
Experience with cloud-based AI/ML platforms and testing environments
Knowledge of data labeling, annotation tools, and dataset management
Experience with performance testing and load testing for AI systems
Understanding of MLOps practices and continuous integration for ML models
Certification in software testing (ISTQB) or quality assurance methodologies
Experience in regulated industries (finance, healthcare, legal services)
Essential Behavioral Competencies
Attention to Detail: Ability to meticulously identify and document defects and inconsistencies
Analytical Thinking: Strong problem-solving skills to analyze test results and model behaviors
Team Collaboration: Effective communication and teamwork with cross-functional teams
Adaptability: Staying updated with latest AI and QA trends and adapting to new methodologies
Integrity: Upholding high standards of honesty and ethical behavior in all quality assessments
Tools & Technologies
Testing Frameworks
pytest, Selenium, unittest
Ragas, DeepEval, TruLens
Postman, REST Assured
AI/ML Tools
OpenAI API, Azure AI, Anthropic
Hugging Face, LangChain
TensorFlow, PyTorch (for model evaluation)
Quality & Monitoring
Weights & Biases, MLflow
Datadog, New Relic
Grafana, Prometheus
Collaboration Tools
Jira, Bugzilla
Confluence, Notion
Labelbox, Scale AI
Category
Salary
Posted
Location
( Remote )
Share or copy
