AI Quality Assurance

AI Quality Assurance

Engineering

|

AI Quality Assurance Engineer

AI Quality Assurance Engineer

Job Description

We are seeking a dedicated AI Quality Assurance Engineer to ensure the accuracy, reliability, and performance of AI-generated outputs across our RAG-based document processing and insight generation systems. This role plays a vital part in maintaining the integrity of AI-driven products and services, contributing to our organization's overall quality standards. You will collaborate closely with data scientists, machine learning engineers, and product managers to develop and implement robust testing strategies that validate AI functionalities and identify areas for improvement.

Responsibilities

AI Testing & Validation

  • Design and execute comprehensive test plans for AI models and systems, focusing on accuracy, completeness, and alignment with business outcomes

  • Develop and implement automated testing frameworks and tools specifically for AI validation and model performance assessment

  • Use quantitative and qualitative methods to assess the quality, relevance, and performance of AI-generated outputs

  • Monitor AI system performance across diverse document types, edge cases, and varying input conditions

Quality Assurance & Process Improvement

  • Analyze test results to identify defects, inconsistencies, and misalignments in AI behavior and model outputs

  • Document bugs, performance issues, and areas for improvement with detailed analysis and recommendations

  • Collaborate with AI Product Analysts to refine prompt designs, system instructions, and output template configurations

  • Assist with regression testing after changes to prompts, system logic, or model updates

Cross-Functional Collaboration

  • Work closely with data scientists and engineers to enhance model performance and address quality issues

  • Partner with product managers to ensure AI models meet user requirements and business expectations

  • Contribute to the development of data quality metrics and monitoring systems for continuous improvement

  • Participate in code reviews and process improvements to maintain high development standards

Compliance & Standards

  • Ensure AI outputs meet business expectations, compliance requirements, and user experience standards

  • Implement and maintain quality standards for AI systems in regulated environments (finance, healthcare, legal)

  • Develop and execute validation processes for responsible AI practices, including bias detection and mitigation

  • Create and maintain comprehensive documentation for quality assurance processes and testing methodologies

Requirements

Education & Experience

  • Bachelor's degree in Computer Science, Engineering, or related field

  • 3+ years of experience in software testing and quality assurance, preferably with AI/ML systems

  • Proven track record in designing and executing test plans for complex software systems

Technical Skills

  • Proficiency in Python with experience using testing frameworks (pytest, Selenium, unittest)

  • Familiarity with machine learning concepts, algorithms, and model evaluation techniques

  • Strong understanding of automated testing frameworks and tools for AI validation

  • Experience with data quality assessment, validation, and monitoring processes

  • Knowledge of statistical analysis and performance metrics for AI systems

AI & Domain Expertise

  • Understanding of prompt engineering principles and experience testing large language models (LLMs)

  • Experience with AI model evaluation methodologies and performance benchmarking

  • Familiarity with RAG systems, vector databases, and retrieval-augmented generation workflows

  • Knowledge of AI bias detection, fairness metrics, and responsible AI practices

Analytical & Communication Skills

  • Strong analytical and problem-solving skills with keen attention to detail

  • Excellent documentation skills with ability to create clear, comprehensive test reports

  • Understanding of business domain and regulatory compliance requirements

  • Comfortable working with collaboration platforms and annotation tools

Preferred Skills

Preferred Qualifications
  • Advanced degree in Computer Science, Engineering, or related field

  • Experience with cloud-based AI/ML platforms and testing environments

  • Knowledge of data labeling, annotation tools, and dataset management

  • Experience with performance testing and load testing for AI systems

  • Understanding of MLOps practices and continuous integration for ML models

  • Certification in software testing (ISTQB) or quality assurance methodologies

  • Experience in regulated industries (finance, healthcare, legal services)

Essential Behavioral Competencies
  • Attention to Detail: Ability to meticulously identify and document defects and inconsistencies

  • Analytical Thinking: Strong problem-solving skills to analyze test results and model behaviors

  • Team Collaboration: Effective communication and teamwork with cross-functional teams

  • Adaptability: Staying updated with latest AI and QA trends and adapting to new methodologies

  • Integrity: Upholding high standards of honesty and ethical behavior in all quality assessments

Tools & Technologies

Testing Frameworks

  • pytest, Selenium, unittest

  • Ragas, DeepEval, TruLens

  • Postman, REST Assured

AI/ML Tools

  • OpenAI API, Azure AI, Anthropic

  • Hugging Face, LangChain

  • TensorFlow, PyTorch (for model evaluation)

Quality & Monitoring

  • Weights & Biases, MLflow

  • Datadog, New Relic

  • Grafana, Prometheus

Collaboration Tools

  • Jira, Bugzilla

  • Confluence, Notion

  • Labelbox, Scale AI

Category

Engineering

Engineering

Salary

80k - 120k/year

80k - 120k/year

Posted

3 months ago
5 months ago
5 months ago

Location

United States

United States

( Remote )

Share or copy

linkedin iconfacebook iconx icon

Not the right job for you?