Senior Data Quality Engineer
WHAT WE DO
Founded in 2007, Growth Acceleration Partners (GAP) is a premier consulting and technology services company built for the AI era. We consult, design, build, and modernize revenue-generating software and data engineering solutions. By leveraging AI-native architectures, advanced data analytics, and deep modernization strategies, we help businesses secure a competitive advantage.
GAP’s remote, integrated engineering teams deliver end-to-end solutions that drive true business innovation. We are a woman-owned, Austin-based company with over 600 English-speaking engineers across Latin America and the U.S. Boasting industry-leading customer satisfaction scores, our core focus is the dual success of our clients and our people. We are a values-based organization deeply invested in the growth of our "GAPsters"—providing continuous education, onsite English classes, and cutting-edge training in AI, machine learning, and next-gen technologies to ensure our communities achieve long-term success.
Summary
We are looking for a Senior Data Quality Engineer with strong expertise in SQL and Python to build and scale automated data validation frameworks for enterprise Data Warehouse environments. In this role, you will be responsible for ensuring the integrity, reliability, and consistency of the data powering our analytics and reporting platforms. This is a highly technical role with a strict focus on automation—traditional manual testing will not scale for our needs. You will work closely with Data Engineers and Analysts to replace manual checks with robust, automated testing frameworks, validating the quality of data as it moves across pipelines, transformations, and Data Lakehouse layers.
Education
Bachelor’s Degree in Computer Science, Software Engineering, Information Systems, Data Engineering, or a related technical field.
Professional Experience
• 5+ years of experience working with data validation, data quality, or data pipeline reliability
• Proven experience designing and building automated testing frameworks from scratch (manual testing backgrounds are not a fit for this role)
• Deep hands-on experience utilizing SQL and Python for data automation and validation
• Proven experience working with Data Warehouses, analytical data models, and ETL pipelines
Key Responsibilities
• Design, build, and maintain scalable automated data validation frameworks using Python and SQL
• Ensure the accuracy, completeness, and consistency of data entering the Data Warehouse by automating validation across all analytical layers
• Review and validate ETL pipelines, data integrations, and transformation logic feeding Data Lakehouses
• Write advanced SQL queries to perform automated source-to-target validation, reconciliation, and data investigation
• Investigate discrepancies, identify anomalies, and perform root cause analysis on data quality issues across systems
• Integrate repeatable automated validation checks into CI/CD pipelines to ensure high-quality data delivery
• Partner cross-functionally with Data Engineers, Analysts, and governance teams to maintain enterprise data quality standards
Required Technical Skills
Strong experience in the following areas:
• Automation & Scripting: Strong Python expertise specifically focused on building automated testing frameworks and data processing
• Data Querying & Analysis: Advanced SQL skills for complex data validation, troubleshooting, and anomaly detection in large datasets
• Pipeline Validation: Hands-on experience automating the validation of ETL pipelines, ingestion workflows, and data integrations
• Data Architecture: Deep understanding of data flows, Data Lakehouses, and analytical data models
• QA Integration: Experience defining automated QA workflows and integrating data validation processes directly into CI/CD pipelines
Nice to Have
• Experience with dbt and Great Expectations for data validation and transformation
• Familiarity with GitHub, particularly in the context of CI/CD pipelines (we are currently migrating our pipelines from Azure DevOps to GitHub)
• Familiarity with Databricks (specifically for orchestration) and Unity Catalog (UC)
• Experience working with Microsoft Fabric
Soft Skills
• Advanced English proficiency (spoken and written)
• Strong analytical, investigative, and problem-solving mindset
• Excellent documentation and communication skills
• Ability to identify risks and proactively address data quality issues
• Collaborative mindset when working with cross-functional data and engineering teams
At Growth Acceleration Partners, we're an equal opportunity employer committed to building a diverse and inclusive team. We value everyone's unique background, and we provide equal opportunities regardless of race, color, creed, religion, sexual orientation, gender identity, age, national origin, disability, marital status, veteran status or any other personal right protected by law. We foster a culture of belonging and strive to provide a welcoming environment where everyone feels safe to contribute and grow.
- Department
- Analytics
- Role
- Senior Data Engineer
- Locations
- Colombia, Costa Rica, Latam
- Remote status
- Hybrid
- Main Technology
- Data Engineering