Data Engineer - Hybrid/Guadalajara, Mexico
Job DescriptionJob DescriptionDescription:
About Brook Inc
Brook Health is a digital health company. Brook offers a set of products and services extending health-care-management beyond the walls of doctors’ offices and into people’s homes and their day-to-day lives. Brook provides people who are living with chronic conditions a highly personalized high-touch support via a smartphone app offering AI tools, data collection through connected devices, and real-time access to health coaches to make smart, daily decisions and to build healthy habits to achieve their long-term health goals. Brook also offers a CDC-approved preventative program for people who are at high risk for diabetes. For primary care providers, Brook offers SaaS tools for continuous remote monitoring, providing insights into their patient’s health needs, enabling a new model of care, and early preventative interventions with our own care delivery team resulting in better health outcomes.
At Brook.ai, we're tackling complex data and engineering challenges to build a platform that empowers people to live healthier lives. We're leveraging cutting-edge technologies like AI/ML, real-time data processing, and scalable cloud infrastructure to create a personalized health support system. Our platform integrates data from connected devices, AI-driven insights, and human coaching to drive meaningful behavior change.
We're a fast-paced, user-focused team that thrives on solving hard problems. We value innovation, continuous improvement, and a data-driven approach. Our goal is to change the approach to pre-condition and chronic condition care management, to use technology to support health-care providers in improving patients’ health outcomes. If you're passionate about building scalable, high-performance systems and making a real-world impact, Brook.ai is the place for you.
Job Overview:
As a Data Engineer based in Guadalajara, you will be responsible for the design, development and operation of our data pipelines, warehouses, and architectures. You’ll partner with analytics, product, and engineering teams to transform raw data into trusted, consumable formats—enabling reporting, dashboards, and downstream services. This is a hands-on role in a flexible hybrid setup, ideal for someone who is passionate about solving complex data challenges and creating scalable data solutions.
Requirements:
Key Responsibilities:
- Data Pipeline Development: Build, maintain, and optimize robust ETL/ELT workflows (batch and real-time) using Python or Java and orchestration tools like Airflow or Prefect.
- Data Warehouse & Modeling: Design and evolve schemas in Snowflake (or similar), ensuring efficient storage, query performance, and maintainability.
- DBT Modeling & Quality: Develop and maintain dbt models and tests in Snowflake to drive standardized data transformations and enforce data quality.
- Infrastructure & Scalability: Plan and monitor capacity, performance-tune data stores (PostgreSQL, MongoDB and MySQL)
- Architecture & Solution Design: Collaborate on solution architecture for new data initiatives—evaluating technologies, defining interfaces, and documenting standards.
- Data Quality & Governance: Implement validation, reconciliation, and lineage tracking to ensure data integrity, security, and compliance with relevant regulations like HIPPA and SOC2.
- Cross-Functional Collaboration: Partner with data analysts, and product managers to translate business requirements into data models, pipelines, and dashboards.
- Troubleshooting & Support: Diagnose and resolve data-related incidents, maintaining high availability and SLAs for data services.
- Process Improvement: Drive continuous improvement through Agile/Scrum practices—refining workflows, CI/CD for data code, and monitoring/alerting.
Required Qualifications:
- Bachelor’s degree in Computer Science, Engineering, or related field (Master’s a plus).
- 5+ years of hands-on data engineering experience.
- Advanced proficiency in Python and/or Java for data pipeline development.
- Deep expertise in SQL and relational database design (PostgreSQL).
- Practical experience with AWS services like EKS and infrastructure-as-code.
- Hands-on with Snowflake (or equivalent cloud data warehouse) and MongoDB.
- Strong understanding of data warehousing principles, ETL/ELT patterns, and performance tuning.
- Experience working in Agile/Scrum teams and using CI/CD tooling for data workflows.
Qualifications:
- Data engineering certifications (e.g., AWS Certified Data Analytics, Google Cloud Data Engineer).
- Prior experience in healthcare or regulated industries (HIPAA, data privacy).
- Familiarity with streaming architectures (Kafka, Kinesis), message queues, such as Redis, for real-time data integration or change-data-capture (CDC) techniques.
- Experience with data transformation frameworks (dbt) or orchestration platforms (Airflow, Prefect).
- Experience with healthcare data standards like FHIR is a plus.
Skills:
- Excellent communication and stakeholder management—translating technical concepts for non-technical audiences.
- Strong problem-solving mindset with attention to detail and data accuracy.
- Self-starter with the ability to work independently and collaborate effectively.
- Passion for building scalable, reliable data systems that directly impact patient care and business decisions.
This role is located in Guadalajara, Mexico. Salary is based on experience.
This role is not eligible for visa sponsorship or relocation. The candidate must live within a commuting distance from the office location. This is a hybrid role, onsite in the office required weekly along with remote work.
Brook Inc is an equal opportunity employer. We are committed to building an inclusive and diverse workforce. Brook does not discriminate on the basis of , , , , , , marital status, , non-disqualifying physical or mental disability, or , military service status, citizenship or any other protected characteristic covered by appropriate law. All employment is decided on the basis of qualifications, merit, and business need.