hero

Portfolio Company Careers

Discover career opportunities across PFG's network of portfolio companies

Data Platform Architect

Baubap

Baubap

IT
Mexico
Posted on Nov 28, 2025

The mission:

Design and implement the foundational data architecture that will power our next stage of growth. This role is hybrid: highly strategic in defining long-term standards and architecture, while also hands-on in building the first pipelines, storage layers, and frameworks that make data reliable, scalable, and accessible.

Unlike roles limited to pipeline maintenance, this position is responsible for the entire backbone of Baubap’s data ecosystem: migrating analytical workloads away from the monolithic architectures, standardizing schema practices, enabling distributed storage, and ensuring financial and behavioral data can be trusted at terabyte scale. This person will also be responsible for mentoring and growing the Data Engineering team, establishing a culture of quality and ownership

The expected outcome:

  • Define and implement the transition from monolithic RDS analytical workloads to a scalable data warehouse/lake.
  • Standardize schema and database practices across the company.
  • Build and maintain pipelines (batch and real-time) that handle terabyte-scale datasets from multiple sources (mobile, financial, behavioral).
  • Implement monitoring, validation, and alerting to ensure reliable data delivery.
  • Ensure financial datasets (payments, collections, reports) are reliable, reconcilable, and auditable, reducing inconsistencies.
  • Own the AWS-based data stack (Aurora, Redshift, S3, Airflow, DMS, Glue), balancing reliability, cost, and scalability.
  • Evaluate and introduce tools for orchestration, observability, and optimization.
  • Mentor backend and data engineers, creating a culture of excellence in data engineering.
  • Define architectural standards, coding practices, and documentation templates.

The day to day tasks:

  • Build and maintain ETL/ELT pipelines (batch + streaming).
  • Optimize large-scale queries and data models for performance.
  • Manage and partition storage solutions (S3, Redshift) for efficiency.
  • Audit and fix bottlenecks in financial data flows.
  • Collaborate with data scientists, analysts, and backend engineers to translate needs into infrastructure.
  • Document architecture, schemas, and pipelines for long-term clarity.
  • Propose and implement improvements in cost efficiency, resilience, and reliability.

Why YOU should apply:

  • 7+ years in data engineering/architecture, with proven experience at TB+ scale.
  • Experience designing and implementing data platforms
  • Strong expertise in AWS (Aurora RDS, Redshift, S3, Glue, DMS).
  • Mastery of SQL and Python for data workflows.
  • Experience with pipeline orchestration (Airflow, Step Functions, etc.).
  • Solid knowledge of data warehousing and data lakes (partitioning, schema design, optimization).
  • Familiarity with distributed systems and performance tuning at scale.
  • Experience balancing strategic design and hands-on execution.
  • Knowledge of CI/CD for data pipelines, access control, monitoring.
  • Ability to communicate trade-offs and designs to both technical and non-technical stakeholders.
  • Experience mentoring engineers or growing data teams.

What we can offer:

  • Being part of a multicultural, highly driven team of professionals
  • 20 vacation days / year + 75% holiday bonus (Prima Vacacional)
  • 1 month (proportional) of Christmas bonus (Aguinaldo)
  • Food vouchers
  • Health & Life insurance
  • Competitive salary