Product Data Engineer

Posted 9ds ago

Employment Information

Education
Salary
Experience
Job Type

Report this job

Job expired or something wrong with this job?

Job Description

Product Data Engineer owning and evolving the data platform for healthcare CRM solution. Design and manage ETL pipelines, ensuring data quality for analytics and product decisions.

Responsibilities:

  • Design, build, and maintain reliable end-to-end ETL pipelines orchestrated with Apache Airflow
  • Integrate data from multiple sources (internal operational databases, third-party APIs, SaaS tools) into the Google Cloud Data Warehouse (BigQuery)
  • Design and evolve data models, warehouse schemas, and transformations to support scalable analytics and KPIs
  • Ensure data quality, reliability, and observability through monitoring, validation, and alerting
  • Own the product data structure, mapping product features and behaviors to analytics-ready data models
  • Define and maintain meaningful KPIs in collaboration with Product and BI
  • Enable analytics for AI-powered product features, ensuring visibility on usage, performance, quality, and business impact
  • Partner with Product, BI, and other stakeholders to gather requirements and deliver dashboards and reports
  • Maintain clear and up-to-date documentation for data models, pipelines, and metrics
  • Act as the primary bridge between Backend Engineering and BI, owning the flow from data production to analytics consumption
  • Triage, analyze, and address BI requests related to data availability, correctness, performance, and modeling
  • Collaborate with Backend Engineers on data contracts, schema evolution, and performance optimization, without owning core backend services
  • Proactively identify and resolve data-related issues impacting BI and Product teams
  • Own first-level monitoring and support for data pipelines and Airflow DAGs, ensuring timely resolution of failures
  • Collaborate with BI and Backend teams to troubleshoot and resolve complex issues
  • Continuously improve the stability, performance, and maintainability of the data platform.

Requirements:

  • 2+ years of experience in Data Engineering or a similar role
  • Hands-on experience designing, scheduling, and maintaining ETL pipelines using Apache Airflow
  • Strong SQL skills and solid understanding of data warehousing concepts (preferably Google BigQuery)
  • Proficiency in Python for ETL development and automation
  • Experience working in AI product environments, supporting data needs for AI features such as experimentation, monitoring, and analytics
  • Experience integrating data from multiple sources (APIs, databases, flat files, external platforms)
  • Experience building dashboards or analytical views using BI tools (preferably Looker)
  • Familiarity with Google Cloud Platform (GCP) services
  • Strong analytical and problem-solving skills
  • Comfortable working in a cross-functional, ambiguous environment
  • Strong communication skills and ability to collaborate with both technical and non-technical stakeholders
  • Strong interest in product data and how data drives product decisions.

Benefits:

  • 100% remote work, with the option to join our offices in Bologna or Barcelona
  • One extra day off for your birthday
  • Access to iFeel – our mental wellbeing platform
  • €8/day meal vouchers – lunch is covered if you're in the Bologna office
  • Private health coverage via Metasalute
  • Comprehensive private health insurance with Adeslas
  • Flexoh – flexible compensation platform
  • Wellhub – gym & wellness network membership
  • Language courses