Senior Data Engineer, Databricks, Azure
Posted 102ds ago
Employment Information
Report this job
Job expired or something wrong with this job?
Job Description
Senior Data Engineer implementing data solutions within a medallion architecture using Databricks and Azure. Collaborating across business intelligence and data analytics teams for data-driven decision-making.
Responsibilities:
- Collaborate with the Product Owner, Business Analyst and other team members to understand domain-specific data requirements and design scalable data pipelines and architectures
- Transform and enrich data from the Bronze layer (raw data) into the Silver layer by performing data quality checks, cleansing, validation, standardization, and modeling activities
- Build and maintain data products and data marts in the Gold layer that are curated, business-ready, and optimized for analytics and reporting purposes
- Develop efficient ETL/ELT workflows to extract, transform, and load data from various sources through the medallion architecture layers
- Optimize and fine-tune data pipelines for performance, reliability, and scalability using PySpark and Databricks
- Implement comprehensive data quality checks and monitoring to ensure data accuracy, consistency, and integrity across all layers
- Develop Power BI reports and dashboards that leverage the Gold layer data products to deliver business insights
- Work with BI developers and data analysts to provide them with the necessary data infrastructure and tools for analysis and reporting
- Troubleshoot and resolve data-related issues, including performance bottlenecks and data inconsistencies
- Stay up to date with the latest trends and technologies in data engineering and recommend improvements to existing systems and processes
- Document data engineering processes, data flows, and system configurations
Requirements:
- Bachelor's degree in Computer Science, Engineering, or a related field
- 5+ years of experience as a Data Engineer on Databricks and Azure, with a focus on designing and building data pipelines within a medallion architecture (Bronze, Silver, Gold layers)
- Strong problem-solving and analytical skills
- Effective communication in English and collaboration in agile environments
- Experience working in global, multicultural teams
- Advanced proficiency in PySpark and SparkSQL for large-scale data transformations and processing
- Hands-on experience with Databricks workspace, Delta Lake, workflows, and notebooks
- Proven experience developing interactive reports and dashboards in Power BI
- Strong programming skills in SQL and Python (Scala knowledge is a plus)
- Hands-on experience with Azure services such as Azure Data Lake Storage (ADLS), Azure Data Factory (ADF), Azure SQL, Synapse Analytics is an advantage
- Proficiency in data modeling, database design, and SQL optimization
- Solid understanding of data integration patterns, ETL/ELT best practices, cloud computing, and security principles
- You should have demonstrable experience in at least one of the following domains: Sales, Finance, Project Business
Benefits:
- Workload: 100%
- Working model: Remote
- Long-term project with option to extend




















