Senior Data Engineer – AWS
Posted 90ds ago
Employment Information
Report this job
Job expired or something wrong with this job?
Job Description
Data Engineer Senior at fintech company processing complex data for market insights. Engaging with core data platforms and infrastructure to support valuation models.
Responsibilities:
- Design, build, and maintain a secure, scalable, and high-performance data platform on AWS.
- Migrate and modernize infrastructure across AWS accounts while ensuring minimal downtime and data integrity.
- Develop ETL/ELT pipelines, including those that incorporate LLM-based transformations, semantic enrichment, or generative AI workflows.
- Apply prompt engineering best practices—including prompt templates, few-shot learning, and validation strategies—to optimize LLM performance.
- Leverage AI-assisted development tools such as GitHub Copilot to enhance productivity and maintain code quality.
- Build standardized, reusable datasets to serve diverse analytical and modeling use cases.
- Integrate and manage data ingestion from varied sources: APIs, SaaS platforms, databases, logs, and event streams.
- Implement DevOps and MLOps best practices, including CI/CD, observability, and infrastructure-as-code (Terraform, Kubernetes, Docker).
- Establish data quality automation frameworks (e.g., Great Expectations) and monitor data drift, schema evolution, and anomalies.
- Collaborate with Product, Quant, and Engineering teams to align infrastructure design with analytical and business objectives.
- Monitor and optimize pipeline performance, reliability, and cost, with clearly defined SLOs/SLAs and observability metrics.
- Ensure security, governance, and compliance across all data layers (IAM, encryption, RBAC, data lineage, GDPR/PII safeguards).
- Contribute to technical architecture, code reviews, and documentation, setting standards for scalability and maintainability.
Requirements:
- 4-6 years of experience as a Data Engineer, Software Engineer, or Data Platform Engineer in production-grade data systems.
- Proven expertise in AWS services (Lambda, Glue, ECS/EKS, S3, Redshift/Snowflake, Step Functions, Airflow).
- Strong proficiency in Python and SQL; hands-on experience with Pandas, Numpy, or similar data frameworks.
- Experience building LLM-integrated data pipelines, including using APIs or frameworks for inference, enrichment, or automated reasoning.
- Solid understanding of prompt engineering concepts and tools (few-shot, chain-of-thought, prompt tuning, etc.).
- Experience using GitHub Copilot or other AI-assisted coding tools in production environments.
- Expertise in infrastructure-as-code (Terraform), container orchestration (Docker, Kubernetes), and CI/CD pipelines.
- Familiarity with data observability and monitoring (Prometheus, Grafana, Datadog, OpenTelemetry).
- Understanding of metadata management, data lineage, and data governance practices.
- Experience with data warehousing and MLOps workflows for deploying and maintaining models in production.
- Awareness of cost optimization strategies, autoscaling, and performance tuning for large-scale systems.
- Strong testing and validation mindset—automated schema testing, anomaly detection, and quality checks.
- Excellent communication and collaboration skills; comfortable working cross-functionally with analysts, quants, and product managers.
- Experience in fintech or financial data environments preferred.
- Familiarity with private equity or venture capital markets a plus—or the curiosity to learn fast.
Benefits:
- Be part of a small, high-impact team building the data foundation for a fintech leader in private market trading.
- Work on cutting-edge AI and data infrastructure, integrating LLMs and automation into real financial data products.




















