You will join a team that fuses data engineering with cutting-edge science, using HPC and AWS to deliver reproducible workflows at scale. From first ingestion to consumption by scientists and AI models, you will set the standard for reliability, speed, and governance across our data foundation.
Do you thrive where learning is continuous and bold ideas are encouraged? You will have the freedom to experiment, the support to grow, and the opportunity to see your work influence breakthroughs as they take shape.
Accountabilities:
- Pipeline Engineering: Design, implement, and operate fit-for-purpose data pipelines for bioinformatics and scientific data, from ingestion to consumption.
- Workflow Orchestration: Build reproducible pipelines using frameworks such as Nextflow (preferred) or Snakemake; integrate with schedulers and HPC/cloud resources.
- Data Platforms: Develop data models, warehousing layers, and metadata/lineage; ensure data quality, reliability, and governance.
- Scalability and Performance: Optimize pipelines for throughput and cost across Unix/Linux HPC and cloud environments (AWS preferred); implement observability and reliability practices.
- Collaboration: Translate scientific and business requirements into technical designs; partner with CPSS stakeholders, R&D IT, and DS&AI to co-create solutions.
- Engineering Excellence: Establish and maintain version control, CI/CD, automated testing, code review, and design patterns to ensure maintainability and compliance.
- Enablement: Produce documentation and reusable components; mentor peers and promote best practices in data engineering and scientific computing.
Essential Skills/Experience:
- Pipeline engineering: Design, implement, and operate fit-for-purpose data pipelines for bioinformatics and scientific data, from ingestion to consumption.
- Workflow orchestration: Build reproducible pipelines using frameworks such as Nextflow (preferred) or Snakemake; integrate with schedulers and HPC/cloud resources.
- Data platforms: Develop data models, warehousing layers, and metadata/lineage; ensure data quality, reliability, and governance.
- Scalability and performance: Optimize pipelines for throughput and cost across Unix/Linux HPC and cloud environments (AWS preferred); implement observability and reliability practices.
- Collaboration: Translate scientific and business requirements into technical designs; partner with CPSS stakeholders, R&D IT, and DS&AI to co-create solutions.
- Engineering excellence: Establish and maintain version control, CI/CD, automated testing, code review, and design patterns to ensure maintainability and compliance.
- Enablement: Produce documentation and reusable components; mentor peers and promote best practices in data engineering and scientific computing.
Desirable Skills/Experience:
- Strong programming in Python and Bash for workflow development and scientific computing.
- Experience with containerization and packaging (Docker, Singularity, Conda) for reproducible pipelines.
- Familiarity with data warehousing and analytics platforms (e.g., Redshift, Snowflake, Databricks) and data catalog/lineage tools.
- Experience with observability and reliability tooling (Prometheus/Grafana, ELK, tracing) in HPC and cloud contexts.
- Knowledge of infrastructure as code and cloud orchestration (Terraform, CloudFormation, Kubernetes).
- Understanding of FAIR data principles and domain-specific bioinformatics formats and standards.
- Track record of mentoring engineers and enabling cross-functional teams with reusable components and documentation.
- Experience optimizing performance and cost on AWS, including spot strategies, autoscaling, and storage tiers.
When we put unexpected teams in the same room, we unleash bold thinking with the power to inspire life-changing medicines. In-person working gives us the platform we need to connect, work at pace and challenge perceptions. That´s why we work, on average, a minimum of three days per week from the office. But that doesn´t mean we´re not flexible. We balance the expectation of being in the office while respecting individual flexibility. Join us in our unique and ambitious world.
Why AstraZeneca:
Your engineering craft will fuel science at the crossroads of biology, data, and technology. You will collaborate with researchers, data scientists, and technologists to tackle complex diseases, using modern platforms and inclusive ways of working to turn uncertainty into insight. We value kindness alongside ambition, nurture resilience and curiosity, and pair the resources of a global leader with the agility to move at pace-from hands-on experimentation to shared learning and tangible impact for patients.
Ver más
¡No te pierdas nada!
Únete a la comunidad de wijobs y recibe por email las mejores ofertas de empleo
Nunca compartiremos tu email con nadie y no te vamos a enviar spam
Suscríbete AhoraÚltimas ofertas de empleo de Ingeniero/a de Datos en Barcelona
Senior Data Engineer
NuevaDeloitte
AWS Cloud Engineer con inglés
20 ene.Aubay
Barcelona, ES
Senior Site Reliability Engineer
19 ene.Factorial
AMH HEADHUNTING
Test Engineer
17 ene.Sener
Engineer Data Engineer
16 ene.Krell Consulting & Training
Barcelona, ES
Ingeniero/a Cloud
15 ene.Krell Consulting & Training
Barcelona, ES
Data Engineer
15 ene.Aubay
Barcelona, ES
DevOps Engineer
12 ene.Seidorcons
Barcelona, ES
DevOps Integration Expert
12 ene.Sanofi
Barcelona, ES