Job Description
Company Culture and Mission
At Roche, you can be yourself, with a culture that fosters personal expression, open dialogue, and genuine connections. We value and respect each individual, supporting both personal and professional growth. Our goal is to prevent, stop, and cure diseases, ensuring healthcare access for all now and in the future.
The Position
About Roche and Galileo
Galileo is Roche's strategic Informatics program focused on enabling high-value AI use cases through specialized platforms, aiming to establish an AI Center of Excellence. The Use Case Delivery (UCD) Team focuses on building innovative Generative AI applications.
Role Overview
We are seeking a Data Engineer to join a new AI solutions development squad working on cutting-edge applications utilizing Large Language Models (LLMs). This end-to-end role involves designing, building, and maintaining data infrastructure to support AI applications.
Responsibilities
- Generative AI Application Co-creation: Collaborate with AI engineers, data scientists, and product owners to develop scalable, ethical, and user-centric AI applications.
- Data Infrastructure & Integration: Design high-performance data pipelines, ensuring efficient data ingestion, transformation, storage, and retrieval.
- Vector Database Management: Manage vector databases like AWS OpenSearch or Azure AI Search.
- Cloud Data Engineering: Develop and maintain cloud-based data solutions (AWS/Azure).
- Snowflake Implementation: Optimize data storage and processing using Snowflake.
- Data Processing & Transformation: Create ETL/ELT pipelines for real-time and batch data processing.
- Support AI Workflow: Collaborate with AI/ML teams to ensure smooth integration of data pipelines.
- Performance Optimization: Enhance data strategies for efficiency and cost-effectiveness.
Requirements
- 5-7+ years in data engineering supporting AI/ML applications.
- Educational background in Computer Science, Data Engineering, or related fields.
- Programming skills in Python, SQL, and vector databases.
- Experience with relational, NoSQL, vector databases, and Snowflake.
- Hands-on experience with AWS (OpenSearch, S3, Lambda) or Azure (AI Search, Blob Storage, Automation).
- Building scalable ETL/ELT workflows with tools like dbt, Apache Airflow.
- Designing RESTful APIs, understanding data security and governance.
- Familiarity with Git, CI/CD, Docker, Kubernetes, Terraform.
- Experience with AI-specific data needs like embeddings, RAG, LLM fine-tuning.
Note: Relocation benefits are not provided.
About Roche
A global leader with over 100,000 employees dedicated to advancing science and healthcare. Our efforts include treating over 26 million people and conducting 30 billion tests annually. We foster innovation, creativity, and high ambitions to deliver impactful healthcare solutions.
Join us in building a healthier future!
Roche is an Equal Opportunity Employer.