By joining Invitae, you’ll work alongside some of the world’s experts in genetics and healthcare at the forefront of genetic medicine. We’ve crafted a culture that empowers our teams and our teammates to have the biggest impact and to explore their interests and capabilities. We prize freedom with accountability and offer significant flexibility, along with excellent benefits and competitive compensation in a fast-growing organization!
We are looking for a reliable and motivated Senior Data Engineer to join our Data Solutions Team in developing the data ingestion pipelines and data platform architecture that supports the analytical and reporting needs of data scientists, our bioinformatics team, customers, and internal stakeholders.
What you’ll do:
- Support and collaborate with multiple teams to gather requirements, design software, and implement features that support multiple teams and use cases across Data Science, Bioinformatics, and Finance.
- Design and implement reliable, scalable and efficient data framework, data driven products and software solutions for external and internal customers.
- Create a secure, flexible and powerful world-class Health Data Platform for medical research and treatment.
- Enhance existing systems to automate and use latest technologies and tools.
- Ability and passion for data to become the Subject Matter Expert working with users on databases, tables, schemas and meta-data.
- Follow and contribute to agile best practices within the organization, looking for ways to streamline, automate and reduce redundancy and costs.
- Support and respond to teammate and user questions in a fast-paced, collaborative environment in a timely manner.
What you bring:
- Minimum of 8 years of related experience with a Bachelor’s degree, 2 years and a Master’s degree or beyond.
- Skilled in one or more high-level languages (Scala preferred and/or Python). Willingness to learn new languages and technologies.
- Proficient in AWS, Azure, or Google Cloud Platform (AWS preferred).
- Experience with Relational and Columnar databases.
- Experience with one or more containerization tools, especially Docker and Kubernetes.
- Experience with messaging/queuing or stream processing systems (Kafka strongly preferred).
- Experience (or aptitude and interest) in contributing to and maintaining DevOps/Cloud Infrastructure.
Preferred Skills:
- Hands-on experience with troubleshooting, debugging, log collection, and alerting systems.
- Hands-on experience with Snowflake.
- Hands-on experience working with large datasets, ETL pipelines, and modern warehouse technologies.
- Hands-on functional programming in Scala or other language.
- Hands-on parallel programming in Spark or other platforms.
- Experience with maintaining and administering Kubernetes clusters.
- Experience with build automation and CI/CD pipelines (e.g. GitHub Actions).
- Experience with one or more data visualization tools (Looker preferred).
Nice to have:
- Demonstrated experience with data modeling/dimensional modeling.
- Demonstrated experience with database performance tuning.
- Familiarity with data lineage/data governance.
- Demonstrated understanding of security principals including OAuth, Role-Based Access Control and encryption. Experience with Snowflake Security and Data Governance.