Share this Job

Scientific DevOps Engineer

Date: May 3, 2021

Location: Philadelphia, PA, US

Company: Spark Therapeutics

Join the Spark Team


We were born of innovation, springing from the curiosity, imagination and dedication of remarkable scientists and healthcare visionaries. Our shared mission is to challenge the inevitability of genetic disease by discovering, developing, and delivering treatments in ways unimaginable – until now.


We don’t follow footsteps. We create the path.



The Data Science Group at Spark Therapeutics is seeking a DevOps Engineer for the Scientific Computing function. SDOE serves as a key group member to support software development and data engineering needs of the department. Primary responsibilities of the SDOE include:


  • Work closely with statisticians, bioinformatics and machine learning scientists to convert and deploy development grade scripts, bioinformatics pipelines, and machine learning models into production ready software
  • Linux systems administration and VM management
  • In collaboration with the data science team and IT, implement a cloud-based infrastructure for software development, data storage/management, and bioinformatics pipeline design. Present practical input for strategic data science technology decisions.  
  • Day-to-day operation and support of the Data Science cloud infrastructure (e.g., AWS, Azure)




  • B.S. in Computer science, or related engineering field
  • 3+ years of DevOps, infrastructure management
  • 3+ years of experience with Linux environments
  • Work experience in biotech/pharma is preferred




  • Deep understanding of Linux/RHEL environments and VM management best practices
  • Strong coding skills in one or more of modern programming languages such as Python, Go
  • Experience with Cloud infrastructure (e.g., AWS, Azure) to provide IaaS/IaC, including compute, storage & networking (EC2, EBS, S3, Glacier, Gateway, VPC)
  • Strong knowledge of database admin fundamentals and best practices (Postgres preferred)
  • Infrastructure-as-code (IaC) and deployment automation tools e.g., Terraform, Ansible
  • Experience with using variety of APIs and web services technologies such as GraphQL, REST
  • Experience with building solutions for high-dimensional datasets such as genomic data using Big Data technologies (e.g., HDFS, Impala, Spark, etc.)
  • Knowledge of enterprise software patterns and how they may be leveraged in modern infrastructure management
  • Strong comprehension of Agile/Scrum methodologies, Software Development Life Cycle, Source Control systems, and CI/CD pipelines
  • Experience in LIMS systems, next-generation sequencing, and managing high performance computing systems and bioinformatics is preferred
  • Proven ability to work independently, self-motivated to learn and develop new methodologies, manage multiple projects simultaneously, keep accurate records, follow instructions, and comply with company policies
  • Ability to ramp up quickly and learn tools and technologies
  • Excellent communication skills (both oral and written)
  • Works with colleagues as a team player


Additional requirements


Travel may be required, but infrequently and for short periods of time. Responsible and detail-oriented working practice is needed. Ability to prioritize work stream to meet complex deadlines for multiple simultaneous projects.

Nearest Major Market: Philadelphia