Bioinformatics Data Scientist

Job ID:
Job date: 2018-06-23
End Date: 2018-08-22

Company : UCLA Health 

Country : United States 

Role : Research Scientist 


[Click Here to Access the Original Job Post]

Job Description:
UCLA Institute of Precision Health and DGIT are looking for an experienced and creative bioinformatician with solid scripting skills.

This team designs, implements, operates and maintains the high-performance compute infrastructure that is tailor designed to address the substantial processing, networking and data moving challenges arising from genomic science. Utilizing the elasticity of the AWS Cloud and innovative Big Data technologies, the team is to develop highly optimized state-of-the-art bioinformatics solutions. We are looking for a candidate with strong AWS, IaaS, Linux systems administration, HPC experience, who can leverage their expertise to provide investigators and data scientist the computing environment they need to make an impact in medical care by advancing our understanding of the role of the human genome in disease. While the primary responsibilities will focus around the compute infrastructure, you will have opportunities to contribute in the development of new bioinformatics software and genomic analysis pipelines.

The bioinformatician will work closely with a team of investigators, staff researchers and research computing IT in, but not limited to, the Clinical Exome Sequencing (CES) laboratory, Undiagnosed Disease Network (UDN) and Institute of Precision Health Clinical and Translational Science Institute (CTSI) under the direction of Dr. Stanley Nelson and Dr. Hane Lee to perform the following:

1. Building and running various NGS analysis pipeline such as whole exome sequencing, whole genome sequencing and RNAseq;

2. automating the bioinformatics pipeline;

3. building variant database;

4. assisting investigators to query the variant and medical record database;

5. variant annotation;

6. genomics data-set integration

Qualifications:

BS or higher in computational biology, human genetics, bioinformatics, computer science, or related field, or equivalent experience. The candidate must have 2+ years experience analyzing big data on the order of tens of millions of records. The candidate must have 2+ years experience working with genomic data and building genomic and bioinformatics pipelines. The candidate must have demonstrated proficiency working in AWS and distributed computing environments. The successful applicant must have strong programming skills: shell scripting, Python, Perl, C++, SQL and Java are preferred. Proficient in one of the programming language such as PHP, Perl, Python, R; And proficient in Unix/Linux OS and shell scripting. Strong problem solving abilities.

Excellent verbal and written communication skills are required. Strong organizational and interpersonal skills will be needed in our collaborative and fast-paced team. Research, test, recommend, and place orders for new or alternative computing equipment as enhancement to existing systems as needed and in consideration of budget constraints and cost/benefit analysis; design and set-up pre-production test environment for new systems. Investigate performance and process failures and work to improve them. Provide consultation for researchers on hardware and software solutions for their research needs, including design and maintenance of research-based pipeline and coordination with vendors of specialty computing systems required for specific projects.

Document and work with other members of the IT team to cross-train on AWS deployments as they occur. Liaise with vendors and other IT personnel for problem resolution. Experience with WES/WGS variant analysis and interpretation. Experience with RNAseq data analysis and interpretation. Experience with bioinformatics technologies such as sequence alignment, clustering and blasting, Splicing, Next Generation Sequencing . Strong database experience and big data structure design and querying along with utilizing tools/platform: Rstudio, Matlab, Zeppelin, Jupyter, iPython, Hadoop, Spark, Hive..etc.


Requeriments :

Skills :

Areas :


Additional Info:

[Click Here to Access the Original Job Post]