Skip to the content.

Welcome to the HBC Training Program

We are delighted to have you here!

The training team at the Harvard Chan Bioinformatics Core provides training to help biologists become comfortable with using bioinformatics tools to analyse high-throughput sequencing (HTS) data.

We offer courses and skills at three different levels starting at the basics and building upwards. We focus on the two most commonly used HTS interfaces, R and Bash/Shell.

Are you feeling lost and unsure about bioinformatic anaysis?

Click on the following questions to expand them for the answers:

What the heck is 'omics?
Over the last 10-15 years many technological advances allow us to assess the entirety of a certain type of molecule(s) in an organsim. The resulting high-throughput data are called 'omics data. We can break 'omics down into 4 specific categories:


High-throughput data from even a single sample is considered 'omics data. However, we usually are looking at data from large number of biological samples (individuals, cell lines, etc).


What is High-throughput Sequencing (HTS) or Next-generation Sequencing (NGS) data?
Both, genomes and transcriptomes, contain hundreds of millions or billions of nucleic acid units or bases/base pairs (A,T,G,C). Compare that to the average length of a book, which is 375,000 characters. To "read" the sequence of As, Ts, Gs and Cs, we use different methods (a lot of which are PCR-based). The most basic way to sequence DNA is using Sanger Sequencing. Reading those bases one at a time using the Sanger method takes a very long time with high per-base costs, but it was creatively utilized to complete the Human Genome Project (HGP) 1990 - 2003. With the massive advancements spurred by the HGP, the field of "next-generation" sequencing exploded and had rapidly advanced such that now we are able to sequence a whole genome within a day, at a nominal cost. The analyses of these big data generated by HTS is the challenge at present.
Over the last few years the community is slowly replacing the term NGS (Next-generation Sequencing) with the more descriptive HTS (High-throughput Sequencing).
There are hundreds of assays that have been developed for HTS that have enabled us to gain deep insights into the working of a cell. The most commonly used HTS applications that you will encounter are:
How do clusters and HPC relate to analysis of HTS data?
Let's return to our book example. If one book is 375,000 characters then 3.2 billion characters (the size of the human genome) translates to 8,533 books! While we might keep tens or even hundreds of books at our house, most people will never have thousands.

Can you imagine dusting this?

It's the same with our local computer. While we might keep small data files on our laptop, we don't want to clutter it up with huge data files. And this is just thinking about storage! Books or data sets need to be organized and kept track of as well. You might be able to alphabetize or organize a hundred books on your own but working with >8,000 books would be overwhelming! The same goes for our computer. To organize billions of base pairs and make sense of our sequencing data we simply need more power. The Mac laptop I am writing this on has 10 cores (a single unit of processing available in our CPU; see below for more information). In comparison, a high perfomance computing (HPC) cluster might have hundreds or thousands of cores. That is a lot more processing capacity, more in line with the large amount of computational work we want to do! Let's take a quick look at the basic architecture of a cluster environment and some cluster-specific jargon.

The above image reflects the many computers that make up a "cluster" of computers. Each individual computer in the cluster is usually a lot more powerful than any laptop or desktop computer we are used to working with, and is referred to as a "node" (instead of computer). Each node has a designated role, either for logging in or for performing computational analysis/work. A given cluster will usually have a few login nodes and several compute nodes. Each individual node in an HPC environment is a lot more powerful than any laptop or desktop computer we are used to working with. What we mean by powerful here is that each of these nodes have:
E.g. A cluster "Node" that has eight "quad-core" CPUs, means that node has 32 cores (ability to process 32 computations at a time). The data on a cluster is also stored differently than what we are used to with our laptops and desktops, in that it is not computer- or node-specific storage, but it is external and is available to all the nodes in a cluster. This ensures that you don't have to worry about which node is working on your analysis.

Why use the cluster or an HPC environment?

  1. A lot of software is designed to work with the resources on an HPC environment and is either unavailable for, or unusable on, a personal computer.
  2. If you are performing analysis on large data files (e.g. high-throughput sequencing data), you should work on the cluster to avoid issues with memory and to get the analysis done a lot faster with the superior processing capacity. Essentially, a cluster has:
    • 100s of cores for processing!
    • 100s of Gigabytes or Petabytes of storage!
    • 100s of Gigabytes of memory!

Parallelization

Point #2 in the last section brings us to the idea of parallelization or parallel computing that enables us to efficiently use the resources available on the cluster.

One input file

Let's start with the most basic idea of processing 1 input file to generate 1 output (result) file. On a personal computer this would happen with a single core in the CPU.

On a cluster we have access to many cores on a single node, so in theory we could split up the analysis of a single file into multiple distinct processes and use as many cores to speed up the generation of an output file. This is called multithreading, i.e. using multiple threads or cores. As you can imagine, multithreading can speed up how fast the analysis is performed! In the example below, the input file is analyzed using 8 cores, likely resulting in an 8-fold speed up!

Note: Multithreading is done internally by analysis tools being employed, and not by manually splitting the input (except in very unusual circumstances).

Three input files

Now, what if we had 3 input files? Well, we could process these files in serial, i.e. use the same core(s) over and over again, as shown in the image below.

This is great, but it is not as efficient as multithreading each analysis, and using a set of 8 cores for each of the three input samples. This is actually considered to be true parallelization.

With parallelization, several samples can be analysed at the same time!
What is shell and how does it relate to clusters?
So how might you actually use a cluster? Unfortunately you can't just walk up to where the cluster is stored and start using it. Clusters are accessed remotely, that means that you connect to the cluster from your own computer. You will do this from the command line or a text-based user interface. We are used to clicking on applications we want to use and selecting various commands from dropdown menus. Clusters do not work this way. Any task that you want a cluster to do has to be communicated through a text command.

The FAS-RC Cluster

If you have never taken a computer science course or worked with clusters before this will all be brand new to you. But don't worry, we have courses for that! For now let's just review the basics. To look at command line on your own computer you can open the Terminal program on Macs or for Windows download Git BASH or similar application. The shell is what runs in these programs to interpret your commands. These programs all use Bash, a command language. As you get into HTS and computational work you will encounter a lot of languages such as Python, Perl, Fortran, R, C++, Java and more. You can think of these as being akin to human languages; French and English sound very different and have different syntax (the order of words) but can be used to convey the same message. At HBC training we recommend that you become familiar (or fluent) in bash and R to begin with.
What is R and what can it do?
Why do we recommend R instead of other languages? According to R-project, "R is a language and environment for statistical computing and graphics." R is also a well developed and relatively simple language that is widely used among data scientists and people in STEM. Compelling arguements for learning R include:
Where do I go from here?
Hopefully you now feel like you have a grasp on some of these terms. If you want to start getting your hands wet, we recommend that you take our Intro to R Course and the appropriate shell intro for the cluster you will use, either O2 or FAS-RC. You are free to take a workshop with us or work through the lessons yourself at your own pace. See our below for all of our offerings.

See our current workshop schedule on our training website. More detailed information about our courses is found below.

What are the basic skills I need?

Skill Who needs it Overview Courses
A1 - Using the command line interface Anyone planning on doing scientific computing using the command-line. Understanding the need for shell and master basic commands
A2 - Using a HPC cluster Anyone who wants to efficiently run analyses on large datasets (requiring more computational resources than a laptop can provide). Understanding the components of a high performance compute cluster (HPC), and learning to navigate and properly use available HPCs at Harvard.
B - Using R Anyone who wants to learn a programming language that is especially useful for data wrangling and statistics. Learning the R and RStudio interface, Basic R syntax, and data visualization

How can I apply the basic skills?

Skill Who needs it Overview Courses
C - Analysis of HTS data in the HPC environment Anyone planning on doing genomic or transcriptomic next-generation sequencing and is interested in analyzing their own data. - Analysis of bulk RNAseq data, Variant Analysis, and sequencing data related to Chromatin biology starting with raw data.

- Automating the workflow with advanced shell scripts.
D - Statistical Analysis of HTS data in R Anyone who wants to use popular R packages for downstream analysis of HTS data. Main focuses include Seurat and DESeq2. - Using R to implement best practices workflows for the analysis of various forms of HTS data.

- Clear explanations of the theory behind each step in of the workflow.

How can I build my skillset further?

Skill Who needs it Overview Courses
E - Advanced programming with the bash command line Anyone who wants to create custom shell scripts and utilize bash for various tasks. Learning to include version control in your projects and advanced bash scripting
F - Advanced R for generating complex plots and reports Anyone who wants to make publication quality figures

Anyone who wants to make high level HTML reports of analyses
Exploring additional R features such as reports and publication perfect figures

Additional Courses


Contact us:

Email: hbctraining@hsph.harvard.edu

Webpage: http://bioinformatics.sph.harvard.edu/training/

These materials have been developed by members of the teaching team at the Harvard Chan Bioinformatics Core (HBC) RRID:SCR_025373. These are open access materials distributed under the terms of the Creative Commons Attribution license (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

A lot of time and effort went into the preparation of these materials. Citations help us understand the needs of the community, gain recognition for our work, and attract further funding to support our teaching activities. Thank you for citing the corresponding course (as suggested in its “Read Me” section) if it helped you in your data analysis.