May 01, 2023

Public workspaceSARS-CoV-2 consensus genome reconstruction, quality control, and lineage analysis

  • 1Australian National University
Icon indicating open access to content
QR code linking to this content
Protocol Citation: Benjamin Schwessinger 2023. SARS-CoV-2 consensus genome reconstruction, quality control, and lineage analysis. protocols.io https://dx.doi.org/10.17504/protocols.io.14egn2kp6g5d/v1
License: This is an open access protocol distributed under the terms of the Creative Commons Attribution License,  which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited
Protocol status: Working
We use this protocol and it's working
Created: April 30, 2023
Last Modified: February 06, 2024
Protocol Integer ID: 81199
Abstract
This protocols is part of the ANU Biosecurity mini-research project #2 "An SARS-COV2 incursion scenario: Genomics, phylogenetics, and incursions." This mini-research project is modeled on the yearly Quality Assurance Program of The Royal College of Pathologists of Australia (RCPAQAP), we take part in together with ACT Pathology.
This research project is split into two major parts, identical to how the official RCPAQAP is run every year.
Part #1 is focusing on the 'wet- lab' by sequencing SARS-COV2 from real world RNA samples provided by ACT Pathology especially for our ANU biosecurity course (Thank YOU!). Here you will amplify and sequence five (5) RNA samples per research group. You will assess the SARS-COV2 genome sequences for their lineage assignments using online programs, put sequences into a global context, estimate the collection date based on genetic information, and describe mutations in the spike protein.
Part #2 is focusing on the 'dry-lab' by investigating a hypothetical incursion scenario in the so-called city Fantastica. You will combine genomic surveillance of SARS-COV2 with case interview data to trace the spread into of SARS-COV2 in the community and into high risk settings. We will provide you with real publicly available SARS-COV2 genome and fantasized case interviews. You will put these two together to trace the spread and suggest potential improvements in containment strategies with a focus on high risk settings.
This protocol describes the analysis component of Part #1. The metrics you are suppose to report for each of your samples are mostly borrowed from the official SARS-CoV-2 QAP. Don't worry if not all of these mean something to you at the moment as we will explain them again during the prac. In case all/most of your samples have < 50% genome coverage please also include the analysis of MakeUp for points 1 to 7 and TimeMakeUp for point 8. You can access the MakeUp data here (ANU only). Make sure to read the README file so you understand what each item relates to.

The metrics you have to report for each of our samples (or MakeUps) include the following.

  1. Consensus genome coverage.
  2. Average read depth; You might want to include detailed read depth plots here as well.
  3. Pangolin Lineage.
  4. NextClade Lineage.
  5. Base pair differences relative to the original SARS-CoV-2 genome.
  6. Amino acid replacements and deletions in the S (spike) protein sequence.
  7. Evaluation if your and/or the MakeUp samples would make the QC cut-off 90% genome coverage and other metrics that you deem important for QC. Would you 'flag' any of your samples as standing out e.g. being negative control?
  8. Approximate sampling date of your sequences for month and year.

You must report the versions of all tools used in your report and the day the analysis was performed. This is extremely important for reporting as lineage naming and such change VERY frequently during the pandemic and any outbreak.
This protocol is applicable for week 9.

The following links might be useful for your report:

The original publication that describes the sequencing protocol is here: https://academic.oup.com/biomethods/article/5/1/bpaa014/5873518?login=true


Guidelines
You must have read, understood, and follow the health and safety instructions provided in the "Overview Mini-Research Project #2 BIOL3106/6106" provided on Wattle (ANU learning portal).
You must have signed and returned one copy of the "Student Safety Declaration Form For Practical Class Work" before starting any laboratory work.
You must have read and understood the Hazard Sheets (Risk assessment) of all chemicals listed bellow in the "Safety Warnings" section. These Hazard Sheets are provided on Wattle as part of the "Overview Mini-Research Project #2 BIOL3106/6106" document.
Safety warnings
This protocol does not require any hazardous substances or infectious agents. However, maintain a proper posture while working on your computer.
Before start
You must study the protocol carefully before you start. If anything is unclear post questions directly here on protocols.io.
Section I: Consensus genome reconstruction using Epi2Me and the 'wf-artic'.
Section I: Consensus genome reconstruction using Epi2Me and the 'wf-artic'.
This section aims to reconstruct each samples consensus genome based on the publication here https://academic.oup.com/biomethods/article/5/1/bpaa014/5873518?login=true and an adaptation of the artic Sars-CoV-2 analytics workflow https://artic.network/ncov-2019

The principle steps are:
  1. Read filtering on quality and length.
  2. Aligning reads to the reference.
  3. Normalizing to 200x per base read depth on average.
  4. Trimming of primer sequences.
  5. Variant calling for both pools independently.
  6. Combining variant calls of both pools.
  7. Generation of sample's consensus genome reconstruction.
  8. Generation of QC statistics like genome coverage, average read coverage etc.
  9. Additional analysis like lineage calling, variant quantification, variant impact, etc.

We will run all of this within Epi2Me till step 8. The input for the Epi2Me analysis will be the "fastq_pass" folder of the MinKNOW basecalling output. This folder contains one "barcodeXX" folder for each barcode identified by the basecalling software.

Overall, the Epi2Me "wf-artic" workflow takes two inputs.
  1. The basecalled fastq output of MinKNOW (or equivalent e.g. guppy).
  2. A sample sheet in csv (comma separated values) format with the columns "barcode", "alias", "type".

You might also want to execute this section on the MakeUp dataset if your data was of poor quality.
Your basecalled and demutiplexed, by barcode, fastq data should be in the 'fastq_pass' folder be in a similar to this C:\data\20230426-SARS-COV2-MakeUpData\no_sample\20230426_1208_MN20227_AOI905_2cab113b\fastq_pass.

fastq_pass folder with all the demultiplexed barcodeXX folders containing fastq files for each barcode
We need to update the sample sheet to reflect your specific run and samples. This sheet should be located in the same folder your 'fastq_pass' barcode folders. It should look similar to the one below. (Done for you.)



Before we start with the workflow we need to make a backup (e.g. copy of the 'fastq_pass' folder on the desktop for now) and then delete all the barcode folders we have not specified in the sample sheet. We also need to delete all folders of barcodes we have used but have not received any data.

Fastq_pass folder after clean-up and with sample sheet.

Now we are ready to set-up and run the Epi2Me workflow for artic analysis.
Start the "Docker App" and make sure it is running.



Start Epi2Me.



Click the "Installed workflows" bottom.



This should show "wf-artic" to be available.



Now select the "wf-artic".



Site note. Also consider reading the "README" section to get more background for your report.



Now we are ready to set up our "wf-artic" specific to our samples and wet-lab protocols.



Tell the program where to find your 'fastq_pass' folder.

Similar to "C:\data\20230426-SARS-COV2-MakeUpData\no_sample\20230426_1208_MN20227_AOI905_2cab113b\fastq_pass".



Now set up all the other fields.






Hit the "Launch Workflow" bottom.



Now all should be running smoothly.


... and this after 2-5 minutes.



If all goes well... you will see the following complete screen in about 45 minutes.



Now you can open the instance output by hitting the following link at the bottom of the page.


This will open the folder of your specific workflow instance that also contains your "output" folder.



This "output" folder contains LOADS of data. The most important is the "all_consensus.fasta" and the "wf-artic-report.html". We will also make use of some of the other files.


Now rename both files with unique and sensible identifiers for your group and share it with all the RG members.
The "all_consensus.fasta" contains your consensus genomes for each sample and the "wf-artic-report.html" looks like this. We will go over both in more detail during the prac.

Snap shot of html summary report

We will use the "all_consensus.fasta" for most downstream analyses. It contains the consensus fasta sequences of your genomes.
Second II: Visualization whole genome alignment and principle steps 2-7 of section I
Second II: Visualization whole genome alignment and principle steps 2-7 of section I
You will use IGV to look at one of your read sets mapped against the reference genome to get a better understanding of the "raw" data.
Open IGV.


Select "Genomes" > "Load Genomes from File.." and add the "MN908947_3.fasta" reference shared here (ANU only).


The file can end in ".txt" or better ".fasta"

Drag and drop the "MidnightONTv3.primer.bed" into IGV to illustrate where your primer sequences bind.



Load one of the "XXX.primertrimmed.rg.sorted.bam" files from the "output" folder of Epi2Me here (ANU only) or from your own dataset.




Now you see the alignments of all the reads as pileup against the reference, the read coverage and variants.


You can zoom into certain regions e.g. the S-gene 21563-25384 to see how the reads are different to the reference via the different colored bars in the grey area labelled with variants in the picture above. There is loads more to explore. Ask if you have any questions.
Section III: Nextclade for lineage calling, QC measures, amino acid substitutions identification and so much more
Section III: Nextclade for lineage calling, QC measures, amino acid substitutions identification and so much more
We will use Nexclade https://clades.nextstrain.org/ to get some quality assessment metrics, lineages calling, number of mutations, and much more.


Open the webpage https://clades.nextstrain.org/, select SARS-CoV-2 and drag and drop your consensus sequences into it.

Pretty please!

Thanks!

This is an example of a good Midnight consensus sequence set.



This is an example of a not so good Midnight consensus sequence set.



Explore the results page to get to the following:
  1. Pango lineage assignment.
  2. Genome coverage.
  3. Number of mutations relative to reference.
  4. Number of Ns.
  5. Gaps.
  6. Changes to the amino acid sequence of the S protein.
Make sure to download your samples via the Download dialog in the "TSV" (tab separated values) format. You can import this into Excel later.


Ask question during the prac for anything that is unclear.
Section IV: Pangolin lineage assignment
Section IV: Pangolin lineage assignment
Let's assign the latest lineage designation with Pangolin.



Drag and drop your sequences, start analysis, and hope for the best.
Step V: Use Usher to place your samples into a global context and see its closest public neighbor.
Step V: Use Usher to place your samples into a global context and see its closest public neighbor.
Upload your sequences



Explore the webpage for results including:
  1. Neighboring sample in Tree.
  2. That ^^ samples deposition date.
  3. Subtree of each of your sequences.



Ask questions in during the prac.