Steven tasked me with assembling our geoduck metagenomics HiSeqX data. The first part of the process is examining the quality of the sequencing reads, performing quality trimming, and then checking the quality of the trimmed reads. It’s also possible (likely) that I’ll need to run another round of trimming. The process is documented in the Jupyter Notebook linked below. After these reads are cleaned up, I’ll transfer them over to our HPC nodes (Mox) and try assembling them.
Jupyter Notebook (GitHub):
RESULTS
Samples required three rounds of trimming:
Initial quality/adapter trimming.
Remove funky 5’ 10bp from each read.
Remove funky 5’ 10bp from each read (again? maybe I misread the number of bases needing to be trimmed from the previous trimming?)
Now that the reads are cleaned, will transfer triply-trimmed data to Mox for assembly.
Output folder:
Initial FastQC folder:
MultiQC Report (HTML):
TrimGalore! folder (initial qualitry trim):
Post-trimming FastQC folder (first round):
MultiQC Report (HTML):
TrimGalore! folder (second round, 10bp trim):
Post-trimming FastQC folder (second round, 10bp trim):
MultiQC Report (HTML):
TrimGalore! folder (third round, 10bp trim):
Post-trimming FastQC folder (second round, 10bp trim):
MultiQC Report (HTML):