Difference between revisions of "SkimGBS"

From Applied Bioinformatics Group
Jump to: navigation, search
(Calling recombinations)
(FAQ)
Line 196: Line 196:
  
 
== FAQ ==
 
== FAQ ==
To come.
+
No-one has asked any questions yet - if you have any, send me a mail: philipp.bayer@uqconnect.edu.au
 
 
  
 
== Reference ==
 
== Reference ==

Revision as of 06:45, 9 September 2014

This page collects the scripts and a manual for our SkimGBS genotyping by sequencing pipeline.

Dependencies

For the Perl-scripts, just BioPerl. For the Python-scripts, BioPython and at least Python 2.7. we haven't tested the pipeline much with Python 3 but it should work.

Download

How to install

Download the package from above, then just extract the files to a directory of your choosing. The pipeline is a collection of Perl and Python scripts so no need to compile anything.

How to run SkimGBS

As a general overview: First align all reads, then call SNPs for the parental individuals using SGSautoSNP, then call genotypes for the population using snp_genotyping_all.pl, then filter SNPs using removeMonomorphicSNPs.py, then impute using imputeFlapjackSNPs.py, then visualize using Flapjack: http://ics.hutton.ac.uk/flapjack/

Data

As input data, you need BAM alignments for your two parental individuals, and one BAM alignment for each individual of the DH or RIL population.

Alignments

You can filter reads based on quality first, however, since SGSautoSNP doesn't allow heterozygous SNPs, receiving a false-positive SNP from two reads that have exactly the same sequencing error seems unlikely.

Important: SGSautoSNP determines which read comes from which cultivar by checking the read name and expects either to see a P (for paired) or S (for single) followed by the number of the library. In practise, it's enough to use P1- for both cultivars. Example: We used Tapidor and Ningyou, so we added 'TP1-' and 'NP1-' to both read collections. You can do that either before you run the alignments, or after the alignments, here it is after the alignments using SOAPaligner:

   sed 's/^/TP1-/' all.soap > all_renamed.soap

We use SOAPaligner with the option -r 0 to exclude homeologous SNPs from reads aligning in several places, alternatively, filter BWA or Bowtie for quality scores, something above 20 or 30 should work.

   samtools view -q 30 mappings.bam > mappings_filtered.bam

It is best to use only reads that align in pairs in order to increase the accuracy of called SNPs, so depending on your aligner you will have to filter unpaired reads. In SOAPaligner, just use the paired output files.

We merged all files for the two parents:

  samtools merge TapidorAndNingyou.bam Tapidor.bam Ningyou.bam

Remove clones using Picard's MarkDuplicates:

  java -jar `which MarkDuplicates.jar` INPUT=TapidorAndNingyou.bam OUTPUT=TapidorAndNingyou_cleaned.bam METRICS_FILE=stats.txt REMOVE_DUPLICATES=true

This overfilters a bit, since by chance you will find some reads align from different libraries in exactly the same position, so they may be treated as clones. Alternatively, you can run MarkDuplicates for each library itself and merge the results.

To make sure everything worked fine, use Picard again:

  java -jar `which ValidateSamFile.jar` INPUT=your_output_file

SGSautoSNP can either run one process per reference chromosome using multiple threads, or you can run many SGSautoSNP runs, one per chromosome. I prefer the latter, so I split the merged parental reads by chromosomes using bamtools. Both give you exactly the same results, running one SGSautoSNP per chromosome is easier to do when you have job arrays as with PBS.

  bamtools split -in TapidorAndNingyou_cleaned.bam -reference

This will result in one BAM file per chromosome which includes the data of both parental cultivars.

SNP-calling

SGSautoSNP can produce SNPs for FASTA files containing one or several references. To make sure that it offsets correctly, you have to supply a gff3 file of contigs, even if it's only one chromomosome. For that, we have supplied the fast_multiple_to_single_fasta.py script, which will produce a gff3 file for each chromosome.

Usage:

   fast_multiple_to_single_fasta.py C1.fasta

Will create C1_pseudo.gff3, this file is needed for SGSautoSNP.

To run SGSautoSNP:

--bam C1.bam

--chr_offset C1_v4.0_contig.gff3 (from fast_multiple_to_single_fasta.py)

--fasta chromosome_m.fasta (the chromosome - we just need this to extract the sequence ID)

--cultivars T, N (list of your cultivars you renamed to, so you should have NP1-, TP1- or something in your bam-file)

--contig_output Tapidor_Ningyou_SNPs (BA_01_output)

--chr_output Tapidor_Ningyou_SNPs_chr (also output - because SGSautoSNP is slightly buggy this will be identical to the above)

--cpu 1 (amount of CPUs used, 1 per chromosome)

Output files are the SNPs in snp, vcf and gff3 format.

Usage:

   python SGSautoSNP.py --bam BA01.bam --fasta BA01.fasta --contig_output BA_01_output --chr_offset BA01.gff3 --chr_output BA_01_output_chr --cultivars "T,N" --cpu 1

Genotyping

Next, we run the genotyping script, one run per individual of the DH or RIL mapping population and per chromosome.

Usage:

   perl snp_genotyping_all.pl -b TN10_BC0FH8ACXX_GGCTAC_L008_R1.fastq.par_sorted.bam -gff Contigs_IsobellC1_snp.gff3 -off Contigs_IsobellC1_pseudo_contig.gff3 -f Contigs_IsobellC1.fasta -o IsobellC1_TN10_GBS -c1 T -c2 N -h TN -s 1 > TN10_GBS.out

Explanation:

   TN10_BC0FH8ACXX_GGCTAC_L008_R1.fastq.par_sorted.bam - is the BAM  file for the individual TN10 of the DH populatioin mapped against the whole reference.
   Contigs_C1_snp.gff3 - is the gff3 file from SGSautoSNP for chromosome C1 containing all SNPs
   Contigs_C1_pseudo_contig.gff3 - is the offset file to map each chromosome contig to a global chromosome location from 
   Contigs_C1.fasta - the chromosome fasta file
   T - the letter for the first cultivar
   C1_TN10_GBS - the output file name

It's best practise to have one output folder per chromosome since this step will generate many files - 3 files per individual per chromosome, so for 20 chromosomes and 92 individuals that means 92 * 3 * 20 = 5520 files.

Flapjack visualization

To display the genotype data in flapjack, use the script parseGenotype.pl, for example for a directory called GBS_C8 containing the results from the above script:

  parseGenotype.pl -out GBS_C8 GBS_C8/*_csv

This will take all TN samples in the GBS_C8 directory and output both .genotype and .map files for importing to flapjack. The .genotype file is identical to the .dat file Flapjack exports, so you may want to rename it if you like consistency.

Important: This script assumes that the first element of your _csv-file is the name of the individual. Example: chrUnrandom_TN99_GBS_TN.genotyping_csv is correct for the individual TN99.

If your chromosome-names contain underscores (i.e., "chrA01_random") this script will generate garbage output where the names of individuals are replaced by "random" or whatever your second element is. Replace the "_" in your chromosome names first before running.

Example: Rename "./GBS_chrUn_random/chrUn_random_TN99_GBS_TN.genotyping_csv" into "./GBS_chrUn_random/chrUnrandom_TN99_GBS_TN.genotyping_csv"

Removing monomorphic SNPs

Some parental SNPs behave monomorphically in the population, either because they are false positives or because the coverage in the population is too low to observe the second allele.

   python removeMonomorphicSNPs.py map-file dat-file

In case there are individuals that have a too low coverage, use 'grep -v' to remove them:

   grep -v 'TN99' dat_file > cleaned_dat_file

To impute the cleaned results:

  python imputeFlapjackSNPs.py some_flapjack_outputfile.dat

Sideways imputation

To impute the cleaned results:

  python imputeFlapjackSNPs.py some_flapjack_outputfile.dat

This step will put out a .dat-file ending in _imputed.dat.

Imputation follows these simple rules for the two genotypes N and T, assuming that recombinations are so rare so that we can ignore them:

   N, , , N -> N, n, n, N
   T, , , T -> T, t, t, T
   T, , , N -> T, , , N
   , , , T -> , , , T  

Calling recombinations

Usage:

1-makeGCAndLDMapPerIndividual.py dat_file map_file

Will make two files:

   dat_file_crossovers.txt
   dat_file_gene_conversions.txt


Prints a tab-delimited table for each individual like this:

TN21
#Allele	Start SNP	End SNP	Start position	End position	Length	Outer End SNP	Outer End Position	Outer End Length\
T	M1	M3	1	3	2	M5	7	6

dat_file_gene_conversions.txt contains all recombinations >20bp and <10kb, crossovers contains everything bigger 10kb.

Filtering GCs and COs

Since so many GCs and COs overlap, I wrote a script that removes COs whose outer end, inner end or outer start and inner start overlap. Here's some ASCII art to show the difference between outer and inner end:

AAAAA------BBBBBBB--------AAAAAA
     |     |     |       |
     o     i     i       o

The outer start of that B recombination is the first SNP after the last SNP of the first A block (first |, o), the inner start is the first B, the inner end is the last B, the outer end is the last SNP before the next A.

Here's an example for a filtered GC in two individuals:

TN1 AAAAAA-----------BBBBBBBBBBBBBBBBBB
TN3 AAAA-------BBBBBBBBBBBBBBBBBBBBBBBB

This CO overlaps so we ignore it.

python 3-fuzzyFilter.py table_to_filter CO/GC

You have to specify CO or GC - adjacent COs are merged, adjacent GCs are left alone.

This will create output files with the 'filtered' suffix.

FAQ

No-one has asked any questions yet - if you have any, send me a mail: philipp.bayer@uqconnect.edu.au

Reference

Back to Main_Page