Code Monkey home page Code Monkey logo

indelible's Introduction

InDelible: Genomic Structural Variant Caller by Adaptive Training

Table of Contents

  1. Development Team
  2. About InDelible
  3. Installation
  4. Usage
  5. Output

Development Team

Alejandro Sifrim (Creator, Developer)

Eugene Gardner (Developer)

Diana Rajan (Experimental validation)

Elena Prigmore (Experimental validation)

Sarah Lindsay (Experimental validation)

Matthew Hurles (Group Leader)

We are affiliated with the Wellcome Sanger Institute, Cambridge, United Kingdom

About InDelible

Version 1.1.3

Abstract

Structural Variations (SVs) are genetic differences greater than 50bps in size and have been identified as causative of diseases such as rare developmental disorders (DD). Patients presenting with DD are typically referred for chromosomal microarray (CMA) to identify large copy-number variants or for single gene or panel tests based on presenting symptoms. Increasingly, patients for which a diagnosis is not forthcoming are additionally referred for whole exome sequencing (WES) which is used to identify single nucleotide variants or small insertions/deletions (InDels). This leaves a class of intermediate size deletions that are technically difficult to identify, hence patients with SVs undetectable by conventional CMA or WES analysis often remain undiagnosed. To this end, we have developed a novel SV discovery approach, ‘InDelible’, and applied it to 13,438 probands with severe DD recruited as part of the Deciphering Developmental Disorders (DDD) study. InDelible queries WES data to identify split read clusters within a gene set of interest and performs variant quality-control utilizing an active learning methodology. Using InDelible we were able to find 59 previously undetected variants among DDD probands, of which 89.8% (53) were in genes previously associated with DD, had phenotypes which putatively match the conditions of the patient in which they were found, and were thus reported to the referring clinician. InDelible was particularly effective at ascertaining variants between 20-500 bps in size, and increased the total number of causal variants identified by DDD in this size range by 46.4% (n = 26 variants). Of particular interest were seven confirmed de novo SVs in the gene MECP2; these variants represent 35.0% of all de novo PTVs in MECP2 among DDD patients and represent an enrichment of large, causal variants compared to other DD-associated genes. InDelible provides a rapid framework for the discovery of likely pathogenic SVs and has the potential to improve the diagnostic yield of WES.

What is InDelible for?

InDelible was originally designed for the ascertainment of large InDels (>20bp) and Structural Variants from Whole Exome Sequencing (WES) data for which other ascertainment approaches have proven refractory. To reduce the search space, InDelible also makes use of a "target gene list" containing genes that the end user is interested in.

Potential Limitations of InDelible

InDelible is likely to be adaptable to a wide range of sequencing technologies, genetic architectures, and disease/conditions. However, we have not specifically tested InDelible with:

  1. Whole Genome Sequencing (WGS) Data

    • While InDelible should, in theory, be able to identify variants from WGS, the number of reads that InDelible has to process would likely result in significantly increased run-times.
  2. Other Diseases

    • InDelible was designed as part of the Deciphering Developmental Disorders (DDD) study and, as such, was targeted to genes known to contribute to dominant developmental disorders. We have provided for the possibility that end-users may want to identify variants within other gene sets, but have not specifically performed variant discovery among a patient cohort to test this functionality.
  3. Other Genetic Architectures

    • While causal de novo variants play an important role in the genetic architecture of severe DD, recessive causes of DD are likewise a major contributor. While we have focused our primary analysis on de novo variation to try to maximise our discovery potential, we do not preclude the possibility that InDelible could also be used to identify homozygous or compound heterozygous variants which could be plausibly linked to a patient's symptoms.
  4. Higher Allele Frequency Variants

    • Likewise, since InDelible was developed as part of DDD, where the largest contributor to patient symptoms is high-penetrance de novo variation, we focused our filtering to such variants. However, InDelible does report all high-confidence variants identified for each patient and could, in theory, be used to identify population level variation.

How to Cite InDelible

Peer-Reviewed Open Access Manuscript

Eugene J. Gardner, Alejandro Sifrim, Sarah J. Lindsay, Elena Prigmore, Diana Rajan, Petr Danecek, Giuseppe Gallone, Ruth Y. Eberhardt, Hilary C. Martin, Caroline F. Wright, David R. FitzPatrick, Helen V. Firth, Matthew E. Hurles. Detecting cryptic clinically relevant structural variation in exome-sequencing data increases diagnostic yield for developmental disorders. The American Journal of Human Genetics (2021).

https://www.cell.com/ajhg/fulltext/S0002-9297(21)00346-3

Preprint

Eugene J. Gardner, Alejandro Sifrim, Sarah J. Lindsay, Elena Prigmore, Diana Rajan, Petr Danecek, Giuseppe Gallone, Ruth Y. Eberhardt, Hilary C. Martin, Caroline F. Wright, David R. FitzPatrick, Helen V. Firth, Matthew E. Hurles. Detecting cryptic clinically-relevant structural variation in exome sequencing data increases diagnostic yield for developmental disorders. medRxiv (2021).

https://www.medrxiv.org/content/10.1101/2020.10.02.20194241v2

For code used to generate figures/text for these manuscripts, please see the following repository:

https://github.com/HurlesGroupSanger/indelible_paper

Installation

Required Software Dependencies

InDelible is written for Python2.7.* or Python3.7.*

InDelible requires the following software to be installed and in $PATH:

  • bedtools
    • Note: If using CRAM formated files with InDelible, bedtools v2.28 or later is required.
  • tabix
  • bgzip
  • bwa
  • blast
    • The python package Biopython requires a local install of in $PATH in order to function. This needs to be installed prior to installing InDelible.

Installing InDelible on a Local Machine

To install InDelible:

Note: We recommend cloning the latest version of our github repo unless you need to reproduce your analysis with an older version. This should allow us to readily fix issues that you may experience with the software.

  1. Clone the git repo:
git clone https://github.com/eugenegardner/indelible.git
cd indelible/
  1. Create a virtual environment and activate it:
  • for Python2:
virtualenv venv/
source venv/bin/activate
  • for Python3:
python3 -m venv venv/
source venv/bin/activate
  1. Install cython, numpy, and other required packages:
pip install "cython==0.29.13"
pip install "numpy==1.17.2"
pip install -r requirements.txt

Note: If you get error(s) about pysam not being able to load specific libraries (like openssl, libbz2, etc.) that is a pysam problem related to htslib. Please see the pysam website to get help.

Note: We have tested this installation protocol, but other versions of pip may try to install packages out of order. If you get errors pertaining to dependencies stored within requirements.txt, you may need to install them one at a time in the order listed in requirements.txt. In particular, pandas may not properly install unless numpy is installed first, as we have done above.

Indelible was tested with the following package versions:

  • cython v0.29.13
  • numpy v1.17.2
  • pandas v0.25.1
  • pyfaidx v0.5.5.2
  • pysam v0.18.0
  • scipy v1.3.1
  • scikit-learn v0.21.3
    • Note: The random forest model provided in data.zip will likely only work with v0.21.3 of scikit-learn (see this link for an explanation why). Thus, if using a different version of scikit-learn, it is necessary to re-train the model with the provided test set included with this repository (data/observation_data.DDD.17IX2019.txt). Please see documentation below on how to train the random forest used by InDelible.
  • PyYAML v5.4
  • Biopython v1.74
  • intervaltree v3.0.2
  1. Unzip required data files:
cd data/
unzip data.zip
  1. Download required reference resources and create bwa/blast databases:
## Download the GRCh37 (or GRCH38) human reference and create the repeat blast db:
wget ftp://ftp-trace.ncbi.nih.gov/1000genomes/ftp/technical/reference/phase2_reference_assembly_sequence/hs37d5.fa.gz
wget ftp://ftp-trace.ncbi.nih.gov/1000genomes/ftp/technical/reference/phase2_reference_assembly_sequence/hs37d5.fa.gz.fai
gunzip hs37d5.fa
mv hs37d5.fa.gz.fai hs37d5.fa.fai

## Make bwa index
bwa index hs37d5.fa

## Make repeat blast_db
makeblastdb -in repeats.fasta -dbtype nucl
  1. Edit the config file to point to required data files and edit any default parameters:

An example config.yml file (hg19: example_config.hg19.yml, hg38: example_config.hg38.yml) is included in the top level directory. Edit this and save as config.<VERSION>.yml.

cd ..

## Provide the path to each required file at the top of the config.<VERSION>.yml
## Edit parameters at bottom of file.
vim example_config.yml

## cp/mv to config.<VERSION>.yml
cp example_config.<VERSION>.yml config.<VERSION>.yml

Parameters in config.yml are as follows:

HC_THRESHOLD: Quality threshold at which to hard clip from the ends
MINIMUM_LENGTH_SPLIT_READ: Minimum number of bases that need to be soft clipped for a read to be included
MINIMUM_MAPQ: Minimum mapping quality for read to be included
MININUM_AVERAGE_BASE_QUALITY_SR: Minimum average base quality for the soft-clipped segment for read to be included
SHORT_SR_CUTOFF: The cutoff at which a soft-clipped segment is considered "short".
MINIMUM_SR_COVERAGE: Minimum number of reads with soft-clipped segments at a position for that position to be outputted.
SCORE_THRESHOLD: Minimum InDelible score to be considered for denovo calling
SR_THRESHOLD: Maximum number of clipped reads in parental samples to be considered inherited
COV_THRESHOLD: Minimum parental coverage to be able to call event as denovo
WINDOW_SIZE: window around position to look for indels/clipped reads (window_size/2 to the left and to the right)

Big Note: The SR_THRESHOLD parameter has a substantial effect on the runtime of InDelible. The default parameter was designed for the WES data generated by the DDD project and was tuned towards sensitivity rather than specificity to ensure diagnoses were returned to patients. Setting this parameter even slightly higher (e.g. 5) decreases runtime significantly.

Note: The above data resources will only work if you run InDelible using the the human GRCh37/38 reference.

A brief note on Hg38 resources:

  • The InDelible MAF database provided with this distribution is a liftOver of project resources generated using version hg19 of the human genome. When performing liftover, approximately 0.8% of breakpoints were lost and are not included in the resources file. Additionally, the position accuracy of variants contained there-in should not be considered 100% accurate. This is particularly true for variant size estimates – we did not recalculate variant lengths.
  • The training data packaged with InDelible is technically agnostic to genome build. However, differences between alignment methods and/or genome builds may result in slight differences in covariate importance. As such, it may be a good idea to train the random forest with data from study being assessed, if sufficiently large enough.
  • As gnomAD has not generated pLI scores specifically using Hg38-aligned genomes, pLI values are from gnomADv2.1.1.

Note: All commands take the command-line option --config. The user must provide the path to a valid config yml to InDelible at runtime.

Using vr-runner

InDelible comes packages with a script for use with the vr-runner packaged which automates the analysis of multiple samples. For information on how to install and use vr-runner, please see this link.

The InDelible-specific vr-runner script is located at ./vr_runner_scripts/run-indelible. This script requires a list of tab-delimited bam paths, where columns are:

  1. child bam
  2. mum bam
  3. dad bam

For example:

/path/to/child.bam  /path/to/mum.bam    /path/to/dad.bam

If mum or dad bams are not available, substitute a "-" like:

/path/to/child.bam  -   -
/path/to/child.bam  /path/to/mum.bam    -
/path/to/child.bam  - /path/to/dad.bam

First setup the vr-runner config file:

./indelible/vr_runner_scripts/run-indelible +sampleconf > my.conf
## Change ALL parameters to point to the correct files:
## bams = list of bams made above
## ref = reference genome.fa
## config = config.yml
## indelible = path to the main indelible.py script
## database = path to the indelible db

A simple command for running InDelible with vr-runner is as follows:

./indelible/vr_runner_scripts/run-indelible -o ./output/ +config my.conf +maxjobs 1000 +loop 100 +retries -2

Resulting output will be in the ./output/ folder.

Additional commands for run-indelible:

   -b, --bams-list <file>      File with bam files
   -o, --outdir <dir>          Output directory
   -d, --database              Path to a precomputed MAF database [null]

Note: If a filepath to a prebuilt MAF database (e.g. ./Indelible/data/) is not provided for -d/--database, InDelible will rebuild the MAF database based only on samples included in -b!

Using Singularity or Docker

We have developed both a Docker and derived Singularity VM image to enable quick deployment of InDelible to both local and cloud compute platforms. The InDelible docker image is available through Docker Hub here:

https://hub.docker.com/repository/docker/mercury/indelible

This Docker/Singularity image contains both an install of InDelible running on Python3.7 as well as required library files as described in the local install instructions. We describe here a basic command-line for running InDelible via Singularity. Please see the Singularity documentation for more detail instructions for running Singularity itself. Docker can then be run for either Hg38 or Hg19 using the following example command for the complete subcommand:

# pull the docker image:
docker pull mercury/indelible:hg19_feb0a6f

# Example with hg19
docker run -v /home/user/:/scratch/ mercury/indelible python3 indelible.py complete \
    --config config.hg19.yml \
    --i /scratch/indelible/test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam \
    --o /scratch/indelible/test_data/ \
    --r data/hs37d5.fa \
    --priors data/Indelible_db_10k.bed

# Example with hg38
docker run -v /home/user/:/scratch/ mercury/indelible python3 indelible.py complete \
    --config config.hg38.yml \
    --i /scratch/indelible/test_data/DDD_MAIN5194229_Xchrom_subset_sorted.hg38.bam \
    --o /scratch/indelible/test_data/ \
    --r data/GRCh38_full_analysis_set.fa \
    --priors data/Indelible_db_10k.hg38.bed

Additional notes on the above command(s):

  1. The docker pull command will change according to the latest tag on Dockerhub. This tag should align with the current version number, but please check dockerhub to be sure
  2. -v simply mounts the current home directory as a directory within the Docker instance. This will need to be modified according to your local file system
  3. If any of the files provided to config.*.yml need to be modified, you will need to create a new config file outside of the Docker instance and provide modified paths.

The Dockerfile used to create this image is available as part of this repository (./Dockerfile). We have also developed a separate github repository which hosts both this Dockerfile and instructions for converting this Dockerfile into a Singularity image:

https://github.com/wtsi-hgi/indelible-docker/tree/master

An example singlularity command follows:

singularity exec \
    -c \
    --cleanenv \
    --bind /path/to/local/storage/device/ \
    --pwd /usr/src/app/Indelible/ \
    /path/to/indelible_singularity_container.sif \
    indelible.py complete \
    --i /path/to/local/storage/device/proband.cram \
    --o /path/to/local/storage/device/ \
    --r data/hs37d5.fa \
    --d data/Indelible_db_10k.bed \
    --m /path/to/local/storage/device/mum.cram \
    --p /path/to/local/storage/device/dad.cram

The above command will run the complete InDelible SV discovery pipeline on the file proband.cram. Additional notes on the above command:

  1. -c and --cleanenv clean your enviornment variables and home directory prior to running InDelible. These may not be strictly necessary, but are recommended to prevent any conflicts with the environment internal to the Singularity image.
  2. --bind is meant to be used to mount your local storage folder to the Singularity image and should be pointed to the location on your local machine/cluster where your sequence files are stored.
  3. --pwd is the location of the InDelible directory within the Singularity image. This part of the command line must not be changed.
  4. --r and --d point to reference files within the Singularity image. The other references files are also located within the Singularity instance at 4. --rand--d point to reference files within the Singularity image. Prebuilt references files are also located within the Singularity instance at /usr/src/app/Indelible/data.

Big Note: If using InDelible for anything beyond the built-in data files, you will need to generate a local version of the config.hg19.yml file, edit it to point to your own resource files, and then point indelible.py to it with --config /path/to/config.new.yml. The paths in config.new.yml can reflect a mix of both local paths and paths already within the Singularity instance.

Usage

The main help page of the program can be accessed by executing the InDelible script with the -h flag as follows:

./indelible.py -h
usage: indelible [-h] <command> ...

positional arguments:
  <command>   One of the following commands:
    fetch     fetch reads from BAM file
    aggregate aggregate information per position
    score     score positions using Random Forest model
    database  build SR allele frequency database
    annotate  annotate positions with additional information
    denovo    searches for de novo events
    complete  Performs the complete InDelible analysis
    train     trains the Random Forest model on a bunch of examples

optional arguments:
  -h, --help  show this help message and exit

Primary SV Calling Pipeline

The InDelible variant calling process follows several steps:

  1. Fetch – Soft-clipped reads are extracted from the BAM files.
  2. Aggregate - Information is aggregated across reads to find positions where multiple reads are clipped.
  3. Score - Positions are scored using a Random Forest model taking into account the number/quality of clipped reads and the sequence context.
  4. Database - Merges sites across individuals, assigns breakpoint frequency, and determines variant type and breakpoints where possible.
  5. Annotate - Putative SVs are annotated with additional information (e.g. gene annotations) and the positional results from the previous step.
  6. denovo - de novo events are called and inheritance information is appended.

Note: All commands also take the command-line option --config. The user can either change the default config.yml, or provide a path to a different file with this option. See the section above for configuring the config.yml file.

These different steps can be performed by the different sub-commands:

1. Fetch

The fetch command extracts the reads from the BAM file, it takes 2 arguments:

  • --i : path to the input CRAM/BAM file.
  • --o : path to output the clipped reads to.
  • --config : path to the config.yml file.
./indelible.py fetch --i test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam --o test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.sc_reads --config config.yml

2. Aggregate

The aggregate merges information across reads towards a position-level view of the data:

  • --i : path to the input file (the output of the fetch command from previous step).
  • --b : path to the CRAM/BAM file used to generate the input file.
  • --o : the path to the output file.
  • --r : path to reference genome.
  • --config : path to the config.yml file.
./indelible.py aggregate --i test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.sc_reads --b test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam --o DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts --r hs37d5.fasta --config config.yml

Note: It is recommended to retrain the RandomForest following this step with data that you have manually inspected.

3. Score

The score command scores positions based on the read information and sequence context:

  • --i : path to the input file (the output of the aggregate command from previous step).
  • --o : the path to the output file.
  • --config : path to the config.yml file.
./indelible.py score --i test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts --o test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts.scored --config config.yml

4. Database

The database command generates the allele frequency and breakpoint database required for the next step – Annotate. If analysing a small amount of data, it is highly recommended to use the '-p' flag with this command to include breakpoints from the initial DDD study described in our manuscript.

  • --f : file of files to merge to generate split read "allele frequencies"
  • --o : output file to generate
  • --r : path to reference genome.
  • --config : path to the config.yml file.
  • --priors: path to prior MAF database.
  • --tb: threads for multi-threaded bwa mem alignment [default: 1]
  • --old-maf: Use MAF from priors file rather than from sample(s)
ls InDelible_files/*.scored > fofn.txt
./indelible.py database --f fofn.txt --o InDelible_db.tsv --r hs37db.fa --config config.yml

Note The MAF database provided to --prior must conform to the following format:

1   1234    0.0001  1   10000   10.5    1_1256  REALN   DEL 21  15  true    true   1:1234-1256
1   1256    0.0001  1   10000   45.2    1_1234  REALN   DEL 21  15  true    false   1:1234-1256    

Where columns are:

  1. Chromosome
  2. Position
  3. Breakpoint frequency
  4. Breakpoint count (i.e. number of individuals with this breakpoint)
  5. Total individuals assessed
  6. Mean coverage overlapping this locus
  7. Alignment coordinate – where the split read from the score file aligned to in 'chr_position' (can be NA)
  8. Alignment mode – Did we align with bwa (REALN), blast (BLAST_REPEAT), or fail (FAIL_*)?
  9. SV Type (either DEL/DUP/INS/TRANS_SEGDUP/UNK)
  10. SV Size (can be NA)
  11. Alignment Length (can be NA)
  12. Otherside found elsewhere in MAF database?
  13. Is this the left-most breakpoint when '10.' == true? (can be NA)
  14. VCF-like coordinate for this variant (can be NA)

We have provided a priors file within the data.zip file included with InDelible. We recommend using this file when analysing a small number of genomes.

5. Annotate

The annotate command enriches the result with gene/exon annotations and merges the database results with the position file:

  • --i : path to the input file (output of score command after running the score command).
  • --o : path to output the annotated file.
  • --d : path to the InDelible frequency database generated by the database step.
  • --config : path to the config.yml file.
./indelible.py annotate --i test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts.scored --o test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts.scored.annotated --d InDelible_db.tsv --config config.yml

6. Denovo

One can then look for de novo mutation events using the denovo command:

Note: If maternal and/or paternal bam files are not supplied, denovo filtering will not be performed. This behaviour is intended to format non-trio data identically to trio data. If one of maternal or paternal bam is provided, InDelible will count coverage within that sample.

  • --c : path to scored/annotated calls in the proband.
  • --m : path to maternal CRAM/BAM file. [optional]
  • --p : path to paternal CRAM/BAM file. [optional]
  • --o : path to output file.
  • --config : path to the config.yml file.
./indelible.py denovo --c test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.counts.scored.annotated --m maternal.bam --p paternal.bam --o test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam.indelible.denovo.tsv --config config.yml

Additional Commands

InDelible also includes two helper commands:

InDelible Complete

All steps in the InDelible calling pipeline can be performed in succession automatically with the complete command:

  • --i : path to the input CRAM/BAM file.
  • --o : path to directory where output files will be generated.
  • --r : path to reference genome
  • --m : path to maternal CRAM/BAM file. [optional]
  • --p : path to paternal CRAM/BAM file. [optional]
  • --keeptmp : If this flag is given, intermediate files are kept. Otherwise these files will be removed once the analysis is finished.
  • --config : path to the config.yml file.
  • --priors: path to prior MAF database.
  • --old-maf: Use MAF from priors file rather than from sample(s)
  • --tb: threads for multi-threaded bwa mem alignment [default: 1]
./indelible.py complete --i test_data/DDD_MAIN5194229_Xchrom_subset_sorted.bam --o test_data/ --r data/hs37d5.fa --keeptmp --config example_config.hg19.yml --p data/Indelible_db_10k.bed

For InDelible de novo discovery behaviour when not providing maternal or paternal bams, please see de novo.

Train

The train command will run the active learning RandomForest on training data with annotations provided by manual inspection of data.

  • --i input training data
  • --o output random forest
  • --k number of samples to use for initial training and subsequent active selection [50].
  • --s convergence parameter to stop learning at when accuracy does not improve by X% [0.01]
  • --config : path to the config.yml file.

The input data has identical columns to the output generated by Aggregate, except with an additional column appended to the end of the file named "annotation". Column names need to be identical!

See an example training file which can be used to regenerate the RandomForest used on the DDD data in ./data/. Note: This training data is on hg19 aligned data!

./indelible.py train --i input_training_data.txt --o output.pkl

Note: Remember to change the path to the RandomForest from --o in the ./config.yml file!

Output

Primary TSV File

The primary output from InDelible is the output file from the denovo command (e.g. <BAM_NAME>.bam.indelible.denovo.tsv). This file has 41 columns, of which most are only relevant to how InDelible handles internal filtering with the adaptive learning model. What each column contains is listed in the table below:

Column Name Column # Description
chrom 1 Chromosome of breakpoint
position 2 Position of breakpoint
coverage 3 total number of reads covering breakpoint
insertion_context 4 total number of insertions (cigar "I") in reads overlapping this breakpoint
deletion_context 5 total number of deletions (cigar "D") in reads overlapping this breakpoint
sr_total 6 total number of split reads (cigar "S") in reads overlapping this breakpoint
sr_total_long 7 Number of reads with SR length ≥ MINIMUM_LENGTH_SPLIT_READ
sr_total_short 8 Number of reads with SR length < MINIMUM_LENGTH_SPLIT_READ
sr_long_5 9 sr_total_long for 5' end of reads
sr_short_5 10 sr_total_short for 5' end of reads
sr_long_3 11 sr_total_long for 3' end of reads
sr_short_3 12 sr_total_short for 3' end of reads
sr_entropy 13 Sequence entropy of the longest SR sequence given by the formula from Schmitt and Herzel (1997)
context_entropy 14 Sequence entropy of the ±20bp from the breakpoint position
entropy_upstream 15 Sequence entropy of the +20bp from the breakpoint position
entropy_downstream 16 Sequence entropy of the -20bp from the breakpoint position
sr_sw_similarity 17 Smith-Waterman based similarity of split reads from the breakpoint
avg_avg_sr_qual 18 Average sequence quality of split bases
avg_mapq 19 Average mapping quality of reads supporting the breakpoint
seq_longest 20 longest split sequence
pct_double_split 21 Number of reads with both 5' and 3' split reads
prob_N 22 Probability of the breakpoint being a false positive based on the adaptive learning model (1 - prob_Y)
prob_Y 23 Probability of the breakpoint being a true positive based on the adaptive learning model
predicted 24 Is prob_Y > prob_N?
ddg2p 25 Does this breakpoint intersect any genes given by ddg2p_bed file in config.yml
hgnc 26 Does this breakpoint intersect any genes given by hgnc_file in config.yml
hgnc_constrained 27 Does this breakpoint intersect any genes given by hgnc_constrained in config.yml
exonic 28 Does this breakpoint intersect any exons given by ensembl_exons in config.yml
transcripts 29 What transcripts does this breakpoint intersect? If > 10 transcripts, will return 'multiple_transcripts'
maf 30 "Allele Frequency" based on the InDelible database provided with --d
mode 31 How did bwa alignment perform? One of: BLAST_REPEAT (Aligned to a repeat/ME sequence), REALN (Aligned to unique sequence), REALN_CHR (Aligned to unique sequence on another chromosome), REALN_XL (Aligned to unique sequence on the same chromosome, but was flagged as an "improper pair by bwa"), FAIL_ALIGNMENT (split sequence did not align at all), FAIL_LOWMAPQ (split sequence aligned with MAPQ = 0), FAIL_MULTISPLIT (InDelible could not decide whether the sequence was in the 5' or 3' direction), FAIL_REFERENCE (anchoring reference read aligned in the wrong place).
otherside 32 Putative coordinate for alternate breakpoint
svtype 33 Putative SV class. Possible values are DUP (duplication), DEL (deletion), INS (followed by either the assembled sequence OR the type of repeat insertion [i.e. Alu, L1, etc.]), CMPLX (followed by DEL/DUP and assembled additional insertion sequence), or TRANSSEGDUP (segmental duplication or translocation). SEGDUP_TRANS represents either a segmenetal duplication or translocation. As we cannot discern with short read data between a SEGDUP or translocation, we list both here.
size 34 Distance to otherside from 'position'
variant_coord 35 Simply 'chrom' : 'position' - 'otherside', where possible
otherside_found 36 Was InDelible sucessful in identifying the other breakpoint?
is_primary 37 If otherside found or blast_hit = "repeats_hit" potential SV type
aln_length 38 Length of the aligned sequence
mum_sr 39 Number of SRs in the bam/cram provided to --m with the same 'position'
dad_sr 40 Number of SRs in the bam/cram provided to --d with the same 'position'
mum_indel_context 41 Number of reads in the bam/cram provided to --m with cigar 'I/D' values
dad_indel_context 42 Number of reads in the bam/cram provided to --d with cigar 'I/D' values
mum_cov 43 Coverage in in the bam/cram provided to --m
dad_cov 44 Coverage in in the bam/cram provided to --d

Recommended Filtering

As we have described in the InDelible manuscript, we follow a strict filtering regimen to further refine our TSV output. Filters are as follows, and are meant to "funnel" variants down (i.e. each successive filter is applied following the previous filter). Filter names are as above to avoid confusion. Parentheses are applied as pseudocode to demonstrate order of operations.

  1. maf ≤ 0.0004 AND avg_mapq ≥ 20
    • Note: The maf filter is dependent on using the maf data provided in one of the *data.zip files
  2. pct_double_split < 0.5 && (pct_double_split > 0.1 AND svtype != "UNK" AND svtype != "TRANSSEGDUP") OR pct_double_split ≤ 0.1
  3. ddg2p != "NA" AND sr_total ≥ 5 AND exonic = "True"
  4. mom_sr < 2 AND dad_sr < 2

indelible's People

Contributors

eugenegardner avatar asifrim avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.