Difference between revisions of "Mash"

From wiki
Jump to: navigation, search
Line 20: Line 20:
 
== De novo assembly ==
 
== De novo assembly ==
  
We use SPAdes with the --meta option here as we are dealing with metagenomes. First we need the python DRMAA script which will control the where and how the job script will be run:
+
We use SPAdes with the --meta option here as we are dealing with metagenomes. First we need the python DRMAA script which will control the where and how the job script will be run. The keys issues for it will be:
 +
* what job script to run (this need only be a bash script, BUT it must have executable permissions)
 +
* what is the name of the paired up FASTQ filelisting.
 +
* With how many threads/processes which EACH sample paired will run.
 +
* The queue name is not mentioned, so it will launch on all.q.
  
 
  #!/usr/bin/env python2.7
 
  #!/usr/bin/env python2.7
Line 52: Line 56:
 
     # this is an intensive IO job, don't want to whack the FS too much
 
     # this is an intensive IO job, don't want to whack the FS too much
 
     jt.joinFiles=True
 
     jt.joinFiles=True
#
 
#    print eflsz
 
 
     jobid = s.runBulkJobs(jt, 1, pld2, 1)
 
     jobid = s.runBulkJobs(jt, 1, pld2, 1)
 
     print 'Your job has been submitted with id ' + str(jobid)
 
     print 'Your job has been submitted with id ' + str(jobid)
Line 63: Line 65:
 
  if __name__=='__main__':
 
  if __name__=='__main__':
 
     main()
 
     main()
 +
 +
Note the lines:
 +
 +
pld2=len(PL)/2
 +
jt.args =PL
 +
 +
This takes the file listing and sends out the names in pairs as argument to the jobscript. The number of jobs is half the total number of lines, obviously, as a pair represents one sample.
 +
  
 
= Links =
 
= Links =
 
* [https://mash.readthedocs.io/en/latest main Mash documentation]
 
* [https://mash.readthedocs.io/en/latest main Mash documentation]

Revision as of 10:33, 8 March 2017

Introduction

MinHash is a general dimensionality-reduction technique and it is used by Mash to reduce large sequences and sequence sets to small, representative sketches with the result that global mutation distances (Mash distances) can be rapidly estimated.

Other aspects

  • terms itself as an alignment-free method

Usage

Typical analysis

Mash is run on genomes. These will usually be de-novo assembled genomes from tools such as Velvet or SPAdes.

Parallel Usage on gridengine

We'll go through a process here of running Mash on a set of samples, using the DRMAA library to launch Gridengine job arrays.

The scripts will take as argument a file listing of the sample names, and it is assumed there are two pair-ended FASTQ reads per sample. It is also assumed that the paired-ended samples appeared in ordered fashoin in the file-listing: i.e. each consecutive set of two lines represent one sample.

De novo assembly

We use SPAdes with the --meta option here as we are dealing with metagenomes. First we need the python DRMAA script which will control the where and how the job script will be run. The keys issues for it will be:

  • what job script to run (this need only be a bash script, BUT it must have executable permissions)
  • what is the name of the paired up FASTQ filelisting.
  • With how many threads/processes which EACH sample paired will run.
  • The queue name is not mentioned, so it will launch on all.q.
#!/usr/bin/env python2.7
import os, sys, drmaa

def main():
    """Submit an array job."""
    argquan=len(sys.argv)
    if argquan != 4:
        print "This script requires two arguments: 1) the script to run in ja mode  2) filelist of absolute paths and filenames 3) Number of threads/CPU for *each* job array"
        sys.exit(2)

    s = drmaa.Session()
    s.initialize()
    print 'Creating job template'
    jt = s.createJobTemplate()
    jt.workingDirectory=os.getcwd() # means sge job output will be deposited here.
    jt.remoteCommand = jt.workingDirectory + '/' +sys.argv[1]

    with open(sys.argv[2]) as x: fl = x.read().splitlines()
    eflsz=len(fl)
    PL=[]
    for i in xrange(eflsz):
        PL.append(fl[i])
    pld2=len(PL)/2
    jt.args =PL 

    # prepare natSpec
    nm='-N jadrm0'
    jt.nativeSpecification='-V -pe multi ' +sys.argv[3]+ ' '+nm
    # this is an intensive IO job, don't want to whack the FS too much
    jt.joinFiles=True
    jobid = s.runBulkJobs(jt, 1, pld2, 1)
    print 'Your job has been submitted with id ' + str(jobid)

    print 'Cleaning up'
    s.deleteJobTemplate(jt)
    s.exit()
       
if __name__=='__main__':
    main()

Note the lines:

pld2=len(PL)/2
jt.args =PL 

This takes the file listing and sends out the names in pairs as argument to the jobscript. The number of jobs is half the total number of lines, obviously, as a pair represents one sample.


Links