Make plus/minus bigwig filename extension a bit more lax.
[htsworkflow.git] / extra / ucsc_encode_submission / ucsc_gather.py
index e230e5c186d66cd06e74b74dcc2a7ca945bc31ba..7650a9468e96df9486ec98e3e0721d90a46ff3de 100755 (executable)
@@ -1,26 +1,30 @@
 #!/usr/bin/env python
 from ConfigParser import SafeConfigParser
+import fnmatch
 from glob import glob
 import json
 import logging
+import netrc
 from optparse import OptionParser
 import os
 from pprint import pprint, pformat
 import shlex
 from StringIO import StringIO
-import time
+import stat
+from subprocess import Popen, PIPE
 import sys
+import time
 import types
 import urllib
 import urllib2
 import urlparse
 
-
 from htsworkflow.util import api
 from htsworkflow.pipelines.sequences import \
     create_sequence_table, \
     scan_for_sequences
-
+from htsworkflow.pipelines import qseq2fastq
+from htsworkflow.pipelines import srf2fastq
 
 def main(cmdline=None):
     parser = make_parser()
@@ -31,14 +35,16 @@ def main(cmdline=None):
     elif opts.verbose:
         logging.basicConfig(level = logging.INFO )
     else:
-        logging.basicConfig(level = logging.WARNING )
-        
+        logging.basicConfig(level = logging.WARNING )        
     
     apidata = {'apiid': opts.apiid, 'apikey': opts.apikey }
 
     if opts.host is None or opts.apiid is None or opts.apikey is None:
         parser.error("Please specify host url, apiid, apikey")
 
+    if opts.makeddf and opts.daf is None:
+        parser.error("Please specify your daf when making ddf files")
+
     if len(args) == 0:
         parser.error("I need at least one library submission-dir input file")
         
@@ -46,6 +52,9 @@ def main(cmdline=None):
     for a in args:
         library_result_map.extend(read_library_result_map(a))
 
+    if opts.make_tree_from is not None:
+        make_tree_from(opts.make_tree_from, library_result_map)
+            
     if opts.daf is not None:
         link_daf(opts.daf, library_result_map)
 
@@ -54,13 +63,14 @@ def main(cmdline=None):
                      apidata, 
                      opts.sequence, 
                      library_result_map,
-                     not opts.single)
+                     force=opts.force)
 
     if opts.ini:
-        make_submission_ini(opts.host, apidata, library_result_map, not opts.single)
+        make_submission_ini(opts.host, apidata, library_result_map)
 
     if opts.makeddf:
-        make_all_ddfs(library_result_map, opts.daf)
+        make_all_ddfs(library_result_map, opts.daf, force=opts.force)
+
 
 def make_parser():
     # Load defaults from the config files
@@ -82,6 +92,9 @@ def make_parser():
     parser = OptionParser()
 
     # commands
+    parser.add_option('--make-tree-from',
+                      help="create directories & link data files",
+                      default=None)
     parser.add_option('--fastq', help="generate scripts for making fastq files",
                       default=False, action="store_true")
 
@@ -92,6 +105,8 @@ def make_parser():
                       action="store_true")
     
     parser.add_option('--daf', default=None, help='specify daf name')
+    parser.add_option('--force', default=False, action="store_true",
+                      help="Force regenerating fastqs")
 
     # configuration options
     parser.add_option('--apiid', default=apiid, help="Specify API ID")
@@ -100,8 +115,6 @@ def make_parser():
                       help="specify HTSWorkflow host",)
     parser.add_option('--sequence', default=sequence_archive,
                       help="sequence repository")
-    parser.add_option('--single', default=False, action="store_true", 
-                      help="treat the sequences as single ended runs")
 
     # debugging
     parser.add_option('--verbose', default=False, action="store_true",
@@ -111,8 +124,29 @@ def make_parser():
 
     return parser
 
+
+def make_tree_from(source_path, library_result_map):
+    """Create a tree using data files from source path.
+    """
+    for lib_id, lib_path in library_result_map:
+        if not os.path.exists(lib_path):
+            logging.info("Making dir {0}".format(lib_path))
+            os.mkdir(lib_path)
+        source_lib_dir = os.path.join(source_path, lib_path)
+        if os.path.exists(source_lib_dir):
+            pass
+        for filename in os.listdir(source_lib_dir):
+            source_pathname = os.path.join(source_lib_dir, filename)
+            target_pathname = os.path.join(lib_path, filename)
+            if not os.path.exists(source_pathname):
+                raise IOError("{0} does not exist".format(source_pathname))
+            if not os.path.exists(target_pathname):
+                os.symlink(source_pathname, target_pathname)
+                logging.info(
+                    'LINK {0} to {1}'.format(source_pathname, target_pathname))
+    
 def build_fastqs(host, apidata, sequences_path, library_result_map, 
-                 paired=True ):
+                 force=False ):
     """
     Generate condor scripts to build any needed fastq files
     
@@ -121,73 +155,33 @@ def build_fastqs(host, apidata, sequences_path, library_result_map,
       apidata (dict): id & key to post to the server
       sequences_path (str): root of the directory tree to scan for files
       library_result_map (list):  [(library_id, destination directory), ...]
-      paired: should we assume that we are processing paired end records?
-              if False, we will treat this as single ended.
     """
     qseq_condor_header = """
 Universe=vanilla
-executable=/woldlab/rattus/lvol0/mus/home/diane/proj/solexa/gaworkflow/scripts/qseq2fastq
+executable=%(exe)s
 error=log/qseq2fastq.err.$(process).log
 output=log/qseq2fastq.out.$(process).log
 log=log/qseq2fastq.log
 
-"""
+""" % {'exe': sys.executable }
     qseq_condor_entries = []
     srf_condor_header = """
 Universe=vanilla
-executable=/woldlab/rattus/lvol0/mus/home/diane/proj/solexa/gaworkflow/scripts/srf2named_fastq.py
+executable=%(exe)s
 output=log/srf_pair_fastq.out.$(process).log
 error=log/srf_pair_fastq.err.$(process).log
 log=log/srf_pair_fastq.log
 environment="PYTHONPATH=/home/diane/lib/python2.6/site-packages:/home/diane/proj/solexa/gaworkflow PATH=/woldlab/rattus/lvol0/mus/home/diane/bin:/usr/bin:/bin"
 
-"""
+""" % {'exe': sys.executable }
     srf_condor_entries = []
-    fastq_paired_template = '%(lib_id)s_%(flowcell)s_c%(cycle)s_l%(lane)s_r%(read)s.fastq'
-    fastq_single_template = '%(lib_id)s_%(flowcell)s_c%(cycle)s_l%(lane)s.fastq'
     lib_db = find_archive_sequence_files(host, 
                                          apidata, 
                                          sequences_path, 
                                          library_result_map)
 
-    # find what targets we're missing
-    needed_targets = {}
-    for lib_id, result_dir in library_result_map:
-        lib = lib_db[lib_id]
-        for lane_key, sequences in lib['lanes'].items():
-            for seq in sequences:
-                if paired and seq.read is None:
-                    seq.read = 1
-                filename_attributes = { 
-                    'flowcell': seq.flowcell,
-                    'lib_id': lib_id,
-                    'lane': seq.lane,
-                    'read': seq.read,
-                    'cycle': seq.cycle
-                    }
-                # throw out test runs
-                # FIXME: this should probably be configurable
-                if seq.cycle < 50:
-                    continue
-                if seq.flowcell == '30CUUAAXX':
-                    # 30CUUAAXX run sucked
-                    continue
-                if seq.flowcell == '30DY0AAXX':
-                    # 30DY0 only ran for 151 bases instead of 152
-                    # it is actually 76 1st read, 75 2nd read
-                    seq.mid_point = 76
-
-                # end filters
-                if paired:
-                    target_name = fastq_paired_template % filename_attributes
-                else:
-                    target_name = fastq_single_template % filename_attributes
+    needed_targets = find_missing_targets(library_result_map, lib_db, force)
 
-                target_pathname = os.path.join(result_dir, target_name)
-                if not os.path.exists(target_pathname):
-                    t = needed_targets.setdefault(target_pathname, {})
-                    t[seq.filetype] = seq
-                    
     for target_pathname, available_sources in needed_targets.items():
         logging.debug(' target : %s' % (target_pathname,))
         logging.debug(' candidate sources: %s' % (available_sources,))
@@ -196,7 +190,8 @@ environment="PYTHONPATH=/home/diane/lib/python2.6/site-packages:/home/diane/proj
             qseq_condor_entries.append(
                 condor_qseq_to_fastq(source.path, 
                                      target_pathname, 
-                                     source.flowcell)
+                                     source.flowcell,
+                                     force=force)
             )
         elif available_sources.has_key('srf'):
             source = available_sources['srf']
@@ -204,9 +199,10 @@ environment="PYTHONPATH=/home/diane/lib/python2.6/site-packages:/home/diane/proj
             srf_condor_entries.append(
                 condor_srf_to_fastq(source.path, 
                                     target_pathname,
-                                    paired,
+                                    source.paired,
                                     source.flowcell,
-                                    mid)
+                                    mid,
+                                    force=force)
             )
         else:
             print " need file", target_pathname
@@ -221,6 +217,59 @@ environment="PYTHONPATH=/home/diane/lib/python2.6/site-packages:/home/diane/proj
                            qseq_condor_header,
                            qseq_condor_entries)
 
+
+def find_missing_targets(library_result_map, lib_db, force=False):
+    """
+    Check if the sequence file exists.
+    This requires computing what the sequence name is and checking
+    to see if it can be found in the sequence location.
+
+    Adds seq.paired flag to sequences listed in lib_db[*]['lanes']
+    """
+    fastq_paired_template = '%(lib_id)s_%(flowcell)s_c%(cycle)s_l%(lane)s_r%(read)s.fastq'
+    fastq_single_template = '%(lib_id)s_%(flowcell)s_c%(cycle)s_l%(lane)s.fastq'
+    # find what targets we're missing
+    needed_targets = {}
+    for lib_id, result_dir in library_result_map:
+        lib = lib_db[lib_id]
+        lane_dict = make_lane_dict(lib_db, lib_id)
+        
+        for lane_key, sequences in lib['lanes'].items():
+            for seq in sequences:
+                seq.paired = lane_dict[seq.flowcell]['paired_end']
+                lane_status = lane_dict[seq.flowcell]['status']
+
+                if seq.paired and seq.read is None:
+                    seq.read = 1
+                filename_attributes = { 
+                    'flowcell': seq.flowcell,
+                    'lib_id': lib_id,
+                    'lane': seq.lane,
+                    'read': seq.read,
+                    'cycle': seq.cycle
+                    }
+                # skip bad runs
+                if lane_status == 'Failed':
+                    continue
+                if seq.flowcell == '30DY0AAXX':
+                    # 30DY0 only ran for 151 bases instead of 152
+                    # it is actually 76 1st read, 75 2nd read
+                    seq.mid_point = 76
+
+                # end filters
+                if seq.paired:
+                    target_name = fastq_paired_template % filename_attributes
+                else:
+                    target_name = fastq_single_template % filename_attributes
+
+                target_pathname = os.path.join(result_dir, target_name)
+                if force or not os.path.exists(target_pathname):
+                    t = needed_targets.setdefault(target_pathname, {})
+                    t[seq.filetype] = seq
+
+    return needed_targets
+
+
 def link_daf(daf_path, library_result_map):
     if not os.path.exists(daf_path):
         raise RuntimeError("%s does not exist, how can I link to it?" % (daf_path,))
@@ -228,111 +277,65 @@ def link_daf(daf_path, library_result_map):
     base_daf = os.path.basename(daf_path)
     
     for lib_id, result_dir in library_result_map:
+        if not os.path.exists(result_dir):
+            raise RuntimeError("Couldn't find target directory %s" %(result_dir,))
         submission_daf = os.path.join(result_dir, base_daf)
         if not os.path.exists(submission_daf):
+            if not os.path.exists(daf_path):
+                raise RuntimeError("Couldn't find daf: %s" %(daf_path,))
             os.link(daf_path, submission_daf)
 
-def make_submission_ini(host, apidata, library_result_map, paired=True):
-    # ma is "map algorithm"
-    ma = 'TH1014'
 
-    if paired:
-        aligns = "Paired"
-    else:
-        aligns = "Aligns"
-
-    attributes = {
-      # bam index
-     '.bai':                   {'view': None, 'MapAlgorithm': 'NA'},
-     '.bam':                   {'view': aligns, 'MapAlgorithm': ma},
-     '.splices.bam':           {'view': 'Splices', 'MapAlgorithm': ma},
-     '.jnct':                  {'view': 'Junctions', 'MapAlgorithm': ma},
-     '.plus.bigwig':           {'view': 'PlusSignal', 'MapAlgorithm': ma},
-     '.minus.bigwig':          {'view': 'MinusSignal', 'MapAlgorithm': ma},
-     '.bigwig':                {'view': 'Signal', 'MapAlgorithm': ma},
-     '.tar.bz2':               {'view': None},
-     '.condor':                {'view': None},
-     '.daf':                   {'view': None},
-     '.ddf':                   {'view': None},
-     'novel.genes.expr':       {'view': 'GeneDeNovo', 'MapAlgorithm': ma},
-     'novel.transcripts.expr': {'view': 'TranscriptDeNovo', 'MapAlgorithm': ma},
-     '.genes.expr':            {'view': 'GeneFPKM', 'MapAlgorithm': ma},
-     '.transcripts.expr':      {'view': 'TranscriptFPKM', 'MapAlgorithm': ma},
-     '.fastq':                 {'view': 'Fastq', 'MapAlgorithm': 'NA' },
-     '_r1.fastq':              {'view': 'FastqRd1', 'MapAlgorithm': 'NA'},
-     '_r2.fastq':              {'view': 'FastqRd2', 'MapAlgorithm': 'NA'},
-     '.gtf':                   {'view': 'GeneModel', 'MapAlgorithm': ma},
-     '.ini':                   {'view': None},
-     '.log':                   {'view': None},
-     '.stats.txt':             {'view': 'InsLength', 'MapAlgorithm': ma},
-     '.srf':                   {'view': None},
-     '.wig':                   {'view': None},
-     '.zip':                   {'view': None},
-    }
-   
+def make_submission_ini(host, apidata, library_result_map, paired=True):
+    #attributes = get_filename_attribute_map(paired)
+    view_map = NameToViewMap(host, apidata)
+    
     candidate_fastq_src = {}
 
     for lib_id, result_dir in library_result_map:
         order_by = ['order_by=files', 'view', 'replicate', 'cell', 
-                    'readType', 'mapAlgorithm', 'insertLength' ]
+                    'readType', 'mapAlgorithm', 'insertLength', 'md5sum' ]
         inifile =  ['[config]']
         inifile += [" ".join(order_by)]
         inifile += ['']
         line_counter = 1
-        lib_info = get_library_info(host, apidata, lib_id)
         result_ini = os.path.join(result_dir, result_dir+'.ini')
 
-        if lib_info['cell_line'].lower() == 'unknown':
-            logging.warn("Library %s missing cell_line" % (lib_id,))
-
-        standard_attributes = {'cell': lib_info['cell_line'],
-                               'replicate': lib_info['replicate'],
-                               }
-        if paired:
-            if lib_info['insert_size'] is None:
-                errmsg = "Library %s is missing insert_size, assuming 200"
-                logging.warn(errmsg % (lib_id,))
-                insert_size = 200
-            else:
-                insert_size = lib_info['insert_size']
-            standard_attributes['insertLength'] = insert_size
-            standard_attributes['readType'] = '2x75'
-        else:
-            standard_attributes['insertLength'] = 'ilNA'
-            standard_attributes['readType'] = '1x75D'
-
         # write other lines
         submission_files = os.listdir(result_dir)
         fastqs = {}
+        fastq_attributes = {}
         for f in submission_files:
-            best_ext = find_best_extension(attributes, f)
-
-            if best_ext is not None:
-               if attributes[best_ext]['view'] is None:
-                   
-                   continue
-               elif best_ext.endswith('fastq'):
-                   fastqs.setdefault(best_ext, set()).add(f)
-               else:
-                   inifile.extend(
-                       make_submission_section(line_counter,
-                                               [f],
-                                               standard_attributes,
-                                               attributes[best_ext]
-                                               )
-                       )
-                   inifile += ['']
-                   line_counter += 1
-            else:
+            attributes = view_map.find_attributes(f, lib_id)
+            if attributes is None:
                 raise ValueError("Unrecognized file: %s" % (f,))
-
-        # add in fastqs on a single line.
-        for extension, fastq_set in fastqs.items():
+            attributes['md5sum'] = "None"
+            
+            ext = attributes["extension"]
+            if attributes['view'] is None:                   
+                continue               
+            elif attributes.get("type", None) == 'fastq':
+                fastqs.setdefault(ext, set()).add(f)
+                fastq_attributes[ext] = attributes
+            else:
+                md5sum = make_md5sum(os.path.join(result_dir,f))
+                if md5sum is not None:
+                    attributes['md5sum']=md5sum
+                inifile.extend(
+                    make_submission_section(line_counter,
+                                            [f],
+                                            attributes
+                                            )
+                    )
+                inifile += ['']
+                line_counter += 1
+                # add in fastqs on a single line.
+
+        for extension, fastq_files in fastqs.items():
             inifile.extend(
                 make_submission_section(line_counter, 
-                                        fastq_set,
-                                        standard_attributes,
-                                        attributes[extension])
+                                        fastq_files,
+                                        fastq_attributes[extension])
             )
             inifile += ['']
             line_counter += 1
@@ -340,8 +343,19 @@ def make_submission_ini(host, apidata, library_result_map, paired=True):
         f = open(result_ini,'w')
         f.write(os.linesep.join(inifile))
 
+        
+def make_lane_dict(lib_db, lib_id):
+    """
+    Convert the lane_set in a lib_db to a dictionary
+    indexed by flowcell ID
+    """
+    result = []
+    for lane in lib_db[lib_id]['lane_set']:
+        result.append((lane['flowcell'], lane))
+    return dict(result)
+
 
-def make_all_ddfs(library_result_map, daf_name, make_condor=True):
+def make_all_ddfs(library_result_map, daf_name, make_condor=True, force=False):
     dag_fragment = []
     for lib_id, result_dir in library_result_map:
         ininame = result_dir+'.ini'
@@ -353,7 +367,7 @@ def make_all_ddfs(library_result_map, daf_name, make_condor=True):
 
     if make_condor and len(dag_fragment) > 0:
         dag_filename = 'submission.dagman'
-        if os.path.exists(dag_filename):
+        if not force and os.path.exists(dag_filename):
             logging.warn("%s exists, please delete" % (dag_filename,))
         else:
             f = open(dag_filename,'w')
@@ -378,6 +392,9 @@ def make_ddf(ininame,  daf_name, guess_ddf=False, make_condor=False, outdir=None
         output = open(ddf_name,'w')
 
     file_list = read_ddf_ini(ininame, output)
+    logging.info(
+        "Read config {0}, found files: {1}".format(
+            ininame, ", ".join(file_list)))
 
     file_list.append(daf_name)
     if ddf_name is not None:
@@ -395,6 +412,7 @@ def make_ddf(ininame,  daf_name, guess_ddf=False, make_condor=False, outdir=None
     
     return dag_fragments
 
+
 def read_ddf_ini(filename, output=sys.stdout):
     """
     Read a ini file and dump out a tab delmited text file
@@ -423,7 +441,8 @@ def read_ddf_ini(filename, output=sys.stdout):
         output.write("\t".join(values))
         output.write(os.linesep)
     return file_list
-    
+
+
 def read_library_result_map(filename):
     """
     Read a file that maps library id to result directory.
@@ -436,21 +455,25 @@ def read_library_result_map(filename):
 
     results = []
     for line in stream:
-        if not line.startswith('#'):
-            library_id, result_dir = line.strip().split()
+        line = line.rstrip()
+        if not line.startswith('#') and len(line) > 0 :
+            library_id, result_dir = line.split()
             results.append((library_id, result_dir))
     return results
-            
+
+
 def make_condor_archive_script(ininame, files):
     script = """Universe = vanilla
 
 Executable = /bin/tar
-arguments = czvf ../%(archivename)s %(filelist)s
+arguments = czvhf ../%(archivename)s %(filelist)s
 
 Error = compress.err.$(Process).log
 Output = compress.out.$(Process).log
-Log = /tmp/submission-compress.log
+Log = /tmp/submission-compress-%(user)s.log
 initialdir = %(initialdir)s
+environment="GZIP=-3"
+request_memory = 20
 
 queue 
 """
@@ -460,7 +483,8 @@ queue
 
     context = {'archivename': make_submission_name(ininame),
                'filelist': " ".join(files),
-               'initialdir': os.getcwd()}
+               'initialdir': os.getcwd(),
+               'user': os.getlogin()}
 
     condor_script = make_condor_name(ininame, 'archive')
     condor_stream = open(condor_script,'w')
@@ -468,28 +492,41 @@ queue
     condor_stream.close()
     return condor_script
 
+
 def make_condor_upload_script(ininame):
     script = """Universe = vanilla
 
 Executable = /usr/bin/lftp
-arguments = -c put ../%(archivename)s -o ftp://detrout@encodeftp.cse.ucsc.edu/
+arguments = -c put ../%(archivename)s -o ftp://%(ftpuser)s:%(ftppassword)s@%(ftphost)s/%(archivename)s
 
 Error = upload.err.$(Process).log
 Output = upload.out.$(Process).log
-Log = /tmp/submission-upload.log
+Log = /tmp/submission-upload-%(user)s.log
 initialdir = %(initialdir)s
 
 queue 
 """
+    auth = netrc.netrc(os.path.expanduser("~diane/.netrc"))
+    
+    encodeftp = 'encodeftp.cse.ucsc.edu'
+    ftpuser = auth.hosts[encodeftp][0]
+    ftppassword = auth.hosts[encodeftp][2]
     context = {'archivename': make_submission_name(ininame),
-               'initialdir': os.getcwd()}
+               'initialdir': os.getcwd(),
+               'user': os.getlogin(),
+               'ftpuser': ftpuser,
+               'ftppassword': ftppassword,
+               'ftphost': encodeftp}
 
     condor_script = make_condor_name(ininame, 'upload')
     condor_stream = open(condor_script,'w')
     condor_stream.write(script % context)
     condor_stream.close()
+    os.chmod(condor_script, stat.S_IREAD|stat.S_IWRITE)
+
     return condor_script
 
+
 def make_dag_fragment(ininame, archive_condor, upload_condor):
     """
     Make the couple of fragments compress and then upload the data.
@@ -506,14 +543,17 @@ def make_dag_fragment(ininame, archive_condor, upload_condor):
 
     return fragments
 
+
 def get_library_info(host, apidata, library_id):
     url = api.library_url(host, library_id)
     contents = api.retrieve_info(url, apidata)
     return contents
 
+
 def condor_srf_to_fastq(srf_file, target_pathname, paired, flowcell=None,
-                        mid=None):
-    args = ['-c', srf_file, ]
+                        mid=None, force=False):
+    py = srf2fastq.__file__
+    args = [ py, srf_file, ]
     if paired:
         args.extend(['--left', target_pathname])
         # this is ugly. I did it because I was pregenerating the target
@@ -534,6 +574,9 @@ def condor_srf_to_fastq(srf_file, target_pathname, paired, flowcell=None,
     if mid is not None:
         args.extend(['-m', str(mid)])
 
+    if force:
+        args.extend(['--force'])
+
     script = """
 arguments="%s"
 queue
@@ -541,8 +584,10 @@ queue
     
     return  script 
 
-def condor_qseq_to_fastq(qseq_file, target_pathname, flowcell=None):
-    args = ['-i', qseq_file, '-o', target_pathname ]
+
+def condor_qseq_to_fastq(qseq_file, target_pathname, flowcell=None, force=False):
+    py = qseq2fastq.__file__
+    args = [py, '-i', qseq_file, '-o', target_pathname ]
     if flowcell is not None:
         args.extend(['-f', flowcell])
     script = """
@@ -566,6 +611,7 @@ def find_archive_sequence_files(host, apidata, sequences_path,
     candidate_lanes = {}
     for lib_id, result_dir in library_result_map:
         lib_info = get_library_info(host, apidata, lib_id)
+        lib_info['lanes'] = {}
         lib_db[lib_id] = lib_info
 
         for lane in lib_info['lane_set']:
@@ -586,72 +632,228 @@ def find_archive_sequence_files(host, apidata, sequences_path,
         lib_id = candidate_lanes.get(lane_key, None)
         if lib_id is not None:
             lib_info = lib_db[lib_id]
-            lanes = lib_info.setdefault('lanes', {})
-            lanes.setdefault(lane_key, set()).add(seq)
+            lib_info['lanes'].setdefault(lane_key, set()).add(seq)
     
     return lib_db
 
-def find_best_extension(extension_map, filename):
-    """
-    Search through extension_map looking for the best extension
-    The 'best' is the longest match
 
-    :Args:
-      extension_map (dict): '.ext' -> { 'view': 'name' or None }
-      filename (str): the filename whose extention we are about to examine
+class NameToViewMap(object):
+    """Determine view attributes for a given submission file name
     """
-    best_ext = None
-    path, last_ext = os.path.splitext(filename)
-
-    for ext in extension_map.keys():
-        if filename.endswith(ext):
-            if best_ext is None:
-                best_ext = ext
-            elif len(ext) > len(best_ext):
-                best_ext = ext
-    return best_ext
-    
-def make_submission_section(line_counter, files, standard_attributes, file_attributes):
+    def __init__(self, root_url, apidata):
+        self.root_url = root_url
+        self.apidata = apidata
+        
+        self.lib_cache = {}
+        self.lib_paired = {}
+        # ma is "map algorithm"
+        ma = 'TH1014'
+
+        self.patterns = [
+            ('*.bai',                   None),
+            ('*.splices.bam',           'Splices'),
+            ('*.bam',                   self._guess_bam_view),
+            ('junctions.bed',           'Junctions'),
+            ('*.jnct',                  'Junctions'),
+            ('*unique.bigwig',         None),
+            ('*plus.bigwig',           'PlusSignal'),
+            ('*minus.bigwig',          'MinusSignal'),
+            ('*.bigwig',                'Signal'),
+            ('*.tar.bz2',               None),
+            ('*.condor',                None),
+            ('*.daf',                   None),
+            ('*.ddf',                   None),
+            ('*.?ufflinks-0.9.0?genes.expr',       'GeneDeNovo'),
+            ('*.?ufflinks-0.9.0?transcripts.expr', 'TranscriptDeNovo'),
+            ('*.?ufflinks-0.9.0?transcripts.gtf',  'GeneModel'),
+            ('*.GENCODE-v3c?genes.expr',       'GeneGCV3c'),
+            ('*.GENCODE-v3c?transcript*.expr', 'TranscriptGCV3c'),
+            ('*.GENCODE-v3c?transcript*.gtf',  'TranscriptGencV3c'),
+            ('*.GENCODE-v4?genes.expr',        None), #'GeneGCV4'),
+            ('*.GENCODE-v4?transcript*.expr',  None), #'TranscriptGCV4'),
+            ('*.GENCODE-v4?transcript*.gtf',   None), #'TranscriptGencV4'),
+            ('*_1.75mers.fastq',              'FastqRd1'),
+            ('*_2.75mers.fastq',              'FastqRd2'),
+            ('*_r1.fastq',              'FastqRd1'),
+            ('*_r2.fastq',              'FastqRd2'),
+            ('*.fastq',                 'Fastq'),
+            ('*.gtf',                   'GeneModel'),
+            ('*.ini',                   None),
+            ('*.log',                   None),
+            ('*.md5',                   None),
+            ('paired-end-distribution*', 'InsLength'),
+            ('*.stats.txt',              'InsLength'),
+            ('*.srf',                   None),
+            ('*.wig',                   None),
+            ('*.zip',                   None),
+            ]
+
+        self.views = {
+            None: {"MapAlgorithm": "NA"},
+            "Paired": {"MapAlgorithm": ma},
+            "Aligns": {"MapAlgorithm": ma},
+            "Single": {"MapAlgorithm": ma},
+            "Splices": {"MapAlgorithm": ma},
+            "Junctions": {"MapAlgorithm": ma},
+            "PlusSignal": {"MapAlgorithm": ma},
+            "MinusSignal": {"MapAlgorithm": ma},
+            "Signal": {"MapAlgorithm": ma},
+            "GeneModel": {"MapAlgorithm": ma},
+            "GeneDeNovo": {"MapAlgorithm": ma},
+            "TranscriptDeNovo": {"MapAlgorithm": ma},
+            "GeneGCV3c": {"MapAlgorithm": ma},
+            "TranscriptGCV3c": {"MapAlgorithm": ma},
+            "TranscriptGencV3c": {"MapAlgorithm": ma},
+            "GeneGCV4": {"MapAlgorithm": ma},
+            "TranscriptGCV4": {"MapAlgorithm": ma},
+            "FastqRd1": {"MapAlgorithm": "NA", "type": "fastq"},
+            "FastqRd2": {"MapAlgorithm": "NA", "type": "fastq"},
+            "Fastq": {"MapAlgorithm": "NA", "type": "fastq" },
+            "InsLength": {"MapAlgorithm": ma},
+            }
+        # view name is one of the attributes
+        for v in self.views.keys():
+            self.views[v]['view'] = v
+            
+    def find_attributes(self, pathname, lib_id):
+        """Looking for the best extension
+        The 'best' is the longest match
+        
+        :Args:
+        filename (str): the filename whose extention we are about to examine
+        """
+        path, filename = os.path.splitext(pathname)
+        if not self.lib_cache.has_key(lib_id):
+            self.lib_cache[lib_id] = get_library_info(self.root_url,
+                                                      self.apidata, lib_id)
+
+        lib_info = self.lib_cache[lib_id]
+        if lib_info['cell_line'].lower() == 'unknown':
+            logging.warn("Library %s missing cell_line" % (lib_id,))
+        attributes = {
+            'cell': lib_info['cell_line'],
+            'replicate': lib_info['replicate'],
+            }
+        is_paired = self._is_paired(lib_id, lib_info)
+        
+        if is_paired:
+            attributes.update(self.get_paired_attributes(lib_info))
+        else:
+            attributes.update(self.get_single_attributes(lib_info))
+            
+        for pattern, view in self.patterns:
+            if fnmatch.fnmatch(pathname, pattern):
+                if callable(view):
+                    view = view(is_paired=is_paired)
+                    
+                attributes.update(self.views[view])
+                attributes["extension"] = pattern
+                return attributes
+
+
+    def _guess_bam_view(self, is_paired=True):
+        """Guess a view name based on library attributes
+        """
+        if is_paired:
+            return "Paired"
+        else:
+            return "Aligns"
+
+
+    def _is_paired(self, lib_id, lib_info):
+        """Determine if a library is paired end"""
+        if len(lib_info["lane_set"]) == 0:
+            return False
+
+        if not self.lib_paired.has_key(lib_id):
+            is_paired = 0
+            isnot_paired = 0
+            failed = 0
+            # check to see if all the flowcells are the same.
+            # otherwise we might need to do something complicated
+            for flowcell in lib_info["lane_set"]:
+                # yes there's also a status code, but this comparison 
+                # is easier to read
+                if flowcell["status"].lower() == "failed":
+                    # ignore failed flowcell
+                    failed += 1
+                    pass
+                elif flowcell["paired_end"]:
+                    is_paired += 1
+                else:
+                    isnot_paired += 1
+                    
+            logging.debug("Library %s: %d paired, %d single, %d failed" % \
+                     (lib_info["library_id"], is_paired, isnot_paired, failed))
+
+            if is_paired > isnot_paired:
+                self.lib_paired[lib_id] = True
+            elif is_paired < isnot_paired:
+                self.lib_paired[lib_id] = False
+            else:
+                raise RuntimeError("Equal number of paired & unpaired lanes."\
+                                   "Can't guess library paired status")
+            
+        return self.lib_paired[lib_id]
+
+    def get_paired_attributes(self, lib_info):
+        if lib_info['insert_size'] is None:
+            errmsg = "Library %s is missing insert_size, assuming 200"
+            logging.warn(errmsg % (lib_info["library_id"],))
+            insert_size = 200
+        else:
+            insert_size = lib_info['insert_size']
+        return {'insertLength': insert_size,
+                'readType': '2x75'}
+
+    def get_single_attributes(self, lib_info):
+        return {'insertLength':'ilNA',
+                'readType': '1x75D'
+                }
+
+def make_submission_section(line_counter, files, attributes):
     """
     Create a section in the submission ini file
     """
-    inifile = [ '[line%s]' % (line_counter,) ]
+    inifile = [ "[line%s]" % (line_counter,) ]
     inifile += ["files=%s" % (",".join(files))]
-    cur_attributes = {}
-    cur_attributes.update(standard_attributes)
-    cur_attributes.update(file_attributes)
-    
-    for k,v in cur_attributes.items():
+
+    for k,v in attributes.items():
         inifile += ["%s=%s" % (k,v)]
     return inifile
 
+
 def make_base_name(pathname):
     base = os.path.basename(pathname)
     name, ext = os.path.splitext(base)
     return name
 
+
 def make_submission_name(ininame):
     name = make_base_name(ininame)
-    return name + '.tgz'
+    return name + ".tgz"
+
 
 def make_ddf_name(pathname):
     name = make_base_name(pathname)
-    return name + '.ddf'
+    return name + ".ddf"
+
 
 def make_condor_name(pathname, run_type=None):
     name = make_base_name(pathname)
     elements = [name]
     if run_type is not None:
         elements.append(run_type)
-    elements.append('condor')
+    elements.append("condor")
     return ".".join(elements)
-    
+
+
 def make_submit_script(target, header, body_list):
     """
     write out a text file
 
     this was intended for condor submit scripts
-    
+
     Args:
       target (str or stream): 
         if target is a string, we will open and close the file
@@ -663,7 +865,7 @@ def make_submit_script(target, header, body_list):
         a list of blocks to add to the file.
     """
     if type(target) in types.StringTypes:
-        f = open(target,'w')
+        f = open(target,"w")
     else:
         f = target
     f.write(header)
@@ -673,7 +875,8 @@ def make_submit_script(target, header, body_list):
         f.close()
 
 def parse_filelist(file_string):
-    return file_string.split(',')
+    return file_string.split(",")
+
 
 def validate_filelist(files):
     """
@@ -682,6 +885,47 @@ def validate_filelist(files):
     for f in files:
         if not os.path.exists(f):
             raise RuntimeError("%s does not exist" % (f,))
-        
+
+def make_md5sum(filename):
+    """Quickly find the md5sum of a file
+    """
+    md5_cache = os.path.join(filename+".md5")
+    print md5_cache
+    if os.path.exists(md5_cache):
+        logging.debug("Found md5sum in {0}".format(md5_cache))
+        stream = open(md5_cache,'r')
+        lines = stream.readlines()
+        md5sum = parse_md5sum_line(lines, filename)
+    else:
+        md5sum = make_md5sum_unix(filename, md5_cache)
+    return md5sum
+    
+def make_md5sum_unix(filename, md5_cache):
+    cmd = ["md5sum", filename]
+    logging.debug("Running {0}".format(" ".join(cmd)))
+    p = Popen(cmd, stdout=PIPE)
+    stdin, stdout = p.communicate()
+    retcode = p.wait()
+    logging.debug("Finished {0} retcode {1}".format(" ".join(cmd), retcode))
+    if retcode != 0:
+        logging.error("Trouble with md5sum for {0}".format(filename))
+        return None
+    lines = stdin.split(os.linesep)
+    md5sum = parse_md5sum_line(lines, filename)
+    if md5sum is not None:
+        logging.debug("Caching sum in {0}".format(md5_cache))
+        stream = open(md5_cache, "w")
+        stream.write(stdin)
+        stream.close()
+    return md5sum
+
+def parse_md5sum_line(lines, filename):
+    md5sum, md5sum_filename = lines[0].split()
+    if md5sum_filename != filename:
+        errmsg = "MD5sum and I disagre about filename. {0} != {1}"
+        logging.error(errmsg.format(filename, md5sum_filename))
+        return None
+    return md5sum
+
 if __name__ == "__main__":
     main()