diff data_manager/fetch_refseq.py @ 19:d118e256faca draft default tip

"planemo upload for repository https://github.com/galaxyproject/tools-iuc/tree/master/data_managers/data_manager_fetch_refseq commit 120c6491f4b0888220e432693a9805d8198d7397"
author sanbi-uwc
date Thu, 16 Apr 2020 10:19:57 +0000
parents 75c1817c2ecf
children
line wrap: on
line diff
--- a/data_manager/fetch_refseq.py	Fri Sep 28 23:46:24 2018 -0400
+++ /dev/null	Thu Jan 01 00:00:00 1970 +0000
@@ -1,191 +0,0 @@
-#!/usr/bin/env python
-
-from __future__ import division, print_function
-
-import argparse
-import functools
-import gzip
-import json
-import os
-import os.path
-import sys
-from datetime import date
-from multiprocessing import Process, Queue
-
-import requests
-
-try:
-    from io import StringIO
-except ImportError:
-    from StringIO import StringIO
-# Refseq structure
-# - Release number
-# - Divisions
-#   1. archea
-#   2. bacteria
-#   3. fungi
-#   4. invertebrate
-#   5. mitochondrion
-#   6. other
-#   7. plant
-#   8. plasmid
-#   9. plastid
-#  10. protozoa
-#  11. vertebrate mammalian
-#  12. vertebrate other
-#  13. viral
-# within each division
-# DIVNAME.\d+(.\d+)?.(genomic|protein|rna).(fna|gbff|faa|gpff).gz
-#  where fna and faa are FASTA, gbff and gpff are Genbank
-
-
-def _add_data_table_entry(data_manager_dict, data_table_entry, data_table_name):
-    data_manager_dict['data_tables'] = data_manager_dict.get('data_tables', {})
-    data_manager_dict['data_tables'][data_table_name] = data_manager_dict['data_tables'].get('all_fasta', [])
-    data_manager_dict['data_tables'][data_table_name].append(data_table_entry)
-    return data_manager_dict
-
-
-def unzip_to(conn, out_dir, output_filename, chunk_size=4096, debug=False, compress=False, make_len_file=False):
-    input_filename = conn.get()
-    if compress:
-        open_output = gzip.open
-    else:
-        open_output = open
-    if make_len_file:
-        fa_pos = output_filename.find('.fa')
-        if fa_pos == -1:
-            # this should not happen - filename does not contain '.fa'
-            len_filename = output_filename + '.len'
-        else:
-            len_filename = output_filename[:fa_pos] + '.len'
-        len_output = open(len_filename, 'wb')
-        record_len = 0
-        record_id = ''
-    with open_output(os.path.join(out_dir, output_filename), 'wb') as output_file:
-        while input_filename != 'STOP':
-            if debug:
-                print('Reading', input_filename, file=sys.stderr)
-            with gzip.open(input_filename, 'rb') as input_file:
-                read_chunk = functools.partial(input_file.read, (chunk_size))
-                for data in iter(read_chunk, b''):  # use b'' as a sentinel to stop the loop. note '' != b'' in Python 3
-                    if make_len_file:
-                        # break data into lines and parse as FASTA, perhaps continuing from partial previous record
-                        for line in data.split('\n'):
-                            if line.startswith('>'):
-                                if record_id != '':
-                                    len_output.write('{}\t{}\n'.format(record_id, record_len))
-                                # update record ID of record we are processing, set length to 0
-                                record_len = 0
-                                record_id = line[1:].split()[0]
-                            else:
-                                assert record_id != '', "FASTA data found before FASTA record ID known in {}, data: {}".format(input_filename, line)
-                                record_len += len(line.strip())
-                    output_file.write(data)
-                if make_len_file:
-                    # write last entry to .len file
-                    len_output.write('{}\t{}\n'.format(record_id, record_len))
-            os.unlink(input_filename)
-            input_filename = conn.get()
-    len_output.close()
-
-
-def get_refseq_division(division_name, mol_types, output_directory, debug=False, compress=False):
-    base_url = 'https://ftp.ncbi.nlm.nih.gov/refseq/release/'
-    valid_divisions = set(['archea', 'bacteria', 'complete', 'fungi', 'invertebrate', 'mitochondrion', 'other',
-                          'plant', 'plasmid', 'plastid', 'protozoa', 'vertebrate_mammalian', 'vertebrate_other', 'viral'])
-    ending_mappings = {
-        'genomic': '.genomic.fna.gz',
-        'protein': '.protein.faa.gz',
-        'rna': 'rna.fna.gz'
-    }
-    assert division_name in valid_divisions, "Unknown division name ({})".format(division_name)
-    for mol_type in mol_types:
-        assert mol_type in ending_mappings, "Unknown molecule type ({})".format(mol_type)
-    if not os.path.exists(output_directory):
-        os.mkdir(output_directory)
-    release_num_file = base_url + 'RELEASE_NUMBER'
-    r = requests.get(release_num_file)
-    release_num = str(int(r.text.strip()))
-    division_base_url = base_url + division_name
-    if debug:
-        print('Retrieving {}'.format(division_base_url), file=sys.stderr)
-    r = requests.get(division_base_url)
-    listing_text = r.text
-
-    unzip_queues = {}
-    unzip_processes = []
-    final_output_filenames = []
-    for mol_type in mol_types:
-        q = unzip_queues[mol_type] = Queue()
-        output_filename = division_name + '.' + release_num + '.' + mol_type + '.fasta'
-        if compress:
-            output_filename += '.gz'
-        final_output_filenames.append(output_filename)
-        unzip_processes.append(Process(target=unzip_to, args=(q, output_directory, output_filename),
-                                       kwargs=dict(debug=debug, compress=compress)))
-        unzip_processes[-1].start()
-
-    # sample line: <a href="vertebrate_other.86.genomic.gbff.gz">vertebrate_other.86.genomic.gbff.gz</a>   2018-07-13 00:59   10M
-    for line in StringIO(listing_text):
-        if '.gz' not in line:
-            continue
-        parts = line.split('"')
-        assert len(parts) == 3, "Unexpected line format: {}".format(line.rstrip())
-        filename = parts[1]
-        for mol_type in mol_types:
-            ending = ending_mappings[mol_type]
-            if filename.endswith(ending):
-                if debug:
-                    print('Downloading:', filename, ending, mol_type, file=sys.stderr)
-                output_filename = os.path.join(output_directory, filename)
-                with open(output_filename, 'wb') as output_file:
-                    r = requests.get(division_base_url + '/' + filename)
-                    for chunk in r.iter_content(chunk_size=4096):
-                        output_file.write(chunk)
-                conn = unzip_queues[mol_type]
-                conn.put(output_filename)
-
-    for mol_type in mol_types:
-        conn = unzip_queues[mol_type]
-        conn.put('STOP')
-
-    return [release_num, final_output_filenames]
-
-
-if __name__ == '__main__':
-    parser = argparse.ArgumentParser(description='Download RefSeq databases')
-    parser.add_argument('--debug', default=False, action='store_true', help='Print debugging output to stderr (verbose)')
-    parser.add_argument('--compress', default=False, action='store_true', help='Compress output files')
-    parser.add_argument('--output_directory', default='tmp', help='Directory to write output to')
-    parser.add_argument('--galaxy_datamanager_filename', help='Galaxy JSON format file describing data manager inputs')
-    parser.add_argument('--division_names', help='RefSeq divisions to download')
-    parser.add_argument('--mol_types', help='Molecule types (genomic, rna, protein) to fetch')
-    parser.add_argument('--pin_date', help='Force download date to this version string')
-    args = parser.parse_args()
-
-    division_names = args.division_names.split(',')
-    mol_types = args.mol_types.split(',')
-    if args.galaxy_datamanager_filename is not None:
-        dm_opts = json.loads(open(args.galaxy_datamanager_filename).read())
-        output_directory = dm_opts['output_data'][0]['extra_files_path']  # take the extra_files_path of the first output parameter
-        data_manager_dict = {}
-    else:
-        output_directory = args.output_directory
-    for division_name in division_names:
-        if args.pin_date is not None:
-            today_str = args.pin_date
-        else:
-            today_str = date.today().strftime('%Y-%m-%d')  # ISO 8601 date format
-        [release_num, fasta_files] = get_refseq_division(division_name, mol_types, output_directory, args.debug, args.compress)
-        if args.galaxy_datamanager_filename is not None:
-            for i, mol_type in enumerate(mol_types):
-                assert mol_type in fasta_files[i], "Filename does not contain expected mol_type ({}, {})".format(mol_type, fasta_files[i])
-                unique_key = 'refseq_' + division_name + '.' + release_num + '.' + mol_type  # note: this is now same as dbkey
-                dbkey = unique_key
-                desc = 'RefSeq ' + division_name + ' Release ' + release_num + ' ' + mol_type + ' (' + today_str + ')'
-                path = os.path.join(output_directory, fasta_files[i])
-                _add_data_table_entry(data_manager_dict=data_manager_dict,
-                                      data_table_entry=dict(value=unique_key, dbkey=dbkey, name=desc, path=path),
-                                      data_table_name='all_fasta')
-            open(args.galaxy_datamanager_filename, 'wb').write(json.dumps(data_manager_dict).encode())