b48da38d4aa8dcb9dc75cc1a37b528a5b42b3265
max
  Thu Jul 18 03:41:35 2024 -0700
adding user suggested dash fix, refs #34121

diff --git src/utils/expMatrixToBarchartBed/expMatrixToBarchartBed src/utils/expMatrixToBarchartBed/expMatrixToBarchartBed
index 3f0f45e..9b756bf 100755
--- src/utils/expMatrixToBarchartBed/expMatrixToBarchartBed
+++ src/utils/expMatrixToBarchartBed/expMatrixToBarchartBed
@@ -1,364 +1,364 @@
 #!/usr/bin/env python2.7
 # expMatrixToBarchartBed
 """
 Generate a barChart bed6+5 file from a matrix, meta data, and coordinates.
 """
 import os
 import sys
 import argparse
 import tempfile
 
 def parseArgs(args):
     """
     Parse the command line arguments.
     """
     parser= argparse.ArgumentParser(description = __doc__)
     parser.add_argument ("sampleFile",
         help = " Two column no header, the first column is the samples which should match " + \
                 "the matrix, the second is the grouping (cell type, tissue, etc)",
         type = argparse.FileType("r"))
     parser.add_argument ("matrixFile",
             help = " The input matrix file. The samples in the first row should exactly match the ones in " + \
                 "the sampleFile. The labels (ex ENST*****) in the first column should exactly match " + \
                 "the ones in the bed file.",
         type = argparse.FileType("r"))
     parser.add_argument ("bedFile",
         help = " Bed6+1 format. File that maps the column labels from the matrix to coordinates. Tab " + \
                 "separated; chr, start coord, end coord, label, score, strand, gene name. The score " + \
                 "column is ignored.",
         action = "store")
     parser.add_argument ("outputFile",
         help = " The output file, bed 6+5 format. See the schema in kent/src/hg/lib/barChartBed.as. ",
         type =argparse.FileType("w"))
     # Optional arguments.
     parser.add_argument ("--autoSql",
         help = " Optional autoSql description of extra fields in the input bed.", action = "store",
         default=None)
     parser.add_argument ("--groupOrderFile",
         help = " Optional file to define the group order, list the groups in a single column in " + \
                 "the order desired. The default ordering is alphabetical.",
         action = "store")
     parser.add_argument ("--useMean",
     help = " Calculate the group values using mean rather than median.",
     action = "store_true")
     parser.add_argument ("--verbose",
     help = " Show runtime messages.",
     action = "store_true")
 
     parser.set_defaults(verbose = False)
     parser.set_defaults(groupOrderFile = None)
 
     if (len(sys.argv) == 1):
         parser.print_help()
         exit(1)
     options = parser.parse_args()
     return options
 
 def parseExtraFields(autoSqlFile):
     """
     Obtain a list of field names to allow extra fields in the final bed.
     Enforce the bed6+1 format since extra fields must come after the
     expCount, expScores, _dataOffset and _dataLen fields this script will add.
     """
     fields = []
     requiredFieldList = ["chrom", "chromStart", "chromEnd", "name", "score", "strand", "name2",
                         "expCount", "expScores", "_dataOffset", "_dataLen"]
     with open(autoSqlFile, "r") as f:
         for inputLine in f.readlines():
             line = inputLine.strip().split()
             if line[0].startswith("table") or line[0].startswith("\"") or line[0].startswith("("):
                 continue
             elif line[0].startswith(")"):
                 return fields
             else:
                 clean = line[1].rstrip(";")
                 if clean in requiredFieldList:
                     if len(fields) == requiredFieldList.index(clean):
                         fields.append(clean)
                     else:
                         raise Exception("autoSql fields out of order, must be in %s, extraFields order\n" % \
                             ", ".join(requiredFieldList))
                 else:
                     if len(fields) >= 11:
                         fields.append(clean)
                     else:
                         raise Exception("first 11 barChart fields must be %s\n" % ", ".join(requiredFieldList))
 
 def median(lst):
     lst = sorted(lst)
     if len(lst) < 1:
         return None
     if len(lst) %2 == 1:
         return lst[((len(lst)+1)/2)-1]
     else:
         return float(sum(lst[(len(lst)/2)-1:(len(lst)/2)+1]))/2.0
 
 def determineScore(tpmCutoffs, tpm):
     """
     Cast the tpm to a score between 0-1000. Since there are only
     9 visual blocks cast them to be in one of the 9 blocks.
     tpmCutoffs - A list of integers
     tpm - An integer
     """
     count = 0
     for val in tpmCutoffs:
         if (val > tpm):
             return count*111
         count = count + 1
     return 999
 
 def condenseMatrixIntoBedCols(matrix, groupOrder, autoSql, sampleToGroup, validTpms, bedLikeFile, useMean):
     """
     Take an expression matrix and a dictionary that maps the samples to groups.
     Go through the expression matrix and calculate the average for each group, outputting
     it to an intermediate file as they are calculated. The intermediate file has
     three columns, the first is the average tpm for the entire gene, next is the
     number of groups and finally the average tpm for each group as a comma separated list.
 
     matrix - An expression matrix, samples are the x rows, transcripts the y rows.
     groupOrder - Optional order of categories
     autoSql - Optional description of extra fields
     sampleToGroup - A dictionary that maps string samples to string groups.
     validTpms - An empty list of integers.
     bedLikeFile - An intermediate file, looks slightly like a bed.
     """
     # Store some information on the bed file, most important is the order
     # of the 8th column.
     bedInfo = ""
     firstLine = True
     getBedInfo = True
     # Use the first line of the matrix and the sampleToGroup dict to create a dictionary that maps
     # the column to a group.
     columnToGroup = dict()
 
     # Go through the matrix line by line. The first line is used to build an index mapping columns
     # to group blocks, then for each line with TPM values merge the values based on group blocks.
     for line in matrix:
         splitLine = line.strip("\n").split("\t")
 
         # The first line is the word 'transcript' followed by a list of the sample names.
         if firstLine:
             firstLine = False
             count = 1
             firstCol = True
             for col in splitLine:
                 if firstCol:
                     firstCol = False
                     continue
                 group = sampleToGroup[col]
                 columnToGroup.setdefault(count, group)
                 count += 1
             continue
 
         # Handle the tpm rows, calculating the average for each group per row.
         groupAverages = dict()
         groupCounts = dict()
         firstCol = True
         count = 1
         for col in splitLine:
             if firstCol:
                 firstCol = False
                 continue
             if (useMean):
                 # First time this group is seen, add it to the groupCounts dict.
                 if (groupAverages.get(columnToGroup[count]) == None):
                     groupAverages.setdefault(columnToGroup[count], float(col))
                     groupCounts.setdefault(columnToGroup[count], 1)
                 # This group has already been seen, update the TPM average.
                 else:
                     groupCounts[columnToGroup[count]] += 1
                     # Average calculation
                     normal = float(groupCounts[columnToGroup[count]])
                     newTpm = (float(col) * (1/normal))
                     oldTpm = (((normal - 1) / normal) * groupAverages[columnToGroup[count]])
                     groupAverages[columnToGroup[count]] = newTpm + oldTpm
             else:
                 # First time this group is seen, add it to the groupCounts dict.
                 if (groupAverages.get(columnToGroup[count]) == None):
                     groupAverages.setdefault(columnToGroup[count], [float(col)])
                     groupCounts.setdefault(columnToGroup[count], 1)
                 # This group has already been seen, update the TPM average.
                 else:
                     groupCounts[columnToGroup[count]] += 1
                     # Median preparation
                     groupAverages[columnToGroup[count]].append(float(col))
 
             count += 1
 
         # Store some information on the bed file. Most important is the groupOrder.
         if getBedInfo:
             getBedInfo = False
             groups = ""
             bedInfo += "#chr\tchromStart\tchromEnd\tname\tscore\tstrand\tname2\texpCount\texpScores;"
             if (groupOrder is not None):
                 for group in open(groupOrder, "r"):
                     groups += group.strip("\n") + " "
             else:
                 for key, value in sorted(groupAverages.iteritems()):
                     groups += key + " "
             if autoSql and len(autoSql) != 11: # parseExtraFields requires first 11 fields to be standard
                 bedInfo += groups[:-1] + "\t_offset\t_lineLength\t" + "\t".join(autoSql[11:])
             else:
                 bedInfo += groups[:-1] + "\t_offset\t_lineLength"
 
         # Write out the transcript name, this is needed to join with coordinates later.
         bedLikeFile.write(splitLine[0] + "\t")
         # Create a list of the average scores per group.
         bedLine = ""
         # The fullAverage is used to assign a tpm score representative of the entire bed row.
         fullAverage = 0.0
         count = 0.0
         if (groupOrder is not None):
             for group in open(groupOrder, "r"):
                 # Averages
                 if (useMean):
                     value = groupAverages[group.strip("\n")]
                 else:
                     value = median(groupAverages[group.strip("\n")])
                 bedLine = bedLine + "," + "%0.2g" % value
                 count += 1.0
                 fullAverage += value
         else:
             for key, value in sorted(groupAverages.iteritems()):
                 if (useMean):
                     bedLine = bedLine + "," + "%0.2g" % value
                     fullAverage += value
                 else:
                     bedLine = bedLine + "," + "%0.2g" % median(value)
                     fullAverage += median(value)
 
                 count += 1.0
         # Create what will be columns 5, 7 and 8 of the final bed.
         bedLine = str(fullAverage/count) + "\t" + str(int(count)) +  "\t" + bedLine[1:] + "\n"
         # If the fullAverage tpm is greater than 0 then consider it in the validTpm list.
         if (fullAverage > 0.0):
             validTpms.append((fullAverage/count))
         # Write the bedLine to the intermediate bed-like file.
         bedLikeFile.write(bedLine)
     # Return the bedInfo so it can be printed right before the script ends.
     return bedInfo
 
 def expMatrixToBarchartBed(options):
     """
     Convert the expression matrix into a barchart bed file.
     options - The command line options (file names, etc).
         Use the meta data to map the sample names to their groups, then create a dict that maps
     the columns to the groups.  Go through the matrix line by line and get the median or average for each
     group. Print this to an intermediate file, then use the unix 'join' command to link with
     the coordinates file via the first matrix column. This creates a file with many of the bed fields
     just in the wrong order.  Go through this file to re arrange the columns, check for and remove entries
     where chromsomes names include "_" and chr start > chr end.  Finally run Max's bedJoinTabOffset to
     index the matrix adding the dataOffset and dataLen columns and creating a bed 6+5 file.
     """
 
     # Create a dictionary that maps the sample names to their group.
     sampleToGroup = dict()
     count = 0
     for item in options.sampleFile:
         count +=1
         splitLine = item.strip("\n").split("\t")
         if (len(splitLine) is not 2):
             print ("There was an error reading the sample file at line " + str(count))
             exit(1)
         sampleToGroup.setdefault(splitLine[0], splitLine[1])
 
     autoSql = parseExtraFields(options.autoSql) if (options.autoSql) else None
 
     # Use an intermediate file to hold the average values for each group, format: name,mean/median,expCount,expScores
     bedLikeFile = tempfile.NamedTemporaryFile( mode = "w+", bufsize = 1)
     # Keep a list of TPM scores greater than 0. This will be used later
     # to assign bed scores.
     validTpms = []
 
     # Go through the matrix and condense it into a bed like file. Populate
     # the validTpms array and the bedInfo string.
     bedInfo = condenseMatrixIntoBedCols(options.matrixFile, options.groupOrderFile, autoSql, sampleToGroup, \
                     validTpms, bedLikeFile, options.useMean)
     # Find the number which divides the list of non 0 TPM scores into ten blocks.
     tpmMedian = sorted(validTpms)
     blockSizes = len(tpmMedian)/10
 
     # Create a list of the ten TPM values at the edge of each block.
     # These used to cast a TPM score to one of ten value between 0-1000.
     tpmCutoffs = []
     for i in range(1,10):
         tpmCutoffs.append(tpmMedian[blockSizes*i])
 
     # Sort the bed like file to prepare it for the join.
     sortedBedLikeFile = tempfile.NamedTemporaryFile( mode = "w+", bufsize = 1)
     cmd = "sort -k1 " + bedLikeFile.name + " > " + sortedBedLikeFile.name
     os.system(cmd)
 
     # Sort the coordinate file to prepare it for the join.
     sortedCoords = tempfile.NamedTemporaryFile( mode = "w+", bufsize = 1)
     cmd = "sort -k4 " + options.bedFile + " > " + sortedCoords.name
     os.system(cmd)
 
     # Join the bed-like file and the coordinate file, the awk accounts for any extra
     # fields that may be included, and keeps the file in standard bed 6+5 format
     joinedFile = tempfile.NamedTemporaryFile(mode="w+", bufsize=1)
     cmd = "join -t '	' -1 4 -2 1 " + sortedCoords.name + " " + sortedBedLikeFile.name + \
             " | awk -F'\\t' -v OFS=\"\\t\" '{printf \"%s\\t%s\\t%s\\t%s\", $2,$3,$4,$1; " + \
             "for (i=5;i<=NF;i++) {printf \"\\t%s\", $i}; printf\"\\n\";}' > " + joinedFile.name
     os.system(cmd)
 
     # Go through the joined file and re arrange the columns creating a bed 6+5+ file.
     # Also assign a scaled score 0 - 1000 to each tpm value.
     bedFile = tempfile.NamedTemporaryFile(mode="w+", bufsize=1)
     for line in joinedFile:
         splitLine = line.strip("\n").split("\t")
         # Drop sequences where start is greater than end.
         if (float(splitLine[1]) > float(splitLine[2])):
             sys.stderr.write("This transcript: " + splitLine[0] + " was dropped since chr end, " + \
                     splitLine[2] + ", is smaller than chr start, " + splitLine[1] + ".\n")
             continue
         score = str(determineScore(tpmCutoffs, float(splitLine[-3])))
         if autoSql:
             #skip the 4th field since we recalculated it
             #need a different ordering to account for possible extraFields
             bedLine = "\t".join(splitLine[:4] + [score] + splitLine[5:7] + splitLine[-2:] + splitLine[7:-3]) + "\n"
         else:
             #skip the 4th field since we recalculate it
             bedLine = "\t".join(splitLine[:4] + [score] + splitLine[5:7] + splitLine[8:]) + "\n"
 
         bedFile.write(bedLine)
 
     # Run Max's indexing script: TODO: add verbose options
     indexedBedFile = tempfile.NamedTemporaryFile(mode="w+", bufsize=1)
     cmd = "bedJoinTabOffset " + options.matrixFile.name + " " + bedFile.name + " " + indexedBedFile.name
-    if not options.verbose: cmd += " &>/dev/null"
+    if not options.verbose: cmd += " 2>&1 >/dev/null"
     os.system(cmd)
 
     # Prepend the bed info to the start of the file.
     cmd = "echo '" + bedInfo + "' > " + options.outputFile.name
     os.system(cmd)
 
     # any extra fields must come after the fields added by bedJoinTabOffset
     if autoSql:
         reorderedBedFile = tempfile.NamedTemporaryFile(mode="w+", bufsize=1)
         # first print the standard bed6+3 barChart fields
         # then print the two fields added by bedJoinTabOffset
         # then any extra fields at the end:
         cmd = "awk -F'\\t' -v \"OFS=\\t\" '" + \
                 "{for (i = 1; i < 10; i++) {if (i > 1) printf \"\\t\"; printf \"%s\", $i;}; " + \
                 "for (i = NF-1; i <= NF; i++) {printf \"\\t%s\", $i;} " + \
                 "for (i = 10; i < NF - 1; i++) {printf \"\\t%s\", $i;} " + \
                 "printf \"\\n\";}' " + indexedBedFile.name + " > " + reorderedBedFile.name
         os.system(cmd)
         cmd = "cat " + reorderedBedFile.name + " >> " + options.outputFile.name
         os.system(cmd)
     else:
         cmd = "cat " + indexedBedFile.name + " >> " + options.outputFile.name
         os.system(cmd)
 
     if options.verbose: print ("The columns and order of the groups are; \n" + bedInfo)
 
 def main(args):
     """
     Initialized options and calls other functions.
     """
     options = parseArgs(args)
     expMatrixToBarchartBed(options)
 
 if __name__ == "__main__":
     sys.exit(main(sys.argv))