d822406ab76774a681c2509b4232dd3a31f459ea hiram Fri May 14 12:49:29 2021 -0700 completed documentation for cactus 241 way track refs #27519 diff --git src/hg/makeDb/doc/hg38/cactus241way.txt src/hg/makeDb/doc/hg38/cactus241way.txt index 95d3fcc..8c2344e 100644 --- src/hg/makeDb/doc/hg38/cactus241way.txt +++ src/hg/makeDb/doc/hg38/cactus241way.txt @@ -9,953 +9,30 @@ # download from https://drive.google.com/drive/u/0/folders/1Xc215oxu0cvBZOarxFv2L3HAjPmiZ-0A # to directory uppsala # format is, so convert with awk chr1 10074 10075 id-1 0.053000 chr1 10075 10076 id-2 0.064000 chr1 10076 10077 id-3 0.064000 # convert to wig and wib zcat uppsala/chr*.bed.gz | tawk '{print $1,$2,$3,$5}' | wigEncode stdin phyloP241way.wig phyloP241way.wib ln -s $(pwd)/phyloP241way.wib /gbdb/hg38/cactus241way/ hgLoadWiggle -pathPrefix=/gbdb/hg38/cactus241way hg38 phyloP241way phyloP241way.wig - -######################################################################## -# phastCons 241-way (TBD - 2015-05-07 - Hiram) - # split 241way mafs into 10M chunks and generate sufficient statistics - # files for # phastCons - ssh ku - mkdir -p /hive/data/genomes/hg38/bed/cactus241way/cons/ss - mkdir -p /hive/data/genomes/hg38/bed/cactus241way/cons/msa.split - cd /hive/data/genomes/hg38/bed/cactus241way/cons/msa.split - - cat << '_EOF_' > doSplit.csh -#!/bin/csh -ef -set c = $1 -set MAF = /hive/data/genomes/hg38/bed/cactus241way/ucscNames/$c.maf -set WINDOWS = /hive/data/genomes/hg38/bed/cactus241way/cons/ss/$c -set WC = `cat $MAF | wc -l` -set NL = `grep "^#" $MAF | wc -l` -if ( -s $2 ) then - exit 0 -endif -if ( -s $2.running ) then - exit 0 -endif - -date >> $2.running - -rm -fr $WINDOWS -mkdir $WINDOWS -pushd $WINDOWS > /dev/null -if ( $WC != $NL ) then -/cluster/bin/phast.build/cornellCVS/phast.2010-12-30/bin/msa_split \ - $MAF -i MAF -o SS -r $WINDOWS/$c -w 3000000,0 -I 300 -B 5000 -endif -popd > /dev/null -date >> $2 -rm -f $2.running -'_EOF_' - # << happy emacs - chmod +x doSplit.csh - - printf '#LOOP -doSplit.csh $(root1) {check out line+ $(root1).done} -#ENDLOOP -' > template - -# do the easy ones first to see some immediate results - ls -1S -r ../../ucscNames | sed -e "s/.maf//;" > maf.list - # all can finish OK at a 64Gb memory limit - gensub2 maf.list single template jobList - para -ram=64g create jobList - para try ... check ... etc - para push -# Completed: 54 of 54 jobs -# CPU time in finished jobs: 24295s 404.91m 6.75h 0.28d 0.001 y -# IO & Wait Time: 1111s 18.52m 0.31h 0.01d 0.000 y -# Average job time: 470s 7.84m 0.13h 0.01d -# Longest finished job: 2724s 45.40m 0.76h 0.03d -# Submission to last job: 2765s 46.08m 0.77h 0.03d - - # Run phastCons - # This job is I/O intensive in its output files, beware where this - # takes place or do not run too many at once. - ssh ku - mkdir -p /hive/data/genomes/hg38/bed/cactus241way/cons/run.cons - cd /hive/data/genomes/hg38/bed/cactus241way/cons/run.cons - - # This is setup for multiple runs based on subsets, but only running - # the 'all' subset here. - # It triggers off of the current working directory - # $cwd:t which is the "grp" in this script. Running: - # all and vertebrates - - cat << '_EOF_' > doPhast.csh -#!/bin/csh -fe -set PHASTBIN = /cluster/bin/phast.build/cornellCVS/phast.2010-12-30/bin -set c = $1 -set f = $2 -set len = $3 -set cov = $4 -set rho = $5 -set grp = $cwd:t -set cons = /hive/data/genomes/hg38/bed/cactus241way/cons -set tmp = $cons/tmp/$f -mkdir -p $tmp -set ssSrc = $cons/ss -set useGrp = "$grp.mod" -if (-s $cons/$grp/$grp.non-inf) then - ln -s $cons/$grp/$grp.mod $tmp - ln -s $cons/$grp/$grp.non-inf $tmp - ln -s $ssSrc/$c/$f.ss $tmp -else - ln -s $ssSrc/$c/$f.ss $tmp - ln -s $cons/$grp/$grp.mod $tmp -endif -pushd $tmp > /dev/null -if (-s $grp.non-inf) then - $PHASTBIN/phastCons $f.ss $useGrp \ - --rho $rho --expected-length $len --target-coverage $cov --quiet \ - --not-informative `cat $grp.non-inf` \ - --seqname $c --idpref $c --most-conserved $f.bed --score > $f.pp -else - $PHASTBIN/phastCons $f.ss $useGrp \ - --rho $rho --expected-length $len --target-coverage $cov --quiet \ - --seqname $c --idpref $c --most-conserved $f.bed --score > $f.pp -endif -popd > /dev/null -mkdir -p pp/$c bed/$c -sleep 4 -touch pp/$c bed/$c -rm -f pp/$c/$f.pp -rm -f bed/$c/$f.bed -mv $tmp/$f.pp pp/$c -mv $tmp/$f.bed bed/$c -rm -fr $tmp -'_EOF_' - # << happy emacs - chmod +x doPhast.csh - - # this template will serve for all runs - # root1 == chrom name, file1 == ss file name without .ss suffix - printf '#LOOP -../run.cons/doPhast.csh $(root1) $(file1) 45 0.3 0.3 {check out line+ pp/$(root1)/$(file1).pp} -#ENDLOOP -' > template - - ls -1S ../ss/chr*/chr* | sed -e "s/.ss$//" > ss.list - wc -l ss.list - # 930 ss.list - - # Create parasol batch and run it - # run for all species - cd /hive/data/genomes/hg38/bed/cactus241way/cons - mkdir -p all - cd all - # Using the .mod tree - cp -p ../../4d/all.mod ./all.mod - - gensub2 ../run.cons/ss.list single ../run.cons/template jobList - # beware overwhelming the cluster with these fast running high I/O jobs - para -ram=32g create jobList - para try ... check ... - para -maxJob=16 push -# Completed: 930 of 930 jobs -# CPU time in finished jobs: 22874s 381.23m 6.35h 0.26d 0.001 y -# IO & Wait Time: 6698s 111.64m 1.86h 0.08d 0.000 y -# Average job time: 32s 0.53m 0.01h 0.00d -# Longest finished job: 49s 0.82m 0.01h 0.00d -# Submission to last job: 200s 3.33m 0.06h 0.00d - - # create Most Conserved track - cd /hive/data/genomes/hg38/bed/cactus241way/cons/all - time cut -f1 ../../../../chrom.sizes | while read C -do - echo $C 1>&2 - ls -d bed/${C} 2> /dev/null | while read D - do - cat ${D}/${C}*.bed - done | sort -k1,1 -k2,2n \ - | awk '{printf "%s\t%d\t%d\tlod=%d\t%s\n", "'${C}'", $2, $3, $5, $5;}' -done > tmpMostConserved.bed - # real 0m43.805s - - # -rw-rw-r-- 1 164925477 Dec 22 14:02 tmpMostConserved.bed - - time /cluster/bin/scripts/lodToBedScore tmpMostConserved.bed \ - > mostConserved.bed - # real 0m26.890s - - # -rw-rw-r-- 1 169120947 Dec 22 14:04 mostConserved.bed - - # load into database - ssh hgwdev - cd /hive/data/genomes/hg38/bed/cactus241way/cons/all - time hgLoadBed hg38 phastConsElements241way mostConserved.bed - # Read 4829247 elements of size 5 from mostConserved.bed - # real 0m36.697s - - # Try for 5% overall cov, and 70% CDS cov - # --rho 0.3 --expected-length 45 --target-coverage 0.3 - time featureBits hg38 -enrichment ncbiRefSeq:cds phastConsElements241way -# ncbiRefSeq:cds 1.408%, phastConsElements241way 5.679%, both 0.993%, cover 70.51%, enrich 12.42x -# real 0m31.631s - - - time featureBits hg38 -enrichment refGene:cds phastConsElements241way -# refGene:cds 1.333%, phastConsElements241way 5.679%, both 0.972%, cover 72.87%, enrich 12.83x -# real 0m24.183s - - # Create merged posterier probability file and wiggle track data files - cd /hive/data/genomes/hg38/bed/cactus241way/cons/all - mkdir downloads - - time for D in `ls -d pp/chr* | sed -e 's#pp/##'` -do - echo "working: $D" 1>&2 - find ./pp/${D} -type f | sed -e "s#^./##; s#\.# d #g; s#-# m #;" \ - | sort -k1,1 -k3,3n | sed -e "s# d #.#g; s# m #-#g;" | xargs cat \ - | gzip -c > downloads/${D}.phastCons241way.wigFix.gz -done -XXX - running - Tue Dec 22 14:11:26 PST 2020 - # real 32m29.089s - - - # encode those files into wiggle data - time (zcat downloads/*.wigFix.gz \ - | wigEncode stdin phastCons241way.wig phastCons241way.wib) - # Converted stdin, upper limit 1.00, lower limit 0.00 - # real 15m40.010s - - du -hsc *.wi? - # 2.8G phastCons241way.wib - # 283M phastCons241way.wig - - # encode into a bigWig file: - # (warning wigToBigWig process may be too large for memory limits - # in bash, to avoid the 32 Gb memory limit, set 180 Gb here: -export sizeG=188743680 -ulimit -d $sizeG -ulimit -v $sizeG - time (zcat downloads/*.wigFix.gz \ - | wigToBigWig -verbose=2 stdin \ - ../../../../chrom.sizes phastCons241way.bw) > bigWig.log 2>&1 - egrep "VmPeak|real" bigWig.log - # pid=37111: VmPeak: 33886864 kB - # real 42m13.614s - - # -rw-rw-r-- 1 7077152013 Nov 6 08:52 phastCons241way.bw - - - bigWigInfo phastCons241way.bw -version: 4 -isCompressed: yes -isSwapped: 0 -primaryDataSize: 5,097,637,987 -primaryIndexSize: 93,372,648 -zoomLevels: 10 -chromCount: 355 -basesCovered: 2,955,660,600 -mean: 0.128025 -min: 0.000000 -max: 1.000000 -std: 0.247422 - - # if you wanted to use the bigWig file, loading bigWig table: - # but we don't use the bigWig file - mkdir /gbdb/hg38/bbi - ln -s `pwd`/phastCons241way.bw /gbdb/hg38/bbi - hgsql hg38 -e 'drop table if exists phastCons241way; \ - create table phastCons241way (fileName varchar(255) not null); \ - insert into phastCons241way values - ("/gbdb/hg38/bbi/phastCons241way.bw");' - - # Load gbdb and database with wiggle. - ssh hgwdev - cd /hive/data/genomes/hg38/bed/cactus241way/cons/all - ln -s `pwd`/phastCons241way.wib /gbdb/hg38/cactus241way/phastCons241way.wib - time hgLoadWiggle -pathPrefix=/gbdb/hg38/cactus241way hg38 \ - phastCons241way phastCons241way.wig - # real 0m32.272s - - time wigTableStats.sh hg38 phastCons241way -# db.table min max mean count sumData -# hg38.phastCons241way 0 1 0.128025 2955660600 3.78397e+08 -# stdDev viewLimits -# 0.247422 viewLimits=0:1 -# real 0m13.507s - - # Create histogram to get an overview of all the data - ssh hgwdev - cd /hive/data/genomes/hg38/bed/cactus241way/cons/all - time hgWiggle -doHistogram -db=hg38 \ - -hBinSize=0.001 -hBinCount=300 -hMinVal=0.0 -verbose=2 \ - phastCons241way > histogram.data 2>&1 - # real 2m38.952s - - # create plot of histogram: - - printf 'set terminal png small x000000 xffffff xc000ff x66ff66 xffff00 x00ffff font \ -"/usr/share/fonts/default/Type1/n022004l.pfb" -set size 1.4, 0.8 -set key left box -set grid noxtics -set grid ytics -set title " Human Mm39 Histogram phastCons241way track" -set xlabel " phastCons241way score" -set ylabel " Relative Frequency" -set y2label " Cumulative Relative Frequency (CRF)" -set y2range [0:1] -set y2tics -set yrange [0:0.02] - -plot "histogram.data" using 2:5 title " RelFreq" with impulses, \ - "histogram.data" using 2:7 axes x1y2 title " CRF" with lines -' | gnuplot > histo.png - - # take a look to see if it is sane: - - display histo.png & - -######################################################################### -# phyloP for 241-way (TBD - 2017-11-06 - Hiram) -# - # split SS files into 1M chunks, this business needs smaller files - # to complete - - ssh ku - mkdir /hive/data/genomes/hg38/bed/cactus241way/consPhyloP - cd /hive/data/genomes/hg38/bed/cactus241way/consPhyloP - mkdir ss run.split - cd run.split - - printf '#!/bin/csh -ef -set c = $1 -set MAF = /hive/data/genomes/hg38/bed/cactus241way/ucscNames/$c.maf -set WINDOWS = /hive/data/genomes/hg38/bed/cactus241way/consPhyloP/ss/$c -set WC = `cat $MAF | wc -l` -set NL = `grep "^#" $MAF | wc -l` -if ( -s $2 ) then - exit 0 -endif -if ( -s $2.running ) then - exit 0 -endif - -date >> $2.running - -rm -fr $WINDOWS -mkdir -p $WINDOWS -pushd $WINDOWS > /dev/null -if ( $WC != $NL ) then -/cluster/bin/phast.build/cornellCVS/phast.2010-12-30/bin/msa_split \ - $MAF -i MAF -o SS -r $WINDOWS/$c -w 1000000,0 -I 1000 -B 5000 -endif -popd > /dev/null -date >> $2 -rm -f $2.running -' > doSplit.csh - - chmod +x doSplit.csh - - # do the easy ones first to see some immediate results - ls -1S -r ../../ucscNames | sed -e "s/.maf//;" > maf.list - - # this needs a {check out line+ $(root1.done)} test for verification: - printf '#LOOP -./doSplit.csh $(root1) $(root1).done -#ENDLOOP -' > template - - gensub2 maf.list single template jobList - # all can complete successfully at the 64Gb memory limit - para -ram=64g create jobList - para try ... check ... push ... etc... -# Completed: 54 of 54 jobs -# CPU time in finished jobs: 25321s 422.01m 7.03h 0.29d 0.001 y -# IO & Wait Time: 843s 14.06m 0.23h 0.01d 0.000 y -# Average job time: 485s 8.08m 0.13h 0.01d -# Longest finished job: 2873s 47.88m 0.80h 0.03d -# Submission to last job: 2931s 48.85m 0.81h 0.03d - - # run phyloP with score=LRT - ssh ku - mkdir /cluster/data/hg38/bed/cactus241way/consPhyloP - cd /cluster/data/hg38/bed/cactus241way/consPhyloP - - mkdir run.phyloP - cd run.phyloP - # Adjust model file base composition background and rate matrix to be - # representative of the chromosomes in play - grep BACK ../../4d/all.mod - # BACKGROUND: 0.219062 0.338786 0.207231 0.234921 - - grep BACKGROUND ../../4d/all.mod | awk '{printf "%0.3f\n", $3 + $4}' - # 0.546 - /cluster/bin/phast.build/cornellCVS/phast.2010-12-30/bin/modFreqs \ - ../../4d/all.mod 0.546 > all.mod - # verify, the BACKGROUND should now be paired up: - grep BACK all.mod - # BACKGROUND: 0.217500 0.282500 0.282500 0.217500 - - printf '#!/bin/csh -fe -set PHASTBIN = /cluster/bin/phast.build/cornellCVS/phast.2010-12-30/bin -set f = $1 -set ssFile = $1:t -set out = $2 -set cName = $f:h -set n = $f:r:e -set grp = $cwd:t -set cons = /hive/data/genomes/hg38/bed/cactus241way/consPhyloP -set tmp = $cons/tmp/$grp/$f -/bin/rm -fr $tmp -/bin/mkdir -p $tmp -set ssSrc = "$cons/ss/$cName/$ssFile" -set useGrp = "$grp.mod" -/bin/ln -s $cons/run.phyloP/$grp.mod $tmp -pushd $tmp > /dev/null -echo source: $ssSrc.ss -$PHASTBIN/phyloP --method LRT --mode CONACC --wig-scores --chrom $cName \ - -i SS $useGrp $ssSrc.ss > $ssFile.wigFix -popd > /dev/null -/bin/mkdir -p $out:h -sleep 4 -/bin/touch $out:h -/bin/mv $tmp/$ssFile.wigFix $out -/bin/rm -fr $tmp -/bin/rmdir --ignore-fail-on-non-empty $cons/tmp/$grp -/bin/rmdir --ignore-fail-on-non-empty $cons/tmp -' > doPhyloP.csh - - chmod +x doPhyloP.csh - - # Create list of chunks - find ../ss -type f | sed -e "s/.ss$//; s#../ss/##;" > ss.list - # make sure the list looks good - wc -l ss.list - # 3308 ss.list - - # Create template file - # file1 == $chr/$chunk/file name without .ss suffix - printf '#LOOP -../run.phyloP/doPhyloP.csh $(path1) {check out line+ wigFix/$(dir1)/$(file1).wigFix} -#ENDLOOP -' > template - - ###################### Running all species ####################### - # setup run for all species - mkdir /hive/data/genomes/hg38/bed/cactus241way/consPhyloP/all - cd /hive/data/genomes/hg38/bed/cactus241way/consPhyloP/all - rm -fr wigFix - mkdir wigFix - - gensub2 ../run.phyloP/ss.list single ../run.phyloP/template jobList - # beware overloading the cluster with these quick and high I/O jobs - para -ram=32g create jobList - para try ... check ... - para -maxJob=16 push - para time > run.time -XXX - running - Tue Dec 22 13:59:08 PST 2020 - -# Completed: 3308 of 3308 jobs -# CPU time in finished jobs: 647954s 10799.23m 179.99h 7.50d 0.021 y -# IO & Wait Time: 22374s 372.90m 6.22h 0.26d 0.001 y -# Average job time: 203s 3.38m 0.06h 0.00d -# Longest finished job: 349s 5.82m 0.10h 0.00d -# Submission to last job: 3226s 53.77m 0.90h 0.04d - - mkdir downloads - time for D in `ls -d wigFix/chr* | sed -e 's#wigFix/##'` -do - echo "working: $D" 1>&2 - find ./wigFix/${D} -type f | sed -e "s#^./##; s#\.# d #g; s#-# m #;" \ - | sort -k1,1 -k3,3n | sed -e "s# d #.#g; s# m #-#g;" | xargs cat \ - | gzip -c > downloads/${D}.phyloP241way.wigFix.gz -done - # real 48m50.219s - - du -hsc downloads - # 4.6G downloads - - # check integrity of data with wigToBigWig - time (zcat downloads/*.wigFix.gz \ - | wigToBigWig -verbose=2 stdin /hive/data/genomes/hg38/chrom.sizes \ - phyloP241way.bw) > bigWig.log 2>&1 - - - egrep "real|VmPeak" bigWig.log - # pid=66292: VmPeak: 33751268 kB - # real 43m40.194s - - - bigWigInfo phyloP241way.bw | sed -e 's/^/# /;' -# version: 4 -# isCompressed: yes -# isSwapped: 0 -# primaryDataSize: 6,304,076,591 -# primaryIndexSize: 93,404,704 -# zoomLevels: 10 -# chromCount: 355 -# basesCovered: 2,955,660,581 -# mean: 0.097833 -# min: -20.000000 -# max: 1.312000 -# std: 0.727453 - - # encode those files into wiggle data - time (zcat downloads/*.wigFix.gz \ - | wigEncode stdin phyloP241way.wig phyloP241way.wib) - -# Converted stdin, upper limit 1.31, lower limit -20.00 -# real 17m36.880s -# -rw-rw-r-- 1 2955660581 Nov 6 14:10 phyloP241way.wib -# -rw-rw-r-- 1 304274846 Nov 6 14:10 phyloP241way.wig - - du -hsc *.wi? - # 2.8G phyloP241way.wib - # 291M phyloP241way.wig - - # Load gbdb and database with wiggle. - ln -s `pwd`/phyloP241way.wib /gbdb/hg38/cactus241way/phyloP241way.wib - time hgLoadWiggle -pathPrefix=/gbdb/hg38/cactus241way hg38 \ - phyloP241way phyloP241way.wig - # real 0m30.538s - - # use to set trackDb.ra entries for wiggle min and max - # and verify table is loaded correctly - - wigTableStats.sh hg38 phyloP241way -# db.table min max mean count sumData -# hg38.phyloP241way -20 1.312 0.0978331 2955660581 2.89162e+08 -# stdDev viewLimits -# 0.727453 viewLimits=-3.53943:1.312 - - # that range is: 20+1.312= 21.312 for hBinSize=0.021312 - - # Create histogram to get an overview of all the data - time hgWiggle -doHistogram \ - -hBinSize=0.021312 -hBinCount=1000 -hMinVal=-20 -verbose=2 \ - -db=hg38 phyloP241way > histogram.data 2>&1 - # real 2m43.313s - - # xaxis range: - grep -v chrom histogram.data | grep "^[0-9]" | ave -col=2 stdin \ - | sed -e 's/^/# /;' -# Q1 -10.953050 -# median -6.861155 -# Q3 -2.769245 -# average -6.875971 -# min -20.000000 -# max 1.312000 -# count 768 -# total -5280.745380 -# standard deviation 4.757034 - - # find out the range for the 2:5 graph - grep -v chrom histogram.data | grep "^[0-9]" | ave -col=5 stdin \ - | sed -e 's/^/# /;' -# Q1 0.000000 -# median 0.000001 -# Q3 0.000140 -# average 0.001302 -# min 0.000000 -# max 0.023556 -# count 768 -# total 0.999975 -# standard deviation 0.003490 - - # create plot of histogram: - printf 'set terminal png small x000000 xffffff xc000ff x66ff66 xffff00 x00ffff font \ -"/usr/share/fonts/default/Type1/n022004l.pfb" -set size 1.4, 0.8 -set key left box -set grid noxtics -set grid ytics -set title " Human hg38 Histogram phyloP241way track" -set xlabel " phyloP241way score" -set ylabel " Relative Frequency" -set y2label " Cumulative Relative Frequency (CRF)" -set y2range [0:1] -set y2tics -set xrange [-5:1.5] -set yrange [0:0.04] - -plot "histogram.data" using 2:5 title " RelFreq" with impulses, \ - "histogram.data" using 2:7 axes x1y2 title " CRF" with lines -' | gnuplot > histo.png - - # verify it looks sane - display histo.png & - -############################################################################# -# construct download files for 241-way (TBD - 2015-04-15 - Hiram) - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phastCons241way - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phyloP241way - mkdir /hive/data/genomes/hg38/bed/cactus241way/downloads - cd /hive/data/genomes/hg38/bed/cactus241way/downloads - mkdir cactus241way phastCons241way phyloP241way - - ######################################################################### - ## create upstream refGene maf files - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/cactus241way - # bash script - -#!/bin/sh -export geneTbl="refGene" -for S in 300 2000 5000 -do - echo "making upstream${S}.maf" - featureBits hg38 ${geneTbl}:upstream:${S} -fa=/dev/null -bed=stdout \ - | perl -wpe 's/_up[^\t]+/\t0/' | sort -k1,1 -k2,2n \ - | /cluster/bin/$MACHTYPE/mafFrags hg38 cactus241way \ - stdin stdout \ - -orgs=/hive/data/genomes/hg38/bed/cactus241way/species.list \ - | gzip -c > upstream${S}.${geneTbl}.maf.gz - echo "done upstream${S}.${geneTbl}.maf.gz" -done - - # real 88m40.730s - --rw-rw-r-- 1 52659159 Nov 6 11:46 upstream300.ncbiRefSeq.maf.gz --rw-rw-r-- 1 451126665 Nov 6 12:15 upstream2000.ncbiRefSeq.maf.gz --rw-rw-r-- 1 1080533794 Nov 6 12:55 upstream5000.ncbiRefSeq.maf.gz - - ###################################################################### - ## compress the maf files - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/cactus241way - mkdir maf - rsync -a -P ../../ucscNames/ ./maf/ - du -hsc maf/ - # 156G maf - cd maf - time gzip *.maf & - # real 135m1.784s - - du -hscL maf ../../ucscNames/ - # 18G maf - - cd maf - md5sum *.maf.gz *.nh > md5sum.txt - - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/maf - cd maf - ln -s `pwd`/* /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/maf - cd -- - ln -s `pwd`/*.maf.gz `pwd`/*.nh `pwd`/*.txt \ - /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ - - ########################################################################### - - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/cactus241way - grep TREE ../../4d/all.mod | awk '{print $NF}' \ - | ~/kent/src/hg/utils/phyloTrees/asciiTree.pl /dev/stdin \ - > hg38.241way.nh - ~/kent/src/hg/utils/phyloTrees/commonNames.sh hg38.241way.nh \ - | ~/kent/src/hg/utils/phyloTrees/asciiTree.pl /dev/stdin \ - > hg38.241way.commonNames.nh - ~/kent/src/hg/utils/phyloTrees/scientificNames.sh hg38.241way.nh \ - | $HOME/kent/src/hg/utils/phyloTrees/asciiTree.pl /dev/stdin \ - > hg38.241way.scientificNames.nh - time md5sum *.nh *.maf.gz > md5sum.txt - # real 0m3.147s - - ln -s `pwd`/*.maf.gz `pwd`/*.nh \ - /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way - - du -hsc ./maf ../../ucscNames - # 18G ./maf - # 156G ../../ucscNames - - # obtain the README.txt from hg38/multiz20way and update for this - # situation - ln -s `pwd`/*.txt \ - /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ - - ##################################################################### - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/phastCons241way - - mkdir hg38.241way.phastCons - cd hg38.241way.phastCons - ln -s ../../../cons/all/downloads/*.wigFix.gz . - md5sum *.gz > md5sum.txt - - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/phastCons241way - ln -s ../../cons/all/phastCons241way.bw ./hg38.phastCons241way.bw - ln -s ../../cons/all/all.mod ./hg38.phastCons241way.mod - time md5sum *.mod *.bw > md5sum.txt - # real 0m20.354s - - # obtain the README.txt from hg38/phastCons20way and update for this - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phastCons241way/hg38.241way.phastCons - cd hg38.241way.phastCons - ln -s `pwd`/* /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phastCons241way/hg38.241way.phastCons - - cd .. - # situation - ln -s `pwd`/*.mod `pwd`/*.bw `pwd`/*.txt \ - /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phastCons241way - - ##################################################################### - cd /hive/data/genomes/hg38/bed/cactus241way/downloads/phyloP241way - - mkdir hg38.241way.phyloP - cd hg38.241way.phyloP - - ln -s ../../../consPhyloP/all/downloads/*.wigFix.gz . - md5sum *.wigFix.gz > md5sum.txt - - cd .. - - ln -s ../../consPhyloP/run.phyloP/all.mod hg38.phyloP241way.mod - ln -s ../../consPhyloP/all/phyloP241way.bw hg38.phyloP241way.bw - - md5sum *.mod *.bw > md5sum.txt - - # Obtain The README.txt from hg38/phyloP20way and update for this - mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phyloP241way/hg38.241way.phyloP - cd hg38.241way.phyloP - ln -s `pwd`/* \ -/usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phyloP241way/hg38.241way.phyloP - - cd .. - - # situation - ln -s `pwd`/*.mod `pwd`/*.bw `pwd`/*.txt \ - /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/phyloP241way - -############################################################################# -# hgPal downloads (TBD - 2017-11-06 - Hiram) -# FASTA from 241-way for ncbiRefSeq, refGene and knownCanonical - - ssh hgwdev - screen -S hg38HgPal - mkdir /hive/data/genomes/hg38/bed/cactus241way/pal - cd /hive/data/genomes/hg38/bed/cactus241way/pal - cat ../species.list | tr '[ ]' '[\n]' > order.list - - ### knownCanonical with full CDS - cd /hive/data/genomes/hg38/bed/cactus241way/pal - export mz=cactus241way - export gp=knownCanonical - export db=hg38 - mkdir exonAA exonNuc knownCanonical - - time cut -f1 ../../../chrom.sizes | while read C - do - echo $C 1>&2 - hgsql hg38 -N -e "select chrom, chromStart, chromEnd, transcript from knownCanonical where chrom='$C'" > knownCanonical/$C.known.bed - done - - ls knownCanonical/*.known.bed | while read F - do - if [ -s $F ]; then - echo $F | sed -e 's#knownCanonical/##; s/.known.bed//' - fi - done | while read C - do - echo "date" - echo "mafGene -geneBeds=knownCanonical/$C.known.bed -noTrans $db $mz ncbiRefSeq order.list stdout | \ - gzip -c > protNuc/$C.protNuc.fa.gz" - echo "mafGene -geneBeds=knownCanonical/$C.known.bed $db $mz ncbiRefSeq order.list stdout | \ - gzip -c > protAA/$C.protAA.fa.gz" - done > $gp.$mz.prot.jobs - - time sh -x $gp.$mz.jobs > $gp.$mz.job.log 2>&1 - # 267m58.813s - - rm *.known.bed - export mz=cactus241way - export gp=knownCanonical - export db=hg38 - zcat protAA/c*.gz | gzip -c > $gp.$mz.protAA.fa.gz & - zcat protNuc/c*.gz | gzip -c > $gp.$mz.protNuc.fa.gz & - # about 6 minutes - - ### knownCanonical broken up by exon - cd /hive/data/genomes/hg38/bed/multiz100way/pal - export mz=multiz100way - export gp=knownCanonical - export db=hg38 - mkdir exonAA exonNuc knownCanonical - - time cut -f1 ../../../chrom.sizes | while read C - do - echo $C 1>&2 - hgsql hg38 -N -e "select chrom, chromStart, chromEnd, transcript from knownCanonical where chrom='$C'" > knownCanonical/$C.known.bed - done - # real 0m15.897s - - ls knownCanonical/*.known.bed | while read F - do - if [ -s $F ]; then - echo $F | sed -e 's#knownCanonical/##; s/.known.bed//' - fi - done | while read C - do - echo "date" - echo "mafGene -geneBeds=knownCanonical/$C.known.bed -exons -noTrans $db $mz ncbiRefSeq order.list stdout | \ - gzip -c > exonNuc/$C.exonNuc.fa.gz" - echo "mafGene -geneBeds=knownCanonical/$C.known.bed -exons $db $mz ncbiRefSeq order.list stdout | \ - gzip -c > exonAA/$C.exonAA.fa.gz" - done > $gp.$mz.jobs - - time sh -x $gp.$mz.jobs > $gp.$mz.job.log 2>&1 - # 267m58.813s - - rm *.known.bed - export mz=cactus241way - export gp=knownCanonical - export db=hg38 - zcat exonAA/c*.gz | gzip -c > $gp.$mz.exonAA.fa.gz & - zcat exonNuc/c*.gz | gzip -c > $gp.$mz.exonNuc.fa.gz & - # about 6 minutes - - rm -rf exonAA exonNuc - - export mz=multiz100way - export gp=knownCanonical - export db=hg38 - export pd=/usr/local/apache/htdocs-hgdownload/goldenPath/$db/$mz/alignments - mkdir -p $pd - ln -s `pwd`/$gp.$mz.exonAA.fa.gz $pd/$gp.exonAA.fa.gz - ln -s `pwd`/$gp.$mz.exonNuc.fa.gz $pd/$gp.exonNuc.fa.gz - ln -s `pwd`/$gp.$mz.protAA.fa.gz $pd/$gp.protAA.fa.gz - ln -s `pwd`/$gp.$mz.protNuc.fa.gz $pd/$gp.protNuc.fa.gz - cd $pd - md5sum *.fa.gz > md5sum.txt - - rm -rf exonAA exonNuc - - export mz=cactus241way - export gp=knownCanonical - export db=hg38 - export pd=/usr/local/apache/htdocs-hgdownload/goldenPath/$db/$mz/alignments - mkdir -p $pd - ln -s `pwd`/$gp.$mz.exonAA.fa.gz $pd/$gp.exonAA.fa.gz - ln -s `pwd`/$gp.$mz.exonNuc.fa.gz $pd/$gp.exonNuc.fa.gz - - # ncbiRefSeq - export mz=cactus241way - export gp=ncbiRefSeq - export db=hg38 - export I=0 - export D=0 - mkdir exonAA exonNuc - for C in `sort -nk2 ../../../chrom.sizes | cut -f1` - do - I=`echo $I | awk '{print $1+1}'` - D=`echo $D | awk '{print $1+1}'` - dNum=`echo $D | awk '{printf "%03d", int($1/300)}'` - mkdir -p exonNuc/${dNum} > /dev/null - mkdir -p exonAA/${dNum} > /dev/null - echo "mafGene -chrom=$C -exons -noTrans $db $mz $gp order.list stdout | gzip -c > exonNuc/${dNum}/$C.exonNuc.fa.gz &" - echo "mafGene -chrom=$C -exons $db $mz $gp order.list stdout | gzip -c > exonAA/${dNum}/$C.exonAA.fa.gz &" - if [ $I -gt 16 ]; then - echo "date" - echo "wait" - I=0 - fi - done > $gp.jobs - echo "date" >> $gp.jobs - echo "wait" >> $gp.jobs - - time (sh -x ./$gp.jobs) > $gp.jobs.log 2>&1 - # real 79m18.323s - - export mz=cactus241way - export gp=ncbiRefSeq - time find ./exonAA -type f | grep exonAA.fa.gz | xargs zcat \ - | gzip -c > $gp.$mz.exonAA.fa.gz - # real 1m28.841s - - time find ./exonNuc -type f | grep exonNuc.fa.gz | xargs zcat \ - | gzip -c > $gp.$mz.exonNuc.fa.gz - # real 3m56.370s - - # -rw-rw-r-- 1 397928833 Nov 6 18:44 ncbiRefSeq.cactus241way.exonAA.fa.gz - # -rw-rw-r-- 1 580377720 Nov 6 18:49 ncbiRefSeq.cactus241way.exonNuc.fa.gz - - export mz=cactus241way - export gp=ncbiRefSeq - export db=hg38 - export pd=/usr/local/apache/htdocs-hgdownload/goldenPath/$db/$mz/alignments - mkdir -p $pd - ln -s `pwd`/$gp.$mz.exonAA.fa.gz $pd/$gp.exonAA.fa.gz - ln -s `pwd`/$gp.$mz.exonNuc.fa.gz $pd/$gp.exonNuc.fa.gz - ln -s `pwd`/md5sum.txt $pd/ - - cd $pd - md5sum *.fa.gz > md5sum.txt - - rm -rf exonAA exonNuc - -############################################################################# -# wiki page for 241-way (TBD - 2017-11-06 - Hiram) - mkdir /hive/users/hiram/bigWays/hg38.241way - cd /hive/users/hiram/bigWays - echo "hg38" > hg38.241way/ordered.list - awk '{print $1}' /hive/data/genomes/hg38/bed/cactus241way/241way.distances.txt \ - >> hg38.241way/ordered.list - - # sizeStats.sh catches up the cached measurements required for data - # in the tables. They are usually already mostly done, only new - # assemblies will have updates. - ./sizeStats.sh hg38.241way/ordered.list - # dbDb.sh constructs hg38.241way/XenTro9_241-way_conservation_alignment.html - # may need to add new assembly references to srcReference.list and - # urlReference.list - ./dbDb.sh hg38 241way - # sizeStats.pl constructs hg38.241way/XenTro9_241-way_Genome_size_statistics.html - # this requires entries in coverage.list for new sequences - ./sizeStats.pl hg38 241way - - # defCheck.pl constructs XenTro9_241-way_conservation_lastz_parameters.html - ./defCheck.pl hg38 241way - - # this constructs the html pages in hg38.241way/: -# -rw-rw-r-- 1 6247 May 2 17:07 XenTro9_241-way_conservation_alignment.html -# -rw-rw-r-- 1 8430 May 2 17:09 XenTro9_241-way_Genome_size_statistics.html -# -rw-rw-r-- 1 5033 May 2 17:10 XenTro9_241-way_conservation_lastz_parameters.html - - # add those pages to the genomewiki. Their page names are the - # names of the .html files without the .html: -# XenTro9_241-way_conservation_alignment -# XenTro9_241-way_Genome_size_statistics -# XenTro9_241-way_conservation_lastz_parameters - - # when you view the first one you enter, it will have links to the - # missing two. - -############################################################################ -# pushQ readmine (TBD - 2017-11-07 - Hiram) - - cd /usr/local/apache/htdocs-hgdownload/goldenPath/hg38 - find -L `pwd`/cactus241way `pwd`/phastCons241way `pwd`/phyloP241way \ - /gbdb/hg38/cactus241way -type f \ - > /hive/data/genomes/hg38/bed/cactus241way/downloads/redmine.20216.fileList - wc -l /hive/data/genomes/hg38/bed/cactus241way/downloads/redmine.20216.fileList -# 1450 /hive/data/genomes/hg38/bed/cactus241way/downloads/redmine.20216.fileList - - cd /hive/data/genomes/hg38/bed/cactus241way/downloads - hgsql -e 'show tables;' hg38 | grep 241way \ - | sed -e 's/^/hg38./;' > redmine.20216.table.list - ############################################################################ ## adding frames ( DONE - 2021-04-29 - Hiram ) mkdir /hive/data/genomes/hg38/bed/cactus241way/frames cd /hive/data/genomes/hg38/bed/cactus241way/frames mkdir genes hgsql -N -e "select name,chrom,strand,txStart,txEnd,cdsStart,cdsEnd,exonCount,exonStarts,exonEnds from knownGene" hg38 \ | genePredSingleCover stdin stdout | gzip -2c \ > genes/hg38.gp.gz genePredCheck -db=hg38 genes/hg38.gp.gz # checked: 19328 failed: 0 @@ -999,15 +76,135 @@ do echo "${F}" 1>&2 zcat ${F} done | sort -k1,1 -k2,2n | gzip -c > cactus241wayFrames.bed.gz # real 0m3.178s hgLoadMafFrames hg38 cactus241wayFrames cactus241wayFrames.bed.gz featureBits -countGaps hg38 cactus241wayFrames # 33621579 bases of 3272116950 (1.028%) in intersection ############################################################################ + +############################################################################# +# construct download files for 241-way (TBD - 2015-04-15 - Hiram) + mkdir /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way + mkdir /hive/data/genomes/hg38/bed/cactus241way/downloads + cd /hive/data/genomes/hg38/bed/cactus241way/downloads + mkdir maf + for F in `ls ../ucscNames` +do + gzip -c ../ucscNames/$F > maf/${F}.gz +done + # running 10 of those simultaneously in an hgwdev perlPara.pl job: + # real 537m54.983s + + time md5sum *.maf.gz > md5sum.txt + +real 46m43.423s +user 39m29.707s +sys 6m41.942s + + ln -s ../cactus241way.nh hg38.cactus241way.nh + ln -s ../before.nameChange.nh hg38.cactus241way.scientificNames.nh + + # convert the phyloP bed files to bedGraph file: + + zcat ../phyloP/uppsala/*.bed.gz \ + | awk '{printf "%s\t%d\t%d\t%s\n", $1,$2,$3,$5}' \ + | sort -k1,1 -k2,2n > phyloP.bedGraph + + # Convert the bedGraph phyloP data to wigFixedStep via a perl script: + +#!/usr/bin/env perl + +use strict; +use warnings; + +my $chr = ""; +my $start = 0; +my $prevStart = 0; + +open (FH, "<phyloP.bedGraph") or die "can not open phyloP.bedGraph"; + while (my $line = <FH>) { + chomp $line; + my ($name, $chrStart, $chrEnd, $value) = split('\s+', $line); + if ( ($name ne $chr) || (($chrStart - $prevStart) > 1) ) { + printf "fixedStep chrom=%s start=%d step=1 span=1\n", $name, $chrStart+1; + } + printf "%s\n", $value; + $prevStart = $chrStart; + $chr = $name; + } +close (FH); + + time (./phyloP.sh > phyloP.bedGraph) >> phyloP.log 2>&1 + # real 232m54.489s + + # verify the wigEncode and wigToBigWig have the same result: + wigEncode cactus241way.phyloP.wigFix.gz hg38.cactus241way.wig \ + hg38.cactus241way.wib + # Converted cactus241way.phyloP.wigFix.gz, upper limit 9.28, + # lower limit -20.00 +# -rw-rw-r-- 1 2852623265 May 10 12:07 hg38.cactus241way.wib +# -rw-rw-r-- 1 307726912 May 10 12:07 hg38.cactus241way.wig + + wigToBigWig cactus241way.phyloP.wigFix.gz ../../../chrom.sizes \ + cactus241way.phyloP.bw +# -rw-rw-r-- 1 9644660543 May 10 12:28 cactus241way.phyloP.bw + + bigWigInfo cactus241way.phyloP.bw | sed -e 's/^/ # /;' + # version: 4 + # isCompressed: yes + # isSwapped: 0 + # primaryDataSize: 7,469,351,053 + # primaryIndexSize: 89,516,524 + # zoomLevels: 10 + # chromCount: 24 + # basesCovered: 2,852,623,265 + # mean: 0.128264 + # min: -20.000000 + # max: 9.280000 + # std: 1.252659 + + # compared to the loaded table: + time wigTableStats.sh hg38 phyloP241way +# db.table min max mean count sumData stdDev +# hg38.phyloP241way -20 9.28 0.128264 2852623265 3.6589e+08 1.25266 +# viewLimits=-6.13503:6.39156 + +# real 0m15.185s + + md5sum hg38.cactus* > md5sum.txt + + # construct a README.txt file + + # link to download: + mkdir -p /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/maf + ln -s `pwd`/hg38.*.nh \ + /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ + ln -s `pwd`/md5sum.txt \ + /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ + ln -s `pwd`/README.txt \ + /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ + ln -s `pwd`/hg38*.wigFix.gz \ + /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/ + + ln -s `pwd`/maf/* \ + /usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/maf + + # file list for redmine 27519 +/usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/*.txt +/usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/hg38.cactus241way.* +/usr/local/apache/htdocs-hgdownload/goldenPath/hg38/cactus241way/maf/* +/gbdb/hg38/cactus241way/maf/* +/gbdb/hg38/cactus241way/phyloP241way.wib + + # table list for redmine 27519 +hg38.cactus241way +hg38.cactus241wayFrames +hg38.cactus241waySummary +hg38.phyloP241way