4898794edd81be5285ea6e544acbedeaeb31bf78 max Tue Nov 23 08:10:57 2021 -0800 Fixing pointers to README file for license in all source code files. refs #27614 diff --git src/hg/hgTables/wikiTrack.c src/hg/hgTables/wikiTrack.c index 22c5eb5..8a5bd76 100644 --- src/hg/hgTables/wikiTrack.c +++ src/hg/hgTables/wikiTrack.c @@ -1,109 +1,109 @@ /* hgTables - Main and utility functions for table browser. */ /* Copyright (C) 2013 The Regents of the University of California - * See README in this or parent directory for licensing information. */ + * See kent/LICENSE or http://genome.ucsc.edu/license/ for licensing information. */ #include "common.h" #include "trackDb.h" #include "hui.h" #include "hdb.h" #include "wikiTrack.h" #include "hgTables.h" struct hTableInfo *wikiHti() /* Create an hTableInfo for the wikiTrack. */ { struct hTableInfo *hti; AllocVar(hti); hti->rootName = cloneString(WIKI_TRACK_TABLE); hti->isPos = TRUE; hti->isSplit = FALSE; hti->hasBin = TRUE; hti->type = cloneString("none"); strncpy(hti->chromField, "chrom", 32); strncpy(hti->startField, "chromStart", 32); strncpy(hti->endField, "chromEnd", 32); strncpy(hti->nameField, "name", 32); strncpy(hti->scoreField, "score", 32); strncpy(hti->strandField, "strand", 32); return(hti); } void doSummaryStatsWikiTrack(struct sqlConnection *conn) /* Put up page showing summary stats for wikiTrack. */ { struct sqlConnection *wikiConn = wikiConnect(); doSummaryStatsBed(wikiConn); wikiDisconnect(&wikiConn); } static void wikiTrackFilteredBedOnRegion( struct region *region, /* Region to get data from. */ struct hash *idHash, /* Hash of identifiers or NULL */ struct bedFilter *bf, /* Filter or NULL */ struct lm *lm, /* Local memory pool. */ struct bed **pBedList /* Output get's appended to this list */ ) /* Get the wikiTrack items passing filter on a single region. */ { struct bed *bed; int fieldCount = 6; char query[512]; int rowOffset; char **row; struct sqlConnection *wikiConn = wikiConnect(); struct sqlResult *sr = NULL; char where[512]; char *filter = filterClause(wikiDbName(), WIKI_TRACK_TABLE, region->chrom, NULL); if (filter) sqlSafefFrag(where, sizeof(where), "db='%s' AND %-s", database, filter); else sqlSafefFrag(where, sizeof(where), "db='%s'", database); sqlSafef(query, sizeof(query), "select * from %s", WIKI_TRACK_TABLE); sr = hRangeQuery(wikiConn, WIKI_TRACK_TABLE, region->chrom, region->start, region->end, where, &rowOffset); while ((row = sqlNextRow(sr)) != NULL) { bed = bedLoadN(row+rowOffset, fieldCount); if ((idHash == NULL || hashLookup(idHash, bed->name)) && (bf == NULL || bedFilterOne(bf, bed))) { struct bed *copy = lmCloneBed(bed, lm); slAddHead(pBedList, copy); } } sqlFreeResult(&sr); wikiDisconnect(&wikiConn); } struct bed *wikiTrackGetFilteredBeds(char *name, struct region *regionList, struct lm *lm, int *retFieldCount) /* Get list of beds from the wikiTrack * in current regions and that pass * filters. You can bedFree this when done. */ { struct bed *bedList = NULL; struct hash *idHash = NULL; struct bedFilter *bf = NULL; struct region *region = NULL; /* Figure out how to filter things. */ bf = bedFilterForCustomTrack(name); idHash = identifierHash(database, name); /* Grab filtered beds for each region. */ for (region = regionList; region != NULL; region = region->next) wikiTrackFilteredBedOnRegion(region, idHash, bf, lm, &bedList); /* clean up. */ hashFree(&idHash); slReverse(&bedList); return bedList; }