func digestFromIntermediate(test, digest string, inter *intermediate, e *expstorage.Expectations, tile *tiling.Tile, tallies *tally.Tallies, blamer *blame.Blamer, diffStore diff.DiffStore, paramset *paramsets.Summary, includeIgnores bool) *Digest { traceTally := tallies.ByTrace() ret := &Digest{ Test: test, Digest: digest, Status: e.Classification(test, digest).String(), Traces: buildTraces(test, digest, inter, e, tile, traceTally, paramset, includeIgnores), Diff: buildDiff(test, digest, inter, e, tile, tallies.ByTest(), blamer, diffStore, paramset, includeIgnores), } return ret }
// searchTile queries across a tile. func searchTile(q *Query, e *expstorage.Expectations, parsedQuery url.Values, storages *storage.Storage, tile *tiling.Tile, tallies *tally.Tallies, blamer *blame.Blamer, paramset *paramsets.Summary) ([]*Digest, []*tiling.Commit, error) { // TODO Use CommitRange to create a trimmed tile. traceTally := tallies.ByTrace() lastCommitIndex := tile.LastCommitIndex() // Loop over the tile and pull out all the digests that match // the query, collecting the matching traces as you go. Build // up a set of intermediate's that can then be used to calculate // Digest's. // map [test:digest] *intermediate inter := map[string]*intermediate{} for id, tr := range tile.Traces { if tiling.Matches(tr, parsedQuery) { test := tr.Params()[types.PRIMARY_KEY_FIELD] // Get all the digests digests := digestsFromTrace(id, tr, q.Head, lastCommitIndex, traceTally) for _, digest := range digests { cl := e.Classification(test, digest) if q.excludeClassification(cl) { continue } // Fix blamer to make this easier. if q.BlameGroupID != "" { if cl == types.UNTRIAGED { b := blamer.GetBlame(test, digest, tile.Commits) if q.BlameGroupID != blameGroupID(b, tile.Commits) { continue } } else { continue } } key := fmt.Sprintf("%s:%s", test, digest) if i, ok := inter[key]; !ok { inter[key] = newIntermediate(test, digest, id, tr, digests) } else { i.addTrace(id, tr, digests) } } } } // Now loop over all the intermediates and build a Digest for each one. ret := make([]*Digest, 0, len(inter)) for key, i := range inter { parts := strings.Split(key, ":") ret = append(ret, digestFromIntermediate(parts[0], parts[1], i, e, tile, tallies, blamer, storages.DiffStore, paramset, q.IncludeIgnores)) } return ret, tile.Commits, nil }
// oneStep does a single step, calculating all the paramsets from the latest tile and tallies. // // Returns the paramsets for both the tile with and without ignored traces included. func oneStep(tallies *tally.Tallies, storages *storage.Storage) (map[string]map[string][]string, map[string]map[string][]string, error) { defer timer.New("paramsets").Stop() tile, err := storages.GetLastTileTrimmed(false) if err != nil { return nil, nil, fmt.Errorf("Failed to get tile: %s", err) } byTrace := byTraceForTile(tile, tallies.ByTrace()) tile, err = storages.GetLastTileTrimmed(true) if err != nil { return nil, nil, fmt.Errorf("Failed to get tile: %s", err) } byTraceIncludeIgnored := byTraceForTile(tile, tallies.ByTrace()) return byTrace, byTraceIncludeIgnored, nil }
// Search returns a slice of Digests that match the input query, and the total number of Digests // that matched the query. It also returns a slice of Commits that were used in the calculations. func Search(q *Query, storages *storage.Storage, tallies *tally.Tallies, blamer *blame.Blamer, paramset *paramsets.Summary) ([]*Digest, int, []*tiling.Commit, error) { tile, err := storages.GetLastTileTrimmed(q.IncludeIgnores) if err != nil { return nil, 0, nil, fmt.Errorf("Couldn't retrieve tile: %s", err) } // TODO Use CommitRange to create a trimmed tile. parsedQuery, err := url.ParseQuery(q.Query) if err != nil { return nil, 0, nil, fmt.Errorf("Failed to parse Query in Search: %s", err) } e, err := storages.ExpectationsStore.Get() if err != nil { return nil, 0, nil, fmt.Errorf("Couldn't get expectations: %s", err) } traceTally := tallies.ByTrace() lastCommitIndex := tile.LastCommitIndex() // Loop over the tile and pull out all the digests that match // the query, collecting the matching traces as you go. Build // up a set of intermediate's that can then be used to calculate // Digest's. // map [test:digest] *intermediate inter := map[string]*intermediate{} for id, tr := range tile.Traces { if tiling.Matches(tr, parsedQuery) { test := tr.Params()[types.PRIMARY_KEY_FIELD] // Get all the digests digests := digestsFromTrace(id, tr, q.Head, lastCommitIndex, traceTally) for _, digest := range digests { cl := e.Classification(test, digest) switch { case cl == types.NEGATIVE && !q.Neg: continue case cl == types.POSITIVE && !q.Pos: continue case cl == types.UNTRIAGED && !q.Unt: continue } // Fix blamer to make this easier. if q.BlameGroupID != "" { if cl == types.UNTRIAGED { b := blamer.GetBlame(test, digest, tile.Commits) if q.BlameGroupID != blameGroupID(b, tile.Commits) { continue } } else { continue } } key := fmt.Sprintf("%s:%s", test, digest) if i, ok := inter[key]; !ok { inter[key] = newIntermediate(test, digest, id, tr, digests) } else { i.addTrace(id, tr, digests) } } } } // Now loop over all the intermediates and build a Digest for each one. ret := make([]*Digest, 0, len(inter)) for key, i := range inter { parts := strings.Split(key, ":") ret = append(ret, digestFromIntermediate(parts[0], parts[1], i, e, tile, tallies, blamer, storages.DiffStore, paramset, q.IncludeIgnores)) } sort.Sort(DigestSlice(ret)) fullLength := len(ret) if fullLength > q.Limit { ret = ret[0:q.Limit] } return ret, fullLength, tile.Commits, nil }