269 lines
6.5 KiB
Go

// Package gtr defines a standard test report format and provides convenience
// methods to create and convert reports.
package gtr
import (
"fmt"
"strings"
"time"
"github.com/jstemmer/go-junit-report/v2/pkg/junit"
)
var (
propPrefixes = map[string]bool{"goos": true, "goarch": true, "pkg": true}
propFieldsFunc = func(r rune) bool { return r == ':' || r == ' ' }
)
type Report struct {
Packages []Package
}
func (r *Report) IsSuccessful() bool {
for _, pkg := range r.Packages {
if pkg.BuildError.Name != "" || pkg.RunError.Name != "" {
return false
}
for _, t := range pkg.Tests {
if t.Result != Pass && t.Result != Skip {
return false
}
}
}
return true
}
type Package struct {
Name string
Duration time.Duration
Coverage float64
Output []string
Tests []Test
Benchmarks []Benchmark
BuildError Error
RunError Error
}
type Test struct {
Name string
Duration time.Duration
Result Result
Level int
Output []string
}
type Benchmark struct {
Name string
Result Result
Output []string
Iterations int64
NsPerOp float64
MBPerSec float64
BytesPerOp int64
AllocsPerOp int64
}
type Error struct {
Name string
Duration time.Duration
Cause string
Output []string
}
// FromEvents creates a Report from the given list of events.
// TODO: make packageName optional option
func FromEvents(events []Event, packageName string) Report {
report := NewReportBuilder(packageName)
for _, ev := range events {
switch ev.Type {
case "run_test":
report.CreateTest(ev.Name)
case "pause_test":
report.PauseTest(ev.Name)
case "cont_test":
report.ContinueTest(ev.Name)
case "end_test":
report.EndTest(ev.Name, ev.Result, ev.Duration, ev.Indent)
case "benchmark":
report.Benchmark(ev.Name, ev.Iterations, ev.NsPerOp, ev.MBPerSec, ev.BytesPerOp, ev.AllocsPerOp)
case "status":
report.End()
case "summary":
report.CreatePackage(ev.Name, ev.Result, ev.Duration, ev.Data)
case "coverage":
report.Coverage(ev.CovPct, ev.CovPackages)
case "build_output":
report.CreateBuildError(ev.Name)
case "output":
report.AppendOutput(ev.Data)
default:
fmt.Printf("unhandled event type: %v\n", ev.Type)
}
}
return report.Build()
}
// JUnit converts the given report to a collection of JUnit Testsuites.
func JUnit(report Report, hostname string, now time.Time) junit.Testsuites {
timestamp := now.Format(time.RFC3339)
var suites junit.Testsuites
for _, pkg := range report.Packages {
var duration time.Duration
suite := junit.Testsuite{
Name: pkg.Name,
Timestamp: timestamp,
Hostname: hostname,
}
if len(pkg.Output) > 0 {
suite.SystemOut = &junit.Output{Data: formatOutput(pkg.Output, 0)}
}
if pkg.Coverage > 0 {
suite.AddProperty("coverage.statements.pct", fmt.Sprintf("%.2f", pkg.Coverage))
}
for _, line := range pkg.Output {
if fields := strings.FieldsFunc(line, propFieldsFunc); len(fields) == 2 && propPrefixes[fields[0]] {
suite.AddProperty(fields[0], fields[1])
}
}
for _, test := range pkg.Tests {
duration += test.Duration
tc := junit.Testcase{
Classname: pkg.Name,
Name: test.Name,
Time: junit.FormatDuration(test.Duration),
}
if test.Result == Fail {
tc.Failure = &junit.Result{
Message: "Failed",
Data: formatOutput(test.Output, test.Level),
}
} else if test.Result == Skip {
tc.Skipped = &junit.Result{
Message: formatOutput(test.Output, test.Level),
}
} else if test.Result == Unknown {
tc.Error = &junit.Result{
Message: "No test result found",
Data: formatOutput(test.Output, test.Level),
}
}
suite.AddTestcase(tc)
}
for _, bm := range mergeBenchmarks(pkg.Benchmarks) {
tc := junit.Testcase{
Classname: pkg.Name,
Name: bm.Name,
Time: junit.FormatBenchmarkTime(time.Duration(bm.NsPerOp)),
}
if bm.Result == Fail {
tc.Failure = &junit.Result{
Message: "Failed",
}
}
suite.AddTestcase(tc)
}
// JUnit doesn't have a good way of dealing with build or runtime
// errors that happen before a test has started, so we create a single
// failing test that contains the build error details.
if pkg.BuildError.Name != "" {
tc := junit.Testcase{
Classname: pkg.BuildError.Name,
Name: pkg.BuildError.Cause,
Time: junit.FormatDuration(0),
Error: &junit.Result{
Message: "Build error",
Data: strings.Join(pkg.BuildError.Output, "\n"),
},
}
suite.AddTestcase(tc)
}
if pkg.RunError.Name != "" {
tc := junit.Testcase{
Classname: pkg.RunError.Name,
Name: "Failure",
Time: junit.FormatDuration(0),
Error: &junit.Result{
Message: "Run error",
Data: strings.Join(pkg.RunError.Output, "\n"),
},
}
suite.AddTestcase(tc)
}
if (pkg.Duration) == 0 {
suite.Time = junit.FormatDuration(duration)
} else {
suite.Time = junit.FormatDuration(pkg.Duration)
}
suites.AddSuite(suite)
}
return suites
}
func formatOutput(output []string, level int) string {
var lines []string
for _, line := range output {
lines = append(lines, trimOutputPrefix(line, level))
}
return strings.Join(lines, "\n")
}
func trimOutputPrefix(line string, level int) string {
// We only want to trim the whitespace prefix if it was part of the test
// output. Test output is usually prefixed by a series of 4-space indents,
// so we'll check for that to decide whether this output was likely to be
// from a test.
prefixLen := strings.IndexFunc(line, func(r rune) bool { return r != ' ' })
if prefixLen%4 == 0 {
// Use the subtest level to trim a consistenly sized prefix from the
// output lines.
for i := 0; i <= level; i++ {
line = strings.TrimPrefix(line, " ")
}
}
return strings.TrimPrefix(line, "\t")
}
func mergeBenchmarks(benchmarks []Benchmark) []Benchmark {
var merged []Benchmark
benchmap := make(map[string][]Benchmark)
for _, bm := range benchmarks {
if _, ok := benchmap[bm.Name]; !ok {
merged = append(merged, Benchmark{Name: bm.Name})
}
benchmap[bm.Name] = append(benchmap[bm.Name], bm)
}
for i, bm := range merged {
for _, b := range benchmap[bm.Name] {
bm.NsPerOp += b.NsPerOp
bm.MBPerSec += b.MBPerSec
bm.BytesPerOp += b.BytesPerOp
bm.AllocsPerOp += b.AllocsPerOp
}
n := len(benchmap[bm.Name])
merged[i].NsPerOp = bm.NsPerOp / float64(n)
merged[i].MBPerSec = bm.MBPerSec / float64(n)
merged[i].BytesPerOp = bm.BytesPerOp / int64(n)
merged[i].AllocsPerOp = bm.AllocsPerOp / int64(n)
}
return merged
}