-
Notifications
You must be signed in to change notification settings - Fork 17.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
proposal: testing: store test artifacts #71287
Comments
Speaking from the perspective of Go's CI infrastructure, it would be helpful to have some standard convention to indicate output artifacts at the standard library level, so I'm in support of this proposal (or something like it). We share the tooling for processing Go tests with several other teams, and without a standard for this sort of thing, we would be stuck having to define our own amongst ourselves. That may turn out to be OK in practice, but having a single way to do it would save us all the time of trying to define our own convention, since it wouldn't even be a question. As far as test metadata goes, I think this would be one of the most valuable forms, for the Go team anyway, and IMO more valuable than
I don't have many thoughts on the specifics of how this should look in the output or in the |
How would this interact with fuzzing which currently stores new inputs that cause crashes in |
No changes to fuzzing, unless I'm missing something. Is there some interaction that should happen there? |
I was thinking they're similar in that they generate test artifacts, and if you're running in some remote CI system, you may want a (relatively) easy way to extract the new additions from the other fuzz inputs, such as redirecting them to a specified directory as in this proposal. |
I don't think we can have -outputdir stop us from writing fuzz crashes to Perhaps we could write fuzz crashes to both I think a -fuzzoutputdir flag is probably a separate proposal, though. |
Instead of returning a string, could it return an os.Root? |
It could, but that'd be a gratuitous divergence from TB.TempDir. If you want a Root, it's easy enough to open the directory as one. And sometimes you do want just the directory name (when passing it as a flag to some other process, say). |
If I run such tests for an unfamiliar package with just Where will the paths likely be? Under /tmp, under the package, under GOPATH? Do tests get unique paths over multiple test runs, or does one test run overwrite the artifacts of another for the same test? If overwrite, are all previous artifacts deleted first?
Perhaps fuzz artifacts should be thought of as generated inputs, not outputs, so they shouldn't be included in this. |
If you don't pass the --outputdir flag to "go test", artifacts get written to a temporary directory (exactly the same as if they were written to a t.TempDir) and are deleted after the test finishes. If you do pass --outputdir, artifacts get written to that directory and you can decide what to do with them from there. Persistent artifacts are only produced when you ask for them, in the location you specify.
That is an excellent question. I'm going to have to think about it. |
When a test using I see three possible behaviors here:
|
This proposal has been added to the active column of the proposals project |
This is currently documented as
This will muddy the "default" value. I think under this proposal, if There's also an implementation issue with using One option would be for
This sounds reasonable. If we give each iteration the same output directory, there's bound to be a confusing mix of fixed file names that get overwritten each time and generated file names that accumulate. |
If muddying the default value for Another option might be to add |
For profiling, there are two separate concerns: "enable profiling" and "where do profiles get written" (outputdir). It seems reasonable to have one outputdir and just have a flag for "write artifacts" (default off). Then we can still keep the new OutputDir method, which may be useful regardless, but also have a "Artifacts() bool" method or something like that. Maybe it's not a method but is a top-level function like testing.Short. So you'd do go test -artifacts to drop them in the current directory, and by default you wouldn't get them at all. Although I'm not sure why it's one flag to enable all possible artifacts. It seems like you might want each artifact writer to define its own flag, or at least have some kind of sub-selection. In the explanation:
If one test is failing, it doesn't seem like you want to save all the artifacts from all the tests. Maybe -artifacts takes a test pattern? Or something else? artifacts is also such a long string to type. It would be nice if we had a shorter name for these. |
If you don't want artifacts from all tests, you can just run the tests that you want artifacts for, no? "Run these tests, but only save the artifacts from those tests" seems overly complicated. A separate "write artifacts (to the output dir, which might be the current dir)" flag sounds reasonable to me. I don't really like an For example, one place I might use artifact storage is in a test for a code generator, like the protobuf compiler. These tests often compare the output of the generator to a golden reference. Storing the current output in the test artifacts is useful when debugging changes. With the current proposal, that's a simple matter of putting the files in |
If we were starting from scratch and didn't already pass -test.outputdir to existing test invocations, @neild's previous comment would be spot on. Given that we have to add a different knob than outputdir (or else all tests will write artifacts to the current directory just because you're profiling), it feels like unnecessary expense to write files to a temporary directory that will be deleted. It seems to come down to:
For (1), especially if artifacts can be big, it seems very wrong to me to write them 100% of the time when 99.9% of the time the go command will just delete them right away. For (2), -artifactdir doesn't seem much worse than -artifacts, and maybe people will want to set -artifactdir separately from -outputdir, so probably ArtifactDir is better. And it can return an empty string if artifacts aren't being kept? |
This is a bit of a philosophical question, I think. My opinion is that tests should, by default, endeavor to always have consistent behavior. In the common case, there's no real cost to writing artifacts and then deleting them. There are also several benefits:
In the event that artifacts are expensive to generate, a test can define its own flag to control generation. I don't think that's the common case, though. |
Perhaps this is implied by the existing proposal and I've missed it, but what happens if I do this?
Does each package being tested get its own output directory within The other thing I'm thinking here is that sometimes (usually?) we only care about test artifacts from failing tests. Both the use cases mentioned in the initial proposal seem to come into that category. How about this: if we haven't printed the name of the artifact directory anywhere (i.e. the test has failed and we're not running in verbose mode) the artifacts for that test are removed after the test completes? When there are thousands of tests, this would vastly improve the signal-to-noise ratio of the artifact directory contents, and A flag could be added to request that all artifacts are written regardless (or just running tests |
This is a good point: When testing multiple packages, tests with the same name in different packages must be given unique artifact directories. The simplest way to achieve this is probably to give each package a unique artifact directory and put the tests under that.
An option to store only artifacts from failing tests seems pretty reasonable, but it's another flag for users to think about. I don't have a strong opinion on whether it's worth it. @rsc proposes (#71287 (comment)) an |
Thinking about published versus unpublished artifacts is helping me make sense of this proposal. I wonder if users have the right headroom, the necessary hooks from A first pass at how I could make headway on optionally publishing artifacts on pass/fail logic:
From prior experimentation I know that the implementation details for step 2 can get a bit weedy, and there can be some rough edges. For example, I'm not sure how well this approach helps for Along these lines there might be a "worse-is-better" appeal to just having a flag that, when present, says what directory artifacts are written to, and when not present, everything is ephemeral. |
I hypothesize that:
So I think the right thing to do is to generate and store artifacts for all tests in a user-defined output directory. If users want some more complicated behavior (store only artifacts for failing tests, say), they can use a tool that parses the test JSON output and does whatever is desired. We don't need to do anything more complicated in the go tool, at least not to start with. |
It sounds like we're converging on adding an
Under the output directory, we create a separate directory for each test package, and within that we create a separate directory for each call of a test function or subtest (presumably named after the test plus some uniquifier if necessary). |
Should this be just a concrete method on |
Another question is what to do if an artifact directory already exists. My inclination is to never delete anything, and lean on the path uniquifier we need anyway in case a test runs multiple times. That also has the benefit of making |
Here's what I've got: The proposal is to add an API and command line flags for saving test artifacts. The following method would be added to package testing
type TB interface {
...
// ArtifactDir returns a directory for the test to store output files in.
// When the -artifacts flag is provided, this directory will be located
// under the output directory. Otherwise, ArtifactDir returns a temporary directory
// which is removed after the test completes.
//
// Each test or subtest within each test package has a unique artifact directory.
// Repeated calls to ArtifactDir in the same test or subtest return the same directory.
// Subtest outputs are not located under the parent test's output directory.
func (t *testing.T) ArtifactDir() string
} A new boolean The artifact directory would be named When artifacts are enabled, the first call to
When
|
Is the "path uniquifier" essentially just a call to That would then presumably make them be guaranteed unique across runs while still containing some information to help figure out which subdirectory belonged to which test, in which case I'd find that a fine compromise for allowing |
note that test names can be unreasonably long, as in #71742 |
API in #71287 (comment) LGTM. For the specific directory name, I propose:
We document none of this and make no guarantees that it won't change. (Although Hyrum's Law probably means we can't change any of it.) The use of |
I'd been imagining it would be an actual ascending decimal number (maybe zero-padded?), so it's clear what order the test instances ran in, but whatever. |
A note of caution: test names can not only be very long, but can also contain arbitrary text which might interact poorly with the filesystem (e.g. https://go.dev/play/p/P0LbnZN5V0e) |
If test names seem tricky, a quick thought on another scheme:
|
When I said "somehow derived from" I was being intentionally vague about it but assuming that it would involve stripping out or replacing all but a safe set of characters, and then assuming that any resulting collisions would get taken care of by I expect that in practice those who are writing tests that generate artifacts will choose their test names based on what remains intelligible after that transform, and the names of any preexisting tests that don't generate artifacts don't matter for our purposes here. The length limit is a good point I admittedly hadn't considered, but I think all those same assumptions still apply: there are not yet any tests making use of this new feature, and those who choose to make use of it will presumably choose test names that work well with how the paths get generated. (I don't mean to say that it's okay to generate invalid path names, but rather that it seems okay to turn unreasonable test names into less useful directory names, and assume that those who care about having readable directory names will pick more reasonable test names.) |
Should it also be implemented in |
Running with @apparentlymart 's observation, I think we can leave the exact details of the path-cleaning transformation unspecified and subject to change and specifically document that if people want predictable artifact paths, they should chose simple test names. I think we need to specify that Or we tell people to just use the |
Here's a concrete proposal for the directory naming. I don't think we need to pin down the exact details to move the proposal forward, just the general shape.
|
Have all remaining concerns about this proposal been addressed? The proposal is to add an API and command line flags for saving test artifacts. The following method would be added to package testing
type TB interface {
...
// ArtifactDir returns a directory for the test to store output files in.
// When the -artifacts flag is provided, this directory will be located
// under the output directory. Otherwise, ArtifactDir returns a temporary directory
// which is removed after the test completes.
//
// Each test or subtest within each test package has a unique artifact directory.
// Repeated calls to ArtifactDir in the same test or subtest return the same directory.
// Subtest outputs are not located under the parent test's output directory.
func (t *testing.T) ArtifactDir() string
} A new boolean The artifact directory would be named When artifacts are enabled, the first call to
When
|
If there are, or can be, multiple packages with the same name in your project, then wouldn’t you need to use a unique artifact path per package? If so, having the package name in the artifact path seems redundant. |
On the other hand, package names are often distinct, and using the package name will separate out different tests and make them easier to see and find. Since there is a random string in the directory name, uniqueness is not required. Or, we could use the package path rather than the package name. The package path must be unique. |
I'd agree that it seems quite rare for a single Go module to contain multiple packages with the same name, even though the toolchain technically allows that. AFAICR Again I find myself returning to the idea that this artifact feature is new and so no existing tests could possibly be using it, and when people start using it they are unlikely to intentionally use it in ways cause a confusing directory layout under the artifact directory, so to me it seems safe to assume that package names are "unique enough", and let the unique part of the path deal with the (presumed) rare case where there are two colliding package names. Given that all of the packages in a particular module tend to share a prefix that is much longer than the package name itself, I think it would be annoying to use the entire package path as the directory name. Though if folks disagree with my assumption that package names are unlikely to collide within a single module, perhaps we could compromise and use whatever is left of the package path after trimming the module path prefix from the front of it. 🤔 |
A thin argument for package path over package name: It gives canonical locations relative to module root. Suppose all CI big or small will have opinions about what to do with artifacts, but they don't converge, the upper bound seems like staging things for further automation. Canonical paths relative to module root seems nicer in that sense. |
I imagine in many repos, especially monorepos, package names like “models”, “services”, “auth”, “users”, etc. could be common. It sounds like there’s a desire to enable users to find recent artifacts. Using the package path instead of the name would be more helpful with that. Also, using an incrementing number or a Unix timestamp instead of a random string in the artifact path would make it easier to find the latest run’s artifacts. I imagine for a monorepo with a single module at the root, in CI, you’d just want to do “go test -artifacts $ARTIFACTMOUNT ./…”, and let the go command deal with each package’s artifact path automatically. |
I think there are two use cases with conflicting needs here.
This suggests a way to satisfy both cases: collect all of the package paths on the command line, plus the package path of "." (if in a module), trim the common leading prefix elements, and use what remains as the artifact directory. This can trim off the entire package path, so I propose adding Here are some examples of how this would play out. Suppose we have a module
Some alternatives I don't like as much:
|
I don't have strong opinions here, but my preference is for something simple. I'd personally rather have a long but predictable path than one which varies depending on whether I'm testing one package or two, or what my current working directory is. Package path minus the module path seems like a good balance. For example, golang.org/x/net/dns/dnsmessage would be "_artifacts/dns/dnsmessage". |
Have all remaining concerns about this proposal been addressed? The proposal is to add an API and command line flags for saving test artifacts. The following method would be added to package testing
type TB interface {
...
// ArtifactDir returns a directory for the test to store output files in.
// When the -artifacts flag is provided, this directory will be located
// under the output directory. Otherwise, ArtifactDir returns a temporary directory
// which is removed after the test completes.
//
// Each test or subtest within each test package has a unique artifact directory.
// Repeated calls to ArtifactDir in the same test or subtest return the same directory.
// Subtest outputs are not located under the parent test's output directory.
func (t *testing.T) ArtifactDir() string
} A new boolean The artifact directory would be named When artifacts are enabled, the first call to
When
|
I believe the cleanup part is underrepresented. There are 4 different behaviours necessary
Point 1 and 2 are relevant for this proposal but not addressed. Point 4 is addressed by testing.TB.Tempdir and the matching environment variables for the system. Point 3 is addressed spot on by the current proposal. Especially addressing point 1 would provide alternatives for when using testing.TB.Tempdir would fit but isn't possible, because we can't keep things around on failure. |
Based on the discussion above, this proposal seems like a likely accept. The proposal is to add an API and command line flags for saving test artifacts. The following method would be added to package testing
type TB interface {
...
// ArtifactDir returns a directory for the test to store output files in.
// When the -artifacts flag is provided, this directory will be located
// under the output directory. Otherwise, ArtifactDir returns a temporary directory
// which is removed after the test completes.
//
// Each test or subtest within each test package has a unique artifact directory.
// Repeated calls to ArtifactDir in the same test or subtest return the same directory.
// Subtest outputs are not located under the parent test's output directory.
func (t *testing.T) ArtifactDir() string
} A new boolean The artifact directory would be named When artifacts are enabled, the first call to
When
|
I don't think this case is interesting. The normal outcome is for a test to succeed. If you want to store artifacts when the test succeeds, I don't know why you wouldn't want to when it fails and presumably needs debugging. In the worst case, you have an artifact you'll need to delete.
I think this case actually applies to more than just artifacts. It's not uncommon that when I'm debugging a test failure, I find that I want to examine its temporary files. This is a bit annoying because With the artifacts proposal, I could change the test to use This might imply that if a test creates temporary files that might be useful in debugging failures, it should always use Maybe we should have a test flag that disables deleting temporary directories when a test fails. This would apply to both I think that's a separate proposal than this one, though. |
This is an offshoot of #43936.
Some tests produce output files which the user may want to examine. For example, a test might produce files which are compared to some reference. If the comparison fails, the user will want to examine the generated files. Or a test might produce a packet capture or similar trace which can be used to debug failures.
We call these output files "test artifacts".
This is a proposal to add support for storing test artifacts to the testing package.
We add a new method to testing.TB:
The -outputdir flag already exists, and is currently used to specify a location to put output files from profiling. We're adding an additional meaning to it here: It's now where all your saved test outputs go.
When -outputdir is specified, the first call to OutputDir in a test or subtest will emit a line to the test output consisting of "=== ARTIFACTS ", the test name, and the test artifact directory, separated by spaces:
When -json is specified, this will appear as an entry with an Action of "artifacts", the usual Time, Package, and Test keys, and a "Path" key containing the artifact directory:
That's the proposal.
A few points on the design:
I'm reusing the existing -outputdir flag, on the theory that output files from profiling are just another test artifact. If we don't like that reuse, then we could add a new -artifactdir flag and rename TB.OutputDir to TB.ArtifactDir for consistency.
The test output uses the word "ARTIFACTS" because the JSON already has "output" events.
TB.OutputDir returns a directory, same as TB.TempDir. This seems simpler than asking the testing package to copy files around.
TB.OutputDir returns a directory even if we aren't saving artifacts so test behavior doesn't change depending on the presence or absence of the -outputdir flag.
In simple interactive use, users can pass -outputdir to store test artifacts when debugging a failing test.
Test frameworks that collect artifacts can arrange to pass -outputdir to the tests they run and collect any artifacts after the fact.
As a concrete use case, within Google our testing infrastructure sets an environment variable to the location of a directory. Tests can write files into this directory, and those files will be stored and associated with the test run. If we implement this proposal, we can arrange for the test infrastructure to also pass this directory as an -outputdir flag, and any test using TB.OutputDir will automatically use the right location.
The text was updated successfully, but these errors were encountered: