Skip to content

Latest commit

 

History

History
146 lines (100 loc) · 6.37 KB

HACKING.md

File metadata and controls

146 lines (100 loc) · 6.37 KB

Hacking on composer-cli

This new implementation of composer-cli is broken up into 2 major pieces. The ./weldr/ package contains the library functions for communicating with the API server. These functions are publicly accessable and should be documented so that 3rd party users can use them to communicate with the osbuild-composer server.

The ./cmd/composer-cli/ code is the user interface code, separated out into commands and subcommands. They are for interactive use, and by scripts via the --json argument that switches the output to JSON formated objects.

Adding new library functions

The weldr library should contain functions, separated into descriptively named files, like blueprints.go. The functions should handle communication with the API server and return values to the caller using, if possible, regular Go data types. eg. ListBlueprints returns a slice of strings.

Errors from lower functions should be returned to the called as errors, and API error responses should be returned as APIResponse or slices of APIErrorMsg. The type returned will depend on the response from the server.

The blueprints.go has examples of most of the types of responses you will encounter, use it as a template for adding new functions.

Not every response can use a Go type. For those that need to return more details use structs that match the API structs used in the server. We do not want to import the actual server modules (and cannot because they are all implemented as internal modules) so the structs should be placed into weldr/apischema.go

Try to avoid using interface{} as much as possible. I have only used it for returning the raw blueprint to the caller in the GetBlueprints... functions so that it isn't tightly coupled to the actual blueprint format. The composer-cli isn't manipulating the blueprint data, only passing it through, so it doesn't need to know what's actually there. Blueprints are the thing most likely to change, and they are not versioned like the API is so remaining flexible is important.

Unit tests for new functions

Functions, or parts of functions, that can be tested without an actual API server should be written in the weldr/unit_test.go file, or if there are no integration tests into a foo_test.go file where foo is the name of the file you put the function in.

Integration tests for library functions

Integration tests require an API server to be running, and should be placed into a foo_test.go file alongside the code they are testing. At the top of the file there should be a // +build integration before the package line, ans separated from it with a blank line. See blueprints_test.go for an example.

These tests should be independent from each other. If there is server setup that needs to be done it should be done in a way that works with a clean or a dirty server. Usually it will be run against a clean server, as part of the github action tests that are run when a PR is submitted. But don't depend on that.

If something needs to be setup you can do it in the test function itself, or in the integration_test.go file if necessary. But try to keep changes there to a minimum.

Adding new cli commands

Mention Cobra, point to docs. Point to simple example like 'status show', tell how to hook it up the rootCmd and how to write tests for it.

composer-cli uses the Cobra library to handle command parsing, options, and subcommands.

The ./cmd/composer-cli/ directory is structured so that top level commands go into their own directory and package. They import the root Cobra object from cmd/composer-cli/root and attach themselves to that command using root.AddRootCommand(). Then the subcommands attach themselves to this command. See cmd/composer-cli/blueprints/blueprints.go for an example of a subcommand.

Developer documentation should go into a doc.go file.

Commands are hooked into the root command parser by adding them to the include list in cmd/composer-cli/main.go with a preceeding _ so that the compiler won't complain about unused symbols.

Adding new subcommands

Subcommands should go into their own file, and be members of the command's package. The file should only contain the functions needed to implement the user interface for this command. Any helper functions or API related functions that could be useful to other developers should go into the weldr/ library.

Error handling for cli commands

Commands should use the Cobra RunE method which returns an error to the root command handler. If the command exits immediately on error it should return a root.ExecutionError with a string describing the error, and embedding the err if there was one. See cmd/composer-cli/blueprints/list.go for an example of this type of handling.

If there could be more than one error, eg. when processing a list of blueprints, the command function should print them to os.Stderr as ERROR: ...\n and return a blank root.ExecutionError(cmd, "")to the root handler. Seecmd/composer-cli/blueprints/push.go` for an example.

Unit tests for cli commands

The command line code under ./cmd/ is unit tested against the ./weldr/ library code. They use a mock http client that is setup, by each test, to return the expected JSON or TOML responses. See the tests in ./cmd/composer-cli/blueprints/ for examples of how to set this up.

Basically, you setup a mock http client that stores the query it receives and returns a canned response. This should not be too elaborate, otherwise you end up mocking every detail of the server. The goal is to make sure the cmdline command calls the right library function and that the response from the function is parsed correctly.

The weldr.MockClient function that is used to handle the requests adds a .Req element that can be examined in the tests. eg:

assert.Equal(t, "/api/v1/blueprints/list", mc.Req.URL.Path)

If the test returns an error the MockClient DoFunc needs to return the request so that the apiError function can pass it to the callback function. See common_test.TestRequestMethods404 for an example.

Running tests

Running Unit Tests

Unit tests can be run by running make test.

Running Integration Tests

Integration tests are run by a test binary. You can build the binary by running make integration and then running it on a system with osbuild-composer running. Pass it -test.v to output more verbose details about the tests being run.