|
You are here |
janik6n.net | ||
| | | | |
werat.dev
|
|
| | | | | Benchmarks are often underestimated and don't get the same attention as tests. However "performance is a feature" and when something is not tested it might as well be just broken. If the performance is not measured/tracked regressions are inevitable. Modern tooling makes it really easy to write benchmarks. Some languages have built-in support, for example, Rust comes with cargo bench (docs) and Go has go test -bench (docs). For C++ there is google/benchmark - not as streamlined as having it built into the language infrastructure, but still definitely worth the effort. | |
| | | | |
navendu.me
|
|
| | | | | An exploration of continuous delivery workflows for building and managing APIs at scale. | |
| | | | |
kinoshita.eti.br
|
|
| | | | | Normally when you add test coverage to a GitHub repository, the reporting part boils down to a simple call to some API that will post to a third-party external service like Coveralls or Codecov. Many are already on the GitHub Actions Market Place and a few lines of YAML are enough. A Codecov coverage report We had to report the test coverage of a private GitHub repository of a project I am working on at the moment, that could not have integration with other services besides GitHub. | |
| | | | |
blog.risingstack.com
|
|
| | | Common Kubernetes interview questions and answers about the architecture, deployment, and management of k8s containers. | ||