A build tool for the rest of us
2021-04-13
Imagine a hypothetical company with a codebase ranging between 0 to 10 million lines of code. This codebase spans a number of different languages such as Go/Java/Node.js servers, JS webapps, and Python data pipelines. Is there an existing tool we can use to provide a unified build system? Maybe, but it's probably not worth the effort to implement. Can we do better?
When we talk about build tools it generally illicits one of three archetypes:
- "Execute commands in sequence" tools. e.g. Bash scripts, PHONY Makefiles.
- Language centric systems, e.g. NPM, Cargo, CMake, Meson.
- Polyglot build tools, e.g. Bazel, Buck, Please.
Given the above, our hypotethical company most likely wants the 3rd option, but these existing tools are fairly complex and have a number of shortcomings for companies that don't need to build billions of lines of code.
Language specific tool integration
Bazel was initially designed heavily around Google's core languages: C/++, Java, Python, and pre-npm Javascript. The thing all four of these technologies had in common (at the time the tool was written) was a lack of good language centric build tools. Coupled with a preference for vendoring libraries, this design get brought to a stretch when we look at how it handles languages such as Rust.
If we acknowledge that most modern development use a build tool that is already designed for the language in question, it would make sense for our hypothetical polyglot tool to natively support these language centric tools. In regards to Rust, instead of trying to extract a Cargo.toml
file into a generated BAZEL
file, our build tool could understand what a Rust project looks like, especially as it relates to how the tools already handle dependency management.
For example, consider the following project structure:
src/
internal/sdk/mylib/Cargo.toml (depends on crates.io libraries)
team_a/services/myexe/Cargo.toml (depends on mylib)
We should be able to just run ourtool run team_a/services/myexe
(in addition to build
and test
) and it should "just work". That's to say, it should fetch the crates.io libraries for mylib, build mylib, build myexe, and execute it. This is a dramatic departure from the current build tool migration process which generally involves extensive reworking of build scripts and dependency management.
Pants v2 is looking very promising in this space. It currently only supports Python, but already tries to integrate natively with what Python development is today. For example, you can automatically run any python projects without first having to create a build file. In addition, adding third party dependencies is as simple as adding a requirements.txt
file and having pants push that into the virtual environment (via some Pex magic) for the Python app we're trying to execute. I'm keeping an eye on this to see how it handles the next two languages they seem to be targeting (namely Rust and Go).
Editor / IDE / Language Server Protocol Integration
A follow on from the previous point is that most development smarts (intellisense, refactoring, etc.) all rely on the language centric tools. In the simplest case, having our Javascript project use a real package.json
file with a corresponding node_modules
/.yarn
directory (optimally created and managed by our tool), our editor will natively pick up types and libraries.
There are more complicated cases which are also generally solveable. Python, for example, needs a virtual environment or Pex file created, and code generation from tools such as GRPC needs to actually create the files in the same filesystem as the code (not in a hermetic directory) to ensure the editor picks it up (our tool should probably manage .gitignore
to ensure these files aren't commited).
Make is easy to integrate new tools
While our tool natively supporting language centric build systems works in some environments (e.g. Rust or Go), there are others where it works slightly less well. Javascript web development, for example, doesn't have a standardised build tool. Teams might use Webpack, Rollup, Parcel, Snowpack, Vite, Create-React-App, ESBuild, and so on. Trying to support each of these tools is a recipe for development failure.
The optimal way to approach this should probably two-fold:
- Support the most popular tool and/or the tool that integrates best with the speed and caching principles of the build tool. For example, Webpack for popularity and ESBuild for speed could cover the majority of teams right off the bat.
- Make it easy to integrate new tools.
For example, if we wanted to integrate create react app natively into our build tool, for projects already using npm it should be as simple as:
This should change our ourtool build ./a/path
to "just work".
It should be better for single language projects
Even if we use only a single language, our build tool should add enough on top of the native language tooling to justify it. There are a couple of examples that immediately jump out:
Make it easy to manage a single language monorepo. For example, being able to have a single command to run, build, test, format, lint, etc. no matter where in the repository I am. This also applies in a world where we have many smaller repos that span multiple languages (i.e. cross language command consistency).
Provide opinionated improvements on top of the native language. Pants v2 again here dramatically improves on the native python virtual environment setup, dependency management, and packaging (via Pex) that is makes sense even in repos that are just Python. For example, automatically managing sccache for Rust projects.
Provide dependency management for languages that don't have it. C/++, Zig are natural examples here where having a really simple dependency management system can provide immense benefits. For example:
# .build
// build.zig ... exe.; ...
Fast / Correct - Choose 80%
If you're someone who worked on Bazel I hope you're not reading this post, because at this point you'll be thinking "all of these features will break invariants and make other features such as remote caching more difficult/impossible". The issue is that I agree, but fundamentally, there is most likely an 80% version of these invariants that will enable most of these features while not requiring the same amount of specificity.
Additionally, there's probably a disable_smarts = true
flag that can be conditionally or globally set in situations where an organisation finds the smarts are causing significant enough performance issues. Then again, for these organisations, Bazel already exists.
There are a couple of things we definitely want to keep, however:
Our tool should handle fetching and installing the language tools. There is immense value in being able to rock up to a repository, run
ourtool build
, and have it download the right Rust compiler, Go compiler, Python version, and JVM.File input/output based caching, but we can improve this. A smart tool should be able to read our source files (e.g.
*.py
) or project files (e.g.package.json
) and figure out what the files depend on. This is even possible in C/++ using Clang (Zig uses this internally I believe to track header includes). The tradeoff then becomes, there is a lot of parsing that needs to happen across different languages and for larger repositories disabling this in exchange for explicit configuration might be needed.Concurrent execution and cross-language dependencies. While a tool like this by design locks us into the concurrency of the underlying tools (e.g. using Cargo to concurrently compile Rust projects), the layer of concurrency on top of that can provide benefits as well. For example, if my Go server uses embed to include an esbuild built JavaScript bundle, and a SASS built CSS collection, esbuild and sass should be able to be run in parallel and our Go embed rebuilt (if the JS or scss files changed) when we run
ourtool run my/go/app
.
Conclusion
What is missing here and why does this tool not exist? There are a lot of gaps or issues that should immediately jump out to different people. Some of these issues might disqualify a tool like this from existing entirely, but I suspect that most of these issues can actually be worked around as long as we're willing to accept that we just might not be able to perfectly guarantee certain things (such as pure hermeticism), but is that not still an improvement for teams where Bazel is too complex?