[llvm] [DTLTO][LLVM] Integrated Distributed ThinLTO (DTLTO) (PR #127749)
via llvm-commits
llvm-commits at lists.llvm.org
Thu Mar 6 11:39:58 PST 2025
================
@@ -0,0 +1,224 @@
+===================
+DTLTO
+===================
+.. contents::
+ :local:
+ :depth: 2
+
+.. toctree::
+ :maxdepth: 1
+
+Distributed ThinLTO (DTLTO)
+===========================
+
+Distributed ThinLTO (DTLTO) facilitates the distribution of backend ThinLTO
+compilations via external distribution systems such as Incredibuild.
+
+The existing method of distributing ThinLTO compilations via separate thin-link,
+backend compilation, and link steps often requires significant changes to the
+user's build process to adopt, as it requires using a build system which can
+handle the dynamic dependencies specified by the index files, such as Bazel.
+
+DTLTO eliminates this need by managing distribution internally within the LLD
+linker during the traditional link step. This allows DTLTO to be used with any
+build process that supports in-process ThinLTO.
+
+Limitations
+-----------
+
+The current implementation of DTLTO has the following limitations:
+
+- The ThinLTO cache is not supported.
+- Only ELF and COFF platforms are supported.
+- Archives with bitcode members are not supported.
+- Only a very limited set of LTO configurations are currently supported, e.g.,
+ support for basic block sections is not currently available.
+
+Overview of Operation
+---------------------
+
+For each ThinLTO backend compilation job, LLD:
+
+1. Generates the required summary index shard.
+2. Records a list of input and output files.
+3. Constructs a Clang command line to perform the ThinLTO backend compilation.
+
+This information is supplied, via a JSON file, to a distributor program that
+executes the backend compilations using a distribution system. Upon completion,
+LLD integrates the compiled native object files into the link process.
+
+The design keeps the details of distribution systems out of the LLVM source
+code.
+
+Distributors
+------------
+
+Distributors are programs responsible for:
+
+1. Consuming the JSON backend compilations job description file.
+2. Translating job descriptions into requests for the distribution system.
+3. Blocking execution until all backend compilations are complete.
+
+Distributors must return a non-zero exit code on failure. They can be
+implemented as binaries or in scripting languages, such as Python. An example
+script demonstrating basic local execution is available with the LLVM source
+code.
+
+How Distributors Are Invoked
+----------------------------
+
+Clang and LLD provide options to specify a distributor program for managing
+backend compilations. Distributor options and backend compilation options, can
+also be specified. Such options are transparently forwarded.
+
+The backend compilations are currently performed by invoking Clang. For further
+details, refer to:
+
+- Clang documentation: https://clang.llvm.org/docs/ThinLTO.html
+- LLD documentation: https://lld.llvm.org/DTLTO.html
+
+When invoked with a distributor, LLD generates a JSON file describing the
+backend compilation jobs and executes the distributor passing it this file. The
+JSON file provides the following information to the distributor:
+
+- The **command line** to execute the backend compilations.
+ - DTLTO constructs a Clang command line by translating some of the LTO
+ configuration state into Clang options and forwarding options specified
+ by the user.
+
+- **Link output path**.
+ - A string identifying the output to which this LTO invocation will
+ contribute. Distributors can use this to label build jobs for informational
+ purposes.
+
+- The list of **imports** required for each job.
+ - The per-job list of bitcode files from which importing will occur. This is
+ the same information that is emitted into import files for ThinLTO.
+
+- The **input files** required for each job.
+ - The per-job set of files required for backend compilation, such as bitcode
+ files, summary index files, and profile data.
+
+- The **output files** generated by each job.
+ - The per-job files generated by the backend compilations, such as compiled
+ object files and toolchain metrics.
+
+Temporary Files
+---------------
+
+During its operation, DTLTO generates temporary files. Temporary files are
+created in the same directory as the linker's output file and their filenames
+include the stem of the bitcode module, or the output file that the LTO
+invocation is contributing to, to aid the user in identifying them:
+
+- **JSON Job Description File**:
+ - Format: `dtlto.<PID>.dist-file.json`
+ - Example: `dtlto.77380.dist-file.json` (for output file `dtlto.elf`).
+
+- **Object Files From Backend Compilations**:
+ - Format: `<Module ID stem>.<Task>.<PID>.native.o`
+ - Example: `my.1.77380.native.o` (for bitcode module `my.o`).
+
+- **Summary Index Shard Files**:
+ - Format: `<Module ID stem>.<Task>.<PID>.native.o.thinlto.bc`
+ - Example: `my.1.77380.native.o.thinlto.bc` (for bitcode module `my.o`).
+
+Temporary files are removed, by default, after the backend compilations complete.
+
+JSON Schema
+-----------
+
+Below is an example of a JSON job file for backend compilation of the module
+`dtlto.o`:
+
+.. code-block:: json
+
+ {
+ "common": {
+ "linker_output": "dtlto.elf",
+ "linker_version": "LLD 20.0.0",
+ "args": [
+ "/usr/local/clang",
+ "-O3", "-fprofile-sample-use=my.profdata",
+ "-o", ["primary_output", 0],
+ "-c", "-x", "ir", ["primary_input", 0],
+ ["summary_index", "-fthinlto-index=", 0],
----------------
bd1976bris wrote:
> Is this a standard JSON format currently used by a distributor you targeting? If you have openAPI doc, it will be even better.
It's my own format, although it has been shown to have enough information to drive distribution via real-world distribution systems we have tested with Incredibuild and reclient+buildfarm (both of which are available to test for free).
I'll try creating an OpenAPI doc. My feeling is that the format is too simple to need something like this, though.
> I feel the format is a bit weird. It is an array containing different types, and for the sub-array that provide indexes, they are not same length and the item at the same offset are not always the same thing. Doesn't it make a decoder harder to write?
I don't think that decoding is particularly difficult (see, for example: [local.py](https://github.com/bd1976bris/llvm-project/blob/main/llvm/utils/dtlto/local.py).)
I'm not attached to the format of the JSON as long as it can do what we need. Here are the goals/constraints I came up with for a JSON interface:
- Be as minimal as possible.
- Don't use spaces in field names.
- This allows for programmatic access, such as with member access syntax. e.g., .summary_index.
- Allow the codegen tool to be replaced without a schema update.
- Don't depend on platform-or-shell-specific shlexing of strings
- Think about handling quotes, spaces, etc..
- Support filenames appearing inside arguments
- e.g., -fthinlto-index=summary.index.
- Decouple LLVM from distributors to the largest degree possible.
- This is to try to ensure that if the command line for the codegen tool changes, the distributor(s) shouldn't need updating.
- Explicitly provide the primary input file, summary index files, and primary output file.
- This is to help with supporting distribution systems such as FASTBuild, distcc, and Icecream, which lack sandboxing and might require the paths to be manipulated.
In the current format, the sub-arrays are used to represent references to per-job filenames so that a command line can be stored that doesn't contain filenames. This ensures that the contents of the command line remain transparent to the distributor, allowing another code-generation tool to replace Clang without updating the distributors. It also means that the distributor has access to the filenames (as they are stored separately) and can manipulate them, if required, without needing to change the command line.
https://github.com/llvm/llvm-project/pull/127749
More information about the llvm-commits
mailing list