Skip to content
This repository has been archived by the owner on Dec 16, 2024. It is now read-only.

Latest commit

 

History

History
276 lines (194 loc) · 13 KB

BUILDING.md

File metadata and controls

276 lines (194 loc) · 13 KB

Building Elasticsearch with Gradle

Elasticsearch is built using the Gradle open source build tools.

This document provides a general guidelines for using and working on the Elasticsearch build logic.

Build logic organisation

The Elasticsearch project contains 3 build-related projects that are included into the Elasticsearch build as a composite build.

build-conventions

This project contains build conventions that are applied to all Elasticsearch projects.

build-tools

This project contains all build logic that we publish for third party Elasticsearch plugin authors. We provide the following plugins:

  • elasticsearch.esplugin - A Gradle plugin for building an elasticsearch plugin.
  • elasticsearch.testclusters - A Gradle plugin for setting up es clusters for testing within a build.

This project is published as part of the Elasticsearch release and accessible by org.elasticsearch.gradle:build-tools:<versionNumber>. These build tools are also used by the elasticsearch-hadoop project maintained by elastic.

build-tools-internal

This project contains all Elasticsearch project specific build logic that is not meant to be shared with other internal or external projects.

Build guidelines

This is an intentionally small set of guidelines to build users and authors to ensure we keep the build consistent. We also publish Elasticsearch build logic as build-tools to be usable by thirdparty Elasticsearch plugin authors. This is also used by other elastic teams like elasticsearch-hadoop. Breaking changes should therefore be avoided and an appropriate deprecation cycle should be followed.

Stay up to date

The Elasticsearch build usually uses the latest Gradle GA release. We stay as close to the latest Gradle releases as possible. In certain cases an update is blocked by a breaking behaviour in Gradle. We're usually in contact with the Gradle team here or working on a fix in our build logic to resolve this.

The Elasticsearch build will fail if any deprecated Gradle API is used.

Follow Gradle best practices

Tony Robalik has compiled a good list of rules that aligns with ours when it comes to writing and maintaining Elasticsearch Gradle build logic at http://autonomousapps.com/blog/rules-for-gradle-plugin-authors.html. Our current build does not yet tick off all those rules everywhere but the ultimate goal is to follow these principles. The reasons for following those rules besides better readability or maintenance are also the goal to support newer Gradle features that we will benefit from in terms of performance and reliability. E.g. configuration-cache support, Project Isolation or predictive test selection

Make a change in the build

There are a few guidelines to follow that should make your life easier to make changes to the Elasticsearch build. Please add a member of the es-delivery team as a reviewer if you're making non-trivial changes to the build.

Adding or updating a dependency

We rely on Gradle dependency verification to mitigate the security risks and avoid integrating compromised dependencies.

This requires to have third party dependencies and their checksums listed in gradle/verification-metadata.xml.

For updated or newly added dependencies you need to add an entry to this verification file or update the existing one:

      <component group="asm" name="asm" version="3.1">
         <artifact name="asm-3.1.jar">
            <sha256 value="333ff5369043975b7e031b8b27206937441854738e038c1f47f98d072a20437a" origin="official site"/>
         </artifact>
      </component>

In case of updating a dependency, ensure to remove the unused entry of the outdated dependency manually from the verification-metadata.xml file.

You can also automate the generation of this entry by running your build using the --write-verification-metadata commandline option:

>./gradlew --write-verification-metadata sha256 precommit

The --write-verification-metadata Gradle option is generally able to resolve reachable configurations, but we use detached configurations for a certain set of plugins and tasks. Therefore, please ensure you run this option with a task that uses the changed dependencies. In most cases, precommit or check are good candidates.

We prefer sha256 checksums as md5 and sha1 are not considered safe anymore these days. The generated entry will have the origin attribute been set to Generated by Gradle.

A manual confirmation of the Gradle generated checksums is currently not mandatory. If you want to add a level of verification you can manually confirm the checksum (e.g. by looking it up on the website of the library) Please replace the content of the origin attribute by official site in that case.

Custom plugin and task implementations

Build logic that is used across multiple subprojects should be considered to be moved into a Gradle plugin with according Gradle task implementation. Elasticsearch specific build logic is located in the build-tools-internal subproject including integration tests.

Declaring tasks

The Elasticsearch build makes use of the task avoidance API to keep the configuration time of the build low.

When declaring tasks (in build scripts or custom plugins) this means that we want to register a task like:

tasks.register('someTask') { ... }

instead of eagerly creating the task:

task someTask { ... }

The major difference between these two syntaxes is, that the configuration block of a registered task will only be executed when the task is actually created due to the build requires that task to run. The configuration block of an eagerly created tasks will be executed immediately.

By actually doing less in the Gradle configuration time as only creating tasks that are requested as part of the build and by only running the configurations for those requested tasks, using the task avoidance api contributes a major part in keeping our build fast.

Registering test clusters

When using the Elasticsearch test cluster plugin we want to use (similar to the task avoidance API) a Gradle API to create domain objects lazy or only if required by the build. Therefore we register test cluster by using the following syntax:

def someClusterProvider = testClusters.register('someCluster') { ... }

This registers a potential testCluster named somecluster and provides a provider instance, but doesn't create it yet nor configures it. This makes the Gradle configuration phase more efficient by doing less.

To wire this registered cluster into a TestClusterAware task (e.g. RestIntegTest) you can resolve the actual cluster from the provider instance:

tasks.register('someClusterTest', RestIntegTestTask) {
    useCluster someClusterProvider
    nonInputProperties.systemProperty 'tests.leader_host', "${-> someClusterProvider.get().getAllHttpSocketURI().get(0)}"
}

Adding integration tests

Additional integration tests for a certain Elasticsearch modules that are specific to certain cluster configuration can be declared in a separate so called qa subproject of your module.

The benefit of a dedicated project for these tests are:

  • qa projects are dedicated two specific use-cases and easier to maintain
  • It keeps the specific test logic separated from the common test logic.
  • You can run those tests in parallel to other projects of the build.

Using test fixtures

Sometimes we want to share test fixtures to set up the code under test across multiple projects. There are basically two ways doing so.

Ideally we would use the build-in java-test-fixtures Gradle plugin. This plugin relies on having a separate sourceSet for the test fixtures code.

In the Elasticsearch codebase we have test fixtures and actual tests within the same sourceSet. Therefore we introduced the elasticsearch.internal-test-artifact plugin to provides another build artifact of your project based on the test sourceSet.

This artifact can be resolved by the consumer project as shown in the example below:

dependencies {
  //add the test fixtures of `:providing-project` to testImplementation configuration.
  testImplementation(testArtifact(project(":fixture-providing-project')))
}

This test artifact mechanism makes use of the concept of component capabilities similar to how the Gradle build-in java-test-fixtures plugin works.

testArtifact is a shortcut declared in the Elasticsearch build. Alternatively you can declare the dependency via

dependencies {
  testImplementation(project(":fixture-providing-project')) {
    requireCapabilities("org.elasticsearch.gradle:fixture-providing-project-test-artifacts")
  }
}

FAQ

How do I test a development version of a third party dependency?

To test an unreleased development version of a third party dependency you have several options.

How to use a Maven based third party dependency via mavenlocal?

  1. Clone the third party repository locally
  2. Run mvn install to install copy into your ~/.m2/repository folder.
  3. Add this to the root build script:
   allprojects {
     repositories {
         mavenLocal()
     }
   }
  1. Update the version in your dependency declaration accordingly (likely a snapshot version)
  2. Run the Gradle build as needed

How to use a Maven built based third party dependency with JitPack repository?

https://jitpack.io is an adhoc repository that supports building Maven projects transparently in the background when resolving unreleased snapshots from a GitHub repository. This approach also works as temporally solution and is compliant with our CI builds.

  1. Add the JitPack repository to the root build file:
   allprojects {
     repositories {
        maven { url "https://jitpack.io" }
     }
   }
  1. Add the dependency in the following format
dependencies {
  implementation 'com.github.User:Repo:Tag'
}

As version you could also use a certain short commit hash or main-SNAPSHOT. In addition to snapshot builds JitPack supports building Pull Requests. Simply use PR-SNAPSHOT as the version.

  1. Run the Gradle build as needed. Keep in mind the initial resolution might take a bit longer as this needs to be built by JitPack in the background before we can resolve the adhoc built dependency.

NOTE

You should only use that approach locally or on a developer branch for production dependencies as we do not want to ship unreleased libraries into our releases.

How to use a custom third party artifact?

For third party libraries that are not built with Maven (e.g. Ant) or provided as a plain jar artifact we can leverage a flat directory repository that resolves artifacts from a flat directory on your filesystem.

  1. Put the jar artifact with the format artifactName-version.jar into a directory named localRepo (you have to create this manually)
  2. Declare a flatDir repository in your root build.gradle file (That ensures all projects have the flatDir repository declared and also the projects consuming the project you tweaked can resolve that local dependency)
allprojects {
  repositories {
      flatDir {
          dirs 'localRepo'
      }
  }
}
  1. Update the dependency declaration of the artifact in question to match the custom build version. For a file named e.g. jmxri-1.2.1.jar the dependency definition would be x:jmxri:1.2.1 as it the group information is ignored on flatdir repositories you can replace the x in group name:
dependencies {
    implementation 'x:jmxri:1.2.1'
}
  1. Run the Gradle build as needed with --write-verification-metadata to ensure the Gradle dependency verification does not fail on your custom dependency.

NOTE

As Gradle prefers to use modules whose descriptor has been created from real meta-data rather than being generated, flat directory repositories cannot be used to override artifacts with real meta-data from other repositories declared in the build. For example, if Gradle finds only jmxri-1.2.1.jar in a flat directory repository, but jmxri-1.2.1.pom in another repository that supports meta-data, it will use the second repository to provide the module. Therefore, it is recommended to declare a version that is not resolvable from public repositories we use (e.g. Maven Central)