diff --git a/omnibus/config/patches/datadog-agent-integrations-py3/remove-maxfile-maxpath-psutil.patch b/omnibus/config/patches/datadog-agent-integrations-py3/remove-maxfile-maxpath-psutil.patch deleted file mode 100644 index c92eb2f7f2859..0000000000000 --- a/omnibus/config/patches/datadog-agent-integrations-py3/remove-maxfile-maxpath-psutil.patch +++ /dev/null @@ -1,30 +0,0 @@ -Partially reverts https://github.com/giampaolo/psutil/pull/1863 to remove the maxpath / maxfile fetch -diff --git a/psutil/__init__.py b/psutil/__init__.py -index 1a113bc3..ce962a61 100644 ---- a/psutil/__init__.py -+++ b/psutil/__init__.py -@@ -2012,23 +2012,7 @@ def disk_partitions(all=False): - If *all* parameter is False return physical devices only and ignore - all others. - """ -- def pathconf(path, name): -- try: -- return os.pathconf(path, name) -- except (OSError, AttributeError): -- pass -- -- ret = _psplatform.disk_partitions(all) -- if POSIX: -- new = [] -- for item in ret: -- nt = item._replace( -- maxfile=pathconf(item.mountpoint, 'PC_NAME_MAX'), -- maxpath=pathconf(item.mountpoint, 'PC_PATH_MAX')) -- new.append(nt) -- return new -- else: -- return ret -+ return _psplatform.disk_partitions(all) - - - def disk_io_counters(perdisk=False, nowrap=True): \ No newline at end of file diff --git a/omnibus/config/patches/datadog-agent-integrations-py2/create-regex-at-runtime.patch b/omnibus/config/patches/datadog-agent-integrations/create-regex-at-runtime.patch similarity index 100% rename from omnibus/config/patches/datadog-agent-integrations-py2/create-regex-at-runtime.patch rename to omnibus/config/patches/datadog-agent-integrations/create-regex-at-runtime.patch diff --git a/omnibus/config/patches/datadog-agent-integrations-py2/remove-maxfile-maxpath-psutil.patch b/omnibus/config/patches/datadog-agent-integrations/remove-maxfile-maxpath-psutil.patch similarity index 100% rename from omnibus/config/patches/datadog-agent-integrations-py2/remove-maxfile-maxpath-psutil.patch rename to omnibus/config/patches/datadog-agent-integrations/remove-maxfile-maxpath-psutil.patch diff --git a/omnibus/config/projects/agent.rb b/omnibus/config/projects/agent.rb index 2384e20f0cd00..6b77898b6d3f2 100644 --- a/omnibus/config/projects/agent.rb +++ b/omnibus/config/projects/agent.rb @@ -182,13 +182,13 @@ if with_python_runtime? "2" dependency 'pylint2' - dependency 'datadog-agent-integrations-py2' end -if with_python_runtime? "3" - dependency 'datadog-agent-integrations-py3' +if with_python_runtime? "3" or with_python_runtime? "2" + dependency 'datadog-agent-integrations' end + if linux_target? dependency 'datadog-security-agent-policies' end diff --git a/omnibus/config/software/datadog-agent-integrations-py2.rb b/omnibus/config/software/datadog-agent-integrations-py2.rb deleted file mode 100644 index ff62baa97b2cf..0000000000000 --- a/omnibus/config/software/datadog-agent-integrations-py2.rb +++ /dev/null @@ -1,460 +0,0 @@ -# Unless explicitly stated otherwise all files in this repository are licensed -# under the Apache License Version 2.0. -# This product includes software developed at Datadog (https:#www.datadoghq.com/). -# Copyright 2016-present Datadog, Inc. - -require './lib/ostools.rb' -require 'json' - -name 'datadog-agent-integrations-py2' - -license "BSD-3-Clause" -license_file "./LICENSE" - -dependency 'datadog-agent' -dependency 'datadog-agent-integrations-py2-dependencies' - -relative_path 'integrations-core' -whitelist_file "embedded/lib/python2.7/site-packages/.libsaerospike" -whitelist_file "embedded/lib/python2.7/site-packages/psycopg2" -whitelist_file "embedded/lib/python2.7/site-packages/wrapt" -whitelist_file "embedded/lib/python2.7/site-packages/pymqi" - -source git: 'https://github.com/DataDog/integrations-core.git' - -integrations_core_version = ENV['INTEGRATIONS_CORE_VERSION'] -if integrations_core_version.nil? || integrations_core_version.empty? - integrations_core_version = 'master' -end -default_version integrations_core_version - -# folder names containing integrations from -core that won't be packaged with the Agent -excluded_folders = [ - 'datadog_checks_base', # namespacing package for wheels (NOT AN INTEGRATION) - 'datadog_checks_dev', # Development package, (NOT AN INTEGRATION) - 'datadog_checks_tests_helper', # Testing and Development package, (NOT AN INTEGRATION) - 'docker_daemon', # Agent v5 only -] - -# package names of dependencies that won't be added to the Agent Python environment -excluded_packages = Array.new - - -if suse_target? - # Temporarily exclude Aerospike until builder supports new dependency - excluded_packages.push(/^aerospike==/) - excluded_folders.push('aerospike') -end - -if osx_target? - # exclude aerospike, new version 3.10 is not supported on MacOS yet - excluded_folders.push('aerospike') - - # Temporarily exclude Aerospike until builder supports new dependency - excluded_packages.push(/^aerospike==/) - excluded_folders.push('aerospike') -end - -if arm_target? - # Temporarily exclude Aerospike until builder supports new dependency - excluded_folders.push('aerospike') - excluded_packages.push(/^aerospike==/) - - # This doesn't build on ARM - excluded_folders.push('ibm_mq') - excluded_packages.push(/^pymqi==/) -end - -if arm_target? || !_64_bit? - excluded_packages.push(/^orjson==/) -end - -if linux_target? - excluded_packages.push(/^pyyaml==/) - excluded_packages.push(/^kubernetes==/) -end - -final_constraints_file = 'final_constraints-py2.txt' -agent_requirements_file = 'agent_requirements-py2.txt' -filtered_agent_requirements_in = 'agent_requirements-py2.in' -agent_requirements_in = 'agent_requirements.in' - -build do - # The dir for confs - if osx_target? - conf_dir = "#{install_dir}/etc/conf.d" - else - conf_dir = "#{install_dir}/etc/datadog-agent/conf.d" - end - mkdir conf_dir - - # aliases for pip - if windows_target? - pip = "#{windows_safe_path(python_2_embedded)}\\Scripts\\pip.exe" - python = "#{windows_safe_path(python_2_embedded)}\\python.exe" - else - pip = "#{install_dir}/embedded/bin/pip2" - python = "#{install_dir}/embedded/bin/python2" - end - - # If a python_mirror was set, it's passed through a pip config file so that we're not leaking the API key in the CI Output - # Else the pip config file so pip will act casually - pip_config_file = ENV['PIP_CONFIG_FILE'] - pre_build_env = { - "PIP_CONFIG_FILE" => "#{pip_config_file}" - } - - # Install the checks along with their dependencies - if windows_target? - wheel_build_dir = "#{windows_safe_path(project_dir)}\\.wheels" - build_deps_dir = "#{windows_safe_path(project_dir)}\\.build_deps" - else - wheel_build_dir = "#{project_dir}/.wheels" - build_deps_dir = "#{project_dir}/.build_deps" - end - - # - # Prepare the build env, these dependencies are only needed to build and - # install the core integrations. - # - command "#{pip} download --dest #{build_deps_dir} hatchling==0.25.1", :env => pre_build_env - command "#{pip} download --dest #{build_deps_dir} setuptools==40.9.0", :env => pre_build_env # Version from ./setuptools2.rb - command "#{pip} install wheel==0.37.1", :env => pre_build_env # Pin to the last version that supports Python 2 - command "#{pip} install setuptools-scm==5.0.2", :env => pre_build_env # Pin to the last version that supports Python 2 - command "#{pip} install pip-tools==5.4.0", :env => pre_build_env - uninstall_buildtime_deps = ['rtloader', 'click', 'first', 'pip-tools'] - nix_build_env = { - "PIP_FIND_LINKS" => "#{build_deps_dir}", - "PIP_CONFIG_FILE" => "#{pip_config_file}", - "CFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", - "CXXFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", - "LDFLAGS" => "-L#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", - "LD_RUN_PATH" => "#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", - "PATH" => "#{install_dir}/embedded/bin:#{ENV['PATH']}", - } - win_build_env = { - "PIP_FIND_LINKS" => "#{build_deps_dir}", - "PIP_CONFIG_FILE" => "#{pip_config_file}", - } - # Some libraries (looking at you, aerospike-client-python) need EXT_CFLAGS instead of CFLAGS. - nix_specific_build_env = { - "aerospike" => nix_build_env.merge({"EXT_CFLAGS" => nix_build_env["CFLAGS"] + " -std=gnu99"}), - # Always build pyodbc from source to link to the embedded version of libodbc - "pyodbc" => nix_build_env.merge({"PIP_NO_BINARY" => "pyodbc"}), - } - win_specific_build_env = {} - - - # On Linux & Windows, specify the C99 standard explicitly to avoid issues while building some - # wheels (eg. ddtrace). - # Not explicitly setting that option has caused us problems in the past on SUSE, where the ddtrace - # wheel has to be manually built, as the C code in ddtrace doesn't follow the C89 standard (the default value of std). - # Note: We don't set this on MacOS, as on MacOS we need to build a bunch of packages & C extensions that - # don't have precompiled MacOS wheels. When building C extensions, the CFLAGS variable is added to - # the command-line parameters, even when compiling C++ code, where -std=c99 is invalid. - # See: https://github.com/python/cpython/blob/v2.7.18/Lib/distutils/sysconfig.py#L222 - if linux_target? || windows_target? - nix_build_env["CFLAGS"] += " -std=c99" - end - - # - # Prepare the requirements file containing ALL the dependencies needed by - # any integration. This will provide the "static Python environment" of the Agent. - # We don't use the .in file provided by the base check directly because we - # want to filter out things before installing. - # - if windows_target? - static_reqs_in_file = "#{windows_safe_path(project_dir)}\\datadog_checks_base\\datadog_checks\\base\\data\\#{agent_requirements_in}" - static_reqs_out_folder = "#{windows_safe_path(project_dir)}\\" - static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in - compiled_reqs_file_path = "#{windows_safe_path(install_dir)}\\#{agent_requirements_file}" - else - static_reqs_in_file = "#{project_dir}/datadog_checks_base/datadog_checks/base/data/#{agent_requirements_in}" - static_reqs_out_folder = "#{project_dir}/" - static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in - compiled_reqs_file_path = "#{install_dir}/#{agent_requirements_file}" - end - - specific_build_env = windows_target? ? win_specific_build_env : nix_specific_build_env - build_env = windows_target? ? win_build_env : nix_build_env - cwd = windows_target? ? "#{windows_safe_path(project_dir)}\\datadog_checks_base" : "#{project_dir}/datadog_checks_base" - - # Creating a hash containing the requirements and requirements file path associated to every lib - requirements_custom = Hash.new() - specific_build_env.each do |lib, env| - lib_compiled_req_file_path = (windows_target? ? "#{windows_safe_path(install_dir)}\\" : "#{install_dir}/") + "agent_#{lib}_requirements-py2.txt" - requirements_custom[lib] = { - "req_lines" => Array.new, - "req_file_path" => static_reqs_out_folder + lib + "-py2.in", - "compiled_req_file_path" => lib_compiled_req_file_path, - } - end - - # Remove any excluded requirements from the static-environment req file - requirements = Array.new - - block "Create filtered requirements" do - File.open("#{static_reqs_in_file}", 'r+').readlines().each do |line| - next if excluded_packages.any? { |package_regex| line.match(package_regex) } - - if line.start_with?('psycopg[binary]') && !windows_target? - line.sub! 'psycopg[binary]', 'psycopg[c]' - end - # Keeping the custom env requirements lines apart to install them with a specific env - requirements_custom.each do |lib, lib_req| - if Regexp.new('^' + lib + '==').freeze.match line - lib_req["req_lines"].push(line) - end - end - # In any case we add the lib to the requirements files to avoid inconsistency in the installed versions - # For example if aerospike has dependency A>1.2.3 and a package in the big requirements file has A<1.2.3, the install process would succeed but the integration wouldn't work. - requirements.push(line) - end - - # Adding pympler for memory debug purposes - requirements.push("pympler==0.7") - - end - - # Render the filtered requirements file - erb source: "static_requirements.txt.erb", - dest: "#{static_reqs_out_file}", - mode: 0640, - vars: { requirements: requirements } - - # Render the filtered libraries that are to be built with different env var - requirements_custom.each do |lib, lib_req| - erb source: "static_requirements.txt.erb", - dest: "#{lib_req["req_file_path"]}", - mode: 0640, - vars: { requirements: lib_req["req_lines"] } - end - - # Increasing pip max retries (default: 5 times) and pip timeout (default 15 seconds) to avoid blocking network errors - pip_max_retries = 20 - pip_timeout = 20 - - # Use pip-compile to create the final requirements file. Notice when we invoke `pip` through `python -m pip <...>`, - # there's no need to refer to `pip`, the interpreter will pick the right script. - command "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => build_env, :cwd => cwd - command "#{python} -m pip install datadog_checks_base --no-deps --no-index --find-links=#{wheel_build_dir}" - command "#{python} -m piptools compile --generate-hashes --output-file #{compiled_reqs_file_path} #{static_reqs_out_file} " \ - "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => build_env - # Pip-compiling seperately each lib that needs a custom build installation - specific_build_env.each do |lib, env| - command "#{python} -m piptools compile --generate-hashes --output-file #{requirements_custom[lib]["compiled_req_file_path"]} #{requirements_custom[lib]["req_file_path"]} " \ - "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => env - end - - # - # Install static-environment requirements that the Agent and all checks will use - # - - # First we install the dependencies that need specific flags - specific_build_env.each do |lib, env| - command "#{python} -m pip install --no-deps --require-hashes -r #{requirements_custom[lib]["compiled_req_file_path"]}", :env => env - # Remove the file after use so it is not shipped - delete "#{requirements_custom[lib]["compiled_req_file_path"]}" - end - - # Then we install the rest (already installed libraries will be ignored) with the main flags - command "#{python} -m pip install --no-deps --require-hashes -r #{compiled_reqs_file_path}", :env => build_env - # Remove the file after use so it is not shipped - delete "#{compiled_reqs_file_path}" - - # - # Install Core integrations - # - - # Create a constraint file after installing all the core dependencies and before any integration - # This is then used as a constraint file by the integration command to avoid messing with the agent's python environment - command "#{pip} freeze > #{install_dir}/#{final_constraints_file}" - - if windows_target? - cached_wheels_dir = "#{windows_safe_path(wheel_build_dir)}\\.cached" - else - cached_wheels_dir = "#{wheel_build_dir}/.cached" - end - - checks_to_install = Array.new - - block "Collect integrations to install" do - # Go through every integration package in `integrations-core`, build and install - Dir.glob("#{project_dir}/*").each do |check_dir| - check = check_dir.split('/').last - - # do not install excluded integrations - next if !File.directory?("#{check_dir}") || excluded_folders.include?(check) - - # If there is no manifest file, then we should assume the folder does not - # contain a working check and move onto the next - manifest_file_path = "#{check_dir}/manifest.json" - - # If there is no manifest file, then we should assume the folder does not - # contain a working check and move onto the next - File.exist?(manifest_file_path) || next - - manifest = JSON.parse(File.read(manifest_file_path)) - if manifest.key?("supported_os") - manifest["supported_os"].include?(os) || next - else - if os == "mac_os" - tag = "Supported OS::macOS" - else - tag = "Supported OS::#{os.capitalize}" - end - - manifest["tile"]["classifier_tags"].include?(tag) || next - end - - File.file?("#{check_dir}/setup.py") || File.file?("#{check_dir}/pyproject.toml") || next - # Check if it supports Python 2. - support = `inv agent.check-supports-python-version #{check_dir} 2` - if support == "False" - log.info(log_key) { "Skipping '#{check}' since it does not support Python 2." } - next - end - - checks_to_install.push(check) - end - end - - installed_list = Array.new - cache_bucket = ENV.fetch('INTEGRATION_WHEELS_CACHE_BUCKET', '') - block "Install integrations" do - tasks_dir_in = windows_safe_path(Dir.pwd) - cache_branch = (shellout! "inv release.get-release-json-value base_branch", cwd: File.expand_path('..', tasks_dir_in)).stdout.strip - # On windows, `aws` actually executes Ruby's AWS SDK, but we want the Python one - awscli = if windows_target? then '"c:\Program files\python39\scripts\aws"' else 'aws' end - if cache_bucket != '' - mkdir cached_wheels_dir - shellout! "inv -e agent.get-integrations-from-cache " \ - "--python 2 --bucket #{cache_bucket} " \ - "--branch #{cache_branch || 'main'} " \ - "--integrations-dir #{windows_safe_path(project_dir)} " \ - "--target-dir #{cached_wheels_dir} " \ - "--integrations #{checks_to_install.join(',')} " \ - "--awscli #{awscli}", - :cwd => tasks_dir_in - - # install all wheels from cache in one pip invocation to speed things up - if windows_target? - shellout! "#{python} -m pip install --no-deps --no-index " \ - "--find-links #{windows_safe_path(cached_wheels_dir)} -r #{windows_safe_path(cached_wheels_dir)}\\found.txt" - else - shellout! "#{pip} install --no-deps --no-index " \ - " --find-links #{cached_wheels_dir} -r #{cached_wheels_dir}/found.txt" - end - end - - # get list of integration wheels already installed from cache - if cache_bucket != '' - if windows_target? - installed_out = (shellout! "#{python} -m pip list --format json").stdout - else - installed_out = (shellout! "#{pip} list --format json").stdout - end - if $?.exitstatus == 0 - installed = JSON.parse(installed_out) - installed.each do |package| - package.each do |key, value| - if key == "name" && value.start_with?("datadog-") - installed_list.push(value["datadog-".length..-1]) - end - end - end - else - raise "Failed to list pip installed packages" - end - end - - checks_to_install.each do |check| - check_dir = File.join(project_dir, check) - check_conf_dir = "#{conf_dir}/#{check}.d" - - # For each conf file, if it already exists, that means the `datadog-agent` software def - # wrote it first. In that case, since the agent's confs take precedence, skip the conf - conf_files = ["conf.yaml.example", "conf.yaml.default", "metrics.yaml", "auto_conf.yaml"] - - conf_files.each do |filename| - src = windows_safe_path(check_dir,"datadog_checks", check, "data", filename) - dest = check_conf_dir - if File.exist?(src) and !File.exist?(windows_safe_path(dest, filename)) - FileUtils.mkdir_p(dest) - FileUtils.cp_r(src, dest) - end - end - - # Copy SNMP profiles - profile_folders = ['profiles', 'default_profiles'] - profile_folders.each do |profile_folder| - folder_path = "#{check_dir}/datadog_checks/#{check}/data/#{profile_folder}" - if File.exist? folder_path - FileUtils.cp_r folder_path, "#{check_conf_dir}/" - end - end - - # pip < 21.2 replace underscores by dashes in package names per https://pip.pypa.io/en/stable/news/#v21-2 - # whether or not this might switch back in the future is not guaranteed, so we check for both name - # with dashes and underscores - if installed_list.include?(check) || installed_list.include?(check.gsub('_', '-')) - next - end - - if windows_target? - shellout! "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => win_build_env, :cwd => "#{windows_safe_path(project_dir)}\\#{check}" - shellout! "#{python} -m pip install datadog-#{check} --no-deps --no-index --find-links=#{wheel_build_dir}" - else - shellout! "#{pip} wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => nix_build_env, :cwd => "#{project_dir}/#{check}" - shellout! "#{pip} install datadog-#{check} --no-deps --no-index --find-links=#{wheel_build_dir}" - end - if cache_bucket != '' && ENV.fetch('INTEGRATION_WHEELS_SKIP_CACHE_UPLOAD', '') == '' && cache_branch != nil - shellout! "inv -e agent.upload-integration-to-cache " \ - "--python 2 --bucket #{cache_bucket} " \ - "--branch #{cache_branch} " \ - "--integrations-dir #{windows_safe_path(project_dir)} " \ - "--build-dir #{wheel_build_dir} " \ - "--integration #{check} " \ - "--awscli #{awscli}", - :cwd => tasks_dir_in - end - end - end - - # From now on we don't need piptools anymore, uninstall its deps so we don't include them in the final artifact - uninstall_buildtime_deps.each do |dep| - if windows_target? - command "#{python} -m pip uninstall -y #{dep}" - else - command "#{pip} uninstall -y #{dep}" - end - end - - # Patch applies to only one file: set it explicitly as a target, no need for -p - if windows_target? - patch :source => "create-regex-at-runtime.patch", :target => "#{python_2_embedded}/Lib/site-packages/yaml/reader.py" - patch :source => "remove-maxfile-maxpath-psutil.patch", :target => "#{python_2_embedded}/Lib/site-packages/psutil/__init__.py" - else - patch :source => "create-regex-at-runtime.patch", :target => "#{install_dir}/embedded/lib/python2.7/site-packages/yaml/reader.py" - patch :source => "remove-maxfile-maxpath-psutil.patch", :target => "#{install_dir}/embedded/lib/python2.7/site-packages/psutil/__init__.py" - end - - # Run pip check to make sure the agent's python environment is clean, all the dependencies are compatible - if windows_target? - command "#{python} -m pip check" - else - command "#{pip} check" - end - - # Removing tests that don't need to be shipped in the embedded folder - if windows_target? - delete "#{python_2_embedded}/Lib/site-packages/Cryptodome/SelfTest/" - else - delete "#{install_dir}/embedded/lib/python2.7/site-packages/Cryptodome/SelfTest/" - end - - # Ship `requirements-agent-release.txt` file containing the versions of every check shipped with the agent - # Used by the `datadog-agent integration` command to prevent downgrading a check to a version - # older than the one shipped in the agent - copy "#{project_dir}/requirements-agent-release.txt", "#{install_dir}/" -end diff --git a/omnibus/config/software/datadog-agent-integrations-py3.rb b/omnibus/config/software/datadog-agent-integrations-py3.rb deleted file mode 100644 index ffb437079370d..0000000000000 --- a/omnibus/config/software/datadog-agent-integrations-py3.rb +++ /dev/null @@ -1,490 +0,0 @@ -# Unless explicitly stated otherwise all files in this repository are licensed -# under the Apache License Version 2.0. -# This product includes software developed at Datadog (https:#www.datadoghq.com/). -# Copyright 2016-present Datadog, Inc. - -require './lib/ostools.rb' -require 'json' - -name 'datadog-agent-integrations-py3' - -license "BSD-3-Clause" -license_file "./LICENSE" - -dependency 'datadog-agent' -dependency 'datadog-agent-integrations-py3-dependencies' - -relative_path 'integrations-core' -whitelist_file "embedded/lib/python3.11/site-packages/.libsaerospike" -whitelist_file "embedded/lib/python3.11/site-packages/aerospike.libs" -whitelist_file "embedded/lib/python3.11/site-packages/psycopg2" -whitelist_file "embedded/lib/python3.11/site-packages/pymqi" - -source git: 'https://github.com/DataDog/integrations-core.git' - -gcc_version = ENV['GCC_VERSION'] -if gcc_version.nil? || gcc_version.empty? - gcc_version = '10.4.0' -end - -integrations_core_version = ENV['INTEGRATIONS_CORE_VERSION'] -if integrations_core_version.nil? || integrations_core_version.empty? - integrations_core_version = 'master' -end -default_version integrations_core_version - -# folder names containing integrations from -core that won't be packaged with the Agent -excluded_folders = [ - 'datadog_checks_base', # namespacing package for wheels (NOT AN INTEGRATION) - 'datadog_checks_dev', # Development package, (NOT AN INTEGRATION) - 'datadog_checks_tests_helper', # Testing and Development package, (NOT AN INTEGRATION) - 'docker_daemon', # Agent v5 only -] - -# package names of dependencies that won't be added to the Agent Python environment -excluded_packages = Array.new - -# We build these manually -excluded_packages.push(/^snowflake-connector-python==/) -excluded_packages.push(/^confluent-kafka==/) - -if suse_target? - # Temporarily exclude Aerospike until builder supports new dependency - excluded_packages.push(/^aerospike==/) - excluded_folders.push('aerospike') -end - -if osx_target? - # Temporarily exclude Aerospike until builder supports new dependency - excluded_packages.push(/^aerospike==/) - excluded_folders.push('aerospike') - excluded_folders.push('teradata') -end - -if arm_target? - # This doesn't build on ARM - excluded_folders.push('ibm_ace') - excluded_folders.push('ibm_mq') - excluded_packages.push(/^pymqi==/) -end - -# We explicitly check for redhat builder, not target -# Our centos/redhat builder uses glibc 2.12 while pydantic -# requires glibc 2.17 -if redhat? && !arm_target? - excluded_packages.push(/^pydantic-core==/) -end - -# _64_bit checks the kernel arch. On windows, the builder is 64 bit -# even when doing a 32 bit build. Do a specific check for the 32 bit -# build -if arm_target? || !_64_bit? || (windows_target? && windows_arch_i386?) - excluded_packages.push(/^orjson==/) -end - -if linux_target? - excluded_packages.push(/^oracledb==/) -end - -final_constraints_file = 'final_constraints-py3.txt' -agent_requirements_file = 'agent_requirements-py3.txt' -filtered_agent_requirements_in = 'agent_requirements-py3.in' -agent_requirements_in = 'agent_requirements.in' - -build do - # The dir for confs - if osx_target? - conf_dir = "#{install_dir}/etc/conf.d" - else - conf_dir = "#{install_dir}/etc/datadog-agent/conf.d" - end - mkdir conf_dir - - # aliases for pip - if windows_target? - python = "#{windows_safe_path(python_3_embedded)}\\python.exe" - else - python = "#{install_dir}/embedded/bin/python3" - end - - # If a python_mirror is set, it is set in a pip config file so that we do not leak the token in the CI output - pip_config_file = ENV['PIP_CONFIG_FILE'] - pre_build_env = { - "PIP_CONFIG_FILE" => "#{pip_config_file}" - } - - # Install the checks along with their dependencies - if windows_target? - wheel_build_dir = "#{windows_safe_path(project_dir)}\\.wheels" - build_deps_dir = "#{windows_safe_path(project_dir)}\\.build_deps" - else - wheel_build_dir = "#{project_dir}/.wheels" - build_deps_dir = "#{project_dir}/.build_deps" - end - - # - # Prepare the build env, these dependencies are only needed to build and - # install the core integrations. - # - command "#{python} -m pip download --dest #{build_deps_dir} hatchling==0.25.1", :env => pre_build_env - command "#{python} -m pip download --dest #{build_deps_dir} setuptools==66.1.1", :env => pre_build_env # Version from ./setuptools3.rb - command "#{python} -m pip install wheel==0.38.4", :env => pre_build_env - command "#{python} -m pip install pip-tools==7.3.0", :env => pre_build_env - uninstall_buildtime_deps = ['rtloader', 'click', 'first', 'pip-tools'] - nix_build_env = { - "PIP_FIND_LINKS" => "#{build_deps_dir}", - "PIP_CONFIG_FILE" => "#{pip_config_file}", - # Specify C99 standard explicitly to avoid issues while building some - # wheels (eg. ddtrace) - "CFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", - "CXXFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", - "LDFLAGS" => "-L#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", - "LD_RUN_PATH" => "#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", - "PATH" => "#{install_dir}/embedded/bin:#{ENV['PATH']}", - } - - win_build_env = { - "PIP_FIND_LINKS" => "#{build_deps_dir}", - "PIP_CONFIG_FILE" => "#{pip_config_file}", - } - - # Some libraries (looking at you, aerospike-client-python) need EXT_CFLAGS instead of CFLAGS. - nix_specific_build_env = { - "aerospike" => nix_build_env.merge({"EXT_CFLAGS" => nix_build_env["CFLAGS"] + " -std=gnu99"}), - # Always build pyodbc from source to link to the embedded version of libodbc - "pyodbc" => nix_build_env.merge({"PIP_NO_BINARY" => "pyodbc"}), - } - win_specific_build_env = {} - - # On Linux & Windows, specify the C99 standard explicitly to avoid issues while building some - # wheels (eg. ddtrace). - # Not explicitly setting that option has caused us problems in the past on SUSE, where the ddtrace - # wheel has to be manually built, as the C code in ddtrace doesn't follow the C89 standard (the default value of std). - # Note: We don't set this on MacOS, as on MacOS we need to build a bunch of packages & C extensions that - # don't have precompiled MacOS wheels. When building C extensions, the CFLAGS variable is added to - # the command-line parameters, even when compiling C++ code, where -std=c99 is invalid. - # See: https://github.com/python/cpython/blob/v3.8.8/Lib/distutils/sysconfig.py#L227 - if linux_target? || windows_target? - nix_build_env["CFLAGS"] += " -std=c99" - end - - # We only have gcc 10.4.0 on linux for now - if linux_target? - nix_build_env["CC"] = "/opt/gcc-#{gcc_version}/bin/gcc" - nix_build_env["CXX"] = "/opt/gcc-#{gcc_version}/bin/g++" - end - - # We need to explicitly specify RUSTFLAGS for libssl and libcrypto - # See https://github.com/pyca/cryptography/issues/8614#issuecomment-1489366475 - if redhat_target? && !arm_target? - nix_specific_build_env["cryptography"] = nix_build_env.merge( - { - "RUSTFLAGS" => "-C link-arg=-Wl,-rpath,#{install_dir}/embedded/lib", - "OPENSSL_DIR" => "#{install_dir}/embedded/", - # We have a manually installed dependency (snowflake connector) that already installed cryptography (but without the flags) - # We force reinstall it from source to be sure we use the flag - "PIP_NO_CACHE_DIR" => "off", - "PIP_FORCE_REINSTALL" => "1", - } - ) - end - - # - # Prepare the requirements file containing ALL the dependencies needed by - # any integration. This will provide the "static Python environment" of the Agent. - # We don't use the .in file provided by the base check directly because we - # want to filter out things before installing. - # - if windows_target? - static_reqs_in_file = "#{windows_safe_path(project_dir)}\\datadog_checks_base\\datadog_checks\\base\\data\\#{agent_requirements_in}" - static_reqs_out_folder = "#{windows_safe_path(project_dir)}\\" - static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in - compiled_reqs_file_path = "#{windows_safe_path(install_dir)}\\#{agent_requirements_file}" - else - static_reqs_in_file = "#{project_dir}/datadog_checks_base/datadog_checks/base/data/#{agent_requirements_in}" - static_reqs_out_folder = "#{project_dir}/" - static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in - compiled_reqs_file_path = "#{install_dir}/#{agent_requirements_file}" - end - - specific_build_env = windows_target? ? win_specific_build_env : nix_specific_build_env - build_env = windows_target? ? win_build_env : nix_build_env - cwd_base = windows_target? ? "#{windows_safe_path(project_dir)}\\datadog_checks_base" : "#{project_dir}/datadog_checks_base" - cwd_downloader = windows_target? ? "#{windows_safe_path(project_dir)}\\datadog_checks_downloader" : "#{project_dir}/datadog_checks_downloader" - - # Creating a hash containing the requirements and requirements file path associated to every lib - requirements_custom = Hash.new() - specific_build_env.each do |lib, env| - lib_compiled_req_file_path = (windows_target? ? "#{windows_safe_path(install_dir)}\\" : "#{install_dir}/") + "agent_#{lib}_requirements-py3.txt" - requirements_custom[lib] = { - "req_lines" => Array.new, - "req_file_path" => static_reqs_out_folder + lib + "-py3.in", - "compiled_req_file_path" => lib_compiled_req_file_path, - } - end - - # Remove any excluded requirements from the static-environment req file - requirements = Array.new - - block "Create filtered requirements" do - File.open("#{static_reqs_in_file}", 'r+').readlines().each do |line| - next if excluded_packages.any? { |package_regex| line.match(package_regex) } - - # on non windows OS, we use the c version of the psycopg installation - if line.start_with?('psycopg[binary]') && !windows_target? - line.sub! 'psycopg[binary]', 'psycopg[c]' - end - # Keeping the custom env requirements lines apart to install them with a specific env - requirements_custom.each do |lib, lib_req| - if Regexp.new('^' + lib + '==').freeze.match line - lib_req["req_lines"].push(line) - end - end - # In any case we add the lib to the requirements files to avoid inconsistency in the installed versions - # For example if aerospike has dependency A>1.2.3 and a package in the big requirements file has A<1.2.3, the install process would succeed but the integration wouldn't work. - requirements.push(line) - end - - # Adding pympler for memory debug purposes - requirements.push("pympler==0.7") - end - - # Render the filtered requirements file - erb source: "static_requirements.txt.erb", - dest: "#{static_reqs_out_file}", - mode: 0640, - vars: { requirements: requirements } - - # Render the filtered libraries that are to be built with different env var - requirements_custom.each do |lib, lib_req| - erb source: "static_requirements.txt.erb", - dest: "#{lib_req["req_file_path"]}", - mode: 0640, - vars: { requirements: lib_req["req_lines"] } - end - - # Constraints file for constraining transitive dependencies in those cases where there may be incompatible versions - constraints = [] - if redhat_target? - constraints.push("bcrypt < 4.1.0") - end - - constraints_file = windows_safe_path(project_dir, "constraints.txt") - block "Write constraints file" do - File.open(constraints_file, 'w') { |f| f << constraints.join("\n") } - end - - - # Increasing pip max retries (default: 5 times) and pip timeout (default 15 seconds) to avoid blocking network errors - pip_max_retries = 20 - pip_timeout = 20 - - # Use pip-compile to create the final requirements file. Notice when we invoke `pip` through `python -m pip <...>`, - # there's no need to refer to `pip`, the interpreter will pick the right script. - command "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => build_env, :cwd => cwd_base - command "#{python} -m pip install datadog_checks_base --no-deps --no-index --find-links=#{wheel_build_dir}" - command "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => build_env, :cwd => cwd_downloader - command "#{python} -m pip install datadog_checks_downloader --no-deps --no-index --find-links=#{wheel_build_dir}" - command "#{python} -m piptools compile --generate-hashes -c #{constraints_file} --output-file #{compiled_reqs_file_path} #{static_reqs_out_file} " \ - "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => build_env - # Pip-compiling seperately each lib that needs a custom build installation - specific_build_env.each do |lib, env| - command "#{python} -m piptools compile --generate-hashes -c #{constraints_file} --output-file #{requirements_custom[lib]["compiled_req_file_path"]} #{requirements_custom[lib]["req_file_path"]} " \ - "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => env - end - - # - # Install static-environment requirements that the Agent and all checks will use - # - - # First we install the dependencies that need specific flags - specific_build_env.each do |lib, env| - command "#{python} -m pip install --no-deps --require-hashes -r #{requirements_custom[lib]["compiled_req_file_path"]}", :env => env - # Remove the file after use so it is not shipped - delete "#{requirements_custom[lib]["compiled_req_file_path"]}" - end - # Then we install the rest (already installed libraries will be ignored) with the main flags - command "#{python} -m pip install --no-deps --require-hashes -r #{compiled_reqs_file_path}", :env => build_env - # Remove the file after use so it is not shipped - delete "#{compiled_reqs_file_path}" - - # - # Install Core integrations - # - - # Create a constraint file after installing all the core dependencies and before any integration - # This is then used as a constraint file by the integration command to avoid messing with the agent's python environment - command "#{python} -m pip freeze > #{install_dir}/#{final_constraints_file}" - - if windows_target? - cached_wheels_dir = "#{windows_safe_path(wheel_build_dir)}\\.cached" - else - cached_wheels_dir = "#{wheel_build_dir}/.cached" - end - - checks_to_install = Array.new - - block "Collect integrations to install" do - # Go through every integration package in `integrations-core`, build and install - Dir.glob("#{project_dir}/*").each do |check_dir| - check = check_dir.split('/').last - - # do not install excluded integrations - next if !File.directory?("#{check_dir}") || excluded_folders.include?(check) - - # If there is no manifest file, then we should assume the folder does not - # contain a working check and move onto the next - manifest_file_path = "#{check_dir}/manifest.json" - - # If there is no manifest file, then we should assume the folder does not - # contain a working check and move onto the next - File.exist?(manifest_file_path) || next - - manifest = JSON.parse(File.read(manifest_file_path)) - if manifest.key?("supported_os") - manifest["supported_os"].include?(os) || next - else - if os == "mac_os" - tag = "Supported OS::macOS" - else - tag = "Supported OS::#{os.capitalize}" - end - - manifest["tile"]["classifier_tags"].include?(tag) || next - end - - File.file?("#{check_dir}/setup.py") || File.file?("#{check_dir}/pyproject.toml") || next - # Check if it supports Python 3. - support = `inv agent.check-supports-python-version #{check_dir} 3` - if support == "False" - log.info(log_key) { "Skipping '#{check}' since it does not support Python 3." } - next - end - - checks_to_install.push(check) - end - end - - installed_list = Array.new - cache_bucket = ENV.fetch('INTEGRATION_WHEELS_CACHE_BUCKET', '') - block "Install integrations" do - tasks_dir_in = windows_safe_path(Dir.pwd) - cache_branch = (shellout! "inv release.get-release-json-value base_branch", cwd: File.expand_path('..', tasks_dir_in)).stdout.strip - # On windows, `aws` actually executes Ruby's AWS SDK, but we want the Python one - awscli = if windows_target? then '"c:\Program files\python39\scripts\aws"' else 'aws' end - if cache_bucket != '' - mkdir cached_wheels_dir - shellout! "inv -e agent.get-integrations-from-cache " \ - "--python 3 --bucket #{cache_bucket} " \ - "--branch #{cache_branch || 'main'} " \ - "--integrations-dir #{windows_safe_path(project_dir)} " \ - "--target-dir #{cached_wheels_dir} " \ - "--integrations #{checks_to_install.join(',')} " \ - "--awscli #{awscli}", - :cwd => tasks_dir_in - - # install all wheels from cache in one pip invocation to speed things up - if windows_target? - shellout! "#{python} -m pip install --no-deps --no-index " \ - " --find-links #{windows_safe_path(cached_wheels_dir)} -r #{windows_safe_path(cached_wheels_dir)}\\found.txt" - else - shellout! "#{python} -m pip install --no-deps --no-index " \ - "--find-links #{cached_wheels_dir} -r #{cached_wheels_dir}/found.txt" - end - end - - # get list of integration wheels already installed from cache - if cache_bucket != '' - installed_out = (shellout! "#{python} -m pip list --format json").stdout - if $?.exitstatus == 0 - installed = JSON.parse(installed_out) - installed.each do |package| - package.each do |key, value| - if key == "name" && value.start_with?("datadog-") - installed_list.push(value["datadog-".length..-1]) - end - end - end - else - raise "Failed to list pip installed packages" - end - end - - checks_to_install.each do |check| - check_dir = File.join(project_dir, check) - check_conf_dir = "#{conf_dir}/#{check}.d" - - # For each conf file, if it already exists, that means the `datadog-agent` software def - # wrote it first. In that case, since the agent's confs take precedence, skip the conf - conf_files = ["conf.yaml.example", "conf.yaml.default", "metrics.yaml", "auto_conf.yaml"] - conf_files.each do |filename| - src = windows_safe_path(check_dir,"datadog_checks", check, "data", filename) - dest = check_conf_dir - if File.exist?(src) and !File.exist?(windows_safe_path(dest, filename)) - FileUtils.mkdir_p(dest) - FileUtils.cp_r(src, dest) - end - end - - # Copy SNMP profiles - profile_folders = ['profiles', 'default_profiles'] - profile_folders.each do |profile_folder| - folder_path = "#{check_dir}/datadog_checks/#{check}/data/#{profile_folder}" - if File.exist? folder_path - FileUtils.cp_r folder_path, "#{check_conf_dir}/" - end - end - - # pip < 21.2 replace underscores by dashes in package names per https://pip.pypa.io/en/stable/news/#v21-2 - # whether or not this might switch back in the future is not guaranteed, so we check for both name - # with dashes and underscores - if installed_list.include?(check) || installed_list.include?(check.gsub('_', '-')) - next - end - - if windows_target? - shellout! "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => win_build_env, :cwd => "#{windows_safe_path(project_dir)}\\#{check}" - else - shellout! "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => nix_build_env, :cwd => "#{project_dir}/#{check}" - end - shellout! "#{python} -m pip install datadog-#{check} --no-deps --no-index --find-links=#{wheel_build_dir}" - if cache_bucket != '' && ENV.fetch('INTEGRATION_WHEELS_SKIP_CACHE_UPLOAD', '') == '' && cache_branch != nil - shellout! "inv -e agent.upload-integration-to-cache " \ - "--python 3 --bucket #{cache_bucket} " \ - "--branch #{cache_branch} " \ - "--integrations-dir #{windows_safe_path(project_dir)} " \ - "--build-dir #{wheel_build_dir} " \ - "--integration #{check} " \ - "--awscli #{awscli}", - :cwd => tasks_dir_in - end - end - end - - # From now on we don't need piptools anymore, uninstall its deps so we don't include them in the final artifact - uninstall_buildtime_deps.each do |dep| - command "#{python} -m pip uninstall -y #{dep}" - end - - # Patch applies to only one file: set it explicitly as a target, no need for -p - if windows_target? - patch :source => "remove-maxfile-maxpath-psutil.patch", :target => "#{python_3_embedded}/Lib/site-packages/psutil/__init__.py" - else - patch :source => "remove-maxfile-maxpath-psutil.patch", :target => "#{install_dir}/embedded/lib/python3.11/site-packages/psutil/__init__.py" - end - - # Run pip check to make sure the agent's python environment is clean, all the dependencies are compatible - command "#{python} -m pip check" - - # Removing tests that don't need to be shipped in the embedded folder - if windows_target? - delete "#{python_3_embedded}/Lib/site-packages/Cryptodome/SelfTest/" - else - delete "#{install_dir}/embedded/lib/python3.11/site-packages/Cryptodome/SelfTest/" - end - - # Ship `requirements-agent-release.txt` file containing the versions of every check shipped with the agent - # Used by the `datadog-agent integration` command to prevent downgrading a check to a version - # older than the one shipped in the agent - copy "#{project_dir}/requirements-agent-release.txt", "#{install_dir}/" -end diff --git a/omnibus/config/software/datadog-agent-integrations.rb b/omnibus/config/software/datadog-agent-integrations.rb new file mode 100644 index 0000000000000..fa21fe029973a --- /dev/null +++ b/omnibus/config/software/datadog-agent-integrations.rb @@ -0,0 +1,524 @@ +# Unless explicitly stated otherwise all files in this repository are licensed +# under the Apache License Version 2.0. +# This product includes software developed at Datadog (https:#www.datadoghq.com/). +# Copyright 2016-present Datadog, Inc. + +require './lib/ostools.rb' +require 'json' + +name 'datadog-agent-integrations' + +license "BSD-3-Clause" +license_file "./LICENSE" + +source git: 'https://github.com/DataDog/integrations-core.git' +relative_path 'integrations-core' + +integrations_core_version = ENV['INTEGRATIONS_CORE_VERSION'] +if integrations_core_version.nil? || integrations_core_version.empty? + integrations_core_version = 'master' +end + +default_version integrations_core_version + +dependency 'datadog-agent' + +if with_python_runtime? "2" + dependency 'datadog-agent-integrations-py2-dependencies' + + whitelist_file "embedded/lib/python2.7/site-packages/.libsaerospike" + whitelist_file "embedded/lib/python2.7/site-packages/psycopg2" + whitelist_file "embedded/lib/python2.7/site-packages/wrapt" + whitelist_file "embedded/lib/python2.7/site-packages/pymqi" +end + +if with_python_runtime? "3" + dependency 'datadog-agent-integrations-py3-dependencies' + + whitelist_file "embedded/lib/python3.11/site-packages/.libsaerospike" + whitelist_file "embedded/lib/python3.11/site-packages/aerospike.libs" + whitelist_file "embedded/lib/python3.11/site-packages/psycopg2" + whitelist_file "embedded/lib/python3.11/site-packages/pymqi" +end + +gcc_version = ENV['GCC_VERSION'] +if gcc_version.nil? || gcc_version.empty? + gcc_version = '10.4.0' +end + +# folder names containing integrations from -core that won't be packaged with the Agent +excluded_folders = [ + 'datadog_checks_base', # namespacing package for wheels (NOT AN INTEGRATION) + 'datadog_checks_dev', # Development package, (NOT AN INTEGRATION) + 'datadog_checks_tests_helper', # Testing and Development package, (NOT AN INTEGRATION) + 'docker_daemon', # Agent v5 only +] + +# package names of dependencies that won't be added to the Agent Python environment +excluded_packages = Array.new + +# We build these manually +excluded_packages.push(/^snowflake-connector-python==/) +excluded_packages.push(/^confluent-kafka==/) + +if suse_target? + # Temporarily exclude Aerospike until builder supports new dependency + excluded_packages.push(/^aerospike==/) + excluded_folders.push('aerospike') +end + +if osx_target? + # Temporarily exclude Aerospike until builder supports new dependency + excluded_packages.push(/^aerospike==/) + excluded_folders.push('aerospike') + excluded_folders.push('teradata') +end + +if arm_target? + # This doesn't build on ARM + excluded_folders.push('ibm_ace') + excluded_folders.push('ibm_mq') + excluded_packages.push(/^pymqi==/) +end + +# We explicitly check for redhat builder, not target +# Our centos/redhat builder uses glibc 2.12 while pydantic +# requires glibc 2.17 +if redhat? && !arm_target? + excluded_packages.push(/^pydantic-core==/) +end + +# _64_bit checks the kernel arch. On windows, the builder is 64 bit +# even when doing a 32 bit build. Do a specific check for the 32 bit +# build +if arm_target? || !_64_bit? || (windows_target? && windows_arch_i386?) + excluded_packages.push(/^orjson==/) +end + +if linux_target? + excluded_packages.push(/^oracledb==/) +end + +build do + ["2", "3"].each do |python_major_version| + # Skip python version if not included in this build + next if not with_python_runtime?(python_major_version) + + final_constraints_file = "final_constraints-py#{python_major_version}.txt" + agent_requirements_file = "agent_requirements-py#{python_major_version}.txt" + filtered_agent_requirements_in = "agent_requirements-py#{python_major_version}.in" + agent_requirements_in = 'agent_requirements.in' + + # The dir for confs + if osx_target? + conf_dir = "#{install_dir}/etc/conf.d" + else + conf_dir = "#{install_dir}/etc/datadog-agent/conf.d" + end + mkdir conf_dir + + # aliases for pip + if windows_target? + python = windows_safe_path(install_dir, "embedded#{python_major_version}", "python.exe") + else + python = "#{install_dir}/embedded/bin/python#{python_major_version}" + end + + # If a python_mirror is set, it is set in a pip config file so that we do not leak the token in the CI output + pip_config_file = ENV['PIP_CONFIG_FILE'] + pre_build_env = { + "PIP_CONFIG_FILE" => "#{pip_config_file}" + } + + # Install the checks along with their dependencies + wheel_build_dir = windows_safe_path(project_dir, python_major_version, ".wheels") + build_deps_dir = windows_safe_path(project_dir, python_major_version, ".build_deps") + + # + # Prepare the build env, these dependencies are only needed to build and + # install the core integrations. + # + if python_major_version == "2" + command "#{python} -m pip download --dest #{build_deps_dir} hatchling==0.25.1", :env => pre_build_env + command "#{python} -m pip download --dest #{build_deps_dir} setuptools==40.9.0", :env => pre_build_env # Version from ./setuptools2.rb + command "#{python} -m pip install wheel==0.37.1", :env => pre_build_env # Pin to the last version that supports Python 2 + command "#{python} -m pip install setuptools-scm==5.0.2", :env => pre_build_env # Pin to the last version that supports Python 2 + command "#{python} -m pip install pip-tools==5.4.0", :env => pre_build_env + else + command "#{python} -m pip download --dest #{build_deps_dir} hatchling==0.25.1", :env => pre_build_env + command "#{python} -m pip download --dest #{build_deps_dir} setuptools==66.1.1", :env => pre_build_env # Version from ./setuptools3.rb + command "#{python} -m pip install wheel==0.38.4", :env => pre_build_env + command "#{python} -m pip install pip-tools==7.3.0", :env => pre_build_env + end + uninstall_buildtime_deps = ['rtloader', 'click', 'first', 'pip-tools'] + nix_build_env = { + "PIP_FIND_LINKS" => "#{build_deps_dir}", + "PIP_CONFIG_FILE" => "#{pip_config_file}", + "CFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", + "CXXFLAGS" => "-I#{install_dir}/embedded/include -I/opt/mqm/inc", + "LDFLAGS" => "-L#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", + "LD_RUN_PATH" => "#{install_dir}/embedded/lib -L/opt/mqm/lib64 -L/opt/mqm/lib", + "PATH" => "#{install_dir}/embedded/bin:#{ENV['PATH']}", + } + + win_build_env = { + "PIP_FIND_LINKS" => "#{build_deps_dir}", + "PIP_CONFIG_FILE" => "#{pip_config_file}", + } + + # Some libraries (looking at you, aerospike-client-python) need EXT_CFLAGS instead of CFLAGS. + nix_specific_build_env = { + "aerospike" => nix_build_env.merge({"EXT_CFLAGS" => nix_build_env["CFLAGS"] + " -std=gnu99"}), + # Always build pyodbc from source to link to the embedded version of libodbc + "pyodbc" => nix_build_env.merge({"PIP_NO_BINARY" => "pyodbc"}), + } + win_specific_build_env = {} + + # On Linux & Windows, specify the C99 standard explicitly to avoid issues while building some + # wheels (eg. ddtrace). + # Not explicitly setting that option has caused us problems in the past on SUSE, where the ddtrace + # wheel has to be manually built, as the C code in ddtrace doesn't follow the C89 standard (the default value of std). + # Note: We don't set this on MacOS, as on MacOS we need to build a bunch of packages & C extensions that + # don't have precompiled MacOS wheels. When building C extensions, the CFLAGS variable is added to + # the command-line parameters, even when compiling C++ code, where -std=c99 is invalid. + # See: https://github.com/python/cpython/blob/v3.8.8/Lib/distutils/sysconfig.py#L227 + if linux_target? || windows_target? + nix_build_env["CFLAGS"] += " -std=c99" + end + + # We only have gcc 10.4.0 on linux for now + if linux_target? + nix_build_env["CC"] = "/opt/gcc-#{gcc_version}/bin/gcc" + nix_build_env["CXX"] = "/opt/gcc-#{gcc_version}/bin/g++" + end + + # We need to explicitly specify RUSTFLAGS for libssl and libcrypto + # See https://github.com/pyca/cryptography/issues/8614#issuecomment-1489366475 + if python_major_version == "3" && redhat_target? && !arm_target? + nix_specific_build_env["cryptography"] = nix_build_env.merge( + { + "RUSTFLAGS" => "-C link-arg=-Wl,-rpath,#{install_dir}/embedded/lib", + "OPENSSL_DIR" => "#{install_dir}/embedded/", + # We have a manually installed dependency (snowflake connector) that already installed cryptography (but without the flags) + # We force reinstall it from source to be sure we use the flag + "PIP_NO_CACHE_DIR" => "off", + "PIP_FORCE_REINSTALL" => "1", + } + ) + end + + # + # Prepare the requirements file containing ALL the dependencies needed by + # any integration. This will provide the "static Python environment" of the Agent. + # We don't use the .in file provided by the base check directly because we + # want to filter out things before installing. + # + if windows_target? + static_reqs_in_file = "#{windows_safe_path(project_dir)}\\datadog_checks_base\\datadog_checks\\base\\data\\#{agent_requirements_in}" + static_reqs_out_folder = "#{windows_safe_path(project_dir)}\\" + static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in + compiled_reqs_file_path = "#{windows_safe_path(install_dir)}\\#{agent_requirements_file}" + else + static_reqs_in_file = "#{project_dir}/datadog_checks_base/datadog_checks/base/data/#{agent_requirements_in}" + static_reqs_out_folder = "#{project_dir}/" + static_reqs_out_file = static_reqs_out_folder + filtered_agent_requirements_in + compiled_reqs_file_path = "#{install_dir}/#{agent_requirements_file}" + end + + specific_build_env = windows_target? ? win_specific_build_env : nix_specific_build_env + build_env = windows_target? ? win_build_env : nix_build_env + cwd_base = windows_target? ? "#{windows_safe_path(project_dir)}\\datadog_checks_base" : "#{project_dir}/datadog_checks_base" + cwd_downloader = windows_target? ? "#{windows_safe_path(project_dir)}\\datadog_checks_downloader" : "#{project_dir}/datadog_checks_downloader" + + # Creating a hash containing the requirements and requirements file path associated to every lib + requirements_custom = Hash.new() + specific_build_env.each do |lib, env| + lib_compiled_req_file_path = (windows_target? ? "#{windows_safe_path(install_dir)}\\" : "#{install_dir}/") + "agent_#{lib}_requirements-py3.txt" + requirements_custom[lib] = { + "req_lines" => Array.new, + "req_file_path" => static_reqs_out_folder + lib + "-py#{python_major_version}.in", + "compiled_req_file_path" => lib_compiled_req_file_path, + } + end + + # Remove any excluded requirements from the static-environment req file + requirements = Array.new + + block "Create filtered requirements" do + File.open("#{static_reqs_in_file}", 'r+').readlines().each do |line| + next if excluded_packages.any? { |package_regex| line.match(package_regex) } + + # on non windows OS, we use the c version of the psycopg installation + if line.start_with?('psycopg[binary]') && !windows_target? + line.sub! 'psycopg[binary]', 'psycopg[c]' + end + # Keeping the custom env requirements lines apart to install them with a specific env + requirements_custom.each do |lib, lib_req| + if Regexp.new('^' + lib + '==').freeze.match line + lib_req["req_lines"].push(line) + end + end + # In any case we add the lib to the requirements files to avoid inconsistency in the installed versions + # For example if aerospike has dependency A>1.2.3 and a package in the big requirements file has A<1.2.3, the install process would succeed but the integration wouldn't work. + requirements.push(line) + end + + # Adding pympler for memory debug purposes + requirements.push("pympler==0.7") + + end + + # Render the filtered requirements file + erb source: "static_requirements.txt.erb", + dest: "#{static_reqs_out_file}", + mode: 0640, + vars: { requirements: requirements } + + # Render the filtered libraries that are to be built with different env var + requirements_custom.each do |lib, lib_req| + erb source: "static_requirements.txt.erb", + dest: "#{lib_req["req_file_path"]}", + mode: 0640, + vars: { requirements: lib_req["req_lines"] } + end + + # Constraints file for constraining transitive dependencies in those cases where there may be incompatible versions (only supported for py3) + constraints_flag = "" + if python_major_version == "3" + constraints = [] + if redhat_target? + constraints.push("bcrypt < 4.1.0") + end + + constraints_file = windows_safe_path(project_dir, "constraints.txt") + block "Write constraints file" do + File.open(constraints_file, 'w') { |f| f << constraints.join("\n") } + end + constraints_flag = "-c #{constraints_file}" + end + + # Increasing pip max retries (default: 5 times) and pip timeout (default 15 seconds) to avoid blocking network errors + pip_max_retries = 20 + pip_timeout = 20 + + # Use pip-compile to create the final requirements file. Notice when we invoke `pip` through `python -m pip <...>`, + # there's no need to refer to `pip`, the interpreter will pick the right script. + command "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => build_env, :cwd => cwd_base + command "#{python} -m pip install datadog_checks_base --no-deps --no-index --find-links=#{wheel_build_dir}" + + # We only install the downloader on Python 3 + if python_major_version == "3" + command "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => build_env, :cwd => cwd_downloader + command "#{python} -m pip install datadog_checks_downloader --no-deps --no-index --find-links=#{wheel_build_dir}" + end + + command "#{python} -m piptools compile --generate-hashes #{constraints_flag} --output-file #{compiled_reqs_file_path} #{static_reqs_out_file} " \ + "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => build_env + # Pip-compiling seperately each lib that needs a custom build installation + specific_build_env.each do |lib, env| + command "#{python} -m piptools compile --generate-hashes #{constraints_flag} --output-file #{requirements_custom[lib]["compiled_req_file_path"]} #{requirements_custom[lib]["req_file_path"]} " \ + "--pip-args \"--retries #{pip_max_retries} --timeout #{pip_timeout}\"", :env => env + end + + # + # Install static-environment requirements that the Agent and all checks will use + # + + # First we install the dependencies that need specific flags + specific_build_env.each do |lib, env| + command "#{python} -m pip install --no-deps --require-hashes -r #{requirements_custom[lib]["compiled_req_file_path"]}", :env => env + # Remove the file after use so it is not shipped + delete "#{requirements_custom[lib]["compiled_req_file_path"]}" + end + # Then we install the rest (already installed libraries will be ignored) with the main flags + command "#{python} -m pip install --no-deps --require-hashes -r #{compiled_reqs_file_path}", :env => build_env + # Remove the file after use so it is not shipped + delete "#{compiled_reqs_file_path}" + + # + # Install Core integrations + # + + # Create a constraint file after installing all the core dependencies and before any integration + # This is then used as a constraint file by the integration command to avoid messing with the agent's python environment + command "#{python} -m pip freeze > #{install_dir}/#{final_constraints_file}" + + if windows_target? + cached_wheels_dir = "#{windows_safe_path(wheel_build_dir)}\\.cached" + else + cached_wheels_dir = "#{wheel_build_dir}/.cached" + end + + checks_to_install = Array.new + + block "Collect integrations to install" do + # Go through every integration package in `integrations-core`, build and install + Dir.glob("#{project_dir}/*").each do |check_dir| + check = check_dir.split('/').last + + # do not install excluded integrations + next if !File.directory?("#{check_dir}") || excluded_folders.include?(check) + + # If there is no manifest file, then we should assume the folder does not + # contain a working check and move onto the next + manifest_file_path = "#{check_dir}/manifest.json" + + # If there is no manifest file, then we should assume the folder does not + # contain a working check and move onto the next + File.exist?(manifest_file_path) || next + + manifest = JSON.parse(File.read(manifest_file_path)) + if manifest.key?("supported_os") + manifest["supported_os"].include?(os) || next + else + if os == "mac_os" + tag = "Supported OS::macOS" + else + tag = "Supported OS::#{os.capitalize}" + end + + manifest["tile"]["classifier_tags"].include?(tag) || next + end + + File.file?("#{check_dir}/setup.py") || File.file?("#{check_dir}/pyproject.toml") || next + # Check if it supports Python version we're building for. + support = `inv agent.check-supports-python-version #{check_dir} #{python_major_version}` + if support == "False" + log.info(log_key) { "Skipping '#{check}' since it does not support Python #{python_major_version}." } + next + end + + checks_to_install.push(check) + end + end + + installed_list = Array.new + cache_bucket = ENV.fetch('INTEGRATION_WHEELS_CACHE_BUCKET', '') + block "Install integrations" do + tasks_dir_in = windows_safe_path(Dir.pwd) + cache_branch = (shellout! "inv release.get-release-json-value base_branch", cwd: File.expand_path('..', tasks_dir_in)).stdout.strip + # On windows, `aws` actually executes Ruby's AWS SDK, but we want the Python one + awscli = if windows_target? then '"c:\Program files\python39\scripts\aws"' else 'aws' end + if cache_bucket != '' + mkdir cached_wheels_dir + shellout! "inv -e agent.get-integrations-from-cache " \ + "--python #{python_major_version} --bucket #{cache_bucket} " \ + "--branch #{cache_branch || 'main'} " \ + "--integrations-dir #{windows_safe_path(project_dir)} " \ + "--target-dir #{cached_wheels_dir} " \ + "--integrations #{checks_to_install.join(',')} " \ + "--awscli #{awscli}", + :cwd => tasks_dir_in + + # install all wheels from cache in one pip invocation to speed things up + if windows_target? + shellout! "#{python} -m pip install --no-deps --no-index " \ + " --find-links #{windows_safe_path(cached_wheels_dir)} -r #{windows_safe_path(cached_wheels_dir)}\\found.txt" + else + shellout! "#{python} -m pip install --no-deps --no-index " \ + "--find-links #{cached_wheels_dir} -r #{cached_wheels_dir}/found.txt" + end + end + + # get list of integration wheels already installed from cache + if cache_bucket != '' + installed_out = (shellout! "#{python} -m pip list --format json").stdout + if $?.exitstatus == 0 + installed = JSON.parse(installed_out) + installed.each do |package| + package.each do |key, value| + if key == "name" && value.start_with?("datadog-") + installed_list.push(value["datadog-".length..-1]) + end + end + end + else + raise "Failed to list pip installed packages" + end + end + + checks_to_install.each do |check| + check_dir = File.join(project_dir, check) + check_conf_dir = "#{conf_dir}/#{check}.d" + + # For each conf file, if it already exists, that means the `datadog-agent` software def + # wrote it first. In that case, since the agent's confs take precedence, skip the conf + conf_files = ["conf.yaml.example", "conf.yaml.default", "metrics.yaml", "auto_conf.yaml"] + conf_files.each do |filename| + src = windows_safe_path(check_dir,"datadog_checks", check, "data", filename) + dest = check_conf_dir + if File.exist?(src) and !File.exist?(windows_safe_path(dest, filename)) + FileUtils.mkdir_p(dest) + FileUtils.cp_r(src, dest) + end + end + + # Copy SNMP profiles + profile_folders = ['profiles', 'default_profiles'] + profile_folders.each do |profile_folder| + folder_path = "#{check_dir}/datadog_checks/#{check}/data/#{profile_folder}" + if File.exist? folder_path + FileUtils.cp_r folder_path, "#{check_conf_dir}/" + end + end + + # pip < 21.2 replace underscores by dashes in package names per https://pip.pypa.io/en/stable/news/#v21-2 + # whether or not this might switch back in the future is not guaranteed, so we check for both name + # with dashes and underscores + if installed_list.include?(check) || installed_list.include?(check.gsub('_', '-')) + next + end + + if windows_target? + shellout! "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => win_build_env, :cwd => "#{windows_safe_path(project_dir)}\\#{check}" + else + shellout! "#{python} -m pip wheel . --no-deps --no-index --wheel-dir=#{wheel_build_dir}", :env => nix_build_env, :cwd => "#{project_dir}/#{check}" + end + shellout! "#{python} -m pip install datadog-#{check} --no-deps --no-index --find-links=#{wheel_build_dir}" + if cache_bucket != '' && ENV.fetch('INTEGRATION_WHEELS_SKIP_CACHE_UPLOAD', '') == '' && cache_branch != nil + shellout! "inv -e agent.upload-integration-to-cache " \ + "--python #{python_major_version} --bucket #{cache_bucket} " \ + "--branch #{cache_branch} " \ + "--integrations-dir #{windows_safe_path(project_dir)} " \ + "--build-dir #{wheel_build_dir} " \ + "--integration #{check} " \ + "--awscli #{awscli}", + :cwd => tasks_dir_in + end + end + end + + # From now on we don't need piptools anymore, uninstall its deps so we don't include them in the final artifact + uninstall_buildtime_deps.each do |dep| + command "#{python} -m pip uninstall -y #{dep}" + end + + if windows_target? + site_packages = windows_safe_path(install_dir, "embedded#{python_major_version}", "Lib/site-packages") + else + if python_major_version == '2' + site_packages = File.join(install_dir, "embedded/lib/python2.7/site-packages") + else + site_packages = File.join(install_dir, "embedded/lib/python3.11/site-packages") + end + end + + # Patch applies to only one file: set it explicitly as a target, no need for -p + patch :source => "remove-maxfile-maxpath-psutil.patch", :target => windows_safe_path(site_packages, "psutil/__init__.py") + + if python_major_version == '2' + patch :source => "create-regex-at-runtime.patch", :target => windows_safe_path(site_packages, "yaml/reader.py") + end + + # Run pip check to make sure the agent's python environment is clean, all the dependencies are compatible + command "#{python} -m pip check" + + # Removing tests that don't need to be shipped in the embedded folder + delete windows_safe_path(site_packages, "Cryptodome/SelfTest") + + # Ship `requirements-agent-release.txt` file containing the versions of every check shipped with the agent + # Used by the `datadog-agent integration` command to prevent downgrading a check to a version + # older than the one shipped in the agent + copy "#{project_dir}/requirements-agent-release.txt", "#{install_dir}/" + end +end diff --git a/omnibus/config/templates/datadog-agent-integrations-py3/static_requirements.txt.erb b/omnibus/config/templates/datadog-agent-integrations-py3/static_requirements.txt.erb deleted file mode 100644 index cf827abae0167..0000000000000 --- a/omnibus/config/templates/datadog-agent-integrations-py3/static_requirements.txt.erb +++ /dev/null @@ -1,3 +0,0 @@ -<% requirements.each do |requirement| -%> -<%= requirement -%> -<% end -%> diff --git a/omnibus/config/templates/datadog-agent-integrations-py2/static_requirements.txt.erb b/omnibus/config/templates/datadog-agent-integrations/static_requirements.txt.erb similarity index 100% rename from omnibus/config/templates/datadog-agent-integrations-py2/static_requirements.txt.erb rename to omnibus/config/templates/datadog-agent-integrations/static_requirements.txt.erb