diff --git a/plugins/repository-hdfs/build.gradle b/plugins/repository-hdfs/build.gradle
index df35a93b57e9e..1dc625563e0b0 100644
--- a/plugins/repository-hdfs/build.gradle
+++ b/plugins/repository-hdfs/build.gradle
@@ -32,7 +32,7 @@ esplugin {
apply plugin: 'elasticsearch.vagrantsupport'
versions << [
- 'hadoop2': '2.7.1'
+ 'hadoop2': '2.8.1'
]
configurations {
@@ -45,7 +45,8 @@ dependencies {
compile "org.apache.hadoop:hadoop-annotations:${versions.hadoop2}"
compile "org.apache.hadoop:hadoop-auth:${versions.hadoop2}"
compile "org.apache.hadoop:hadoop-hdfs:${versions.hadoop2}"
- compile 'org.apache.htrace:htrace-core:3.1.0-incubating'
+ compile "org.apache.hadoop:hadoop-hdfs-client:${versions.hadoop2}"
+ compile 'org.apache.htrace:htrace-core4:4.0.1-incubating'
compile 'com.google.guava:guava:11.0.2'
compile 'com.google.protobuf:protobuf-java:2.5.0'
compile 'commons-logging:commons-logging:1.1.3'
@@ -210,12 +211,15 @@ if (secureFixtureSupported) {
thirdPartyAudit.excludes = [
// classes are missing, because we added hadoop jars one by one until tests pass.
'com.google.gson.stream.JsonReader',
- 'com.google.gson.stream.JsonWriter',
- 'com.jcraft.jsch.ChannelExec',
- 'com.jcraft.jsch.JSch',
- 'com.jcraft.jsch.Logger',
- 'com.jcraft.jsch.Session',
- 'com.sun.jersey.api.ParamException',
+ 'com.google.gson.stream.JsonWriter',
+ 'com.jcraft.jsch.ChannelExec',
+ 'com.jcraft.jsch.ChannelSftp',
+ 'com.jcraft.jsch.ChannelSftp$LsEntry',
+ 'com.jcraft.jsch.JSch',
+ 'com.jcraft.jsch.Logger',
+ 'com.jcraft.jsch.Session',
+ 'com.jcraft.jsch.SftpATTRS',
+ 'com.sun.jersey.api.ParamException',
'com.sun.jersey.api.core.HttpContext',
'com.sun.jersey.core.spi.component.ComponentContext',
'com.sun.jersey.core.spi.component.ComponentScope',
@@ -239,6 +243,7 @@ thirdPartyAudit.excludes = [
'io.netty.channel.ChannelHandlerContext',
'io.netty.channel.ChannelInboundHandlerAdapter',
'io.netty.channel.ChannelInitializer',
+ 'io.netty.channel.ChannelOption',
'io.netty.channel.ChannelPipeline',
'io.netty.channel.EventLoopGroup',
'io.netty.channel.SimpleChannelInboundHandler',
@@ -267,7 +272,8 @@ thirdPartyAudit.excludes = [
'io.netty.handler.stream.ChunkedStream',
'io.netty.handler.stream.ChunkedWriteHandler',
'io.netty.util.concurrent.GlobalEventExecutor',
- 'javax.ws.rs.core.Context',
+ 'io.netty.util.ReferenceCountUtil',
+ 'javax.ws.rs.core.Context',
'javax.ws.rs.core.MediaType',
'javax.ws.rs.core.MultivaluedMap',
'javax.ws.rs.core.Response$ResponseBuilder',
@@ -317,8 +323,7 @@ thirdPartyAudit.excludes = [
'org.apache.commons.digester.substitution.MultiVariableExpander',
'org.apache.commons.digester.substitution.VariableSubstitutor',
'org.apache.commons.digester.xmlrules.DigesterLoader',
- 'org.apache.commons.httpclient.util.URIUtil',
- 'org.apache.commons.jxpath.JXPathContext',
+ 'org.apache.commons.jxpath.JXPathContext',
'org.apache.commons.jxpath.ri.JXPathContextReferenceImpl',
'org.apache.commons.jxpath.ri.QName',
'org.apache.commons.jxpath.ri.compiler.NodeNameTest',
@@ -368,7 +373,8 @@ thirdPartyAudit.excludes = [
'org.apache.curator.utils.EnsurePath',
'org.apache.curator.utils.PathUtils',
'org.apache.curator.utils.ThreadUtils',
- 'org.apache.curator.utils.ZKPaths',
+ 'org.apache.curator.utils.ZKPaths',
+ 'org.apache.directory.shared.kerberos.components.EncryptionKey',
'org.apache.directory.server.kerberos.shared.keytab.Keytab',
'org.apache.directory.server.kerberos.shared.keytab.KeytabEntry',
'org.apache.http.NameValuePair',
@@ -402,33 +408,32 @@ thirdPartyAudit.excludes = [
'org.codehaus.jackson.JsonFactory',
'org.codehaus.jackson.JsonGenerator',
'org.codehaus.jackson.JsonGenerator$Feature',
- 'org.codehaus.jackson.JsonNode',
- 'org.codehaus.jackson.map.MappingJsonFactory',
+ 'org.codehaus.jackson.map.MappingJsonFactory',
'org.codehaus.jackson.map.ObjectMapper',
'org.codehaus.jackson.map.ObjectReader',
'org.codehaus.jackson.map.ObjectWriter',
'org.codehaus.jackson.node.ContainerNode',
- 'org.codehaus.jackson.type.TypeReference',
- 'org.codehaus.jackson.util.MinimalPrettyPrinter',
+ 'org.codehaus.jackson.util.MinimalPrettyPrinter',
'org.fusesource.leveldbjni.JniDBFactory',
'org.iq80.leveldb.DB',
'org.iq80.leveldb.Options',
- 'org.iq80.leveldb.WriteBatch',
- 'org.mortbay.jetty.Connector',
- 'org.mortbay.jetty.Handler',
- 'org.mortbay.jetty.InclusiveByteRange',
- 'org.mortbay.jetty.MimeTypes',
- 'org.mortbay.jetty.NCSARequestLog',
- 'org.mortbay.jetty.RequestLog',
- 'org.mortbay.jetty.Server',
- 'org.mortbay.jetty.handler.ContextHandler$SContext',
- 'org.mortbay.jetty.handler.ContextHandler',
- 'org.mortbay.jetty.handler.ContextHandlerCollection',
- 'org.mortbay.jetty.handler.HandlerCollection',
- 'org.mortbay.jetty.handler.RequestLogHandler',
- 'org.mortbay.jetty.nio.SelectChannelConnector',
- 'org.mortbay.jetty.security.SslSocketConnector',
- 'org.mortbay.jetty.servlet.AbstractSessionManager',
+ 'org.iq80.leveldb.WriteBatch',
+ 'org.mortbay.jetty.Connector',
+ 'org.mortbay.jetty.Handler',
+ 'org.mortbay.jetty.InclusiveByteRange',
+ 'org.mortbay.jetty.MimeTypes',
+ 'org.mortbay.jetty.NCSARequestLog',
+ 'org.mortbay.jetty.RequestLog',
+ 'org.mortbay.jetty.Server',
+ 'org.mortbay.jetty.handler.ContextHandler$SContext',
+ 'org.mortbay.jetty.handler.ContextHandler',
+ 'org.mortbay.jetty.handler.ContextHandlerCollection',
+ 'org.mortbay.jetty.handler.HandlerCollection',
+ 'org.mortbay.jetty.handler.RequestLogHandler',
+ 'org.mortbay.jetty.nio.SelectChannelConnector',
+ 'org.mortbay.jetty.security.SslSelectChannelConnector',
+ 'org.mortbay.jetty.security.SslSocketConnector',
+ 'org.mortbay.jetty.servlet.AbstractSessionManager',
'org.mortbay.jetty.servlet.Context',
'org.mortbay.jetty.servlet.DefaultServlet',
'org.mortbay.jetty.servlet.FilterHolder',
@@ -437,8 +442,7 @@ thirdPartyAudit.excludes = [
'org.mortbay.jetty.servlet.ServletHolder',
'org.mortbay.jetty.servlet.SessionHandler',
'org.mortbay.jetty.webapp.WebAppContext',
- 'org.mortbay.log.Log',
- 'org.mortbay.thread.QueuedThreadPool',
+ 'org.mortbay.thread.QueuedThreadPool',
'org.mortbay.util.MultiException',
'org.mortbay.util.ajax.JSON$Convertible',
'org.mortbay.util.ajax.JSON$Output',
@@ -473,9 +477,26 @@ thirdPartyAudit.excludes = [
'org.apache.log4j.AppenderSkeleton',
'org.apache.log4j.AsyncAppender',
'org.apache.log4j.helpers.ISO8601DateFormat',
- 'org.apache.log4j.spi.ThrowableInformation'
+ 'org.apache.log4j.spi.ThrowableInformation',
+
+ // New optional dependencies in 2.8
+ 'com.nimbusds.jose.JWSObject$State',
+ 'com.nimbusds.jose.crypto.RSASSAVerifier',
+ 'com.nimbusds.jwt.ReadOnlyJWTClaimsSet',
+ 'com.nimbusds.jwt.SignedJWT',
+ 'com.squareup.okhttp.Call',
+ 'com.squareup.okhttp.MediaType',
+ 'com.squareup.okhttp.OkHttpClient',
+ 'com.squareup.okhttp.Request$Builder',
+ 'com.squareup.okhttp.RequestBody',
+ 'com.squareup.okhttp.Response',
+ 'com.squareup.okhttp.ResponseBody'
]
+if (project.rootProject.ext.javaVersion == JavaVersion.VERSION_1_9) {
+ thirdPartyAudit.excludes.add('javax.xml.bind.annotation.adapters.HexBinaryAdapter')
+}
+
// Gradle 2.13 bundles org.slf4j.impl.StaticLoggerBinder in its core.jar which leaks into the forbidden APIs ant task
// Gradle 2.14+ does not bundle this class anymore so we need to properly exclude it here.
if (GradleVersion.current() > GradleVersion.version("2.13")) {
diff --git a/plugins/repository-hdfs/licenses/hadoop-annotations-2.7.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-annotations-2.7.1.jar.sha1
deleted file mode 100644
index 660467a4c6e57..0000000000000
--- a/plugins/repository-hdfs/licenses/hadoop-annotations-2.7.1.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-2a77fe74ee056bf45598cf7e20cd624e8388e627
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-annotations-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-annotations-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..bb000d4ddb025
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-annotations-2.8.1.jar.sha1
@@ -0,0 +1 @@
+335a867cf42bf789919bfc3229ff26747124e8f1
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-auth-2.7.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-auth-2.7.1.jar.sha1
deleted file mode 100644
index 0161301ead213..0000000000000
--- a/plugins/repository-hdfs/licenses/hadoop-auth-2.7.1.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-2515f339f97f1d7ba850485e06e395a58586bc2e
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-auth-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-auth-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..67a1c4f3fa546
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-auth-2.8.1.jar.sha1
@@ -0,0 +1 @@
+688ccccc0e0739d8737a93b0039a4a661e52084b
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-client-2.7.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-client-2.7.1.jar.sha1
deleted file mode 100644
index 4c6dca8af4989..0000000000000
--- a/plugins/repository-hdfs/licenses/hadoop-client-2.7.1.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-dbc2faacd210e6a1e3eb7def6e42065c7457d960
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-client-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-client-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..e7be8d468ddfc
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-client-2.8.1.jar.sha1
@@ -0,0 +1 @@
+4812f251f8100fd4722c3cec5d7353f71f69cda9
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-common-2.7.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-common-2.7.1.jar.sha1
deleted file mode 100644
index 64ff368db60bf..0000000000000
--- a/plugins/repository-hdfs/licenses/hadoop-common-2.7.1.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-50580f5ebab60b1b318ad157f668d8e40a1cc0da
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-common-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-common-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..6e6c625b7d6c2
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-common-2.8.1.jar.sha1
@@ -0,0 +1 @@
+a4df18b79e4d0349ce4b58a52d314e7ae1d6be99
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-hdfs-2.7.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-hdfs-2.7.1.jar.sha1
deleted file mode 100644
index 2d4954befaaf7..0000000000000
--- a/plugins/repository-hdfs/licenses/hadoop-hdfs-2.7.1.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-11681de93a4cd76c841e352b7094f839b072a21f
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-hdfs-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-hdfs-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..e419a3124f66b
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-hdfs-2.8.1.jar.sha1
@@ -0,0 +1 @@
+a378f4bc8e6cd779d779c9f512e0e31edd771633
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/hadoop-hdfs-client-2.8.1.jar.sha1 b/plugins/repository-hdfs/licenses/hadoop-hdfs-client-2.8.1.jar.sha1
new file mode 100644
index 0000000000000..20c7963cd3d1b
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/hadoop-hdfs-client-2.8.1.jar.sha1
@@ -0,0 +1 @@
+6b0100e4f58ecf7ce75817fce1ffdfbec947337a
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/htrace-core-3.1.0-incubating.jar.sha1 b/plugins/repository-hdfs/licenses/htrace-core-3.1.0-incubating.jar.sha1
deleted file mode 100644
index c742d8397cf98..0000000000000
--- a/plugins/repository-hdfs/licenses/htrace-core-3.1.0-incubating.jar.sha1
+++ /dev/null
@@ -1 +0,0 @@
-f73606e7c9ede5802335c290bf47490ad6d51df3
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/htrace-core4-4.0.1-incubating.jar.sha1 b/plugins/repository-hdfs/licenses/htrace-core4-4.0.1-incubating.jar.sha1
new file mode 100644
index 0000000000000..a0a562b2a4eb4
--- /dev/null
+++ b/plugins/repository-hdfs/licenses/htrace-core4-4.0.1-incubating.jar.sha1
@@ -0,0 +1 @@
+f4ef727cb4675788ac66f48e217020acc1690960
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/htrace-core-LICENSE.txt b/plugins/repository-hdfs/licenses/htrace-core4-LICENSE.txt
similarity index 97%
rename from plugins/repository-hdfs/licenses/htrace-core-LICENSE.txt
rename to plugins/repository-hdfs/licenses/htrace-core4-LICENSE.txt
index 0befae88d4f56..5e4a6e6fa77e5 100644
--- a/plugins/repository-hdfs/licenses/htrace-core-LICENSE.txt
+++ b/plugins/repository-hdfs/licenses/htrace-core4-LICENSE.txt
@@ -203,8 +203,8 @@ of dependencies that are NOT Apache Licensed.
See the License for the specific language governing permissions and
limitations under the License.
-The HTrace Owl logo is from http://www.clker.com/clipart-13653.html. It is
-public domain.
+Units, a string formatting go library, is Copyright (c) 2014 Alec Thomas
+and MIT licensed: https://github.com/alecthomas/units/blob/master/COPYING
D3, a javascript library for manipulating data, used by htrace-hbase
is Copyright 2010-2014, Michael Bostock and BSD licensed:
@@ -239,4 +239,7 @@ https://github.com/moment/moment/blob/develop/LICENSE
CMP is an implementation of the MessagePack serialization format in
C. It is licensed under the MIT license:
https://github.com/camgunz/cmp/blob/master/LICENSE
-See ./htrace-c/src/util/cmp.c and ./htrace-c/src/util/cmp.h.
+
+go-codec is an implementation of several serialization and deserialization
+codecs in Go. It is licensed under the MIT license:
+https://github.com/ugorji/go/blob/master/LICENSE
\ No newline at end of file
diff --git a/plugins/repository-hdfs/licenses/htrace-core-NOTICE.txt b/plugins/repository-hdfs/licenses/htrace-core4-NOTICE.txt
similarity index 84%
rename from plugins/repository-hdfs/licenses/htrace-core-NOTICE.txt
rename to plugins/repository-hdfs/licenses/htrace-core4-NOTICE.txt
index 845b6965e0441..c96631aab6721 100644
--- a/plugins/repository-hdfs/licenses/htrace-core-NOTICE.txt
+++ b/plugins/repository-hdfs/licenses/htrace-core4-NOTICE.txt
@@ -11,3 +11,6 @@ that are NOT Apache licensed (with pointers to their licensing)
Apache HTrace includes an Apache Thrift connector to Zipkin. Zipkin
is a distributed tracing system that is Apache 2.0 Licensed.
Copyright 2012 Twitter, Inc.
+
+Our Owl logo we took from http://www.clker.com/clipart-13653.html.
+It is public domain/free.
\ No newline at end of file
diff --git a/plugins/repository-hdfs/src/main/java/org/elasticsearch/repositories/hdfs/HdfsPlugin.java b/plugins/repository-hdfs/src/main/java/org/elasticsearch/repositories/hdfs/HdfsPlugin.java
index 4e51ab23b8091..c0b3d805bcc8f 100644
--- a/plugins/repository-hdfs/src/main/java/org/elasticsearch/repositories/hdfs/HdfsPlugin.java
+++ b/plugins/repository-hdfs/src/main/java/org/elasticsearch/repositories/hdfs/HdfsPlugin.java
@@ -62,8 +62,6 @@ private static Void evilHadoopInit() {
Class.forName("org.apache.hadoop.util.StringUtils");
Class.forName("org.apache.hadoop.util.ShutdownHookManager");
Class.forName("org.apache.hadoop.conf.Configuration");
- Class.forName("org.apache.hadoop.hdfs.protocol.HdfsConstants");
- Class.forName("org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck");
} catch (ClassNotFoundException | IOException e) {
throw new RuntimeException(e);
} finally {
diff --git a/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsBlobStoreContainerTests.java b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsBlobStoreContainerTests.java
index 2bfe6843dafff..195dea9810224 100644
--- a/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsBlobStoreContainerTests.java
+++ b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsBlobStoreContainerTests.java
@@ -19,15 +19,6 @@
package org.elasticsearch.repositories.hdfs;
-import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.fs.AbstractFileSystem;
-import org.apache.hadoop.fs.FileContext;
-import org.apache.hadoop.fs.UnsupportedFileSystemException;
-import org.elasticsearch.common.SuppressForbidden;
-import org.elasticsearch.common.blobstore.BlobStore;
-import org.elasticsearch.repositories.ESBlobStoreContainerTestCase;
-
-import javax.security.auth.Subject;
import java.io.IOException;
import java.lang.reflect.Constructor;
import java.lang.reflect.InvocationTargetException;
@@ -38,7 +29,18 @@
import java.security.PrivilegedActionException;
import java.security.PrivilegedExceptionAction;
import java.util.Collections;
+import javax.security.auth.Subject;
+
+import com.carrotsearch.randomizedtesting.annotations.ThreadLeakFilters;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.AbstractFileSystem;
+import org.apache.hadoop.fs.FileContext;
+import org.apache.hadoop.fs.UnsupportedFileSystemException;
+import org.elasticsearch.common.SuppressForbidden;
+import org.elasticsearch.common.blobstore.BlobStore;
+import org.elasticsearch.repositories.ESBlobStoreContainerTestCase;
+@ThreadLeakFilters(filters = {HdfsClientThreadLeakFilter.class})
public class HdfsBlobStoreContainerTests extends ESBlobStoreContainerTestCase {
@Override
diff --git a/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsClientThreadLeakFilter.java b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsClientThreadLeakFilter.java
new file mode 100644
index 0000000000000..5db9635aa8a3a
--- /dev/null
+++ b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsClientThreadLeakFilter.java
@@ -0,0 +1,46 @@
+/*
+ * Licensed to Elasticsearch under one or more contributor
+ * license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright
+ * ownership. Elasticsearch licenses this file to you under
+ * the Apache License, Version 2.0 (the "License"); you may
+ * not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied. See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+
+package org.elasticsearch.repositories.hdfs;
+
+import com.carrotsearch.randomizedtesting.ThreadFilter;
+
+/**
+ * In Hadoop 2.8.0, there is a thread that is started by the filesystem to clean up old execution stats.
+ * This thread ignores all interrupts, catching InterruptedException, logging it, and continuing on
+ * with its work. The thread is a daemon, so it thankfully does not stop the JVM from closing, and it
+ * is started only once in a class's static initialization. This currently breaks our testing as this
+ * thread leaks out of the client and is picked up by the test framework. This thread filter is meant
+ * to ignore the offending thread until a version of Hadoop is released that addresses the incorrect
+ * interrupt handling.
+ *
+ * @see https://issues.apache.org/jira/browse/HADOOP-12829
+ * @see "org.apache.hadoop.fs.FileSystem.Statistics.StatisticsDataReferenceCleaner"
+ * @see "org.apache.hadoop.fs.FileSystem.Statistics"
+ */
+public final class HdfsClientThreadLeakFilter implements ThreadFilter {
+
+ private static final String OFFENDING_THREAD_NAME =
+ "org.apache.hadoop.fs.FileSystem$Statistics$StatisticsDataReferenceCleaner";
+
+ @Override
+ public boolean reject(Thread t) {
+ return t.getName().equals(OFFENDING_THREAD_NAME);
+ }
+}
diff --git a/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsTests.java b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsTests.java
index 511aafdd29b60..9d0520205ed5b 100644
--- a/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsTests.java
+++ b/plugins/repository-hdfs/src/test/java/org/elasticsearch/repositories/hdfs/HdfsTests.java
@@ -23,6 +23,7 @@
import java.util.Collection;
+import com.carrotsearch.randomizedtesting.annotations.ThreadLeakFilters;
import org.elasticsearch.action.admin.cluster.repositories.put.PutRepositoryResponse;
import org.elasticsearch.action.admin.cluster.snapshots.create.CreateSnapshotResponse;
import org.elasticsearch.action.admin.cluster.snapshots.restore.RestoreSnapshotResponse;
@@ -34,6 +35,7 @@
import org.elasticsearch.snapshots.SnapshotState;
import org.elasticsearch.test.ESSingleNodeTestCase;
+@ThreadLeakFilters(filters = {HdfsClientThreadLeakFilter.class})
public class HdfsTests extends ESSingleNodeTestCase {
@Override
diff --git a/test/fixtures/hdfs-fixture/build.gradle b/test/fixtures/hdfs-fixture/build.gradle
index 3d63939f66ecb..70cb2b6bed2cb 100644
--- a/test/fixtures/hdfs-fixture/build.gradle
+++ b/test/fixtures/hdfs-fixture/build.gradle
@@ -20,7 +20,7 @@
apply plugin: 'elasticsearch.build'
versions << [
- 'hadoop2': '2.7.1'
+ 'hadoop2': '2.8.1'
]
// we create MiniHdfsCluster with the hadoop artifact
diff --git a/test/fixtures/hdfs-fixture/src/main/java/hdfs/MiniHDFS.java b/test/fixtures/hdfs-fixture/src/main/java/hdfs/MiniHDFS.java
index f868f4da97a2a..7d41d94e99a3d 100644
--- a/test/fixtures/hdfs-fixture/src/main/java/hdfs/MiniHDFS.java
+++ b/test/fixtures/hdfs-fixture/src/main/java/hdfs/MiniHDFS.java
@@ -26,6 +26,7 @@
import java.nio.file.Paths;
import java.nio.file.StandardCopyOption;
import java.util.ArrayList;
+import java.util.Arrays;
import java.util.List;
import org.apache.hadoop.conf.Configuration;
@@ -49,7 +50,8 @@ public class MiniHDFS {
public static void main(String[] args) throws Exception {
if (args.length != 1 && args.length != 3) {
- throw new IllegalArgumentException("MiniHDFS [ ]");
+ throw new IllegalArgumentException("Expected: MiniHDFS [ ], " +
+ "got: " + Arrays.toString(args));
}
boolean secure = args.length == 3;
@@ -83,6 +85,7 @@ public static void main(String[] args) throws Exception {
cfg.set(DFSConfigKeys.DFS_NAMENODE_KEYTAB_FILE_KEY, keytabFile);
cfg.set(DFSConfigKeys.DFS_DATANODE_KEYTAB_FILE_KEY, keytabFile);
cfg.set(DFSConfigKeys.DFS_NAMENODE_ACLS_ENABLED_KEY, "true");
+ cfg.set(DFSConfigKeys.DFS_BLOCK_ACCESS_TOKEN_ENABLE_KEY, "true");
cfg.set(DFSConfigKeys.IGNORE_SECURE_PORTS_FOR_TESTING_KEY, "true");
}