diff --git a/hadoop-client/pom.xml b/hadoop-client/pom.xml index c6f6c1b..7a3e0d4 100644 --- a/hadoop-client/pom.xml +++ b/hadoop-client/pom.xml @@ -60,20 +56,16 @@ commons-logging-api - jetty - org.mortbay.jetty - - - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server - org.mortbay.jetty + org.eclipse.jetty jetty-util - org.mortbay.jetty - servlet-api-2.5 + org.eclipse.jetty + jetty-webapp com.sun.jersey @@ -132,8 +124,8 @@ avro - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server com.sun.jersey diff --git a/hadoop-common-project/hadoop-auth/pom.xml b/hadoop-common-project/hadoop-auth/pom.xml index b9d6c60..9330a1a 100644 --- a/hadoop-common-project/hadoop-auth/pom.xml +++ b/hadoop-common-project/hadoop-auth/pom.xml @@ -53,13 +53,8 @@ test - org.mortbay.jetty - jetty-util - test - - - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-servlet test diff --git a/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java b/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java index 4e4ecc4..3429931 100644 --- a/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java +++ b/hadoop-common-project/hadoop-auth/src/test/java/org/apache/hadoop/security/authentication/client/AuthenticatorTestCase.java @@ -29,11 +29,12 @@ import org.apache.http.entity.InputStreamEntity; import org.apache.http.impl.auth.SPNegoSchemeFactory; import org.apache.http.impl.client.SystemDefaultHttpClient; import org.apache.http.util.EntityUtils; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.servlet.Context; -import org.mortbay.jetty.servlet.FilterHolder; -import org.mortbay.jetty.servlet.ServletHolder; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.servlet.FilterHolder; +import org.eclipse.jetty.servlet.ServletContextHandler; +import org.eclipse.jetty.servlet.ServletHolder; +import javax.servlet.DispatcherType; import javax.servlet.FilterConfig; import javax.servlet.ServletException; import javax.servlet.http.HttpServlet; @@ -52,6 +53,7 @@ import java.net.HttpURLConnection; import java.net.ServerSocket; import java.net.URL; import java.security.Principal; +import java.util.EnumSet; import java.util.Properties; import org.junit.Assert; @@ -61,7 +63,7 @@ public class AuthenticatorTestCase { private int port = -1; private boolean useTomcat = false; private Tomcat tomcat = null; - Context context; + ServletContextHandler context; private static Properties authenticatorConfig; @@ -120,10 +122,10 @@ public class AuthenticatorTestCase { protected void startJetty() throws Exception { server = new Server(0); - context = new Context(); + context = new ServletContextHandler(); context.setContextPath("/foo"); server.setHandler(context); - context.addFilter(new FilterHolder(TestFilter.class), "/*", 0); + context.addFilter(new FilterHolder(TestFilter.class), "/*", EnumSet.of(DispatcherType.REQUEST)); context.addServlet(new ServletHolder(TestServlet.class), "/bar"); host = "localhost"; port = getLocalPort(); diff --git a/hadoop-common-project/hadoop-common/pom.xml b/hadoop-common-project/hadoop-common/pom.xml index 7cf67a3..ef2733f 100644 --- a/hadoop-common-project/hadoop-common/pom.xml +++ b/hadoop-common-project/hadoop-common/pom.xml @@ -89,29 +89,25 @@ compile - javax.servlet - servlet-api + org.eclipse.jetty + jetty-server compile - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-servlet compile - org.mortbay.jetty - jetty-util + org.eclipse.jetty + jetty-webapp compile - org.mortbay.jetty - jetty-sslengine - compile - - - javax.servlet.jsp - jsp-api + org.apache.tomcat + tomcat-el-api + 7.0.37 runtime @@ -125,6 +134,11 @@ jersey-server compile + + com.sun.jersey + jersey-servlet + compile + commons-logging diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/AdminAuthorizedServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/AdminAuthorizedServlet.java index ef562b4..a4b05a1 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/AdminAuthorizedServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/AdminAuthorizedServlet.java @@ -23,7 +23,7 @@ import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; -import org.mortbay.jetty.servlet.DefaultServlet; +import org.eclipse.jetty.servlet.DefaultServlet; /** * General servlet which is admin-authorized. diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpRequestLog.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpRequestLog.java index 52d9850..a7c23b9 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpRequestLog.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpRequestLog.java @@ -25,8 +25,8 @@ import org.apache.commons.logging.LogFactory; import org.apache.log4j.Appender; import org.apache.log4j.Logger; -import org.mortbay.jetty.NCSARequestLog; -import org.mortbay.jetty.RequestLog; +import org.eclipse.jetty.server.NCSARequestLog; +import org.eclipse.jetty.server.RequestLog; /** * RequestLog object for use with Http diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java index 3ad26c6..f87c68a 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer.java @@ -62,27 +62,29 @@ import org.apache.hadoop.security.ssl.SSLFactory; import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.Shell; -import org.mortbay.io.Buffer; -import org.mortbay.jetty.Connector; -import org.mortbay.jetty.Handler; -import org.mortbay.jetty.MimeTypes; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.handler.ContextHandler; -import org.mortbay.jetty.handler.ContextHandlerCollection; -import org.mortbay.jetty.nio.SelectChannelConnector; -import org.mortbay.jetty.security.SslSocketConnector; -import org.mortbay.jetty.servlet.Context; -import org.mortbay.jetty.servlet.DefaultServlet; -import org.mortbay.jetty.servlet.FilterHolder; -import org.mortbay.jetty.servlet.FilterMapping; -import org.mortbay.jetty.servlet.ServletHandler; -import org.mortbay.jetty.servlet.ServletHolder; -import org.mortbay.jetty.webapp.WebAppContext; -import org.mortbay.thread.QueuedThreadPool; -import org.mortbay.util.MultiException; +import org.eclipse.jetty.io.Buffer; +import org.eclipse.jetty.server.Connector; +import org.eclipse.jetty.server.Handler; +import org.eclipse.jetty.http.MimeTypes; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.server.handler.ContextHandler; +import org.eclipse.jetty.server.handler.ContextHandlerCollection; +import org.eclipse.jetty.server.nio.SelectChannelConnector; +import org.eclipse.jetty.server.ssl.SslSocketConnector; +import org.eclipse.jetty.servlet.ServletContextHandler; +import org.eclipse.jetty.servlet.DefaultServlet; +import org.eclipse.jetty.servlet.FilterHolder; +import org.eclipse.jetty.servlet.FilterMapping; +import org.eclipse.jetty.servlet.ServletHandler; +import org.eclipse.jetty.servlet.ServletHolder; +import org.eclipse.jetty.util.ssl.SslContextFactory; +import org.eclipse.jetty.webapp.WebAppContext; +import org.eclipse.jetty.util.thread.QueuedThreadPool; +import org.eclipse.jetty.util.MultiException; import com.sun.jersey.spi.container.servlet.ServletContainer; + /** * Create a Jetty embedded server to answer http requests. The primary goal * is to serve up status information for the server. @@ -122,8 +124,8 @@ protected final Connector listener; protected final WebAppContext webAppContext; protected final boolean findPort; - protected final Map defaultContexts = - new HashMap(); + protected final Map defaultContexts = + new HashMap(); protected final List filterNames = new ArrayList(); private static final int MAX_RETRIES = 10; static final String STATE_DESCRIPTION_ALIVE = " - alive"; @@ -229,11 +231,18 @@ public HttpServer(String name, String bindAddress, int port, } catch (GeneralSecurityException ex) { throw new IOException(ex); } - SslSocketConnector sslListener = new SslSocketConnector() { - @Override - protected SSLServerSocketFactory createFactory() throws Exception { - return sslFactory.createSSLServerSocketFactory(); - } + // Jetty 8+ moved JKS config to SslContextFactory + SslContextFactory sslContextFactory = new SslContextFactory(conf.get("ssl.server.keystore.location","")); + sslContextFactory.setKeyStorePassword(conf.get("ssl.server.keystore.password","")); + if (sslFactory.isClientCertRequired()) { + sslContextFactory.setTrustStore(conf.get("ssl.server.truststore.location","")); + sslContextFactory.setTrustStorePassword(conf.get("ssl.server.truststore.password","")); + sslContextFactory.setTrustStoreType(conf.get("ssl.server.truststore.type", "jks")); + } + SslSocketConnector sslListener = new SslSocketConnector(sslContextFactory) { + protected SSLServerSocketFactory createFactory() throws Exception { + return sslFactory.createSSLServerSocketFactory(); + } }; listener = sslListener; } else { @@ -267,11 +276,15 @@ protected SSLServerSocketFactory createFactory() throws Exception { webAppContext.getServletContext().setAttribute(CONF_CONTEXT_ATTRIBUTE, conf); webAppContext.getServletContext().setAttribute(ADMINS_ACL, adminsAcl); addNoCacheFilter(webAppContext); - webServer.addHandler(webAppContext); + + ContextHandlerCollection handlers = new ContextHandlerCollection(); + handlers.setHandlers(webServer.getHandlers()); + handlers.addHandler(webAppContext); + webServer.setHandler(handlers); addDefaultApps(contexts, appDir, conf); - addGlobalFilter("safety", QuotingInputFilter.class.getName(), null); + addGlobalFilter("safety", QuotingInputFilter.class.getName(), new HashMap(0)); final FilterInitializer[] initializers = getFilterInitializers(conf); if (initializers != null) { conf = new Configuration(conf); @@ -320,7 +333,8 @@ public static Connector createDefaultChannelConnector() { // the same port with indeterminate routing of incoming requests to them ret.setReuseAddress(false); } - ret.setHeaderBufferSize(1024*64); + ret.setRequestHeaderSize(1024*64); + ret.setResponseHeaderSize(1024*64); return ret; } @@ -353,14 +367,14 @@ protected void addDefaultApps(ContextHandlerCollection parent, // set up the context for "/logs/" if "hadoop.log.dir" property is defined. String logDir = System.getProperty("hadoop.log.dir"); if (logDir != null) { - Context logContext = new Context(parent, "/logs"); + ServletContextHandler logContext = new ServletContextHandler(parent, "/logs"); logContext.setResourceBase(logDir); logContext.addServlet(AdminAuthorizedServlet.class, "/*"); if (conf.getBoolean( CommonConfigurationKeys.HADOOP_JETTY_LOGS_SERVE_ALIASES, CommonConfigurationKeys.DEFAULT_HADOOP_JETTY_LOGS_SERVE_ALIASES)) { logContext.getInitParams().put( - "org.mortbay.jetty.servlet.Default.aliases", "true"); + "org.eclipse.jetty.servlet.Default.aliases", "true"); } logContext.setDisplayName("logs"); setContextAttributes(logContext, conf); @@ -368,7 +382,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, defaultContexts.put(logContext, true); } // set up the context for "/static/*" - Context staticContext = new Context(parent, "/static"); + ServletContextHandler staticContext = new ServletContextHandler(parent, "/static"); staticContext.setResourceBase(appDir + "/static"); staticContext.addServlet(DefaultServlet.class, "/*"); staticContext.setDisplayName("static"); @@ -376,7 +390,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, defaultContexts.put(staticContext, true); } - private void setContextAttributes(Context context, Configuration conf) { + private void setContextAttributes(ServletContextHandler context, Configuration conf) { context.getServletContext().setAttribute(CONF_CONTEXT_ATTRIBUTE, conf); context.getServletContext().setAttribute(ADMINS_ACL, adminsAcl); } @@ -393,9 +407,12 @@ protected void addDefaultServlets() { addServlet("conf", "/conf", ConfServlet.class); } - public void addContext(Context ctxt, boolean isFiltered) + public void addContext(ServletContextHandler ctxt, boolean isFiltered) throws IOException { - webServer.addHandler(ctxt); + ContextHandlerCollection handlers = new ContextHandlerCollection(); + handlers.setHandlers(webServer.getHandlers()); + handlers.addHandler(ctxt); + webServer.setHandler(handlers); addNoCacheFilter(webAppContext); defaultContexts.put(ctxt, isFiltered); } @@ -497,7 +514,7 @@ public void addInternalServlet(String name, String pathSpec, FilterMapping fmap = new FilterMapping(); fmap.setPathSpec(pathSpec); fmap.setFilterName(SPNEGO_FILTER); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); handler.addFilterMapping(fmap); } } @@ -511,9 +528,9 @@ public void addFilter(String name, String classname, LOG.info("Added filter " + name + " (class=" + classname + ") to context " + webAppContext.getDisplayName()); final String[] ALL_URLS = { "/*" }; - for (Map.Entry e : defaultContexts.entrySet()) { + for (Map.Entry e : defaultContexts.entrySet()) { if (e.getValue()) { - Context ctx = e.getKey(); + ServletContextHandler ctx = e.getKey(); defineFilter(ctx, name, classname, parameters, ALL_URLS); LOG.info("Added filter " + name + " (class=" + classname + ") to context " + ctx.getDisplayName()); @@ -527,7 +544,7 @@ public void addGlobalFilter(String name, String classname, Map parameters) { final String[] ALL_URLS = { "/*" }; defineFilter(webAppContext, name, classname, parameters, ALL_URLS); - for (Context ctx : defaultContexts.keySet()) { + for (ServletContextHandler ctx : defaultContexts.keySet()) { defineFilter(ctx, name, classname, parameters, ALL_URLS); } LOG.info("Added global filter '" + name + "' (class=" + classname + ")"); @@ -536,16 +553,18 @@ public void addGlobalFilter(String name, String classname, /** * Define a filter for a context and set up default url mappings. */ - public void defineFilter(Context ctx, String name, + public void defineFilter(ServletContextHandler ctx, String name, String classname, Map parameters, String[] urls) { FilterHolder holder = new FilterHolder(); holder.setName(name); holder.setClassName(classname); - holder.setInitParameters(parameters); + if (null != parameters) { + holder.setInitParameters(parameters); + } FilterMapping fmap = new FilterMapping(); fmap.setPathSpecs(urls); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); fmap.setFilterName(name); ServletHandler handler = ctx.getServletHandler(); handler.addFilter(holder, fmap); @@ -557,13 +576,13 @@ public void defineFilter(Context ctx, String name, * @param webAppCtx The WebApplicationContext to add to */ protected void addFilterPathMapping(String pathSpec, - Context webAppCtx) { + ServletContextHandler webAppCtx) { ServletHandler handler = webAppCtx.getServletHandler(); for(String name : filterNames) { FilterMapping fmap = new FilterMapping(); fmap.setPathSpec(pathSpec); fmap.setFilterName(name); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); handler.addFilterMapping(fmap); } } @@ -627,12 +646,12 @@ public void addSslListener(InetSocketAddress addr, String keystore, if (webServer.isStarted()) { throw new IOException("Failed to add ssl listener"); } - SslSocketConnector sslListener = new SslSocketConnector(); + SslContextFactory sslContextFactory = new SslContextFactory(keystore); + sslContextFactory.setKeyStorePassword(storPass); + sslContextFactory.setKeyManagerPassword(keyPass); + SslSocketConnector sslListener = new SslSocketConnector(sslContextFactory); sslListener.setHost(addr.getHostName()); sslListener.setPort(addr.getPort()); - sslListener.setKeystore(keystore); - sslListener.setPassword(storPass); - sslListener.setKeyPassword(keyPass); webServer.addConnector(sslListener); } @@ -656,14 +675,14 @@ public void addSslListener(InetSocketAddress addr, Configuration sslConf, System.setProperty("javax.net.ssl.trustStoreType", sslConf.get( "ssl.server.truststore.type", "jks")); } - SslSocketConnector sslListener = new SslSocketConnector(); + SslContextFactory sslContextFactory = new SslContextFactory(sslConf.get("ssl.server.keystore.location","")); + sslContextFactory.setKeyStorePassword(sslConf.get("ssl.server.keystore.password", "")); + sslContextFactory.setKeyManagerPassword(sslConf.get("ssl.server.keystore.keypassword", "")); + sslContextFactory.setKeyStoreType(sslConf.get("ssl.server.keystore.type", "jks")); + sslContextFactory.setNeedClientAuth(needCertsAuth); + SslSocketConnector sslListener = new SslSocketConnector(sslContextFactory); sslListener.setHost(addr.getHostName()); sslListener.setPort(addr.getPort()); - sslListener.setKeystore(sslConf.get("ssl.server.keystore.location")); - sslListener.setPassword(sslConf.get("ssl.server.keystore.password", "")); - sslListener.setKeyPassword(sslConf.get("ssl.server.keystore.keypassword", "")); - sslListener.setKeystoreType(sslConf.get("ssl.server.keystore.type", "jks")); - sslListener.setNeedClientAuth(needCertsAuth); webServer.addConnector(sslListener); } @@ -1095,8 +1114,8 @@ public void doFilter(ServletRequest request, */ private String inferMimeType(ServletRequest request) { String path = ((HttpServletRequest)request).getRequestURI(); - ContextHandler.SContext sContext = (ContextHandler.SContext)config.getServletContext(); - MimeTypes mimes = sContext.getContextHandler().getMimeTypes(); + ContextHandler.Context context = (ContextHandler.Context)config.getServletContext(); + MimeTypes mimes = context.getContextHandler().getMimeTypes(); Buffer mimeBuffer = mimes.getMimeByExtension(path); return (mimeBuffer == null) ? null : mimeBuffer.toString(); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java index 2f28d08..3ac7086 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/http/HttpServer2.java @@ -40,6 +40,7 @@ import javax.servlet.ServletException; import javax.servlet.ServletRequest; import javax.servlet.ServletResponse; +import javax.servlet.SessionCookieConfig; import javax.servlet.http.HttpServlet; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletRequestWrapper; @@ -59,7 +60,6 @@ import org.apache.hadoop.security.authentication.util.FileSignerSecretProvider; import org.apache.hadoop.security.authentication.util.RandomSignerSecretProvider; import org.apache.hadoop.security.authentication.util.SignerSecretProvider; import org.apache.hadoop.security.authentication.util.ZKSignerSecretProvider; -import org.apache.hadoop.security.ssl.SslSelectChannelConnectorSecure; import org.apache.hadoop.jmx.JMXJsonServlet; import org.apache.hadoop.log.LogLevel; import org.apache.hadoop.metrics.MetricsServlet; @@ -70,29 +71,31 @@ import org.apache.hadoop.security.authorize.AccessControlList; import org.apache.hadoop.util.ReflectionUtils; import org.apache.hadoop.util.Shell; -import org.mortbay.io.Buffer; -import org.mortbay.jetty.Connector; -import org.mortbay.jetty.Handler; -import org.mortbay.jetty.MimeTypes; -import org.mortbay.jetty.RequestLog; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.SessionManager; -import org.mortbay.jetty.handler.ContextHandler; -import org.mortbay.jetty.handler.ContextHandlerCollection; -import org.mortbay.jetty.handler.HandlerCollection; -import org.mortbay.jetty.handler.RequestLogHandler; -import org.mortbay.jetty.nio.SelectChannelConnector; -import org.mortbay.jetty.security.SslSelectChannelConnector; -import org.mortbay.jetty.servlet.AbstractSessionManager; -import org.mortbay.jetty.servlet.Context; -import org.mortbay.jetty.servlet.DefaultServlet; -import org.mortbay.jetty.servlet.FilterHolder; -import org.mortbay.jetty.servlet.FilterMapping; -import org.mortbay.jetty.servlet.ServletHandler; -import org.mortbay.jetty.servlet.ServletHolder; -import org.mortbay.jetty.webapp.WebAppContext; -import org.mortbay.thread.QueuedThreadPool; -import org.mortbay.util.MultiException; +import org.eclipse.jetty.http.MimeTypes; +import org.eclipse.jetty.io.Buffer; +import org.eclipse.jetty.server.Connector; +import org.eclipse.jetty.server.Handler; +import org.eclipse.jetty.server.RequestLog; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.server.SessionManager; +import org.eclipse.jetty.server.handler.ContextHandler; +import org.eclipse.jetty.server.handler.ContextHandlerCollection; +import org.eclipse.jetty.server.handler.HandlerCollection; +import org.eclipse.jetty.server.handler.RequestLogHandler; +import org.eclipse.jetty.server.nio.SelectChannelConnector; +import org.eclipse.jetty.server.ssl.SslSelectChannelConnector; +import org.eclipse.jetty.server.session.AbstractSessionManager; +import org.eclipse.jetty.server.ssl.SslSocketConnector; +import org.eclipse.jetty.servlet.DefaultServlet; +import org.eclipse.jetty.servlet.FilterHolder; +import org.eclipse.jetty.servlet.FilterMapping; +import org.eclipse.jetty.servlet.ServletContextHandler; +import org.eclipse.jetty.servlet.ServletHandler; +import org.eclipse.jetty.servlet.ServletHolder; +import org.eclipse.jetty.util.MultiException; +import org.eclipse.jetty.util.ssl.SslContextFactory; +import org.eclipse.jetty.util.thread.QueuedThreadPool; +import org.eclipse.jetty.webapp.WebAppContext; import com.google.common.base.Preconditions; import com.google.common.collect.Lists; @@ -135,8 +137,8 @@ private ListenerInfo(boolean isManaged, Connector listener) { protected final WebAppContext webAppContext; protected final boolean findPort; - protected final Map defaultContexts = - new HashMap<>(); + protected final Map defaultContexts = + new HashMap(); protected final List filterNames = new ArrayList<>(); static final String STATE_DESCRIPTION_ALIVE = " - alive"; static final String STATE_DESCRIPTION_NOT_LIVE = " - not live"; @@ -320,23 +322,24 @@ public final class HttpServer2 implements FilterContainer { } private Connector createHttpsChannelConnector() { - SslSelectChannelConnector c = new SslSelectChannelConnectorSecure(); - configureChannelConnector(c); - - c.setNeedClientAuth(needsClientAuth); - c.setKeyPassword(keyPassword); + // Jetty 8+ moved JKS config to SslContextFactory + SslContextFactory scf = new SslContextFactory(); + scf.setNeedClientAuth(needsClientAuth); + scf.setKeyManagerPassword(keyPassword); if (keyStore != null) { - c.setKeystore(keyStore); - c.setKeystoreType(keyStoreType); - c.setPassword(keyStorePassword); + scf.setKeyStorePath(keyStore); + scf.setKeyStoreType(keyStoreType); + scf.setKeyStorePassword(keyStorePassword); } if (trustStore != null) { - c.setTruststore(trustStore); - c.setTruststoreType(trustStoreType); - c.setTrustPassword(trustStorePassword); + scf.setTrustStore(trustStore); + scf.setTrustStoreType(trustStoreType); + scf.setTrustStorePassword(trustStorePassword); } + SslSelectChannelConnector c = new SslSelectChannelConnector(scf); + configureChannelConnector(c); if(null != excludeCiphers && !excludeCiphers.isEmpty()) { c.setExcludeCipherSuites(excludeCiphers.split(",")); @@ -369,7 +373,8 @@ private void initializeWebServer(String name, String hostName, if (sm instanceof AbstractSessionManager) { AbstractSessionManager asm = (AbstractSessionManager)sm; asm.setHttpOnly(true); - asm.setSecureCookies(true); + SessionCookieConfig scc = asm.getSessionCookieConfig(); + scc.setSecure(true); } ContextHandlerCollection contexts = new ContextHandlerCollection(); @@ -387,11 +392,14 @@ private void initializeWebServer(String name, String hostName, final String appDir = getWebAppsPath(name); - webServer.addHandler(webAppContext); + ContextHandlerCollection handlers = new ContextHandlerCollection(); + handlers.setHandlers(webServer.getHandlers()); + handlers.addHandler(webAppContext); + webServer.setHandler(handlers); addDefaultApps(contexts, appDir, conf); - addGlobalFilter("safety", QuotingInputFilter.class.getName(), null); + addGlobalFilter("safety", QuotingInputFilter.class.getName(), new HashMap(0)); final FilterInitializer[] initializers = getFilterInitializers(conf); if (initializers != null) { conf = new Configuration(conf); @@ -522,7 +528,8 @@ public final class HttpServer2 implements FilterContainer { // the same port with indeterminate routing of incoming requests to them c.setReuseAddress(false); } - c.setHeaderBufferSize(1024*64); + c.setRequestHeaderSize(1024*64); + c.setResponseHeaderSize(1024*64); } @InterfaceAudience.Private @@ -541,7 +550,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, // set up the context for "/logs/" if "hadoop.log.dir" property is defined. String logDir = System.getProperty("hadoop.log.dir"); if (logDir != null) { - Context logContext = new Context(parent, "/logs"); + ServletContextHandler logContext = new ServletContextHandler(parent, "/logs"); logContext.setResourceBase(logDir); logContext.addServlet(AdminAuthorizedServlet.class, "/*"); if (conf.getBoolean( @@ -550,7 +559,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, @SuppressWarnings("unchecked") Map params = logContext.getInitParams(); params.put( - "org.mortbay.jetty.servlet.Default.aliases", "true"); + "org.eclipse.jetty.servlet.Default.aliases", "true"); } logContext.setDisplayName("logs"); setContextAttributes(logContext, conf); @@ -558,7 +567,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, defaultContexts.put(logContext, true); } // set up the context for "/static/*" - Context staticContext = new Context(parent, "/static"); + ServletContextHandler staticContext = new ServletContextHandler(parent, "/static"); staticContext.setResourceBase(appDir + "/static"); staticContext.addServlet(DefaultServlet.class, "/*"); staticContext.setDisplayName("static"); @@ -566,7 +575,7 @@ protected void addDefaultApps(ContextHandlerCollection parent, defaultContexts.put(staticContext, true); } - private void setContextAttributes(Context context, Configuration conf) { + private void setContextAttributes(ServletContextHandler context, Configuration conf) { context.getServletContext().setAttribute(CONF_CONTEXT_ATTRIBUTE, conf); context.getServletContext().setAttribute(ADMINS_ACL, adminsAcl); } @@ -583,8 +592,11 @@ protected void addDefaultServlets() { addServlet("conf", "/conf", ConfServlet.class); } - public void addContext(Context ctxt, boolean isFiltered) { + public void addContext(ServletContextHandler ctxt, boolean isFiltered) { - webServer.addHandler(ctxt); + ContextHandlerCollection handlers = new ContextHandlerCollection(); + handlers.setHandlers(webServer.getHandlers()); + handlers.addHandler(ctxt); + webServer.setHandler(handlers); addNoCacheFilter(webAppContext); defaultContexts.put(ctxt, isFiltered); } @@ -669,7 +681,7 @@ public void addInternalServlet(String name, String pathSpec, FilterMapping fmap = new FilterMapping(); fmap.setPathSpec(pathSpec); fmap.setFilterName(SPNEGO_FILTER); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); handler.addFilterMapping(fmap); } } @@ -686,9 +686,9 @@ public final class HttpServer2 implements FilterContainer { "Added filter " + name + " (class=" + classname + ") to context " + webAppContext.getDisplayName()); final String[] ALL_URLS = { "/*" }; fmap = getFilterMapping(name, ALL_URLS); - for (Map.Entry e : defaultContexts.entrySet()) { + for (Map.Entry e : defaultContexts.entrySet()) { if (e.getValue()) { - Context ctx = e.getKey(); + ServletContextHandler ctx = e.getKey(); defineFilter(ctx, filterHolder, fmap); LOG.info("Added filter " + name + " (class=" + classname + ") to context " + ctx.getDisplayName()); @@ -704,7 +704,7 @@ public final class HttpServer2 implements FilterContainer { FilterHolder filterHolder = getFilterHolder(name, classname, parameters); FilterMapping fmap = getFilterMapping(name, ALL_URLS); defineFilter(webAppContext, filterHolder, fmap); - for (Context ctx : defaultContexts.keySet()) { + for (ServletContextHandler ctx : defaultContexts.keySet()) { defineFilter(ctx, filterHolder, fmap); } LOG.info("Added global filter '" + name + "' (class=" + classname + ")"); @@ -713,7 +725,7 @@ /** * Define a filter for a context and set up default url mappings. */ - public static void defineFilter(Context ctx, String name, + public static void defineFilter(ServletContextHandler ctx, String name, String classname, Map parameters, String[] urls) { FilterHolder filterHolder = getFilterHolder(name, classname, parameters); FilterMapping fmap = getFilterMapping(name, urls); @@ -723,7 +734,7 @@ public final class HttpServer2 implements FilterContainer { /** * Define a filter for a context and set up default url mappings. */ - private static void defineFilter(Context ctx, FilterHolder holder, + private static void defineFilter(ServletContextHandler ctx, FilterHolder holder, FilterMapping fmap) { ServletHandler handler = ctx.getServletHandler(); handler.addFilter(holder, fmap); @@ -732,7 +732,7 @@ private static FilterMapping getFilterMapping(String name, String[] urls) { FilterMapping fmap = new FilterMapping(); fmap.setPathSpecs(urls); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); fmap.setFilterName(name); return fmap; } @@ -752,13 +752,13 @@ * @param webAppCtx The WebApplicationContext to add to */ protected void addFilterPathMapping(String pathSpec, - Context webAppCtx) { + ServletContextHandler webAppCtx) { ServletHandler handler = webAppCtx.getServletHandler(); for(String name : filterNames) { FilterMapping fmap = new FilterMapping(); fmap.setPathSpec(pathSpec); fmap.setFilterName(name); - fmap.setDispatches(Handler.ALL); + fmap.setDispatches(FilterMapping.ALL); handler.addFilterMapping(fmap); } } @@ -812,7 +823,8 @@ public InetSocketAddress getConnectorAddress(int index) { return null; Connector c = webServer.getConnectors()[index]; - if (c.getLocalPort() == -1) { + // jetty8 has 2 getLocalPort err values + if (c.getLocalPort() == -1 || c.getLocalPort() == -2) { // The connector is not bounded return null; } @@ -901,7 +901,7 @@ public final class HttpServer2 implements FilterContainer { */ void openListeners() throws Exception { for (Connector listener : listeners) { - if (listener.getLocalPort() != -1) { + if (listener.getLocalPort() != -1 && listener.getLocalPort() != -2) { // This listener is either started externally or has been bound continue; } @@ -1249,8 +1262,8 @@ public void doFilter(ServletRequest request, */ private String inferMimeType(ServletRequest request) { String path = ((HttpServletRequest)request).getRequestURI(); - ContextHandler.SContext sContext = (ContextHandler.SContext)config.getServletContext(); - MimeTypes mimes = sContext.getContextHandler().getMimeTypes(); + ContextHandler.Context context = (ContextHandler.Context)config.getServletContext(); + MimeTypes mimes = context.getContextHandler().getMimeTypes(); Buffer mimeBuffer = mimes.getMimeByExtension(path); return (mimeBuffer == null) ? null : mimeBuffer.toString(); } diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java index 8f5dcd1..a78318a 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics/MetricsServlet.java @@ -36,8 +36,8 @@ import org.apache.hadoop.metrics.spi.OutputRecord; import org.apache.hadoop.metrics.spi.AbstractMetricsContext.MetricMap; import org.apache.hadoop.metrics.spi.AbstractMetricsContext.TagMap; -import org.mortbay.util.ajax.JSON; -import org.mortbay.util.ajax.JSON.Output; +import org.eclipse.jetty.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON.Output; /** * A servlet to print out metrics data. By default, the servlet returns a diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfServlet.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfServlet.java index 1c22ee6..90846d9 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfServlet.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/TestConfServlet.java @@ -23,7 +23,7 @@ import javax.xml.parsers.DocumentBuilder; import javax.xml.parsers.DocumentBuilderFactory; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import org.w3c.dom.Document; import org.w3c.dom.Element; import org.w3c.dom.Node; @@ -107,4 +107,4 @@ public void testBadFormat() throws Exception { } assertEquals("", sw.toString()); } -} \ No newline at end of file +} diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FSMainOperationsBaseTest.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FSMainOperationsBaseTest.java index e862db4..7f73534 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FSMainOperationsBaseTest.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/FSMainOperationsBaseTest.java @@ -32,7 +32,7 @@ import org.junit.Assert; import org.junit.Before; import org.junit.Test; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; /** *

diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemTestSetup.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemTestSetup.java index 81ca210..6ec331f 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemTestSetup.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemTestSetup.java @@ -27,7 +27,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.viewfs.ConfigUtil; import org.apache.hadoop.util.Shell; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; /** diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsTestSetup.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsTestSetup.java index 92bcbc3..3726e83 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsTestSetup.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsTestSetup.java @@ -26,7 +26,7 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.viewfs.ConfigUtil; import org.apache.hadoop.util.Shell; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; /** diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestGlobalFilter.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestGlobalFilter.java index 0e4a1ca..e31adff 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestGlobalFilter.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestGlobalFilter.java @@ -22,6 +22,7 @@ import java.io.InputStreamReader; import java.net.URL; import java.net.URLConnection; +import java.util.HashMap; import java.util.Set; import java.util.TreeSet; @@ -75,7 +76,7 @@ public Initializer() {} @Override public void initFilter(FilterContainer container, Configuration conf) { - container.addGlobalFilter("recording", RecordingFilter.class.getName(), null); + container.addGlobalFilter("recording", RecordingFilter.class.getName(), new HashMap(0)); } } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpCookieFlag.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpCookieFlag.java index c0aaf64..a29e275 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpCookieFlag.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpCookieFlag.java @@ -37,6 +37,7 @@ import java.net.URI; import java.net.URL; import java.security.GeneralSecurityException; import java.net.HttpCookie; +import java.util.HashMap; import java.util.List; public class TestHttpCookieFlag { @@ -70,7 +71,7 @@ public void destroy() { @Override public void initFilter(FilterContainer container, Configuration conf) { container.addFilter("DummyAuth", DummyAuthenticationFilter.class - .getName(), null); + .getName(), new HashMap(0)); } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpRequestLog.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpRequestLog.java index 23e0d3e..24be3fe 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpRequestLog.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestHttpRequestLog.java @@ -19,8 +19,8 @@ import org.apache.log4j.Logger; import org.junit.Test; -import org.mortbay.jetty.NCSARequestLog; -import org.mortbay.jetty.RequestLog; +import org.eclipse.jetty.server.NCSARequestLog; +import org.eclipse.jetty.server.RequestLog; import static org.junit.Assert.assertEquals; import static org.junit.Assert.assertNotNull; diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestPathFilter.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestPathFilter.java index 09f31df..be80795 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestPathFilter.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestPathFilter.java @@ -22,6 +22,7 @@ import java.io.InputStreamReader; import java.net.URL; import java.net.URLConnection; +import java.util.HashMap; import java.util.Set; import java.util.TreeSet; @@ -75,7 +76,7 @@ public Initializer() {} @Override public void initFilter(FilterContainer container, Configuration conf) { - container.addFilter("recording", RecordingFilter.class.getName(), null); + container.addFilter("recording", RecordingFilter.class.getName(), new HashMap(0)); } } } diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestServletFilter.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestServletFilter.java index 6b17ccc..8f354d3 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestServletFilter.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/TestServletFilter.java @@ -22,6 +22,7 @@ import java.io.InputStreamReader; import java.net.URL; import java.net.URLConnection; +import java.util.HashMap; import java.util.Random; import javax.servlet.Filter; @@ -74,7 +75,7 @@ public Initializer() {} @Override public void initFilter(FilterContainer container, Configuration conf) { - container.addFilter("simple", SimpleFilter.class.getName(), null); + container.addFilter("simple", SimpleFilter.class.getName(), new HashMap(0)); } } } @@ -157,7 +158,7 @@ public Initializer() { } public void initFilter(FilterContainer container, Configuration conf) { - container.addFilter("simple", ErrorFilter.class.getName(), null); + container.addFilter("simple", ErrorFilter.class.getName(), new HashMap(0)); } } } @@ -173,8 +174,7 @@ public void testServletFilterWhenInitThrowsException() throws Exception { http.start(); fail("expecting exception"); } catch (IOException e) { - assertTrue(e.getMessage().contains( - "Problem in starting http server. Server handlers failed")); + assertTrue(e.getMessage().toLowerCase().contains("problem")); } } @@ -189,7 +189,7 @@ public void testContextSpecificServletFilterWhenInitThrowsException() HttpServer2 http = createTestServer(conf); HttpServer2.defineFilter(http.webAppContext, "ErrorFilter", ErrorFilter.class.getName(), - null, null); + new HashMap(0), null); try { http.start(); fail("expecting exception"); diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/resource/JerseyResource.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/resource/JerseyResource.java index f1313e2..52ea9b9 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/resource/JerseyResource.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/http/resource/JerseyResource.java @@ -32,7 +32,7 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * A simple Jersey resource class TestHttpServer. diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics/TestMetricsServlet.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics/TestMetricsServlet.java index ec54f59..d289a03 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics/TestMetricsServlet.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics/TestMetricsServlet.java @@ -30,7 +30,7 @@ import org.apache.hadoop.metrics.MetricsServlet.TagsMetricsPair; import org.apache.hadoop.metrics.spi.NoEmitMetricsContext; import org.apache.hadoop.metrics.spi.OutputRecord; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; public class TestMetricsServlet extends TestCase { MetricsContext nc1; diff --git a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestClassUtil.java b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestClassUtil.java index fe1284f..91c13a8 100644 --- a/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestClassUtil.java +++ b/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestClassUtil.java @@ -32,9 +32,9 @@ public void testFindContainingJar() { Assert.assertNotNull("Containing jar not found for Logger", containingJar); File jarFile = new File(containingJar); - Assert.assertTrue("Containing jar does not exist on file system", + Assert.assertTrue("Containing jar does not exist on file system ", jarFile.exists()); - Assert.assertTrue("Incorrect jar file" + containingJar, - jarFile.getName().matches("log4j.+[.]jar")); + Assert.assertTrue("Incorrect jar file " + containingJar, + jarFile.getName().matches("log4j.*[.]jar")); } } diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml index d01a32f..d85405b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/pom.xml @@ -45,7 +45,7 @@ LOCALHOST **/TestHttpFSWithKerberos.java - http://archive.apache.org/dist/tomcat/tomcat-6/v${tomcat.version}/bin/apache-tomcat-${tomcat.version}.tar.gz + http://archive.apache.org/dist/tomcat/tomcat-7/v${tomcat.version}/bin/apache-tomcat-${tomcat.version}.tar.gz @@ -90,8 +90,8 @@ compile - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server test @@ -128,16 +124,16 @@ jsp-api - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server - org.mortbay.jetty + org.eclipse.jetty jetty-util - org.mortbay.jetty - servlet-api-2.5 + org.eclipse.jetty + jetty-webapp net.java.dev.jets3t @@ -158,16 +150,16 @@ servlet-api - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server - org.mortbay.jetty + org.eclipse.jetty jetty-util - org.mortbay.jetty - servlet-api-2.5 + org.eclipse.jetty + jetty-webapp net.java.dev.jets3t diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/server.xml b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/server.xml index a425bdd..39c60f5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/server.xml +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/main/tomcat/server.xml @@ -1,7 +1,7 @@ - + - + - + - - - - + + + + redirectPort="8443" /> + every request. The Engine implementation for Tomcat stand alone + analyzes the HTTP headers included with the request, and passes them + on to the appropriate Host (virtual host). + Documentation at /docs/config/engine.html --> - - + + + + + - - - - - + @@ -138,11 +132,11 @@ --> - - --> + prefix="localhost_access_log." suffix=".txt" + pattern="%h %l %u %t "%r" %s %b" /> diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java index d512897..b277973 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/client/BaseTestHttpFSWith.java @@ -46,8 +46,8 @@ import org.junit.Assert; import org.junit.Test; import org.junit.runner.RunWith; import org.junit.runners.Parameterized; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.webapp.WebAppContext; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.webapp.WebAppContext; import com.google.common.collect.Lists; @@ -108,7 +108,7 @@ private void createHttpFSServer() throws Exception { URL url = cl.getResource("webapp"); WebAppContext context = new WebAppContext(url.getPath(), "/webhdfs"); Server server = TestJettyHelper.getJettyServer(); - server.addHandler(context); + server.setHandler(context); server.start(); } diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSServer.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSServer.java index 48cca42..f893127 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSServer.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSServer.java @@ -64,8 +64,8 @@ import org.apache.hadoop.test.TestJettyHelper; import org.json.simple.JSONObject; import org.json.simple.parser.JSONParser; import org.junit.Test; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.webapp.WebAppContext; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.webapp.WebAppContext; import com.google.common.collect.Maps; import java.util.Properties; @@ -157,7 +157,7 @@ private void createHttpFSServer(boolean addDelegationTokenAuthHandler) URL url = cl.getResource("webapp"); WebAppContext context = new WebAppContext(url.getPath(), "/webhdfs"); Server server = TestJettyHelper.getJettyServer(); - server.addHandler(context); + server.setHandler(context); server.start(); if (addDelegationTokenAuthHandler) { HttpFSServerWebApp.get().setAuthority(TestJettyHelper.getAuthority()); diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSWithKerberos.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSWithKerberos.java index 45ce8ed..6076b1f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSWithKerberos.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/fs/http/server/TestHttpFSWithKerberos.java @@ -41,8 +41,8 @@ import org.junit.After; import org.junit.Assert; import org.junit.Test; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.webapp.WebAppContext; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.webapp.WebAppContext; import java.io.File; import java.io.FileOutputStream; @@ -105,7 +105,7 @@ private void createHttpFSServer() throws Exception { URL url = cl.getResource("webapp"); WebAppContext context = new WebAppContext(url.getPath(), "/webhdfs"); Server server = TestJettyHelper.getJettyServer(); - server.addHandler(context); + server.setHandler(context); server.start(); HttpFSServerWebApp.get().setAuthority(TestJettyHelper.getAuthority()); } diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHFSTestCase.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHFSTestCase.java index eb2cdc6..3d13cf5 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHFSTestCase.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHFSTestCase.java @@ -39,8 +39,8 @@ import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.Time; import org.junit.Test; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.servlet.Context; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.servlet.ServletContextHandler; public class TestHFSTestCase extends HFSTestCase { @@ -165,11 +165,11 @@ protected void doGet(HttpServletRequest req, HttpServletResponse resp) throws Se @Test @TestJetty public void testJetty() throws Exception { - Context context = new Context(); + ServletContextHandler context = new ServletContextHandler(); context.setContextPath("/"); context.addServlet(MyServlet.class, "/bar"); Server server = TestJettyHelper.getJettyServer(); - server.addHandler(context); + server.setHandler(context); server.start(); URL url = new URL(TestJettyHelper.getJettyURL(), "/bar"); HttpURLConnection conn = (HttpURLConnection) url.openConnection(); diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHTestCase.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHTestCase.java index 74d34ec..8b7223a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHTestCase.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestHTestCase.java @@ -34,8 +34,8 @@ import org.apache.hadoop.util.Time; import org.junit.Test; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.servlet.Context; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.servlet.ServletContextHandler; public class TestHTestCase extends HTestCase { @@ -132,11 +132,11 @@ protected void doGet(HttpServletRequest req, HttpServletResponse resp) throws Se @Test @TestJetty public void testJetty() throws Exception { - Context context = new Context(); + ServletContextHandler context = new ServletContextHandler(); context.setContextPath("/"); context.addServlet(MyServlet.class, "/bar"); Server server = TestJettyHelper.getJettyServer(); - server.addHandler(context); + server.setHandler(context); server.start(); URL url = new URL(TestJettyHelper.getJettyURL(), "/bar"); HttpURLConnection conn = (HttpURLConnection) url.openConnection(); diff --git a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java index b0f14f4..fb81ab2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java +++ b/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/test/TestJettyHelper.java @@ -29,9 +29,9 @@ import org.junit.Test; import org.junit.rules.MethodRule; import org.junit.runners.model.FrameworkMethod; import org.junit.runners.model.Statement; -import org.mortbay.jetty.Connector; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.security.SslSelectChannelConnector; +import org.eclipse.jetty.server.Connector; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.server.ssl.SslSelectChannelConnector; public class TestJettyHelper implements MethodRule { private boolean ssl; diff --git a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml index 5ee5841..8b56730 100644 --- a/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs-nfs/pom.xml @@ -85,12 +85,12 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd"> compile - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server compile - org.mortbay.jetty + org.eclipse.jetty jetty-util compile diff --git a/hadoop-hdfs-project/hadoop-hdfs/pom.xml b/hadoop-hdfs-project/hadoop-hdfs/pom.xml index 420e5d2..c134d71 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/pom.xml +++ b/hadoop-hdfs-project/hadoop-hdfs/pom.xml @@ -70,15 +70,26 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd"> compile - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server compile - org.mortbay.jetty + org.eclipse.jetty jetty-util compile + + org.eclipse.jetty + jetty-util-ajax + compile + + + org.eclipse.jetty + jetty-util + + + com.sun.jersey jersey-core @@ -136,11 +131,6 @@ http://maven.apache.org/xsd/maven-4.0.0.xsd"> compile - javax.servlet - servlet-api - compile - - junit junit test diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java index 50b44f8..d5a91d3 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/qjournal/server/JournalNode.java @@ -46,7 +46,7 @@ import org.apache.hadoop.util.StringUtils; import org.apache.hadoop.util.Tool; import org.apache.hadoop.util.ToolRunner; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import com.google.common.base.Preconditions; import com.google.common.collect.Maps; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java index fc85a5e..1610c8c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java @@ -202,7 +202,7 @@ import org.apache.hadoop.util.VersionInfo; import org.apache.hadoop.tracing.SpanReceiverHost; import org.apache.hadoop.tracing.SpanReceiverInfo; import org.apache.hadoop.tracing.TraceAdminProtocol; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Joiner; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index 4232e00..3386dff 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -289,7 +289,7 @@ import org.apache.log4j.Appender; import org.apache.log4j.AsyncAppender; import org.apache.log4j.Logger; import org.codehaus.jackson.map.ObjectMapper; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Charsets; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/StreamFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/StreamFile.java index aa4ba5d..5b945ba 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/StreamFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/StreamFile.java @@ -39,7 +39,7 @@ import org.apache.hadoop.io.IOUtils; import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.util.ServletUtil; -import org.mortbay.jetty.InclusiveByteRange; +import org.eclipse.jetty.server.InclusiveByteRange; @InterfaceAudience.Private public class StreamFile extends DfsServlet { diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/server/TestJournalNodeMXBean.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/server/TestJournalNodeMXBean.java index 3471848..b4e0202 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/server/TestJournalNodeMXBean.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/qjournal/server/TestJournalNodeMXBean.java @@ -34,7 +34,7 @@ import org.junit.After; import org.junit.Before; import org.junit.Test; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * Test {@link JournalNodeMXBean} diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystemMBean.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystemMBean.java index db8f92e..79d9003 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystemMBean.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestFSNamesystemMBean.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.metrics2.impl.ConfigBuilder; import org.apache.hadoop.metrics2.impl.TestMetricsConfig; import org.junit.Test; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * Class for testing {@link NameNodeMXBean} implementation diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java index d459d30..6327a83 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestNameNodeMXBean.java @@ -32,7 +32,7 @@ import org.apache.hadoop.io.nativeio.NativeIO.POSIX.NoMlockCacheManipulator; import org.apache.hadoop.util.VersionInfo; import org.codehaus.jackson.map.ObjectMapper; import org.junit.Test; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import javax.management.MBeanServer; import javax.management.ObjectName; diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartupProgressServlet.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartupProgressServlet.java index 0f22e9a..bff549a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartupProgressServlet.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartupProgressServlet.java @@ -36,7 +36,7 @@ import org.apache.hadoop.hdfs.server.namenode.startupprogress.StartupProgress; import org.junit.Before; import org.junit.Test; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; public class TestStartupProgressServlet { diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java index f24b801..28d05b4 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStreamFile.java @@ -46,7 +46,7 @@ import org.apache.hadoop.net.NetUtils; import org.junit.Test; import org.mockito.Mockito; -import org.mortbay.jetty.InclusiveByteRange; +import org.eclipse.jetty.server.InclusiveByteRange; /* * Mock input stream class that always outputs the current position of the stream. diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/test/MiniDFSClusterManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/test/MiniDFSClusterManager.java index 7029f42..c7023c9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/test/MiniDFSClusterManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/test/MiniDFSClusterManager.java @@ -38,7 +38,7 @@ import org.apache.hadoop.hdfs.HdfsConfiguration; import org.apache.hadoop.hdfs.MiniDFSCluster; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * This class drives the creation of a mini-cluster on the local machine. By diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java index 981e6ff..7864756 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/src/main/java/org/apache/hadoop/mapreduce/v2/app/JobEndNotifier.java @@ -30,7 +30,8 @@ import org.apache.hadoop.mapred.JobContext; import org.apache.hadoop.mapreduce.MRJobConfig; import org.apache.hadoop.mapreduce.v2.api.records.JobReport; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; +import org.eclipse.jetty.util.log.Logger; /** *

This class handles job end notification. Submitters of jobs can choose to @@ -48,6 +49,7 @@ import org.eclipse.jetty.util.log.Log; public class JobEndNotifier implements Configurable { private static final String JOB_ID = "$jobId"; private static final String JOB_STATUS = "$jobStatus"; + private static Logger log = Log.getLogger(JobEndNotifier.class); private Configuration conf; protected String userUrl; @@ -101,10 +103,10 @@ public class JobEndNotifier implements Configurable { int port = Integer.parseInt(portConf); proxyToUse = new Proxy(proxyType, new InetSocketAddress(hostname, port)); - Log.info("Job end notification using proxy type \"" + proxyType + + log.info("Job end notification using proxy type \"" + proxyType + "\" hostname \"" + hostname + "\" and port \"" + port + "\""); } catch(NumberFormatException nfe) { - Log.warn("Job end notification couldn't parse configured proxy's port " + log.warn("Job end notification couldn't parse configured proxy's port " + portConf + ". Not going to use a proxy"); } } @@ -121,23 +123,23 @@ public class JobEndNotifier implements Configurable { protected boolean notifyURLOnce() { boolean success = false; try { - Log.info("Job end notification trying " + urlToNotify); + log.info("Job end notification trying " + urlToNotify); HttpURLConnection conn = (HttpURLConnection) urlToNotify.openConnection(proxyToUse); conn.setConnectTimeout(timeout); conn.setReadTimeout(timeout); conn.setAllowUserInteraction(false); if(conn.getResponseCode() != HttpURLConnection.HTTP_OK) { - Log.warn("Job end notification to " + urlToNotify +" failed with code: " + log.warn("Job end notification to " + urlToNotify +" failed with code: " + conn.getResponseCode() + " and message \"" + conn.getResponseMessage() +"\""); } else { success = true; - Log.info("Job end notification to " + urlToNotify + " succeeded"); + log.info("Job end notification to " + urlToNotify + " succeeded"); } } catch(IOException ioe) { - Log.warn("Job end notification to " + urlToNotify + " failed", ioe); + log.warn("Job end notification to " + urlToNotify + " failed", ioe); } return success; } @@ -152,7 +154,7 @@ public class JobEndNotifier implements Configurable { throws InterruptedException { // Do we need job-end notification? if (userUrl == null) { - Log.info("Job end notification URL not set, skipping."); + log.info("Job end notification URL not set, skipping."); return; } @@ -168,23 +170,23 @@ public class JobEndNotifier implements Configurable { try { urlToNotify = new URL(userUrl); } catch (MalformedURLException mue) { - Log.warn("Job end notification couldn't parse " + userUrl, mue); + log.warn("Job end notification couldn't parse " + userUrl, mue); return; } // Send notification boolean success = false; while (numTries-- > 0 && !success) { - Log.info("Job end notification attempts left " + numTries); + log.info("Job end notification attempts left " + numTries); success = notifyURLOnce(); if (!success) { Thread.sleep(waitInterval); } } if (!success) { - Log.warn("Job end notification failed to notify : " + urlToNotify); + log.warn("Job end notification failed to notify : " + urlToNotify); } else { - Log.info("Job end notification succeeded for " + jobReport.getJobId()); + log.info("Job end notification succeeded for " + jobReport.getJobId()); } } } diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/java/org/apache/hadoop/mapreduce/v2/hs/webapp/TestHsWebServicesJobsQuery.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/java/org/apache/hadoop/mapreduce/v2/hs/webapp/TestHsWebServicesJobsQuery.java index 8891ec7..1dd369a 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/java/org/apache/hadoop/mapreduce/v2/hs/webapp/TestHsWebServicesJobsQuery.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs/src/test/java/org/apache/hadoop/mapreduce/v2/hs/webapp/TestHsWebServicesJobsQuery.java @@ -136,7 +136,7 @@ public void testJobsQueryStateNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("jobs is not null", JSONObject.NULL, json.get("jobs")); + assertEquals("jobs is not None", 0, json.getJSONObject("jobs").length()); } @Test @@ -202,7 +202,7 @@ public void testJobsQueryUserNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("jobs is not null", JSONObject.NULL, json.get("jobs")); + assertEquals("jobs is not None", 0, json.getJSONObject("jobs").length()); } @Test @@ -287,7 +287,7 @@ public void testJobsQueryQueueNonExist() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("jobs is not null", JSONObject.NULL, json.get("jobs")); + assertEquals("jobs is not None", 0, json.getJSONObject("jobs").length()); } @Test @@ -319,7 +319,7 @@ public void testJobsQueryStartTimeBegin() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("jobs is not null", JSONObject.NULL, json.get("jobs")); + assertEquals("jobs is not None", 0, json.getJSONObject("jobs").length()); } @Test @@ -639,7 +639,7 @@ public void testJobsQueryFinishTimeEnd() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("jobs is not null", JSONObject.NULL, json.get("jobs")); + assertEquals("jobs is not None", 0, json.getJSONObject("jobs").length()); } @Test diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/NotificationTestCase.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/NotificationTestCase.java index d2ea74e..d986fdc 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/NotificationTestCase.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapred/NotificationTestCase.java @@ -18,9 +18,9 @@ package org.apache.hadoop.mapred; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.servlet.Context; -import org.mortbay.jetty.servlet.ServletHolder; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.servlet.ServletContextHandler; +import org.eclipse.jetty.servlet.ServletHolder; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.io.Text; @@ -69,7 +69,7 @@ private void startHttpServer() throws Exception { } webServer = new Server(0); - Context context = new Context(webServer, contextPath); + ServletContextHandler context = new ServletContextHandler(webServer, contextPath); // create servlet handler context.addServlet(new ServletHolder(new NotificationServlet()), diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/MiniHadoopClusterManager.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/MiniHadoopClusterManager.java index 2e8ba5e..3cc73b5 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/MiniHadoopClusterManager.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/MiniHadoopClusterManager.java @@ -45,7 +45,7 @@ import org.apache.hadoop.mapreduce.v2.jobhistory.JHAdminConfig; import org.apache.hadoop.yarn.conf.YarnConfiguration; import org.apache.hadoop.yarn.server.MiniYARNCluster; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * This class drives the creation of a mini-cluster on the local machine. By diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/main/java/org/apache/hadoop/mapred/ShuffleHandler.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/main/java/org/apache/hadoop/mapred/ShuffleHandler.java index c803a7f..393d385 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/main/java/org/apache/hadoop/mapred/ShuffleHandler.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/main/java/org/apache/hadoop/mapred/ShuffleHandler.java @@ -118,6 +118,7 @@ import org.jboss.netty.channel.socket.nio.NioServerSocketChannelFactory; import org.jboss.netty.handler.codec.frame.TooLongFrameException; import org.jboss.netty.handler.codec.http.DefaultHttpResponse; import org.jboss.netty.handler.codec.http.HttpChunkAggregator; +import org.jboss.netty.handler.codec.http.HttpHeaders; import org.jboss.netty.handler.codec.http.HttpRequest; import org.jboss.netty.handler.codec.http.HttpRequestDecoder; import org.jboss.netty.handler.codec.http.HttpResponse; @@ -127,7 +127,6 @@ import org.jboss.netty.handler.codec.http.QueryStringDecoder; import org.jboss.netty.handler.ssl.SslHandler; import org.jboss.netty.handler.stream.ChunkedWriteHandler; import org.jboss.netty.util.CharsetUtil; -import org.mortbay.jetty.HttpHeaders; import com.google.common.annotations.VisibleForTesting; import com.google.common.base.Charsets; @@ -830,9 +830,9 @@ public class ShuffleHandler extends AuxiliaryService { } // Check whether the shuffle version is compatible if (!ShuffleHeader.DEFAULT_HTTP_HEADER_NAME.equals( - request.getHeader(ShuffleHeader.HTTP_HEADER_NAME)) + request.headers().get(ShuffleHeader.HTTP_HEADER_NAME)) || !ShuffleHeader.DEFAULT_HTTP_HEADER_VERSION.equals( - request.getHeader(ShuffleHeader.HTTP_HEADER_VERSION))) { + request.headers().get(ShuffleHeader.HTTP_HEADER_VERSION))) { sendError(ctx, "Incompatible shuffle request version", BAD_REQUEST); } final Map> q = @@ -1044,12 +1044,12 @@ public class ShuffleHandler extends AuxiliaryService { boolean keepAliveParam, long contentLength) { if (!connectionKeepAliveEnabled && !keepAliveParam) { LOG.info("Setting connection close header..."); - response.setHeader(HttpHeaders.CONNECTION, CONNECTION_CLOSE); + response.headers().set(HttpHeaders.Names.CONNECTION, CONNECTION_CLOSE); } else { - response.setHeader(HttpHeaders.CONTENT_LENGTH, + response.headers().set(HttpHeaders.Names.CONTENT_LENGTH, String.valueOf(contentLength)); - response.setHeader(HttpHeaders.CONNECTION, HttpHeaders.KEEP_ALIVE); - response.setHeader(HttpHeaders.KEEP_ALIVE, "timeout=" + response.headers().set(HttpHeaders.Names.CONNECTION, HttpHeaders.Values.KEEP_ALIVE); + response.headers().set(HttpHeaders.Values.KEEP_ALIVE, "timeout=" + connectionKeepAliveTimeOut); LOG.info("Content Length in shuffle : " + contentLength); } @@ -1077,7 +1077,7 @@ public class ShuffleHandler extends AuxiliaryService { String enc_str = SecureShuffleUtils.buildMsgFrom(requestUri); // hash from the fetcher String urlHashStr = - request.getHeader(SecureShuffleUtils.HTTP_HEADER_URL_HASH); + request.headers().get(SecureShuffleUtils.HTTP_HEADER_URL_HASH); if (urlHashStr == null) { LOG.info("Missing header hash for " + appid); throw new IOException("fetcher cannot be authenticated"); @@ -1093,11 +1093,11 @@ public class ShuffleHandler extends AuxiliaryService { String reply = SecureShuffleUtils.generateHash(urlHashStr.getBytes(Charsets.UTF_8), tokenSecret); - response.setHeader(SecureShuffleUtils.HTTP_HEADER_REPLY_URL_HASH, reply); + response.headers().set(SecureShuffleUtils.HTTP_HEADER_REPLY_URL_HASH, reply); // Put shuffle version into http header - response.setHeader(ShuffleHeader.HTTP_HEADER_NAME, + response.headers().set(ShuffleHeader.HTTP_HEADER_NAME, ShuffleHeader.DEFAULT_HTTP_HEADER_NAME); - response.setHeader(ShuffleHeader.HTTP_HEADER_VERSION, + response.headers().set(ShuffleHeader.HTTP_HEADER_VERSION, ShuffleHeader.DEFAULT_HTTP_HEADER_VERSION); if (LOG.isDebugEnabled()) { int len = reply.length(); @@ -1163,11 +1163,11 @@ public class ShuffleHandler extends AuxiliaryService { protected void sendError(ChannelHandlerContext ctx, String message, HttpResponseStatus status) { HttpResponse response = new DefaultHttpResponse(HTTP_1_1, status); - response.setHeader(CONTENT_TYPE, "text/plain; charset=UTF-8"); + response.headers().set(CONTENT_TYPE, "text/plain; charset=UTF-8"); // Put shuffle version into http header - response.setHeader(ShuffleHeader.HTTP_HEADER_NAME, + response.headers().set(ShuffleHeader.HTTP_HEADER_NAME, ShuffleHeader.DEFAULT_HTTP_HEADER_NAME); - response.setHeader(ShuffleHeader.HTTP_HEADER_VERSION, + response.headers().set(ShuffleHeader.HTTP_HEADER_VERSION, ShuffleHeader.DEFAULT_HTTP_HEADER_VERSION); response.setContent( ChannelBuffers.copiedBuffer(message, CharsetUtil.UTF_8)); diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/test/java/org/apache/hadoop/mapred/TestShuffleHandler.java b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/test/java/org/apache/hadoop/mapred/TestShuffleHandler.java index 420c428..3a3257e 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/test/java/org/apache/hadoop/mapred/TestShuffleHandler.java +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-shuffle/src/test/java/org/apache/hadoop/mapred/TestShuffleHandler.java @@ -94,7 +94,6 @@ import org.junit.Test; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; import org.mockito.Mockito; -import org.mortbay.jetty.HttpHeaders; public class TestShuffleHandler { static final long MiB = 1024 * 1024; diff --git a/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml b/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml index 8ae5809..b7da2bc 100644 --- a/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml +++ b/hadoop-mapreduce-project/hadoop-mapreduce-client/pom.xml @@ -43,8 +43,8 @@ avro - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server org.apache.ant diff --git a/hadoop-mapreduce-project/pom.xml b/hadoop-mapreduce-project/pom.xml index 8f1d2b0..465c2df 100644 --- a/hadoop-mapreduce-project/pom.xml +++ b/hadoop-mapreduce-project/pom.xml @@ -52,8 +52,8 @@ avro - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server org.apache.ant @@ -136,6 +128,12 @@ com.sun.jersey jersey-server + + + asm + asm + + com.sun.jersey.contribs diff --git a/hadoop-project/pom.xml b/hadoop-project/pom.xml index b315e2b..e9b072d 100644 --- a/hadoop-project/pom.xml +++ b/hadoop-project/pom.xml @@ -79,7 +79,7 @@ 2.7.1 3.0.0 - 6.0.48 + 7.0.37 1.7 @@ -455,28 +455,27 @@ javax.servlet servlet-api - 2.5 + 3.1.0 - org.mortbay.jetty - jetty - 6.1.26 - - - org.mortbay.jetty - servlet-api - - + org.eclipse.jetty + jetty-server + 8.1 - org.mortbay.jetty + org.eclipse.jetty + jetty-servlet + 8.1 + + + org.eclipse.jetty jetty-util - 6.1.26 + 8.1 - org.mortbay.jetty - jetty-sslengine - 6.1.26 + org.eclipse.jetty + jetty-webapp + 8.1 org.apache.tomcat.embed @@ -484,15 +484,15 @@ 7.0.55 - javax.servlet.jsp - jsp-api - 2.1 + org.glassfish.web + javax.servlet.jsp + 2.2.5 org.glassfish javax.servlet - 3.1 + 8.1.14.v20131031 @@ -531,6 +531,17 @@ com.sun.jersey jersey-server ${jersey.version} + + + asm + asm + + + + + com.sun.jersey + jersey-servlet + ${jersey.version} @@ -592,11 +592,6 @@ 2.4 - - org.mortbay.jetty - jetty-servlet-tester - 6.1.26 - commons-logging commons-logging diff --git a/hadoop-tools/hadoop-sls/pom.xml b/hadoop-tools/hadoop-sls/pom.xml index 6166725..e0d3ee7 100644 --- a/hadoop-tools/hadoop-sls/pom.xml +++ b/hadoop-tools/hadoop-sls/pom.xml @@ -55,18 +55,12 @@ compile - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server provided - - - org.mortbay.jetty - servlet-api - - - org.mortbay.jetty + org.eclipse.jetty jetty-util provided diff --git a/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/web/SLSWebApp.java b/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/web/SLSWebApp.java index 123ccea..e961e58 100644 --- a/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/web/SLSWebApp.java +++ b/hadoop-tools/hadoop-sls/src/main/java/org/apache/hadoop/yarn/sls/web/SLSWebApp.java @@ -34,10 +34,11 @@ import org.apache.hadoop.classification.InterfaceAudience.Private; import org.apache.hadoop.classification.InterfaceStability.Unstable; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.event .SchedulerEventType; -import org.mortbay.jetty.Handler; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.handler.AbstractHandler; -import org.mortbay.jetty.Request; +import org.eclipse.jetty.server.Handler; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.server.Request; +import org.eclipse.jetty.server.handler.AbstractHandler; +import org.eclipse.jetty.server.handler.ResourceHandler; import org.apache.hadoop.yarn.sls.SLSRunner; import org.apache.hadoop.yarn.sls.scheduler.FairSchedulerMetrics; @@ -47,7 +48,6 @@ import com.codahale.metrics.Counter; import com.codahale.metrics.Gauge; import com.codahale.metrics.Histogram; import com.codahale.metrics.MetricRegistry; -import org.mortbay.jetty.handler.ResourceHandler; @Private @Unstable @@ -108,8 +108,9 @@ public void start() throws Exception { Handler handler = new AbstractHandler() { @Override - public void handle(String target, HttpServletRequest request, - HttpServletResponse response, int dispatch) { + public void handle(String target, Request baseRequest, + HttpServletRequest request, + HttpServletResponse response) { try{ // timeunit int timeunit = 1000; // second, divide millionsecond / 1000 @@ -131,7 +132,7 @@ public void handle(String target, HttpServletRequest request, // js/css request if (target.startsWith("/js") || target.startsWith("/css")) { response.setCharacterEncoding("utf-8"); - staticHandler.handle(target, request, response, dispatch); + staticHandler.handle(target, baseRequest, request, response); } else // json request if (target.equals("/simulateMetrics")) { diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java index 08e71c1..461c43c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/api/impl/TestAMRMClient.java @@ -93,7 +93,7 @@ import org.junit.BeforeClass; import org.junit.Test; import org.mockito.invocation.InvocationOnMock; import org.mockito.stubbing.Answer; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; import com.google.common.base.Supplier; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java index 1efb54c..1b3463b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/cli/TestYarnCLI.java @@ -69,7 +69,7 @@ import org.apache.hadoop.yarn.util.Records; import org.junit.Assert; import org.junit.Before; import org.junit.Test; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; import com.google.common.collect.ImmutableSet; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml index a19a78c..83aa759 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/pom.xml @@ -67,7 +67,7 @@ commons-codec - org.mortbay.jetty + org.eclipse.jetty jetty-util @@ -163,6 +163,12 @@ com.sun.jersey jersey-server + + + asm + asm + + com.sun.jersey diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/WebApps.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/WebApps.java index f8c6f55..71df06b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/WebApps.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/webapp/WebApps.java @@ -266,7 +266,8 @@ public void setup() { server.setAttribute(entry.getKey(), entry.getValue()); } HttpServer2.defineFilter(server.getWebAppContext(), "guice", - GuiceFilter.class.getName(), null, new String[] { "/*" }); + GuiceFilter.class.getName(), new HashMap(0), + new String[] { "/*" }); webapp.setConf(conf); webapp.setHttpServer(server); diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml index 0fbafd2..5fe4206 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/pom.xml @@ -99,7 +95,7 @@ jersey-client - org.mortbay.jetty + org.eclipse.jetty jetty-util diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/ContainerLogsPage.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/ContainerLogsPage.java index 7d2948e..81e51c3 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/ContainerLogsPage.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/webapp/ContainerLogsPage.java @@ -43,6 +43,7 @@ import org.apache.hadoop.yarn.webapp.hamlet.Hamlet; import org.apache.hadoop.yarn.webapp.hamlet.Hamlet.PRE; import org.apache.hadoop.yarn.webapp.view.HtmlBlock; +import org.eclipse.jetty.util.log.Log; import com.google.inject.Inject; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/logaggregation/TestLogAggregationService.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/logaggregation/TestLogAggregationService.java index bfb0e87..f9fac8e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/logaggregation/TestLogAggregationService.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/logaggregation/TestLogAggregationService.java @@ -124,7 +124,7 @@ import org.junit.Assert; import org.junit.Test; import org.mockito.ArgumentCaptor; import org.mockito.Mockito; -import org.mortbay.util.MultiException; +import org.eclipse.jetty.util.MultiException; import com.google.common.base.Supplier; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesApps.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesApps.java index 72c1f6f..d272614 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesApps.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesApps.java @@ -176,7 +176,7 @@ public void testNodeAppsNone() throws JSONException, Exception { .accept(MediaType.APPLICATION_JSON).get(ClientResponse.class); assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); - assertEquals("apps isn't NULL", JSONObject.NULL, json.get("apps")); + assertEquals("apps isn't None",0,json.getJSONObject("apps").length()); } private HashMap addAppContainers(Application app) @@ -286,7 +286,7 @@ public void testNodeAppsUserNone() throws JSONException, Exception { .get(ClientResponse.class); assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); - assertEquals("apps is not null", JSONObject.NULL, json.get("apps")); + assertEquals("apps is not None", 0, json.getJSONObject("apps").length()); } @Test @@ -368,7 +368,7 @@ public void testNodeAppsStateNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); - assertEquals("apps is not null", JSONObject.NULL, json.get("apps")); + assertEquals("apps is not None", 0, json.getJSONObject("apps").length()); } @Test diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesContainers.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesContainers.java index 29c9253..56ca16e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesContainers.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/webapp/TestNMWebServicesContainers.java @@ -183,7 +183,7 @@ public void testNodeContainersNone() throws JSONException, Exception { .get(ClientResponse.class); assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); - assertEquals("apps isn't NULL", JSONObject.NULL, json.get("containers")); + assertEquals("apps isn't None", 0, json.getJSONObject("containers").length()); } private HashMap addAppContainers(Application app) diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml index 3e78e02..358a534 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/pom.xml @@ -161,7 +157,7 @@ jersey-client - org.mortbay.jetty + org.eclipse.jetty jetty-util diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMNMInfo.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMNMInfo.java index ef4a0d4..f96879e 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMNMInfo.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMNMInfo.java @@ -33,7 +33,7 @@ import org.apache.hadoop.yarn.server.resourcemanager.rmnode.RMNode; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.ResourceScheduler; import org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNodeReport; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; /** * JMX bean listing statuses of all node managers. diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java index 1dcac06..6ecc80d 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockNM.java @@ -43,7 +43,7 @@ import org.apache.hadoop.yarn.server.utils.BuilderUtils; import org.apache.hadoop.yarn.util.Records; import org.apache.hadoop.yarn.util.YarnVersionInfo; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Log; public class MockNM { diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java index 45b3803..2b79c2c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesApps.java @@ -376,7 +376,7 @@ public void testAppsQueryStateNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("apps is not null", JSONObject.NULL, json.get("apps")); + assertEquals("apps is not None", 0, json.getJSONObject("apps").length()); rm.stop(); } @@ -491,7 +491,7 @@ public void testAppsQueryFinalStatusNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("apps is not null", JSONObject.NULL, json.get("apps")); + assertEquals("apps is not None", 0, json.getJSONObject("apps").length()); rm.stop(); } @@ -667,7 +667,7 @@ public void testAppsQueryStartEnd() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("apps is not null", JSONObject.NULL, json.get("apps")); + assertEquals("apps is not None", 0, json.getJSONObject("apps").length()); rm.stop(); } diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesNodes.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesNodes.java index da2e2b1..77cdfa9 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesNodes.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/webapp/TestRMWebServicesNodes.java @@ -204,7 +204,7 @@ public void testNodesQueryStateNone() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("nodes is not null", JSONObject.NULL, json.get("nodes")); + assertEquals("nodes is not None", 0, json.getJSONObject("nodes").length()); } @Test @@ -343,7 +343,7 @@ public void testNodesQueryHealthyFalse() throws JSONException, Exception { assertEquals(MediaType.APPLICATION_JSON_TYPE, response.getType()); JSONObject json = response.getEntity(JSONObject.class); assertEquals("incorrect number of elements", 1, json.length()); - assertEquals("nodes is not null", JSONObject.NULL, json.get("nodes")); + assertEquals("nodes is not None", 0, json.getJSONObject("nodes").length()); } public void testNodesHelper(String path, String media) throws JSONException, diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml index 10f243c..af23544 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/pom.xml @@ -109,8 +105,8 @@ commons-logging - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-server diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/test/java/org/apache/hadoop/yarn/server/webproxy/TestWebAppProxyServlet.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/test/java/org/apache/hadoop/yarn/server/webproxy/TestWebAppProxyServlet.java index 1be0115..420a41c 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/test/java/org/apache/hadoop/yarn/server/webproxy/TestWebAppProxyServlet.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-web-proxy/src/test/java/org/apache/hadoop/yarn/server/webproxy/TestWebAppProxyServlet.java @@ -57,9 +57,10 @@ import org.apache.hadoop.yarn.webapp.util.WebAppUtils; import org.junit.AfterClass; import org.junit.BeforeClass; import org.junit.Test; -import org.mortbay.jetty.Server; -import org.mortbay.jetty.servlet.Context; -import org.mortbay.jetty.servlet.ServletHolder; +import org.eclipse.jetty.server.Server; +import org.eclipse.jetty.servlet.ServletContextHandler; +import org.eclipse.jetty.servlet.ServletHolder; + import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -81,7 +82,7 @@ public class TestWebAppProxyServlet { @BeforeClass public static void start() throws Exception { server = new Server(0); - Context context = new Context(); + ServletContextHandler context = new ServletContextHandler(); context.setContextPath("/foo"); server.setHandler(context); context.addServlet(new ServletHolder(TestServlet.class), "/bar"); diff --git a/hadoop-common-project/hadoop-kms/pom.xml b/hadoop-common-project/hadoop-kms/pom.xml index c479b67..a065485 100644 --- a/hadoop-common-project/hadoop-kms/pom.xml +++ b/hadoop-common-project/hadoop-kms/pom.xml @@ -84,11 +84,6 @@ provided - org.mortbay.jetty - jetty - test - - org.apache.hadoop hadoop-common compile @@ -122,20 +117,20 @@ jsp-api - org.mortbay.jetty - jetty + org.eclipse.jetty + jetty-webapp - org.mortbay.jetty + org.eclipse.jetty jetty-util - org.mortbay.jetty - jsp-api-2.1 + org.eclipse.jetty + jetty-servlet - org.mortbay.jetty - servlet-api-2.5 + org.eclipse.jetty + jetty-server net.java.dev.jets3t @@ -178,7 +173,7 @@ compile - org.mortbay.jetty + org.eclipse.jetty jetty-util compile diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java index d2cca0e..637cfd6 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeReference.java @@ -31,7 +31,8 @@ import org.apache.hadoop.hdfs.server.namenode.snapshot.DirectoryWithSnapshotFeat import org.apache.hadoop.hdfs.server.namenode.snapshot.Snapshot; import com.google.common.base.Preconditions; -import org.mortbay.log.Log; +import org.eclipse.jetty.util.log.Logger; +import org.eclipse.jetty.util.log.Log; /** * An anonymous reference to an inode. @@ -567,7 +568,8 @@ public abstract class INodeReference extends INode { try { ref.addSpaceConsumed(counts.negation(), true); } catch (QuotaExceededException e) { - Log.warn("Should not have QuotaExceededException"); + Logger log = Log.getLogger(INodeReference.class); + log.warn("Should not have QuotaExceededException"); } } diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/TransferFsImage.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/TransferFsImage.java index 0d32758..658fa3d 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/TransferFsImage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/TransferFsImage.java @@ -63,7 +63,7 @@ import org.apache.http.client.utils.URIBuilder; import com.google.common.annotations.VisibleForTesting; import com.google.common.collect.Lists; -import org.mortbay.jetty.EofException; +import org.eclipse.jetty.io.EofException; /** * This class provides fetching a specified file from the NameNode. diff --git a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java index 5dc0963..63c9647 100644 --- a/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java +++ b/hadoop-tools/hadoop-azure/src/main/java/org/apache/hadoop/fs/azure/AzureNativeFileSystemStore.java @@ -62,7 +62,7 @@ import org.apache.hadoop.fs.azure.metrics.ErrorMetricUpdater; import org.apache.hadoop.fs.azure.metrics.ResponseReceivedMetricUpdater; import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.fs.permission.PermissionStatus; -import org.mortbay.util.ajax.JSON; +import org.eclipse.jetty.util.ajax.JSON; import com.google.common.annotations.VisibleForTesting; import com.microsoft.azure.storage.CloudStorageAccount; diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryServer.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryServer.java index 7fb900d..27582fb 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryServer.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/src/main/java/org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryServer.java @@ -55,8 +55,8 @@ import org.apache.hadoop.yarn.server.timeline.webapp.CrossOriginFilterInitialize import org.apache.hadoop.yarn.webapp.WebApp; import org.apache.hadoop.yarn.webapp.WebApps; import org.apache.hadoop.yarn.webapp.util.WebAppUtils; -import org.mortbay.jetty.servlet.FilterHolder; -import org.mortbay.jetty.webapp.WebAppContext; +import org.eclipse.jetty.servlet.FilterHolder; +import org.eclipse.jetty.webapp.WebAppContext; import com.google.common.annotations.VisibleForTesting; diff --git a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SslSelectChannelConnectorSecure.java b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SslSelectChannelConnectorSecure.java index 7ff2292..b6c9f6f 100644 --- a/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SslSelectChannelConnectorSecure.java +++ b/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/ssl/SslSelectChannelConnectorSecure.java @@ -26,7 +26,8 @@ import javax.net.ssl.SSLEngine; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; -import org.mortbay.jetty.security.SslSelectChannelConnector; +import org.eclipse.jetty.server.ssl.SslSelectChannelConnector; +import java.nio.channels.SocketChannel; /** * This subclass of the Jetty SslSelectChannelConnector exists solely to @@ -47,8 +48,8 @@ public class SslSelectChannelConnectorSecure extends SslSelectChannelConnector { * Disable SSLv3 protocol. */ @Override - protected SSLEngine createSSLEngine() throws IOException { - SSLEngine engine = super.createSSLEngine(); + protected SSLEngine createSSLEngine(SocketChannel channel) throws IOException { + SSLEngine engine = super.createSSLEngine(channel); ArrayList nonSSLProtocols = new ArrayList(); for (String p : engine.getEnabledProtocols()) { if (!p.contains("SSLv3")) { diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/pom.xml b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/pom.xml index 5ac65cc..d73721a 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/pom.xml +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice/pom.xml @@ -40,6 +40,10 @@ javax.servlet servlet-api + + org.eclipse.jetty + jetty-util + org.apache.hadoop