Date: Mon, 28 Jul 2014 19:03:07 GMT From: pkg-fallout@FreeBSD.org To: demon@FreeBSD.org Cc: pkg-fallout@FreeBSD.org Subject: [package - 91i386-default-ssp][devel/hadoop2] Failed for apache-hadoop2-2.4.1_2 in build Message-ID: <201407281903.s6SJ3735061198@beefy4.isc.freebsd.org>
next in thread | raw e-mail | index | archive | help
You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build. Maintainer: demon@FreeBSD.org Last committer: demon@FreeBSD.org Ident: $FreeBSD: head/devel/hadoop2/Makefile 361391 2014-07-09 16:13:17Z demon $ Log URL: http://beefy4.isc.freebsd.org/data/91i386-default-ssp/2014-07-28_02h40m33s/logs/apache-hadoop2-2.4.1_2.log Build URL: http://beefy4.isc.freebsd.org/build.html?mastername=91i386-default-ssp&build=2014-07-28_02h40m33s Log: ====>> Building devel/hadoop2 build started at Mon Jul 28 19:01:57 UTC 2014 port directory: /usr/ports/devel/hadoop2 building for: FreeBSD pkg.FreeBSD.org 9.1-RELEASE-p13 FreeBSD 9.1-RELEASE-p13 i386 maintained by: demon@FreeBSD.org Makefile ident: $FreeBSD: head/devel/hadoop2/Makefile 361391 2014-07-09 16:13:17Z demon $ Poudriere version: 3.1-pre Host OSVERSION: 1100027 Jail OSVERSION: 901000 ---Begin Environment--- UNAME_m=i386 UNAME_p=i386 OSVERSION=901000 UNAME_v=FreeBSD 9.1-RELEASE-p13 UNAME_r=9.1-RELEASE-p13 BLOCKSIZE=K MAIL=/var/mail/root STATUS=1 SAVED_TERM= MASTERMNT=/usr/local/poudriere/data/.m/91i386-default-ssp/ref PKG_EXT=txz FORCE_PACKAGE=yes PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin POUDRIERE_BUILD_TYPE=bulk PKGNG=1 PKGNAME=apache-hadoop2-2.4.1_2 PKG_DELETE=/usr/local/sbin/pkg-static delete -y -f PKG_ADD=/usr/local/sbin/pkg-static add OLDPWD=/root PWD=/usr/local/poudriere/data/.m/91i386-default-ssp/ref/.p/pool MASTERNAME=91i386-default-ssp USER=root HOME=/root POUDRIERE_VERSION=3.1-pre LOCALBASE=/usr/local PACKAGE_BUILDING=yes PKG_VERSION=/.p/pkg-static version PKG_BIN=/usr/local/sbin/pkg-static ---End Environment--- ---Begin OPTIONS List--- ---End OPTIONS List--- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" SHELL=/bin/sh CONFIG_SHELL=/bin/sh PATH=/usr/local/libexec/ccache:/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin CCACHE_DIR="/root/.ccache" --End CONFIGURE_ENV-- --MAKE_ENV-- HADOOP_PROTOC_PATH=/usr/local/bin/protoc XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" NO_PIE=yes SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local LIBDIR="/usr/lib" CC="cc" CFLAGS="-O2 -pipe -fno-strict-aliasing" CPP="cpp" CPPFLAGS="" LDFLAGS="" LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fno-strict-aliasing" MANPREFIX="/usr/local" PATH=/usr/local/libexec/ccache:/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin CCACHE_DIR="/root/.ccache" BSD_INSTALL_PROGRAM="install -s -o root -g wheel -m 555" BSD_INSTALL_LIB="install -s -o root -g wheel -m 444" BSD_INSTALL_SCRIPT="install -o root -g wheel -m 555" BSD_INSTALL_DATA="install -o root -g wheel -m 444" BSD_INSTALL_MAN="install -o root -g wheel -m 444" --End MAKE_ENV-- --PLIST_SUB-- PORTVERSION="2.4.1" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" RESETPREFIX=/usr/local JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" OSREL=9.1 PREFIX=%D LOCALBASE=/usr/local PORTDOCS="" PORTEXAMPLES="" PORTDATA="" LIB32DIR=lib DOCSDIR="share/doc/hadoop" EXAMPLESDIR="share/examples/hadoop" DATADIR="share/hadoop" WWWDIR="www/hadoop" ETCDIR="etc/hadoop" --End PLIST_SUB-- --SUB_LIST-- HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVA_HOME="/usr/local/openjdk7" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" JAVA_VERSION="1.7+" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/hadoop DOCSDIR=/usr/local/share/doc/hadoop EXAMPLESDIR=/usr/local/share/examples/hadoop WWWDIR=/usr/local/www/hadoop ETCDIR=/usr/local/etc/hadoop --End SUB_LIST-- ---Begin make.conf--- ARCH=i386 MACHINE=i386 MACHINE_ARCH=i386 USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs ARCH=i386 MACHINE=i386 MACHINE_ARCH=i386 WITH_CCACHE_BUILD=yes CCACHE_DIR=/root/.ccache USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles #### /usr/local/etc/poudriere.d/make.conf #### WITH_PKGNG=yes #### /usr/local/etc/poudriere.d/ssp-make.conf #### WITH_SSP_PORTS=yes DISABLE_MAKE_JOBS=poudriere ---End make.conf--- =======================<phase: check-sanity >============================ ===> License APACHE20 accepted by the user =========================================================================== =======================<phase: pkg-depends >============================ ===> apache-hadoop2-2.4.1_2 depends on file: /usr/local/sbin/pkg - not found ===> Verifying install for /usr/local/sbin/pkg in /usr/ports/ports-mgmt/pkg ===> Installing existing package /packages/All/pkg-1.2.7_4.txz Installing pkg-1.2.7_4... done If you are upgrading from the old package format, first run: # pkg2ng ===> Returning to build of apache-hadoop2-2.4.1_2 =========================================================================== =======================<phase: fetch-depends >============================ =========================================================================== =======================<phase: fetch >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-hadoop2-2.4.1_2 for building =========================================================================== =======================<phase: checksum >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-hadoop2-2.4.1_2 for building => SHA256 Checksum OK for hadoop/hadoop-2.4.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.4.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.36.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: extract-depends>============================ =========================================================================== =======================<phase: extract >============================ ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-hadoop2-2.4.1_2 for building ===> Extracting for apache-hadoop2-2.4.1_2 => SHA256 Checksum OK for hadoop/hadoop-2.4.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.4.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.36.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =======================<phase: patch-depends >============================ =========================================================================== =======================<phase: patch >============================ ===> Patching for apache-hadoop2-2.4.1_2 ===> Applying FreeBSD patches for apache-hadoop2-2.4.1_2 /usr/bin/sed -i.bak -e "s#/bin/bash#/usr/local/bin/bash#" /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DefaultContainerExecutor.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/ContainerLaunch.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java =========================================================================== =======================<phase: build-depends >============================ ===> apache-hadoop2-2.4.1_2 depends on file: /usr/local/share/java/maven3/bin/mvn - not found ===> Verifying install for /usr/local/share/java/maven3/bin/mvn in /usr/ports/devel/maven3 ===> Installing existing package /packages/All/maven3-3.0.5.txz Installing maven3-3.0.5...Installing maven-wrapper-1_2... done Installing openjdk-7.65.17,1...Installing alsa-lib-1.0.27.2_2... done Installing dejavu-2.34_3... done Installing fontconfig-2.11.0_3,1...Installing expat-2.1.0_1... done Installing freetype2-2.5.3_2... done Running fc-cache to build fontconfig cache... /usr/local/share/fonts: skipping, no such directory /usr/local/lib/X11/fonts: caching, new cache contents: 0 fonts, 2 dirs /usr/local/lib/X11/fonts/dejavu: caching, new cache contents: 21 fonts, 0 dirs /usr/local/lib/X11/fonts/local: caching, new cache contents: 0 fonts, 0 dirs /root/.local/share/fonts: skipping, no such directory /root/.fonts: skipping, no such directory /usr/local/share/fonts: skipping, no such directory /usr/local/lib/X11/fonts: caching, new cache contents: 0 fonts, 2 dirs /root/.local/share/fonts: skipping, no such directory /root/.fonts: skipping, no such directory /var/db/fontconfig: cleaning cache directory /root/.cache/fontconfig: not cleaning non-existent cache directory /root/.fontconfig: not cleaning non-existent cache directory fc-cache: succeeded done Installing java-zoneinfo-2014.e... done Installing javavmwrapper-2.5... done Installing libX11-1.6.2_2,1...Installing kbproto-1.0.6... done Installing libXau-1.0.8_2...Installing xproto-7.0.26... done done <snip> Building index for all classes... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-frame.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-noframe.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/index.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/help-doc.html... [INFO] Building jar: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-auth-examples/target/hadoop-auth-examples-2.4.1-javadoc.jar [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop Common 2.4.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.4.1:clean (default-clean) @ hadoop-common --- [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-os) @ hadoop-common --- [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-common --- [INFO] Executing tasks main: [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/test-dir [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/test/data [INFO] Executed tasks [INFO] [INFO] --- hadoop-maven-plugins:2.4.1:protoc (compile-protoc) @ hadoop-common --- [INFO] [INFO] --- hadoop-maven-plugins:2.4.1:version-info (version-info) @ hadoop-common --- [WARNING] [svn, info] failed: java.io.IOException: Cannot run program "svn": error=2, No such file or directory [WARNING] [git, branch] failed: java.io.IOException: Cannot run program "git": error=2, No such file or directory [INFO] SCM: NONE [INFO] Computed MD5: cff2ee55e63d634c896188f0af98819 [INFO] [INFO] --- maven-resources-plugin:2.2:resources (default-resources) @ hadoop-common --- [INFO] Using default encoding to copy filtered resources. [INFO] [INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) @ hadoop-common --- [INFO] Compiling 741 source files to /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.4.1-src/hadoop-common-project/hadoop-common/target/classes [INFO] ------------------------------------------------------------- [ERROR] COMPILATION ERROR : [INFO] ------------------------------------------------------------- [ERROR] Failure executing javac, but could not parse the error: The system is out of resources. Consult the following stack trace for details. java.lang.OutOfMemoryError: Java heap space at com.sun.tools.javac.util.Position$LineMapImpl.build(Position.java:153) at com.sun.tools.javac.util.Position.makeLineMap(Position.java:77) at com.sun.tools.javac.parser.Scanner.getLineMap(Scanner.java:1147) at com.sun.tools.javac.parser.JavacParser.parseCompilationUnit(JavacParser.java:2310) at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:595) at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:632) at com.sun.tools.javac.main.JavaCompiler.parseFiles(JavaCompiler.java:909) at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:824) at com.sun.tools.javac.main.Main.compile(Main.java:439) at com.sun.tools.javac.main.Main.compile(Main.java:353) at com.sun.tools.javac.main.Main.compile(Main.java:342) at com.sun.tools.javac.main.Main.compile(Main.java:333) at com.sun.tools.javac.Main.compile(Main.java:94) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess0(JavacCompiler.java:551) at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess(JavacCompiler.java:526) at org.codehaus.plexus.compiler.javac.JavacCompiler.compile(JavacCompiler.java:167) at org.apache.maven.plugin.AbstractCompilerMojo.execute(AbstractCompilerMojo.java:678) at org.apache.maven.plugin.CompilerMojo.execute(CompilerMojo.java:128) at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:101) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:209) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:84) at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:59) at org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(LifecycleStarter.java:183) at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:161) at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:320) at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156) [INFO] 1 error [INFO] ------------------------------------------------------------- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop Main ................................ SUCCESS [1.687s] [INFO] Apache Hadoop Project POM ......................... SUCCESS [1.488s] [INFO] Apache Hadoop Annotations ......................... SUCCESS [4.830s] [INFO] Apache Hadoop Assemblies .......................... SUCCESS [0.607s] [INFO] Apache Hadoop Project Dist POM .................... SUCCESS [2.484s] [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [5.227s] [INFO] Apache Hadoop MiniKDC ............................. SUCCESS [4.501s] [INFO] Apache Hadoop Auth ................................ SUCCESS [5.439s] [INFO] Apache Hadoop Auth Examples ....................... SUCCESS [3.121s] [INFO] Apache Hadoop Common .............................. FAILURE [12.333s] [INFO] Apache Hadoop NFS ................................. SKIPPED [INFO] Apache Hadoop Common Project ...................... SKIPPED [INFO] Apache Hadoop HDFS ................................ SKIPPED [INFO] Apache Hadoop HttpFS .............................. SKIPPED [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED [INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED [INFO] Apache Hadoop HDFS Project ........................ SKIPPED [INFO] hadoop-yarn ....................................... SKIPPED [INFO] hadoop-yarn-api ................................... SKIPPED [INFO] hadoop-yarn-common ................................ SKIPPED [INFO] hadoop-yarn-server ................................ SKIPPED [INFO] hadoop-yarn-server-common ......................... SKIPPED [INFO] hadoop-yarn-server-nodemanager .................... SKIPPED [INFO] hadoop-yarn-server-web-proxy ...................... SKIPPED [INFO] hadoop-yarn-server-applicationhistoryservice ...... SKIPPED [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED [INFO] hadoop-yarn-server-tests .......................... SKIPPED [INFO] hadoop-yarn-client ................................ SKIPPED [INFO] hadoop-yarn-applications .......................... SKIPPED [INFO] hadoop-yarn-applications-distributedshell ......... SKIPPED [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SKIPPED [INFO] hadoop-yarn-site .................................. SKIPPED [INFO] hadoop-yarn-project ............................... SKIPPED [INFO] hadoop-mapreduce-client ........................... SKIPPED [INFO] hadoop-mapreduce-client-core ...................... SKIPPED [INFO] hadoop-mapreduce-client-common .................... SKIPPED [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED [INFO] hadoop-mapreduce-client-app ....................... SKIPPED [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED [INFO] hadoop-mapreduce-client-hs-plugins ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] hadoop-mapreduce .................................. SKIPPED [INFO] Apache Hadoop MapReduce Streaming ................. SKIPPED [INFO] Apache Hadoop Distributed Copy .................... SKIPPED [INFO] Apache Hadoop Archives ............................ SKIPPED [INFO] Apache Hadoop Rumen ............................... SKIPPED [INFO] Apache Hadoop Gridmix ............................. SKIPPED [INFO] Apache Hadoop Data Join ........................... SKIPPED [INFO] Apache Hadoop Extras .............................. SKIPPED [INFO] Apache Hadoop Pipes ............................... SKIPPED [INFO] Apache Hadoop OpenStack support ................... SKIPPED [INFO] Apache Hadoop Client .............................. SKIPPED [INFO] Apache Hadoop Mini-Cluster ........................ SKIPPED [INFO] Apache Hadoop Scheduler Load Simulator ............ SKIPPED [INFO] Apache Hadoop Tools Dist .......................... SKIPPED [INFO] Apache Hadoop Tools ............................... SKIPPED [INFO] Apache Hadoop Distribution ........................ SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 44.357s [INFO] Finished at: Mon Jul 28 19:03:07 GMT 2014 [INFO] Final Memory: 37M/48M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:2.5.1:compile (default-compile) on project hadoop-common: Compilation failure [ERROR] Failure executing javac, but could not parse the error: [ERROR] [ERROR] [ERROR] The system is out of resources. [ERROR] Consult the following stack trace for details. [ERROR] java.lang.OutOfMemoryError: Java heap space [ERROR] at com.sun.tools.javac.util.Position$LineMapImpl.build(Position.java:153) [ERROR] at com.sun.tools.javac.util.Position.makeLineMap(Position.java:77) [ERROR] at com.sun.tools.javac.parser.Scanner.getLineMap(Scanner.java:1147) [ERROR] at com.sun.tools.javac.parser.JavacParser.parseCompilationUnit(JavacParser.java:2310) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:595) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parse(JavaCompiler.java:632) [ERROR] at com.sun.tools.javac.main.JavaCompiler.parseFiles(JavaCompiler.java:909) [ERROR] at com.sun.tools.javac.main.JavaCompiler.compile(JavaCompiler.java:824) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:439) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:353) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:342) [ERROR] at com.sun.tools.javac.main.Main.compile(Main.java:333) [ERROR] at com.sun.tools.javac.Main.compile(Main.java:94) [ERROR] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) [ERROR] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) [ERROR] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) [ERROR] at java.lang.reflect.Method.invoke(Method.java:606) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess0(JavacCompiler.java:551) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compileInProcess(JavacCompiler.java:526) [ERROR] at org.codehaus.plexus.compiler.javac.JavacCompiler.compile(JavacCompiler.java:167) [ERROR] at org.apache.maven.plugin.AbstractCompilerMojo.execute(AbstractCompilerMojo.java:678) [ERROR] at org.apache.maven.plugin.CompilerMojo.execute(CompilerMojo.java:128) [ERROR] at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:101) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:209) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153) [ERROR] at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:84) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:59) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleStarter.singleThreadedBuild(LifecycleStarter.java:183) [ERROR] at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:161) [ERROR] at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:320) [ERROR] at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:156) [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-common *** [do-build] Error code 1 Stop in /usr/ports/devel/hadoop2.
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?201407281903.s6SJ3735061198>