From owner-freebsd-pkg-fallout@freebsd.org Sat Dec 12 05:48:09 2015 Return-Path: Delivered-To: freebsd-pkg-fallout@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id 3B5F49D8C53 for ; Sat, 12 Dec 2015 05:48:09 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from mailman.ysv.freebsd.org (mailman.ysv.freebsd.org [IPv6:2001:1900:2254:206a::50:5]) by mx1.freebsd.org (Postfix) with ESMTP id 2B4BD1A24 for ; Sat, 12 Dec 2015 05:48:09 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: by mailman.ysv.freebsd.org (Postfix) id 29D669D8C52; Sat, 12 Dec 2015 05:48:09 +0000 (UTC) Delivered-To: pkg-fallout@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id 0F6CA9D8C51 for ; Sat, 12 Dec 2015 05:48:09 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from beefy2.nyi.freebsd.org (beefy2.nyi.freebsd.org [IPv6:2610:1c1:1:6080::16:e5]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by mx1.freebsd.org (Postfix) with ESMTPS id A1D5F1A23; Sat, 12 Dec 2015 05:48:08 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from beefy2.nyi.freebsd.org (localhost [127.0.0.1]) by beefy2.nyi.freebsd.org (8.15.2/8.15.2) with ESMTP id tBC5m7uE048967; Sat, 12 Dec 2015 05:48:07 GMT (envelope-from pkg-fallout@FreeBSD.org) Received: (from root@localhost) by beefy2.nyi.freebsd.org (8.15.2/8.15.2/Submit) id tBC5m7Mi048963; Sat, 12 Dec 2015 05:48:07 GMT (envelope-from pkg-fallout@FreeBSD.org) Date: Sat, 12 Dec 2015 05:48:07 GMT From: pkg-fallout@FreeBSD.org Message-Id: <201512120548.tBC5m7Mi048963@beefy2.nyi.freebsd.org> To: demon@FreeBSD.org Subject: [package - 93amd64-default][devel/hadoop2] Failed for hadoop2-2.7.1_1 in build Cc: pkg-fallout@FreeBSD.org X-BeenThere: freebsd-pkg-fallout@freebsd.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: Fallout logs from package building List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sat, 12 Dec 2015 05:48:09 -0000 You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build. Maintainer: demon@FreeBSD.org Last committer: demon@FreeBSD.org Ident: $FreeBSD: head/devel/hadoop2/Makefile 401919 2015-11-19 09:27:33Z demon $ Log URL: http://beefy2.nyi.freebsd.org/data/93amd64-default/403578/logs/hadoop2-2.7.1_1.log Build URL: http://beefy2.nyi.freebsd.org/build.html?mastername=93amd64-default&build=403578 Log: ====>> Building devel/hadoop2 build started at Sat Dec 12 05:46:13 UTC 2015 port directory: /usr/ports/devel/hadoop2 building for: FreeBSD 93amd64-default-job-15 9.3-RELEASE-p31 FreeBSD 9.3-RELEASE-p31 amd64 maintained by: demon@FreeBSD.org Makefile ident: $FreeBSD: head/devel/hadoop2/Makefile 401919 2015-11-19 09:27:33Z demon $ Poudriere version: 3.1.10 Host OSVERSION: 1100085 Jail OSVERSION: 903000 ---Begin Environment--- SHELL=/bin/csh OSVERSION=903000 UNAME_v=FreeBSD 9.3-RELEASE-p31 UNAME_r=9.3-RELEASE-p31 BLOCKSIZE=K MAIL=/var/mail/root STATUS=1 SAVED_TERM= MASTERMNT=/usr/local/poudriere/data/.m/93amd64-default/ref PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin POUDRIERE_BUILD_TYPE=bulk PKGNAME=hadoop2-2.7.1_1 OLDPWD=/ PWD=/usr/local/poudriere/data/.m/93amd64-default/ref/.p/pool MASTERNAME=93amd64-default SCRIPTPREFIX=/usr/local/share/poudriere USER=root HOME=/root POUDRIERE_VERSION=3.1.10 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh LIBEXECPREFIX=/usr/local/libexec/poudriere LOCALBASE=/usr/local PACKAGE_BUILDING=yes ---End Environment--- ---Begin OPTIONS List--- ===> The following configuration options are available for hadoop2-2.7.1_1: EXAMPLES=on: Build and/or install examples ===> Use 'make config' to modify these settings ---End OPTIONS List--- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- JAVA_HOME=/usr/local/openjdk7 HADOOP_PROTOC_PATH=/usr/local/protobuf25/bin/protoc XDG_DATA_HOME=/wrkdirs/usr/ports/devel/hadoop2/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/hadoop2/work HOME=/wrkdirs/usr/ports/devel/hadoop2/work TMPDIR="/tmp" NO_PIE=yes NO_DEBUG_FILES=yes SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local LIBDIR="/usr/lib" CC="cc" CFLAGS="-O2 -pipe -fstack-protector -fno-strict-aliasing" CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector" LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector -fno-strict-aliasing" MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 444" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- PORTVERSION="2.7.1" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" OSREL=9.3 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local PORTDOCS="" PORTEXAMPLES="" LIB32DIR=lib DOCSDIR="share/doc/hadoop" EXAMPLESDIR="share/examples/hadoop" DATADIR="share/hadoop" WWWDIR="www/hadoop" ETCDIR="etc/hadoop" --End PLIST_SUB-- --SUB_LIST-- HDFS_USER="hdfs" MAPRED_USER="mapred" HADOOP_GROUP="hadoop" JAVA_HOME="/usr/local/openjdk7" HADOOP_LOGDIR="/var/log/hadoop" HADOOP_RUNDIR="/var/run/hadoop" JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" JAVA_VERSION="1.7" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/hadoop DOCSDIR=/usr/local/share/doc/hadoop EXAMPLESDIR=/usr/local/share/examples/hadoop WWWDIR=/usr/local/www/hadoop ETCDIR=/usr/local/etc/hadoop --End SUB_LIST-- ---Begin make.conf--- USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles #### /usr/local/etc/poudriere.d/make.conf #### DISABLE_MAKE_JOBS=poudriere ---End make.conf--- =================================================== ===> License APACHE20 accepted by the user =========================================================================== =================================================== ===> hadoop2-2.7.1_1 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.6.2.txz [93amd64-default-job-15] Installing pkg-1.6.2... [93amd64-default-job-15] Extracting pkg-1.6.2: .......... done Message from pkg-1.6.2: If you are upgrading from the old package format, first run: # pkg2ng ===> hadoop2-2.7.1_1 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of hadoop2-2.7.1_1 =========================================================================== =================================================== =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1_1 for building =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1_1 for building => SHA256 Checksum OK for hadoop/hadoop-2.7.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.7.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.41.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =================================================== =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by hadoop2-2.7.1_1 for building ===> Extracting for hadoop2-2.7.1_1 => SHA256 Checksum OK for hadoop/hadoop-2.7.1-src.tar.gz. => SHA256 Checksum OK for hadoop/FreeBSD-hadoop2-2.7.1-maven-repository.tar.gz. => SHA256 Checksum OK for hadoop/apache-tomcat-6.0.41.tar.gz. => SHA256 Checksum OK for hadoop/jetty-6.1.14.zip. =========================================================================== =================================================== =========================================================================== =================================================== ===> Patching for hadoop2-2.7.1_1 ===> Applying FreeBSD patches for hadoop2-2.7.1_1 /usr/bin/sed -i.bak -e "s#/bin/bash#/usr/local/bin/bash#" /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/DefaultContainerExecutor.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRJobConfig.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/ContainerLaunch.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/Shell.java /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh =========================================================================== =================================================== ===> hadoop2-2.7.1_1 depends on file: /usr/local/share/java/maven3/bin/mvn - not found ===> Installing existing package /packages/All/maven3-3.0.5.txz [93amd64-default-job-15] Installing maven3-3.0.5... [93amd64-default-job-15] `-- Installing maven-wrapper-1_2... [93amd64-default-job-15] `-- Extracting maven-wrapper-1_2: . done [93amd64-default-job-15] `-- Installing openjdk8-8.60.24_2... [93amd64-default-job-15] | `-- Installing giflib-5.1.1... [93amd64-default-job-15] | `-- Extracting giflib-5.1.1: .......... done [93amd64-default-job-15] | `-- Installing libXt-1.1.5,1... [93amd64-default-job-15] | | `-- Installing xproto-7.0.28... [93amd64-default-job-15] | | `-- Extracting xproto-7.0.28: .......... done [93amd64-default-job-15] | | `-- Installing libSM-1.2.2_3,1... [93amd64-default-job-15] | | `-- Installing libICE-1.0.9_1,1... [93amd64-default-job-15] | | `-- Extracting libICE-1.0.9_1,1: .......... done [93amd64-default-job-15] | | `-- Extracting libSM-1.2.2_3,1: .......... done [93amd64-default-job-15] | | `-- Installing libX11-1.6.3,1... [93amd64-default-job-15] | | `-- Installing kbproto-1.0.7... [93amd64-default-job-15] | | `-- Extracting kbproto-1.0.7: .......... done [93amd64-default-job-15] | | `-- Installing libXdmcp-1.1.2... [93amd64-default-job-15] | | `-- Extracting libXdmcp-1.1.2: ......... done [93amd64-default-job-15] | | `-- Installing libxcb-1.11.1... [93amd64-default-job-15] | | | `-- Installing libxml2-2.9.3... [93amd64-default-job-15] | | | `-- Installing libiconv-1.14_9... [93amd64-default-job-15] | | | `-- Extracting libiconv-1.14_9: .......... done [93amd64-default-job-15] | | | `-- Extracting libxml2-2.9.3: .......... done [93amd64-default-job-15] | | | `-- Installing libpthread-stubs-0.3_6... [93amd64-default-job-15] | | | `-- Extracting libpthread-stubs-0.3_6: ..... done [93amd64-default-job-15] | | | `-- Installing libXau-1.0.8_3... [93amd64-default-job-15] | | | `-- Extracting libXau-1.0.8_3: .......... done [93amd64-default-job-15] | | `-- Extracting libxcb-1.11.1: .......... done [93amd64-default-job-15] | | `-- Extracting libX11-1.6.3,1: .......... done [93amd64-default-job-15] | `-- Extracting libXt-1.1.5,1: .......... done [93amd64-default-job-15] | `-- Installing libXtst-1.2.2_3... [93amd64-default-job-15] | | `-- Installing libXext-1.3.3_1,1... [93amd64-default-job-15] | | `-- Installing xextproto-7.3.0... [93amd64-default-job-15] | | `-- Extracting xextproto-7.3.0: .......... done [93amd64-default-job-15] | | `-- Extracting libXext-1.3.3_1,1: .......... done [93amd64-default-job-15] | | `-- Installing inputproto-2.3.1... [93amd64-default-job-15] | | `-- Extracting inputproto-2.3.1: ..... done [93amd64-default-job-15] | | `-- Installing libXi-1.7.5,1... [93amd64-default-job-15] | | `-- Installing libXfixes-5.0.1_3... [93amd64-default-job-15] | | | `-- Installing fixesproto-5.0... [93amd64-default-job-15] | | | `-- Extracting fixesproto-5.0: .... done [93amd64-default-job-15] | | `-- Extracting libXfixes-5.0.1_3: .......... done Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-frame.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/allclasses-noframe.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/index.html... Generating /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/help-doc.html... [INFO] Building jar: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth-examples/target/hadoop-auth-examples-2.7.1-javadoc.jar [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop Common 2.7.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-os) @ hadoop-common --- [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-common --- [INFO] Executing tasks main: [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/test-dir [mkdir] Created dir: /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/test/data [INFO] Executed tasks [INFO] [INFO] --- hadoop-maven-plugins:2.7.1:protoc (compile-protoc) @ hadoop-common --- [INFO] [INFO] --- hadoop-maven-plugins:2.7.1:version-info (version-info) @ hadoop-common --- [WARNING] [svn, info] failed: java.io.IOException: Cannot run program "svn": error=2, No such file or directory [WARNING] [git, branch] failed: java.io.IOException: Cannot run program "git": error=2, No such file or directory [INFO] SCM: NONE [INFO] Computed MD5: 4e4e024af7121717eca2e5756702dc5 [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hadoop-common --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 7 resources [INFO] Copying 1 resource [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ hadoop-common --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 849 source files to /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/classes [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[46,16] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[50,19] sun.net.dns.ResolverConfiguration is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[51,20] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[25,16] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[21,16] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[22,16] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[44,43] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[331,39] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[332,17] sun.misc.Cleaner is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[333,25] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[687,17] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[689,7] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java:[689,24] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[507,9] sun.net.dns.ResolverConfiguration is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[525,11] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[527,21] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[529,18] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/SecurityUtil.java:[531,21] sun.net.util.IPAddressUtil is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[136,20] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[142,22] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/FastByteComparisons.java:[147,29] sun.misc.Unsafe is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[39,37] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[40,21] sun.misc.Cleaner is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/crypto/CryptoStreamUtils.java:[41,23] sun.nio.ch.DirectBuffer is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[46,19] sun.misc.SignalHandler is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[50,39] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[50,21] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/SignalLogger.java:[59,24] sun.misc.Signal is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/LdapGroupsMapping.java:[265,28] com.sun.jndi.ldap.LdapCtxFactory is internal proprietary API and may be removed in a future release [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/compiler/ant/RccTask.java: Some input files use or override a deprecated API. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/record/compiler/ant/RccTask.java: Recompile with -Xlint:deprecation for details. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java: Some input files use unchecked or unsafe operations. [WARNING] /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystem.java: Recompile with -Xlint:unchecked for details. [INFO] [INFO] --- native-maven-plugin:1.0-alpha-8:javah (default) @ hadoop-common --- [INFO] /bin/sh -c cd /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common && /usr/local/openjdk7/bin/javah -d /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/native/javah -classpath /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/classes:/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-annotations/target/hadoop-annotations-2.7.1.jar:/usr/local/openjdk7/jre/../lib/tools.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/guava/guava/11.0.2/guava-11.0.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-cli/commons-cli/1.2/commons-cli-1.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/commons/commons-math3/3.1.1/commons-math3-3.1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/xmlenc/xmlenc/0.52/xmlenc-0.52.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-httpclient/commons-httpclient/3.1/commons-h ttpclient-3.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-codec/commons-codec/1.4/commons-codec-1.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-io/commons-io/2.4/commons-io-2.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-net/commons-net/3.1/commons-net-3.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-collections/commons-collections/3.2.1/commons-collections-3.2.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/servlet/servlet-api/2.5/servlet-api-2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/mortbay/jetty/jetty/6.1.26/jetty-6.1.26.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/mortbay/jetty/jetty-util/6.1.26/jetty-util-6.1.26.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-core/1.9/jersey-core-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-json/1.9/jersey-json-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jettison/jettison/1.1/jettison-1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/wor k/m2/com/sun/xml/bind/jaxb-impl/2.2.3-1/jaxb! -impl-2.2.3-1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/xml/bind/jaxb-api/2.2.2/jaxb-api-2.2.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/xml/stream/stax-api/1.0-2/stax-api-1.0-2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/javax/activation/activation/1.1/activation-1.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-jaxrs/1.9.13/jackson-jaxrs-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-xc/1.9.13/jackson-xc-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/sun/jersey/jersey-server/1.9/jersey-server-1.9.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/asm/asm/3.2/asm-3.2.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-logging/commons-logging/1.1.3/commons-logging-1.1.3.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/log4j/log4j/1.2.17/log4j-1.2.17.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/net/java/dev/jets3t/jets3t/0.9.0/jets3t-0.9.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/httpcompo nents/httpclient/4.2.5/httpclient-4.2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/httpcomponents/httpcore/4.2.5/httpcore-4.2.5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/jamesmurty/utils/java-xmlbuilder/0.4/java-xmlbuilder-0.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-lang/commons-lang/2.6/commons-lang-2.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-configuration/commons-configuration/1.6/commons-configuration-1.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-digester/commons-digester/1.8/commons-digester-1.8.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-beanutils/commons-beanutils/1.7.0/commons-beanutils-1.7.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/commons-beanutils/commons-beanutils-core/1.8.0/commons-beanutils-core-1.8.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/slf4j/slf4j-api/1.7.10/slf4j-api-1.7.10.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/codehaus/jackson/jackson-core-asl/1.9.13/jackson-core-asl-1.9.13. jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2! /org/code! haus/jackson/jackson-mapper-asl/1.9.13/jackson-mapper-asl-1.9.13.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/avro/avro/1.7.4/avro-1.7.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/thoughtworks/paranamer/paranamer/2.3/paranamer-2.3.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/xerial/snappy/snappy-java/1.0.4.1/snappy-java-1.0.4.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/protobuf/protobuf-java/2.5.0/protobuf-java-2.5.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/code/gson/gson/2.2.4/gson-2.2.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-auth/target/hadoop-auth-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/server/apacheds-kerberos-codec/2.0.0-M15/apacheds-kerberos-codec-2.0.0-M15.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/server/apacheds-i18n/2.0.0-M15/apacheds-i18n-2.0.0-M15.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api -asn1-api/1.0.0-M20/api-asn1-api-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-util/1.0.0-M20/api-util-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/curator/curator-framework/2.7.1/curator-framework-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-i18n/1.0.0-M20/api-i18n-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-ldap-model/1.0.0-M20/api-ldap-model-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/mina/mina-core/2.0.0-M5/mina-core-2.0.0-M5.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/net/sf/ehcache/ehcache-core/2.4.4/ehcache-core-2.4.4.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/antlr/antlr/2.7.7/antlr-2.7.7.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/directory/api/api-asn1-ber/1.0.0-M20/api-asn1-ber-1.0.0-M20.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/jcraft/jsch/0.1.42/jsch-0.1.42.jar:/wrkdirs/usr/ports/devel/hadoop2/w ork/m2/org/apache/curator/curator-client/2.7! .1/curato! r-client-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/curator/curator-recipes/2.7.1/curator-recipes-2.7.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/com/google/code/findbugs/jsr305/3.0.0/jsr305-3.0.0.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/htrace/htrace-core/3.1.0-incubating/htrace-core-3.1.0-incubating.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/zookeeper/zookeeper/3.4.6/zookeeper-3.4.6.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/io/netty/netty/3.6.2.Final/netty-3.6.2.Final.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/apache/commons/commons-compress/1.4.1/commons-compress-1.4.1.jar:/wrkdirs/usr/ports/devel/hadoop2/work/m2/org/tukaani/xz/1.0/xz-1.0.jar org.apache.hadoop.io.compress.zlib.ZlibCompressor org.apache.hadoop.io.compress.zlib.ZlibDecompressor org.apache.hadoop.io.compress.bzip2.Bzip2Compressor org.apache.hadoop.io.compress.bzip2.Bzip2Decompressor org.apache.hadoop.security.JniBasedUnixGroupsMapping org.apache.hadoop.io.na tiveio.NativeIO org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory org.apache.hadoop.security.JniBasedUnixGroupsNetgroupMapping org.apache.hadoop.io.compress.snappy.SnappyCompressor org.apache.hadoop.io.compress.snappy.SnappyDecompressor org.apache.hadoop.io.compress.lz4.Lz4Compressor org.apache.hadoop.io.compress.lz4.Lz4Decompressor org.apache.hadoop.crypto.OpensslCipher org.apache.hadoop.crypto.random.OpensslSecureRandom org.apache.hadoop.util.NativeCrc32 org.apache.hadoop.net.unix.DomainSocket org.apache.hadoop.net.unix.DomainSocketWatcher [INFO] [INFO] --- maven-antrun-plugin:1.7:run (make) @ hadoop-common --- [INFO] Executing tasks main: [exec] -- The C compiler identification is GNU 4.2.1 [exec] -- The CXX compiler identification is GNU 4.2.1 [exec] -- Check for working C compiler: /usr/bin/cc [exec] -- Check for working C compiler: /usr/bin/cc -- works [exec] -- Detecting C compiler ABI info [exec] -- Detecting C compiler ABI info - done [exec] -- Detecting C compile features [exec] -- Detecting C compile features - done [exec] -- Check for working CXX compiler: /usr/bin/c++ [exec] -- Check for working CXX compiler: /usr/bin/c++ -- works [exec] -- Detecting CXX compiler ABI info [exec] -- Detecting CXX compiler ABI info - done [exec] -- Detecting CXX compile features [exec] -- Detecting CXX compile features - done [exec] -- Found JNI: /usr/local/openjdk7/jre/lib/amd64/libjawt.so [exec] -- Found ZLIB: /usr/lib/libz.so (found version "1.2.8") [exec] -- Looking for sync_file_range [exec] -- Looking for sync_file_range - not found [exec] -- Looking for posix_fadvise [exec] -- Looking for posix_fadvise - found [exec] -- Looking for dlopen in dl [exec] CUSTOM_OPENSSL_PREFIX = [exec] -- Looking for dlopen in dl - not found [exec] -- Performing Test HAS_NEW_ENOUGH_OPENSSL [exec] The OpenSSL library installed at /usr/lib/libcrypto.so is too old. You need a version at least new enough to have EVP_aes_256_ctr. [exec] Cannot find a usable OpenSSL library. OPENSSL_LIBRARY=/usr/lib/libcrypto.so, OPENSSL_INCLUDE_DIR=/usr/include, CUSTOM_OPENSSL_LIB=, CUSTOM_OPENSSL_PREFIX=, CUSTOM_OPENSSL_INCLUDE= [exec] CMake Error at CMakeLists.txt:207 (MESSAGE): [exec] Terminating build because require.openssl was specified. [exec] [exec] [exec] -- Performing Test HAS_NEW_ENOUGH_OPENSSL - Failed [exec] -- Configuring incomplete, errors occurred! [exec] See also "/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/native/CMakeFiles/CMakeOutput.log". [exec] See also "/wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/native/CMakeFiles/CMakeError.log". [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop Main ................................ SUCCESS [2.460s] [INFO] Apache Hadoop Project POM ......................... SUCCESS [1.508s] [INFO] Apache Hadoop Annotations ......................... SUCCESS [7.518s] [INFO] Apache Hadoop Assemblies .......................... SUCCESS [0.434s] [INFO] Apache Hadoop Project Dist POM .................... SUCCESS [2.077s] [INFO] Apache Hadoop Maven Plugins ....................... SUCCESS [6.871s] [INFO] Apache Hadoop MiniKDC ............................. SUCCESS [5.544s] [INFO] Apache Hadoop Auth ................................ SUCCESS [8.194s] [INFO] Apache Hadoop Auth Examples ....................... SUCCESS [6.664s] [INFO] Apache Hadoop Common .............................. FAILURE [29.018s] [INFO] Apache Hadoop NFS ................................. SKIPPED [INFO] Apache Hadoop KMS ................................. SKIPPED [INFO] Apache Hadoop Common Project ...................... SKIPPED [INFO] Apache Hadoop HDFS ................................ SKIPPED [INFO] Apache Hadoop HttpFS .............................. SKIPPED [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED [INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED [INFO] Apache Hadoop HDFS Project ........................ SKIPPED [INFO] hadoop-yarn ....................................... SKIPPED [INFO] hadoop-yarn-api ................................... SKIPPED [INFO] hadoop-yarn-common ................................ SKIPPED [INFO] hadoop-yarn-server ................................ SKIPPED [INFO] hadoop-yarn-server-common ......................... SKIPPED [INFO] hadoop-yarn-server-nodemanager .................... SKIPPED [INFO] hadoop-yarn-server-web-proxy ...................... SKIPPED [INFO] hadoop-yarn-server-applicationhistoryservice ...... SKIPPED [INFO] hadoop-yarn-server-resourcemanager ................ SKIPPED [INFO] hadoop-yarn-server-tests .......................... SKIPPED [INFO] hadoop-yarn-client ................................ SKIPPED [INFO] hadoop-yarn-server-sharedcachemanager ............. SKIPPED [INFO] hadoop-yarn-applications .......................... SKIPPED [INFO] hadoop-yarn-applications-distributedshell ......... SKIPPED [INFO] hadoop-yarn-applications-unmanaged-am-launcher .... SKIPPED [INFO] hadoop-yarn-site .................................. SKIPPED [INFO] hadoop-yarn-registry .............................. SKIPPED [INFO] hadoop-yarn-project ............................... SKIPPED [INFO] hadoop-mapreduce-client ........................... SKIPPED [INFO] hadoop-mapreduce-client-core ...................... SKIPPED [INFO] hadoop-mapreduce-client-common .................... SKIPPED [INFO] hadoop-mapreduce-client-shuffle ................... SKIPPED [INFO] hadoop-mapreduce-client-app ....................... SKIPPED [INFO] hadoop-mapreduce-client-hs ........................ SKIPPED [INFO] hadoop-mapreduce-client-jobclient ................. SKIPPED [INFO] hadoop-mapreduce-client-hs-plugins ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] hadoop-mapreduce .................................. SKIPPED [INFO] Apache Hadoop MapReduce Streaming ................. SKIPPED [INFO] Apache Hadoop Distributed Copy .................... SKIPPED [INFO] Apache Hadoop Archives ............................ SKIPPED [INFO] Apache Hadoop Rumen ............................... SKIPPED [INFO] Apache Hadoop Gridmix ............................. SKIPPED [INFO] Apache Hadoop Data Join ........................... SKIPPED [INFO] Apache Hadoop Ant Tasks ........................... SKIPPED [INFO] Apache Hadoop Extras .............................. SKIPPED [INFO] Apache Hadoop Pipes ............................... SKIPPED [INFO] Apache Hadoop OpenStack support ................... SKIPPED [INFO] Apache Hadoop Amazon Web Services support ......... SKIPPED [INFO] Apache Hadoop Azure support ....................... SKIPPED [INFO] Apache Hadoop Client .............................. SKIPPED [INFO] Apache Hadoop Mini-Cluster ........................ SKIPPED [INFO] Apache Hadoop Scheduler Load Simulator ............ SKIPPED [INFO] Apache Hadoop Tools Dist .......................... SKIPPED [INFO] Apache Hadoop Tools ............................... SKIPPED [INFO] Apache Hadoop Distribution ........................ SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 1:14.824s [INFO] Finished at: Sat Dec 12 05:48:06 GMT 2015 [INFO] Final Memory: 78M/2014M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7:run (make) on project hadoop-common: An Ant BuildException has occured: exec returned: 1 [ERROR] around Ant part ...... @ 4:158 in /wrkdirs/usr/ports/devel/hadoop2/work/hadoop-2.7.1-src/hadoop-common-project/hadoop-common/target/antrun/build-main.xml [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :hadoop-common *** [do-build] Error code 1 Stop in /usr/ports/devel/hadoop2.