From owner-freebsd-pkg-fallout@freebsd.org Sun Jul 26 20:04:51 2015 Return-Path: Delivered-To: freebsd-pkg-fallout@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id A4C769AB527 for ; Sun, 26 Jul 2015 20:04:51 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from mailman.ysv.freebsd.org (mailman.ysv.freebsd.org [IPv6:2001:1900:2254:206a::50:5]) by mx1.freebsd.org (Postfix) with ESMTP id 8F125EF4 for ; Sun, 26 Jul 2015 20:04:51 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: by mailman.ysv.freebsd.org (Postfix) id 8BDAC9AB525; Sun, 26 Jul 2015 20:04:51 +0000 (UTC) Delivered-To: pkg-fallout@mailman.ysv.freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:1900:2254:206a::19:1]) by mailman.ysv.freebsd.org (Postfix) with ESMTP id 8B40C9AB523 for ; Sun, 26 Jul 2015 20:04:51 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from beefy1.nyi.freebsd.org (beefy1.nyi.freebsd.org [IPv6:2610:1c1:1:6080::16:e4]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (Client did not present a certificate) by mx1.freebsd.org (Postfix) with ESMTPS id 4EA1FEF3; Sun, 26 Jul 2015 20:04:51 +0000 (UTC) (envelope-from pkg-fallout@FreeBSD.org) Received: from beefy1.nyi.freebsd.org (localhost [127.0.0.1]) by beefy1.nyi.freebsd.org (8.15.2/8.15.2) with ESMTP id t6QK4osJ065448; Sun, 26 Jul 2015 20:04:50 GMT (envelope-from pkg-fallout@FreeBSD.org) Received: (from root@localhost) by beefy1.nyi.freebsd.org (8.15.2/8.15.2/Submit) id t6QK4ov1065427; Sun, 26 Jul 2015 20:04:50 GMT (envelope-from pkg-fallout@FreeBSD.org) Date: Sun, 26 Jul 2015 20:04:50 GMT From: pkg-fallout@FreeBSD.org Message-Id: <201507262004.t6QK4ov1065427@beefy1.nyi.freebsd.org> To: demon@FreeBSD.org Subject: [package - 93i386-default][devel/spark] Failed for apache-spark-1.2.1 in build Cc: pkg-fallout@FreeBSD.org X-BeenThere: freebsd-pkg-fallout@freebsd.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: Fallout logs from package building List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Sun, 26 Jul 2015 20:04:51 -0000 You are receiving this mail as a port that you maintain is failing to build on the FreeBSD package build server. Please investigate the failure and submit a PR to fix build. Maintainer: demon@FreeBSD.org Last committer: mat@FreeBSD.org Ident: $FreeBSD: head/devel/spark/Makefile 386097 2015-05-11 18:34:57Z mat $ Log URL: http://beefy1.nyi.freebsd.org/data/93i386-default/392908/logs/apache-spark-1.2.1.log Build URL: http://beefy1.nyi.freebsd.org/build.html?mastername=93i386-default&build=392908 Log: ====>> Building devel/spark build started at Sun Jul 26 19:58:28 UTC 2015 port directory: /usr/ports/devel/spark building for: FreeBSD 93i386-default-job-03 9.3-RELEASE-p20 FreeBSD 9.3-RELEASE-p20 i386 maintained by: demon@FreeBSD.org Makefile ident: $FreeBSD: head/devel/spark/Makefile 386097 2015-05-11 18:34:57Z mat $ Poudriere version: 3.1.7 Host OSVERSION: 1100077 Jail OSVERSION: 903000 ---Begin Environment--- SHELL=/bin/csh UNAME_p=i386 UNAME_m=i386 OSVERSION=903000 UNAME_v=FreeBSD 9.3-RELEASE-p20 UNAME_r=9.3-RELEASE-p20 BLOCKSIZE=K MAIL=/var/mail/root STATUS=1 SAVED_TERM= MASTERMNT=/usr/local/poudriere/data/.m/93i386-default/ref PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/games:/usr/local/sbin:/usr/local/bin:/root/bin POUDRIERE_BUILD_TYPE=bulk PKGNAME=apache-spark-1.2.1 OLDPWD=/ PWD=/usr/local/poudriere/data/.m/93i386-default/ref/.p/pool MASTERNAME=93i386-default SCRIPTPREFIX=/usr/local/share/poudriere USER=root HOME=/root POUDRIERE_VERSION=3.1.7 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh LIBEXECPREFIX=/usr/local/libexec/poudriere LOCALBASE=/usr/local PACKAGE_BUILDING=yes ---End Environment--- ---Begin OPTIONS List--- ---End OPTIONS List--- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- PYTHON="/usr/local/bin/python2.7" XDG_DATA_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/spark/work HOME=/wrkdirs/usr/ports/devel/spark/work TMPDIR="/tmp" SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m" XDG_DATA_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/spark/work HOME=/wrkdirs/usr/ports/devel/spark/work TMPDIR="/tmp" NO_PIE=yes SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local LIBDIR="/usr/lib" CC="cc" CFLAGS="-O2 -pipe -fno-strict-aliasing" CPP="cpp" CPPFLAGS="" LDFLAGS="" LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fno-strict-aliasing" MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 444" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- SPARK_USER=spark SPARK_GROUP=spark VER=1.2.1 JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" PYTHON_INCLUDEDIR=include/python2.7 PYTHON_LIBDIR=lib/python2.7 PYTHON_PLATFORM=freebsd9 PYTHON_SITELIBDIR=lib/python2.7/site-packages PYTHON_VERSION=python2.7 PYTHON_VER=2.7 OSREL=9.3 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local PORTDOCS="" PORTEXAMPLES="" LIB32DIR=lib DOCSDIR="share/doc/spark" EXAMPLESDIR="share/examples/spark" DATADIR="share/spark" WWWDIR="www/spark" ETCDIR="etc/spark" --End PLIST_SUB-- --SUB_LIST-- SPARK_USER=spark SPARK_GROUP=spark JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" JAVA_VERSION="1.7+" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/spark DOCSDIR=/usr/local/share/doc/spark EXAMPLESDIR=/usr/local/share/examples/spark WWWDIR=/usr/local/www/spark ETCDIR=/usr/local/etc/spark --End SUB_LIST-- ---Begin make.conf--- MACHINE=i386 MACHINE_ARCH=i386 ARCH=${MACHINE_ARCH} USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles #### /usr/local/etc/poudriere.d/make.conf #### MASTER_SITE_FREEBSD=yes DISABLE_MAKE_JOBS=poudriere ---End make.conf--- =================================================== ===> License APACHE20 accepted by the user =========================================================================== =================================================== ===> apache-spark-1.2.1 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.5.5.txz [93i386-default-job-03] Installing pkg-1.5.5... [93i386-default-job-03] Extracting pkg-1.5.5: .......... done Message for pkg-1.5.5: If you are upgrading from the old package format, first run: # pkg2ng ===> apache-spark-1.2.1 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of apache-spark-1.2.1 =========================================================================== =================================================== =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-1.2.1 for building =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-1.2.1 for building => SHA256 Checksum OK for hadoop/spark-1.2.1.tgz. => SHA256 Checksum OK for hadoop/FreeBSD-spark-1.2.1-maven-repository.tar.gz. =========================================================================== =================================================== =========================================================================== =================================================== ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-1.2.1 for building ===> Extracting for apache-spark-1.2.1 => SHA256 Checksum OK for hadoop/spark-1.2.1.tgz. => SHA256 Checksum OK for hadoop/FreeBSD-spark-1.2.1-maven-repository.tar.gz. =========================================================================== =================================================== =========================================================================== =================================================== ===> Patching for apache-spark-1.2.1 ===> Applying FreeBSD patches for apache-spark-1.2.1 =========================================================================== =================================================== ===> apache-spark-1.2.1 depends on file: /usr/local/share/java/maven3/bin/mvn - not found ===> Installing existing package /packages/All/maven3-3.0.5.txz [93i386-default-job-03] Installing maven3-3.0.5... [93i386-default-job-03] `-- Installing maven-wrapper-1_2... [93i386-default-job-03] `-- Extracting maven-wrapper-1_2: . done [93i386-default-job-03] `-- Installing openjdk-7.80.15_1,1... [93i386-default-job-03] | `-- Installing alsa-lib-1.0.29... [93i386-default-job-03] | `-- Extracting alsa-lib-1.0.29: .......... done [93i386-default-job-03] | `-- Installing dejavu-2.35... [93i386-default-job-03] | | `-- Installing fontconfig-2.11.1,1... [93i386-default-job-03] | | `-- Installing expat-2.1.0_2... [93i386-default-job-03] | | `-- Extracting expat-2.1.0_2: .......... done [93i386-default-job-03] | | `-- Installing freetype2-2.6_1... [93i386-default-job-03] | | `-- Extracting freetype2-2.6_1: .......... done [93i386-default-job-03] | | `-- Extracting fontconfig-2.11.1,1: .......... done Running fc-cache to build fontconfig cache... /usr/local/share/fonts: skipping, no such directory /usr/local/lib/X11/fonts: skipping, no such directory /root/.local/share/fonts: skipping, no such directory /root/.fonts: skipping, no such directory /var/db/fontconfig: cleaning cache directory /root/.cache/fontconfig: not cleaning non-existent cache directory /root/.fontconfig: not cleaning non-existent cache directory fc-cache: succeeded [93i386-default-job-03] | | `-- Installing mkfontdir-1.0.7... [93i386-default-job-03] | | `-- Installing mkfontscale-1.1.2... [93i386-default-job-03] | | | `-- Installing libfontenc-1.1.2_3... [93i386-default-job-03] | | | `-- Installing xproto-7.0.27... [93i386-default-job-03] | | | `-- Extracting xproto-7.0.27: .......... done [93i386-default-job-03] | | | `-- Extracting libfontenc-1.1.2_3: ...... done [93i386-default-job-03] | | `-- Extracting mkfontscale-1.1.2: .. done [93i386-default-job-03] | | `-- Extracting mkfontdir-1.0.7: .. done [93i386-default-job-03] | `-- Extracting dejavu-2.35: .......... done [93i386-default-job-03] | `-- Installing java-zoneinfo-2015.e_1... [93i386-default-job-03] | `-- Extracting java-zoneinfo-2015.e_1: .......... done [93i386-default-job-03] | `-- Installing javavmwrapper-2.5... [93i386-default-job-03] | `-- Extracting javavmwrapper-2.5: .......... done [93i386-default-job-03] | `-- Installing libX11-1.6.2_3,1... [93i386-default-job-03] | | `-- Installing kbproto-1.0.6... [93i386-default-job-03] | | `-- Extracting kbproto-1.0.6: .......... done [93i386-default-job-03] | | `-- Installing libXau-1.0.8_3... [93i386-default-job-03] | | `-- Extracting libXau-1.0.8_3: .......... done [93i386-default-job-03] | | `-- Installing libXdmcp-1.1.2... [93i386-default-job-03] | | `-- Extracting libXdmcp-1.1.2: ......... done [93i386-default-job-03] | | `-- Installing libxcb-1.11_1... [93i386-default-job-03] | | `-- Installing libpthread-stubs-0.3_6... [93i386-default-job-03] | | `-- Extracting libpthread-stubs-0.3_6: ..... done [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ spark-core_2.10 --- [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (enforce-versions) @ spark-core_2.10 --- [INFO] [INFO] --- build-helper-maven-plugin:1.8:add-source (add-scala-sources) @ spark-core_2.10 --- [INFO] Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala added. [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process (default) @ spark-core_2.10 --- [INFO] [INFO] --- maven-antrun-plugin:1.7:run (default) @ spark-core_2.10 --- [WARNING] Parameter tasks is deprecated, use target instead [INFO] Executing tasks main: [unzip] Expanding: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/python/lib/py4j-0.8.2.1-src.zip into /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/python/build [INFO] Executed tasks [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ spark-core_2.10 --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 11 resources [INFO] Copying 23 resources [INFO] Copying 7 resources [INFO] Copying 3 resources [INFO] [INFO] --- scala-maven-plugin:3.2.0:compile (scala-compile-first) @ spark-core_2.10 --- [WARNING] Zinc server is not available at port 3030 - reverting to normal incremental compile [INFO] Using incremental compilation [INFO] compiler plugin: BasicArtifact(org.scalamacros,paradise_2.10.4,2.0.1,null) [INFO] Compiling 403 Scala sources and 33 Java sources to /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/scala-2.10/classes... [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/scheduler/TaskResultGetter.scala:50: inferred existential type (org.apache.spark.scheduler.DirectTaskResult[_$1], Int) forSome { type _$1 }, which cannot be expressed by wildcards, should be enabled by making the implicit value scala.language.existentials visible. This can be achieved by adding the import clause 'import scala.language.existentials' or by setting the compiler option -language:existentials. See the Scala docs for value scala.language.existentials for a discussion why the feature should be explicitly enabled. [WARNING] val (result, size) = serializer.get().deserialize[TaskResult[_]](serializedData) match { [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/SparkContext.scala:591: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new NewHadoopJob(hadoopConfiguration) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/SparkContext.scala:637: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new NewHadoopJob(hadoopConfiguration) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/SparkContext.scala:796: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new NewHadoopJob(conf) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala:168: constructor TaskID in class TaskID is deprecated: see corresponding Javadoc for more information. [WARNING] new TaskAttemptID(new TaskID(jID.value, true, splitID), attemptID)) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/SparkHadoopWriter.scala:189: method makeQualified in class Path is deprecated: see corresponding Javadoc for more information. [WARNING] outputPath.makeQualified(fs) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:103: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] if (!fs.getFileStatus(path).isDir) { [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/deploy/history/FsHistoryProvider.scala:157: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] val logDirs = if (logStatus != null) logStatus.filter(_.isDir).toSeq else Seq[FileStatus]() [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/input/PortableDataStream.scala:49: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] if (file.isDir) 0L else file.getLen [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/input/WholeTextFileInputFormat.scala:63: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] if (file.isDir) 0L else file.getLen [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/mapred/SparkHadoopMapRedUtil.scala:56: constructor TaskAttemptID in class TaskAttemptID is deprecated: see corresponding Javadoc for more information. [WARNING] new TaskAttemptID(jtIdentifier, jobId, isMap, taskId, attemptId) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/rdd/CheckpointRDD.scala:110: method getDefaultReplication in class FileSystem is deprecated: see corresponding Javadoc for more information. [WARNING] fs.create(tempOutputPath, false, bufferSize, fs.getDefaultReplication, blockSize) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/rdd/HadoopRDD.scala:349: constructor TaskID in class TaskID is deprecated: see corresponding Javadoc for more information. [WARNING] val taId = new TaskAttemptID(new TaskID(jobID, true, splitId), attemptId) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/rdd/PairRDDFunctions.scala:888: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new NewAPIHadoopJob(hadoopConf) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/rdd/PairRDDFunctions.scala:956: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new NewAPIHadoopJob(hadoopConf) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/scheduler/EventLoggingListener.scala:202: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] fileStatuses.filter(!_.isDir).map(_.getPath).toSeq [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/main/scala/org/apache/spark/scheduler/InputFormatInfo.scala:106: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new Job(conf) [WARNING] ^ [WARNING] 17 warnings found [WARNING] warning: [options] bootstrap class path not set in conjunction with -source 1.6 [WARNING] 1 warning [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ spark-core_2.10 --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 33 source files to /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/scala-2.10/classes [INFO] [INFO] --- build-helper-maven-plugin:1.8:add-test-source (add-scala-test-sources) @ spark-core_2.10 --- [INFO] Test Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala added. [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ spark-core_2.10 --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 4 resources [INFO] Copying 3 resources [INFO] [INFO] --- scala-maven-plugin:3.2.0:testCompile (scala-test-compile-first) @ spark-core_2.10 --- [WARNING] Zinc server is not available at port 3030 - reverting to normal incremental compile [INFO] Using incremental compilation [INFO] compiler plugin: BasicArtifact(org.scalamacros,paradise_2.10.4,2.0.1,null) [INFO] Compiling 125 Scala sources and 4 Java sources to /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/scala-2.10/test-classes... [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala/org/apache/spark/FileSuite.scala:497: constructor Job in class Job is deprecated: see corresponding Javadoc for more information. [WARNING] val job = new Job(sc.hadoopConfiguration) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala/org/apache/spark/metrics/InputOutputMetricsSuite.scala:34: trait ShouldMatchers in package matchers is deprecated: Please use org.scalatest.Matchers instead. [WARNING] class InputOutputMetricsSuite extends FunSuite with SharedSparkContext with ShouldMatchers { [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala/org/apache/spark/scheduler/EventLoggingListenerSuite.scala:177: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] assert(logDir.isDir) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala/org/apache/spark/scheduler/ReplayListenerSuite.scala:124: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] assert(eventLogDir.isDir) [WARNING] ^ [WARNING] /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/scala/org/apache/spark/util/FileLoggerSuite.scala:106: method isDir in class FileStatus is deprecated: see corresponding Javadoc for more information. [WARNING] assert(fileSystem.getFileStatus(logDirPath).isDir) [WARNING] ^ [WARNING] 5 warnings found [WARNING] warning: [options] bootstrap class path not set in conjunction with -source 1.6 [WARNING] Note: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/java/org/apache/spark/JavaAPISuite.java uses or overrides a deprecated API. [WARNING] Note: Recompile with -Xlint:deprecation for details. [WARNING] Note: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/src/test/java/org/apache/spark/JavaAPISuite.java uses unchecked or unsafe operations. [WARNING] Note: Recompile with -Xlint:unchecked for details. [WARNING] 1 warning [INFO] [INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ spark-core_2.10 --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 4 source files to /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/scala-2.10/test-classes [INFO] [INFO] --- maven-dependency-plugin:2.9:build-classpath (default) @ spark-core_2.10 --- [INFO] Wrote classpath file '/wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/spark-test-classpath.txt'. [INFO] [INFO] --- gmavenplus-plugin:1.2:execute (default) @ spark-core_2.10 --- [INFO] Using Groovy 2.3.7 to perform execute. [INFO] [INFO] --- maven-surefire-plugin:2.17:test (default-test) @ spark-core_2.10 --- [INFO] Tests are skipped. [INFO] [INFO] --- scalatest-maven-plugin:1.0:test (test) @ spark-core_2.10 --- [INFO] Tests are skipped. [INFO] [INFO] --- maven-jar-plugin:2.4:jar (default-jar) @ spark-core_2.10 --- [INFO] Building jar: /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/spark-core_2.10-1.2.1.jar [INFO] [INFO] --- maven-site-plugin:3.3:attach-descriptor (attach-descriptor) @ spark-core_2.10 --- [INFO] [INFO] --- maven-dependency-plugin:2.9:copy-dependencies (copy-dependencies) @ spark-core_2.10 --- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Spark Project Parent POM .......................... SUCCESS [8.402s] [INFO] Spark Project Networking .......................... SUCCESS [37.553s] [INFO] Spark Project Shuffle Streaming Service ........... SUCCESS [7.244s] [INFO] Spark Project Core ................................ FAILURE [4:49.374s] [INFO] Spark Project Bagel ............................... SKIPPED [INFO] Spark Project GraphX .............................. SKIPPED [INFO] Spark Project Streaming ........................... SKIPPED [INFO] Spark Project Catalyst ............................ SKIPPED [INFO] Spark Project SQL ................................. SKIPPED [INFO] Spark Project ML Library .......................... SKIPPED [INFO] Spark Project Tools ............................... SKIPPED [INFO] Spark Project Hive ................................ SKIPPED [INFO] Spark Project REPL ................................ SKIPPED [INFO] Spark Project YARN Parent POM ..................... SKIPPED [INFO] Spark Project YARN Stable API ..................... SKIPPED [INFO] Spark Project Assembly ............................ SKIPPED [INFO] Spark Project External Twitter .................... SKIPPED [INFO] Spark Project External Flume Sink ................. SKIPPED [INFO] Spark Project External Flume ...................... SKIPPED [INFO] Spark Project External MQTT ....................... SKIPPED [INFO] Spark Project External ZeroMQ ..................... SKIPPED [INFO] Spark Project External Kafka ...................... SKIPPED [INFO] Spark Project Examples ............................ SKIPPED [INFO] Spark Project YARN Shuffle Service ................ SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 5:43.911s [INFO] Finished at: Sun Jul 26 20:04:49 GMT 2015 [INFO] Final Memory: 57M/640M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-dependency-plugin:2.9:copy-dependencies (copy-dependencies) on project spark-core_2.10: Error copying artifact from /wrkdirs/usr/ports/devel/spark/work/m2/com/google/guava/guava/14.0.1/guava-14.0.1.jar to /wrkdirs/usr/ports/devel/spark/work/spark-1.2.1/core/target/jars/guava-14.0.1.jar: Map failed -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn -rf :spark-core_2.10 *** [do-build] Error code 1 Stop in /usr/ports/devel/spark.