=>> Building devel/spark build started at Sat Dec 4 19:48:06 MSK 2021 port directory: /usr/ports/devel/spark package name: apache-spark-2.1.1_2 building for: FreeBSD 13-amd64-default-py311-job-02 13.0-RELEASE-p4 FreeBSD 13.0-RELEASE-p4 amd64 maintained by: yuri@FreeBSD.org Makefile datestamp: -rw-r--r-- 1 root wheel 2737 Oct 18 19:18 /usr/ports/devel/spark/Makefile Poudriere version: poudriere-git-3.3.99.20211130 Host OSVERSION: 1300139 Jail OSVERSION: 1300139 Job Id: 02 ---Begin Environment--- SHELL=/bin/csh OSVERSION=1300139 UNAME_v=FreeBSD 13.0-RELEASE-p4 UNAME_r=13.0-RELEASE-p4 BLOCKSIZE=K MAIL=/var/mail/root MM_CHARSET=UTF-8 LANG=C.UTF-8 STATUS=1 HOME=/root PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin LOCALBASE=/usr/local USER=root POUDRIERE_NAME=poudriere-git LIBEXECPREFIX=/usr/local/libexec/poudriere POUDRIERE_VERSION=3.3.99.20211130 MASTERMNT=/usr/local/poudriere/data/.m/13-amd64-default-py311/ref LC_COLLATE=C POUDRIERE_BUILD_TYPE=bulk SAVED_TERM=screen GID=0 OUTPUT_REDIRECTED_STDERR=4 OUTPUT_REDIRECTED=1 UID=0 PWD=/usr/local/poudriere/data/.m/13-amd64-default-py311/02/.p OUTPUT_REDIRECTED_STDOUT=3 P_PORTS_FEATURES=FLAVORS SELECTED_OPTIONS MASTERNAME=13-amd64-default-py311 SCRIPTPREFIX=/usr/local/share/poudriere SCRIPTNAME=bulk.sh OLDPWD=/usr/local/poudriere/data/.m/13-amd64-default-py311/ref/.p/pool POUDRIERE_PKGNAME=poudriere-git-3.3.99.20211130 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh POUDRIEREPATH=/usr/local/bin/poudriere ---End Environment--- ---Begin Poudriere Port Flags/Env--- PORT_FLAGS= PKGENV= FLAVOR= DEPENDS_ARGS= MAKE_ARGS= ---End Poudriere Port Flags/Env--- ---Begin OPTIONS List--- ---End OPTIONS List--- --MAINTAINER-- yuri@FreeBSD.org --End MAINTAINER-- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- PYTHON="/usr/local/bin/python3.11" XDG_DATA_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CACHE_HOME=/wrkdirs/usr/ports/devel/spark/work/.cache HOME=/wrkdirs/usr/ports/devel/spark/work PATH=/wrkdirs/usr/ports/devel/spark/work/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- JAVA_HOME=/usr/local/openjdk8 MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m" XDG_DATA_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CACHE_HOME=/wrkdirs/usr/ports/devel/spark/work/.cache HOME=/wrkdirs/usr/ports/devel/spark/work PATH=/wrkdirs/usr/ports/devel/spark/work/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin NO_PIE=yes MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector-strong " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- SPARK_USER=spark SPARK_GROUP=spark VER=2.1.1 JAVASHAREDIR="share/java" JAVAJARDIR="share/java/classes" PYTHON_INCLUDEDIR=include/python3.11 PYTHON_LIBDIR=lib/python3.11 PYTHON_PLATFORM=freebsd13 PYTHON_SITELIBDIR=lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_EXT_SUFFIX=.cpython-311 PYTHON_VER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" OSREL=13.0 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local LIB32DIR=lib DOCSDIR="share/doc/spark" EXAMPLESDIR="share/examples/spark" DATADIR="share/spark" WWWDIR="www/spark" ETCDIR="etc/spark" --End PLIST_SUB-- --SUB_LIST-- SPARK_USER=spark SPARK_GROUP=spark JAVASHAREDIR="/usr/local/share/java" JAVAJARDIR="/usr/local/share/java/classes" JAVALIBDIR="/usr/local/share/java/classes" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/spark DOCSDIR=/usr/local/share/doc/spark EXAMPLESDIR=/usr/local/share/examples/spark WWWDIR=/usr/local/www/spark ETCDIR=/usr/local/etc/spark --End SUB_LIST-- ---Begin make.conf--- USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles FORCE_PACKAGE=yes #### /usr/local/etc/poudriere.d/make.conf #### PETITECHEZ_AGREE_LICENSE= yes _LICENSE_STATUS= accepted #OPTIONS_UNSET+= DOXYGEN BOOTSTRAP OPTIONS_UNSET+= LATEX LTO BUILD_ALL_PYTHON_FLAVORS= yes .if ${.CURDIR:T} == "llvm11" || ${.CURDIR:T} == "llvm12" || ${.CURDIR:T} == "llvm13" || ${.CURDIR:T} == "gcc-arm-embedded" || ${.CURDIR:T} == "llvm-devel" || ${.CURDIR:T} == "llvm-cheri" BROKEN= no heavt crap from ports .endif #### /usr/local/etc/poudriere.d/py311-make.conf #### DEFAULT_VERSIONS+= python=3.11 python3=3.11 WITH_CCACHE_BUILD=yes CCACHE_DIR=/root/.ccache #### Misc Poudriere #### .include "/etc/make.conf.ports_env" ---End make.conf--- --Resource limits-- cpu time (seconds, -t) unlimited file size (512-blocks, -f) unlimited data seg size (kbytes, -d) 33554432 stack size (kbytes, -s) 524288 core file size (512-blocks, -c) unlimited max memory size (kbytes, -m) unlimited locked memory (kbytes, -l) unlimited max user processes (-u) 34275 open files (-n) 1024 virtual mem size (kbytes, -v) 4194304 swap limit (kbytes, -w) unlimited socket buffer size (bytes, -b) unlimited pseudo-terminals (-p) unlimited kqueues (-k) unlimited umtx shared locks (-o) unlimited --End resource limits-- =================================================== ===== env: NO_DEPENDS=yes USER=nobody UID=65534 GID=65534 ===> License APACHE20 accepted by the user =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> apache-spark-2.1.1_2 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-1.17.5.pkg [13-amd64-default-py311-job-02] Installing pkg-1.17.5... [13-amd64-default-py311-job-02] Extracting pkg-1.17.5: .......... done ===> apache-spark-2.1.1_2 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of apache-spark-2.1.1_2 =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-2.1.1_2 for building =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-2.1.1_2 for building => SHA256 Checksum OK for hadoop/spark-2.1.1.tgz. => SHA256 Checksum OK for hadoop/FreeBSD-spark-2.1.1-maven-repository.tar.gz. =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=nobody UID=65534 GID=65534 ===> License APACHE20 accepted by the user ===> Fetching all distfiles required by apache-spark-2.1.1_2 for building ===> Extracting for apache-spark-2.1.1_2 => SHA256 Checksum OK for hadoop/spark-2.1.1.tgz. => SHA256 Checksum OK for hadoop/FreeBSD-spark-2.1.1-maven-repository.tar.gz. =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=nobody UID=65534 GID=65534 ===> Patching for apache-spark-2.1.1_2 ===> Applying FreeBSD patches for apache-spark-2.1.1_2 from /usr/ports/devel/spark/files =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> apache-spark-2.1.1_2 depends on file: /usr/local/share/java/maven33/bin/mvn - not found ===> Installing existing package /packages/All/maven33-3.3.9_1.pkg [13-amd64-default-py311-job-02] Installing maven33-3.3.9_1... [13-amd64-default-py311-job-02] `-- Installing maven-wrapper-1_2... [13-amd64-default-py311-job-02] `-- Extracting maven-wrapper-1_2: .. done [13-amd64-default-py311-job-02] `-- Installing openjdk8-8.312.07.1... [13-amd64-default-py311-job-02] | `-- Installing alsa-lib-1.2.2_1... [13-amd64-default-py311-job-02] | `-- Extracting alsa-lib-1.2.2_1: .......... done [13-amd64-default-py311-job-02] | `-- Installing dejavu-2.37_1... [13-amd64-default-py311-job-02] | | `-- Installing fontconfig-2.13.94_1,1... [13-amd64-default-py311-job-02] | | `-- Installing expat-2.4.1... [13-amd64-default-py311-job-02] | | `-- Extracting expat-2.4.1: .......... done [13-amd64-default-py311-job-02] | | `-- Installing freetype2-2.11.0... [13-amd64-default-py311-job-02] | | | `-- Installing png-1.6.37_1... [13-amd64-default-py311-job-02] | | | `-- Extracting png-1.6.37_1: .......... done [13-amd64-default-py311-job-02] | | `-- Extracting freetype2-2.11.0: .......... done [13-amd64-default-py311-job-02] | | `-- Extracting fontconfig-2.13.94_1,1: .......... done Running fc-cache to build fontconfig cache... Font directories: /usr/local/share/fonts /usr/local/lib/X11/fonts /usr/local/share/fonts: skipping, no such directory /usr/local/lib/X11/fonts: skipping, no such directory /var/db/fontconfig: cleaning cache directory fc-cache: succeeded [13-amd64-default-py311-job-02] | | `-- Installing mkfontscale-1.2.1... [13-amd64-default-py311-job-02] | | `-- Installing libfontenc-1.1.4... [13-amd64-default-py311-job-02] | | `-- Extracting libfontenc-1.1.4: ......... done [13-amd64-default-py311-job-02] | | `-- Extracting mkfontscale-1.2.1: ....... done [13-amd64-default-py311-job-02] | `-- Extracting dejavu-2.37_1: .......... done [13-amd64-default-py311-job-02] | `-- Installing giflib-5.2.1... [13-amd64-default-py311-job-02] | `-- Extracting giflib-5.2.1: .......... done [13-amd64-default-py311-job-02] | `-- Installing java-zoneinfo-2021.e... [13-amd64-default-py311-job-02] | `-- Extracting java-zoneinfo-2021.e: .......... done [13-amd64-default-py311-job-02] | `-- Installing javavmwrapper-2.7.8... [13-amd64-default-py311-job-02] | `-- Extracting javavmwrapper-2.7.8: .......... done [13-amd64-default-py311-job-02] | `-- Installing libX11-1.7.2,1... [13-amd64-default-py311-job-02] | | `-- Installing libxcb-1.14_1... [13-amd64-default-py311-job-02] | | `-- Installing libXau-1.0.9... [13-amd64-default-py311-job-02] | | `-- Extracting libXau-1.0.9: .......... done [13-amd64-default-py311-job-02] | | `-- Installing libXdmcp-1.1.3... [13-amd64-default-py311-job-02] | | | `-- Installing xorgproto-2021.5... [13-amd64-default-py311-job-02] | | | `-- Extracting xorgproto-2021.5: .......... done [13-amd64-default-py311-job-02] | | `-- Extracting libXdmcp-1.1.3: ......... done [13-amd64-default-py311-job-02] | | `-- Installing libpthread-stubs-0.4... [13-amd64-default-py311-job-02] | | `-- Extracting libpthread-stubs-0.4: .... done [13-amd64-default-py311-job-02] | | `-- Extracting libxcb-1.14_1: .......... done [13-amd64-default-py311-job-02] | `-- Extracting libX11-1.7.2,1: .......... done [13-amd64-default-py311-job-02] | `-- Installing libXext-1.3.4,1... [13-amd64-default-py311-job-02] | `-- Extracting libXext-1.3.4,1: .......... done [13-amd64-default-py311-job-02] | `-- Installing libXi-1.8,1... [13-amd64-default-py311-job-02] | | `-- Installing libXfixes-6.0.0... [13-amd64-default-py311-job-02] | | `-- Extracting libXfixes-6.0.0: .......... done [13-amd64-default-py311-job-02] | `-- Extracting libXi-1.8,1: .......... done [13-amd64-default-py311-job-02] | `-- Installing libXrender-0.9.10_2... [13-amd64-default-py311-job-02] | `-- Extracting libXrender-0.9.10_2: .......... done [13-amd64-default-py311-job-02] | `-- Installing libXt-1.2.1,1... [13-amd64-default-py311-job-02] | | `-- Installing libICE-1.0.10,1... [13-amd64-default-py311-job-02] | | `-- Extracting libICE-1.0.10,1: .......... done [13-amd64-default-py311-job-02] | | `-- Installing libSM-1.2.3,1... [13-amd64-default-py311-job-02] | | `-- Extracting libSM-1.2.3,1: .......... done [13-amd64-default-py311-job-02] | `-- Extracting libXt-1.2.1,1: .......... done [13-amd64-default-py311-job-02] | `-- Installing libXtst-1.2.3_2... [13-amd64-default-py311-job-02] | `-- Extracting libXtst-1.2.3_2: .......... done [13-amd64-default-py311-job-02] | `-- Installing libinotify-20180201_2... [13-amd64-default-py311-job-02] | `-- Extracting libinotify-20180201_2: .......... done [13-amd64-default-py311-job-02] `-- Extracting openjdk8-8.312.07.1: .......... done [13-amd64-default-py311-job-02] Extracting maven33-3.3.9_1: .......... done ===== Message from maven-wrapper-1_2: -- ===> NOTICE: The maven-wrapper port currently does not have a maintainer. As a result, it is more likely to have unresolved issues, not be up-to-date, or even be removed in the future. To volunteer to maintain this port, please create an issue at: https://bugs.freebsd.org/bugzilla More information about port maintainership is available at: https://docs.freebsd.org/en/articles/contributing/#ports-contributing ===== Message from alsa-lib-1.2.2_1: -- ===> NOTICE: The alsa-lib port currently does not have a maintainer. As a result, it is more likely to have unresolved issues, not be up-to-date, or even be removed in the future. To volunteer to maintain this port, please create an issue at: https://bugs.freebsd.org/bugzilla More information about port maintainership is available at: https://docs.freebsd.org/en/articles/contributing/#ports-contributing ===== Message from freetype2-2.11.0: -- The 2.7.x series now uses the new subpixel hinting mode (V40 port's option) as the default, emulating a modern version of ClearType. This change inevitably leads to different rendering results, and you might change port's options to adapt it to your taste (or use the new "FREETYPE_PROPERTIES" environment variable). The environment variable "FREETYPE_PROPERTIES" can be used to control the driver properties. Example: FREETYPE_PROPERTIES=truetype:interpreter-version=35 \ cff:no-stem-darkening=1 \ autofitter:warping=1 This allows to select, say, the subpixel hinting mode at runtime for a given application. If LONG_PCF_NAMES port's option was enabled, the PCF family names may include the foundry and information whether they contain wide characters. For example, "Sony Fixed" or "Misc Fixed Wide", instead of "Fixed". This can be disabled at run time with using pcf:no-long-family-names property, if needed. Example: FREETYPE_PROPERTIES=pcf:no-long-family-names=1 How to recreate fontconfig cache with using such environment variable, if needed: # env FREETYPE_PROPERTIES=pcf:no-long-family-names=1 fc-cache -fsv The controllable properties are listed in the section "Controlling FreeType Modules" in the reference's table of contents (/usr/local/share/doc/freetype2/reference/index.html, if documentation was installed). ===== Message from dejavu-2.37_1: -- Make sure that the freetype module is loaded. If it is not, add the following line to the "Modules" section of your X Windows configuration file: Load "freetype" Add the following line to the "Files" section of X Windows configuration file: FontPath "/usr/local/share/fonts/dejavu/" Note: your X Windows configuration file is typically /etc/X11/XF86Config if you are using XFree86, and /etc/X11/xorg.conf if you are using X.Org. ===== Message from libinotify-20180201_2: -- Libinotify functionality on FreeBSD is missing support for - detecting a file being moved into or out of a directory within the same filesystem - certain modifications to a symbolic link (rather than the file it points to.) in addition to the known limitations on all platforms using kqueue(2) where various open and close notifications are unimplemented. This means the following regression tests will fail: Directory notifications: IN_MOVED_FROM IN_MOVED_TO Open/close notifications: IN_OPEN IN_CLOSE_NOWRITE IN_CLOSE_WRITE Symbolic Link notifications: IN_DONT_FOLLOW IN_ATTRIB IN_MOVE_SELF IN_DELETE_SELF Kernel patches to address the missing directory and symbolic link notifications are available from: https://github.com/libinotify-kqueue/libinotify-kqueue/tree/master/patches You might want to consider increasing the kern.maxfiles tunable if you plan to use this library for applications that need to monitor activity of a lot of files. ===== Message from openjdk8-8.312.07.1: -- This OpenJDK implementation requires fdescfs(5) mounted on /dev/fd and procfs(5) mounted on /proc. If you have not done it yet, please do the following: mount -t fdescfs fdesc /dev/fd mount -t procfs proc /proc To make it permanent, you need the following lines in /etc/fstab: fdesc /dev/fd fdescfs rw 0 0 proc /proc procfs rw 0 0 ===== Message from maven33-3.3.9_1: -- ===> NOTICE: This port is deprecated; you may wish to reconsider installing it: Outdated, unsupported upstream. It is scheduled to be removed on or after 2022-01-15. ===> apache-spark-2.1.1_2 depends on file: /usr/local/share/java/maven33/bin/mvn - found ===> Returning to build of apache-spark-2.1.1_2 ===> apache-spark-2.1.1_2 depends on file: /usr/local/lib/libsnappyjava.so - not found ===> Installing existing package /packages/All/snappyjava-1.1.7.5.pkg [13-amd64-default-py311-job-02] Installing snappyjava-1.1.7.5... [13-amd64-default-py311-job-02] Extracting snappyjava-1.1.7.5: ..... done ===> apache-spark-2.1.1_2 depends on file: /usr/local/lib/libsnappyjava.so - found ===> Returning to build of apache-spark-2.1.1_2 ===> apache-spark-2.1.1_2 depends on executable: bash - not found ===> Installing existing package /packages/All/bash-5.1.12.pkg [13-amd64-default-py311-job-02] Installing bash-5.1.12... [13-amd64-default-py311-job-02] `-- Installing gettext-runtime-0.21... [13-amd64-default-py311-job-02] | `-- Installing indexinfo-0.3.1... [13-amd64-default-py311-job-02] | `-- Extracting indexinfo-0.3.1: .... done [13-amd64-default-py311-job-02] `-- Extracting gettext-runtime-0.21: .......... done [13-amd64-default-py311-job-02] `-- Installing readline-8.1.1... [13-amd64-default-py311-job-02] `-- Extracting readline-8.1.1: .......... done [13-amd64-default-py311-job-02] Extracting bash-5.1.12: .......... done ===> apache-spark-2.1.1_2 depends on executable: bash - found ===> Returning to build of apache-spark-2.1.1_2 ===> apache-spark-2.1.1_2 depends on file: /usr/local/openjdk8/bin/java - found ===> apache-spark-2.1.1_2 depends on file: /usr/local/bin/python3.11 - not found ===> Installing existing package /packages/All/python311-3.11.0.a2.pkg [13-amd64-default-py311-job-02] Installing python311-3.11.0.a2... [13-amd64-default-py311-job-02] `-- Installing libffi-3.3_1... [13-amd64-default-py311-job-02] `-- Extracting libffi-3.3_1: .......... done [13-amd64-default-py311-job-02] `-- Installing mpdecimal-2.5.1... [13-amd64-default-py311-job-02] `-- Extracting mpdecimal-2.5.1: .......... done [13-amd64-default-py311-job-02] Extracting python311-3.11.0.a2: .......... done ===== Message from python311-3.11.0.a2: -- Note that some standard Python modules are provided as separate ports as they require additional dependencies. They are available as: py311-gdbm databases/py-gdbm@py311 py311-sqlite3 databases/py-sqlite3@py311 py311-tkinter x11-toolkits/py-tkinter@py311 ===> apache-spark-2.1.1_2 depends on file: /usr/local/bin/python3.11 - found ===> Returning to build of apache-spark-2.1.1_2 ===> apache-spark-2.1.1_2 depends on file: /usr/local/bin/ccache - not found ===> Installing existing package /packages/All/ccache-3.7.12_2.pkg [13-amd64-default-py311-job-02] Installing ccache-3.7.12_2... [13-amd64-default-py311-job-02] Extracting ccache-3.7.12_2: .......... done Create compiler links... create symlink for cc create symlink for cc (world) create symlink for c++ create symlink for c++ (world) create symlink for CC create symlink for CC (world) create symlink for clang create symlink for clang (world) create symlink for clang++ create symlink for clang++ (world) ===== Message from ccache-3.7.12_2: -- NOTE: Please read /usr/local/share/doc/ccache/ccache-howto-freebsd.txt for information on using ccache with FreeBSD ports and src. ===> apache-spark-2.1.1_2 depends on file: /usr/local/bin/ccache - found ===> Returning to build of apache-spark-2.1.1_2 =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=nobody UID=65534 GID=65534 ===> Configuring for apache-spark-2.1.1_2 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=nobody UID=65534 GID=65534 ===> Building for apache-spark-2.1.1_2 /bin/mkdir -p /wrkdirs/usr/ports/devel/spark/work/snappy/org/xerial/snappy/native/FreeBSD/x86_64 cp -f /usr/local/lib/libsnappyjava.so /wrkdirs/usr/ports/devel/spark/work/snappy/org/xerial/snappy/native/FreeBSD/x86_64 cd /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1 && /usr/bin/env JAVA_HOME=/usr/local/openjdk8 MAVEN_OPTS="-Xmx2g -XX:MaxPermSize=512M -XX:ReservedCodeCacheSize=512m" XDG_DATA_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CONFIG_HOME=/wrkdirs/usr/ports/devel/spark/work XDG_CACHE_HOME=/wrkdirs/usr/ports/devel/spark/work/.cache HOME=/wrkdirs/usr/ports/devel/spark/work PATH=/usr/local/libexec/ccache:/wrkdirs/usr/ports/devel/spark/work/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/nonexistent/bin NO_PIE=yes MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" -fstack-protector-strong " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " MANPREFIX="/usr/local" CCACHE_DIR="/root/.ccache" BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" /usr/local/share/java/maven33/bin/mvn --offline -Dmaven.repo.local=/wrkdirs/usr/ports/devel/spark/work/m2 clean package -Dhadoop.version=2.7.2 -Pyarn -Phadoop-2.6 -DskipTests -Duser.home=/wrkdirs/usr/ports/devel/spark/work OpenJDK 64-Bit Server VM warning: ignoring option MaxPermSize=512M; support was removed in 8.0 [INFO] Scanning for projects... [INFO] ------------------------------------------------------------------------ [INFO] Reactor Build Order: [INFO] [INFO] Spark Project Parent POM [INFO] Spark Project Tags [INFO] Spark Project Sketch [INFO] Spark Project Networking [INFO] Spark Project Shuffle Streaming Service [INFO] Spark Project Unsafe [INFO] Spark Project Launcher [INFO] Spark Project Core [INFO] Spark Project ML Local Library [INFO] Spark Project GraphX [INFO] Spark Project Streaming [INFO] Spark Project Catalyst [INFO] Spark Project SQL [INFO] Spark Project ML Library [INFO] Spark Project Tools [INFO] Spark Project Hive [INFO] Spark Project REPL [INFO] Spark Project YARN Shuffle Service [INFO] Spark Project YARN [INFO] Spark Project Assembly [INFO] Spark Project External Flume Sink [INFO] Spark Project External Flume [INFO] Spark Project External Flume Assembly [INFO] Spark Integration for Kafka 0.8 [INFO] Spark Project Examples [INFO] Spark Project External Kafka Assembly [INFO] Spark Integration for Kafka 0.10 [INFO] Spark Integration for Kafka 0.10 Assembly [INFO] Kafka 0.10 Source for Structured Streaming [INFO] Spark Project Java 8 Tests [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Spark Project Parent POM 2.1.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:3.0.0:clean (default-clean) @ spark-parent_2.11 --- [INFO] [INFO] --- maven-enforcer-plugin:1.4.1:enforce (enforce-versions) @ spark-parent_2.11 --- [INFO] [INFO] --- scala-maven-plugin:3.2.2:add-source (eclipse-add-source) @ spark-parent_2.11 --- [INFO] Add Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/src/main/scala [INFO] Add Test Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/src/test/scala [INFO] [INFO] --- maven-dependency-plugin:2.10:build-classpath (default-cli) @ spark-parent_2.11 --- [INFO] Dependencies classpath: /wrkdirs/usr/ports/devel/spark/work/m2/org/spark-project/spark/unused/1.0.0/unused-1.0.0.jar [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process (default) @ spark-parent_2.11 --- [INFO] [INFO] --- scala-maven-plugin:3.2.2:compile (scala-compile-first) @ spark-parent_2.11 --- [INFO] No sources to compile [INFO] [INFO] --- maven-antrun-plugin:1.8:run (create-tmp-dir) @ spark-parent_2.11 --- [INFO] Executing tasks main: [mkdir] Created dir: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/target/tmp [INFO] Executed tasks [INFO] [INFO] --- scala-maven-plugin:3.2.2:testCompile (scala-test-compile-first) @ spark-parent_2.11 --- [INFO] No sources to compile [INFO] [INFO] --- maven-dependency-plugin:2.10:build-classpath (generate-test-classpath) @ spark-parent_2.11 --- [INFO] [INFO] --- maven-surefire-plugin:2.19.1:test (test) @ spark-parent_2.11 --- [INFO] Tests are skipped. [INFO] [INFO] --- scalatest-maven-plugin:1.0:test (test) @ spark-parent_2.11 --- [INFO] Tests are skipped. [INFO] [INFO] --- maven-jar-plugin:2.6:test-jar (prepare-test-jar) @ spark-parent_2.11 --- [INFO] Building jar: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/target/spark-parent_2.11-2.1.1-tests.jar [INFO] [INFO] --- maven-site-plugin:3.3:attach-descriptor (attach-descriptor) @ spark-parent_2.11 --- [INFO] [INFO] --- maven-shade-plugin:2.4.3:shade (default) @ spark-parent_2.11 --- [INFO] Including org.spark-project.spark:unused:jar:1.0.0 in the shaded jar. [INFO] Replacing original artifact with shaded artifact. [INFO] [INFO] --- maven-source-plugin:2.4:jar-no-fork (create-source-jar) @ spark-parent_2.11 --- [INFO] [INFO] --- maven-source-plugin:2.4:test-jar-no-fork (create-source-jar) @ spark-parent_2.11 --- [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Spark Project Tags 2.1.1 [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:3.0.0:clean (default-clean) @ spark-tags_2.11 --- [INFO] [INFO] --- maven-enforcer-plugin:1.4.1:enforce (enforce-versions) @ spark-tags_2.11 --- [INFO] [INFO] --- scala-maven-plugin:3.2.2:add-source (eclipse-add-source) @ spark-tags_2.11 --- [INFO] Add Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/common/tags/src/main/scala [INFO] Add Test Source directory: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/common/tags/src/test/scala [INFO] [INFO] --- maven-dependency-plugin:2.10:build-classpath (default-cli) @ spark-tags_2.11 --- [INFO] Dependencies classpath: /wrkdirs/usr/ports/devel/spark/work/m2/org/spark-project/spark/unused/1.0.0/unused-1.0.0.jar:/wrkdirs/usr/ports/devel/spark/work/m2/org/scala-lang/scala-library/2.11.8/scala-library-2.11.8.jar [INFO] [INFO] --- maven-remote-resources-plugin:1.5:process (default) @ spark-tags_2.11 --- [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ spark-tags_2.11 --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/common/tags/src/main/resources [INFO] Copying 3 resources [INFO] [INFO] --- scala-maven-plugin:3.2.2:compile (scala-compile-first) @ spark-tags_2.11 --- [WARNING] Zinc server is not available at port 3030 - reverting to normal incremental compile [INFO] Using incremental compilation [INFO] 'compiler-interface' not yet compiled for Scala 2.11.8. Compiling... [INFO] Compilation completed in 34.038 s [INFO] Compiling 2 Scala sources and 6 Java sources to /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/common/tags/target/scala-2.11/classes... # # A fatal error has been detected by the Java Runtime Environment: # # SIGSEGV (0xb) at pc=0x00000008010a7737, pid=58848, tid=0x000000000006b979 # # JRE version: OpenJDK Runtime Environment (8.0_312-b07) (build 1.8.0_312-b07) # Java VM: OpenJDK 64-Bit Server VM (25.312-b07 mixed mode bsd-amd64 compressed oops) # Problematic frame: # C [libthr.so.3+0x15737] [error occurred during error reporting (printing problematic frame), id 0xb] # Core dump written. Default location: /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/java.core # # An error report file with more information is saved as: # /wrkdirs/usr/ports/devel/spark/work/spark-2.1.1/hs_err_pid58848.log # # If you would like to submit a bug report, please visit: # https://bugs.freebsd.org/bugzilla/enter_bug.cgi?product=Ports%20%26%20Packages&component=Individual%20Port(s)&short_desc=java/openjdk8%3A%20 # *** Signal 6 Stop. make: stopped in /usr/ports/devel/spark =>> Cleaning up wrkdir ===> Cleaning for apache-spark-2.1.1_2 build of devel/spark | apache-spark-2.1.1_2 ended at Sat Dec 4 19:50:34 MSK 2021 build time: 00:02:29 !!! build failure encountered !!!