Skip site navigation (1)Skip section navigation (2)
Date:      Sun, 17 Mar 2019 06:16:24 GMT
From:      pkg-fallout@FreeBSD.org
To:        skreuzer@FreeBSD.org
Cc:        pkg-fallout@FreeBSD.org
Subject:   [exp - 112amd64-default-build-as-user][www/py-scrapy] Failed for py27-Scrapy-1.6.0 in run-depends
Message-ID:  <201903170616.x2H6GO0A090446@package19.nyi.freebsd.org>

next in thread | raw e-mail | index | archive | help
You are receiving this mail as a port that you maintain
is failing to build on the FreeBSD package build server.
Please investigate the failure and submit a PR to fix
build.

Maintainer:     skreuzer@FreeBSD.org
Last committer: skreuzer@FreeBSD.org
Ident:          $FreeBSD: head/www/py-scrapy/Makefile 495093 2019-03-09 00:55:41Z skreuzer $
Log URL:        http://package19.nyi.freebsd.org/data/112amd64-default-build-as-user/495839/logs/py27-Scrapy-1.6.0.log
Build URL:      http://package19.nyi.freebsd.org/build.html?mastername=112amd64-default-build-as-user&build=495839
Log:

=>> Building www/py-scrapy
build started at Sun Mar 17 06:16:16 UTC 2019
port directory: /usr/ports/www/py-scrapy
package name: py27-Scrapy-1.6.0
building for: FreeBSD 112amd64-default-build-as-user-job-24 11.2-RELEASE-p9 FreeBSD 11.2-RELEASE-p9 amd64
maintained by: skreuzer@FreeBSD.org
Makefile ident:      $FreeBSD: head/www/py-scrapy/Makefile 495093 2019-03-09 00:55:41Z skreuzer $
Poudriere version: 3.2.8-3-g02cc9753
Host OSVERSION: 1300009
Jail OSVERSION: 1102000
Job Id: 24

---Begin Environment---
SHELL=/bin/csh
OSVERSION=1102000
UNAME_v=FreeBSD 11.2-RELEASE-p9
UNAME_r=11.2-RELEASE-p9
BLOCKSIZE=K
MAIL=/var/mail/root
STATUS=1
HOME=/root
PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin
LOCALBASE=/usr/local
USER=root
LIBEXECPREFIX=/usr/local/libexec/poudriere
POUDRIERE_VERSION=3.2.8-3-g02cc9753
MASTERMNT=/poudriere/data/.m/112amd64-default-build-as-user/ref
POUDRIERE_BUILD_TYPE=bulk
PACKAGE_BUILDING=yes
SAVED_TERM=
GID=0
UID=0
PWD=/poudriere/data/.m/112amd64-default-build-as-user/ref/.p/pool
P_PORTS_FEATURES=FLAVORS SELECTED_OPTIONS
MASTERNAME=112amd64-default-build-as-user
SCRIPTPREFIX=/usr/local/share/poudriere
OLDPWD=/poudriere/data/.m/112amd64-default-build-as-user/ref/.p
SCRIPTPATH=/usr/local/share/poudriere/bulk.sh
POUDRIEREPATH=/usr/local/bin/poudriere
---End Environment---

---Begin Poudriere Port Flags/Env---
PORT_FLAGS=
PKGENV=
FLAVOR=py27
DEPENDS_ARGS=
MAKE_ARGS= FLAVOR=py27
---End Poudriere Port Flags/Env---

---Begin OPTIONS List---
===> The following configuration options are available for py27-Scrapy-1.6.0:
     SSL=on: SSL protocol support
===> Use 'make config' to modify these settings
---End OPTIONS List---

--MAINTAINER--
skreuzer@FreeBSD.org
--End MAINTAINER--

--CONFIGURE_ARGS--

--End CONFIGURE_ARGS--

--CONFIGURE_ENV--
PYTHON="/usr/local/bin/python2.7" XDG_DATA_HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27  XDG_CONFIG_HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27  HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/www/py-scrapy/work-py27/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin SHELL=/bin/sh CONFIG_SHELL=/bin/sh
--End CONFIGURE_ENV--

--MAKE_ENV--
XDG_DATA_HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27  XDG_CONFIG_HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27  HOME=/wrkdirs/usr/ports/www/py-scrapy/work-py27 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/www/py-scrapy/work-py27/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin NO_PIE=yes MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES LDSHARED="cc -shared" PYTHONDONTWRITEBYTECODE= PYTHONOPTIMIZE= PREFIX=/usr/local  LOCALBASE=/usr/local  CC="cc" CFLAGS="-O2 -pipe  -fstack-protector -fno-strict-aliasing "  CPP="cpp" CPPFLAGS=""  LDFLAGS=" -fstack-protector " LIBS=""  CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector -fno-strict-aliasing  "  MANPREFIX="/usr/local" BSD_INSTALL_PROGRAM="install  -s -m 555"  BSD_INSTALL_LIB="install  -s -m 0644"  BSD_INSTALL_SCRIPT="install  -m 555"  BSD_INSTALL_DATA="install  -m 0644"  BSD_INSTALL_MAN="install  -m 444"
--End MAKE_ENV--

--PLIST_SUB--
PYTHON_INCLUDEDIR=include/python2.7  PYTHON_LIBDIR=lib/python2.7  PYTHON_PLATFORM=freebsd11  PYTHON_PYOEXTENSION=pyo  PYTHON_SITELIBDIR=lib/python2.7/site-packages  PYTHON_SUFFIX=27  PYTHON_VER=2.7  PYTHON_VERSION=python2.7 PYTHON2="" PYTHON3="@comment " OSREL=11.2 PREFIX=%D LOCALBASE=/usr/local  RESETPREFIX=/usr/local LIB32DIR=lib DOCSDIR="share/doc/py27-Scrapy"  EXAMPLESDIR="share/examples/py27-Scrapy"  DATADIR="share/py27-Scrapy"  WWWDIR="www/py27-Scrapy"  ETCDIR="etc/py27-Scrapy"
--End PLIST_SUB--

--SUB_LIST--
PREFIX=/usr/local LOCALBASE=/usr/local  DATADIR=/usr/local/share/py27-Scrapy DOCSDIR=/usr/local/share/doc/py27-Scrapy EXAMPLESDIR=/usr/local/share/examples/py27-Scrapy  WWWDIR=/usr/local/www/py27-Scrapy ETCDIR=/usr/local/etc/py27-Scrapy
--End SUB_LIST--

---Begin make.conf---
USE_PACKAGE_DEPENDS=yes
BATCH=yes
WRKDIRPREFIX=/wrkdirs
PORTSDIR=/usr/ports
PACKAGES=/packages
DISTDIR=/distfiles
FORCE_PACKAGE=yes
PACKAGE_BUILDING=yes
PACKAGE_BUILDING_FLAVORS=yes
#### /usr/local/etc/poudriere.d/make.conf ####
# Build ALLOW_MAKE_JOBS_PACKAGES with 2 jobs
MAKE_JOBS_NUMBER=2

# stable/10 includes src.conf too late but make.conf is in sys.mk
.if ${.CURDIR:M/poudriere/jails/10*/usr/src/usr.bin/xlint*}
# Disable build of llib now that head no longer has lint(1)
LINT=	true
.endif
#### /usr/ports/Mk/Scripts/ports_env.sh ####
_CCVERSION_921dbbb2=FreeBSD clang version 6.0.0 (tags/RELEASE_600/final 326565) (based on LLVM 6.0.0) Target: x86_64-unknown-freebsd11.2 Thread model: posix InstalledDir: /usr/bin
_ALTCCVERSION_921dbbb2=none
_CXXINTERNAL_acaad9ca=FreeBSD clang version 6.0.0 (tags/RELEASE_600/final 326565) (based on LLVM 6.0.0) Target: x86_64-unknown-freebsd11.2 Thread model: posix InstalledDir: /usr/bin "/usr/bin/ld" "--eh-frame-hdr" "-dynamic-linker" "/libexec/ld-elf.so.1" "--hash-style=both" "--enable-new-dtags" "-o" "a.out" "/usr/lib/crt1.o" "/usr/lib/crti.o" "/usr/lib/crtbegin.o" "-L/usr/lib" "/dev/null" "-lc++" "-lm" "-lgcc" "--as-needed" "-lgcc_s" "--no-as-needed" "-lc" "-lgcc" "--as-needed" "-lgcc_s" "--no-as-needed" "/usr/lib/crtend.o" "/usr/lib/crtn.o"
CC_OUTPUT_921dbbb2_58173849=yes
CC_OUTPUT_921dbbb2_9bdba57c=yes
CC_OUTPUT_921dbbb2_6a4fe7f5=yes
CC_OUTPUT_921dbbb2_6bcac02b=yes
CC_OUTPUT_921dbbb2_67d20829=yes
CC_OUTPUT_921dbbb2_bfa62e83=yes
CC_OUTPUT_921dbbb2_f0b4d593=yes
CC_OUTPUT_921dbbb2_308abb44=yes
CC_OUTPUT_921dbbb2_f00456e5=yes
CC_OUTPUT_921dbbb2_65ad290d=yes
CC_OUTPUT_921dbbb2_f2776b26=yes
CC_OUTPUT_921dbbb2_b2657cc3=yes
CC_OUTPUT_921dbbb2_380987f7=yes
CC_OUTPUT_921dbbb2_160933ec=yes
CC_OUTPUT_921dbbb2_fb62803b=yes
_OBJC_CCVERSION_921dbbb2=FreeBSD clang version 6.0.0 (tags/RELEASE_600/final 326565) (based on LLVM 6.0.0) Target: x86_64-unknown-freebsd11.2 Thread model: posix InstalledDir: /usr/bin
_OBJC_ALTCCVERSION_921dbbb2=none
ARCH=amd64
OPSYS=FreeBSD
_OSRELEASE=11.2-RELEASE-p9
OSREL=11.2
OSVERSION=1102000
PYTHONBASE=/usr/local
HAVE_COMPAT_IA32_KERN=YES
CONFIGURE_MAX_CMD_LEN=262144
HAVE_PORTS_ENV=1
#### Misc Poudriere ####
DISABLE_MAKE_JOBS=poudriere
---End make.conf---
--Resource limits--
cpu time               (seconds, -t)  unlimited
file size           (512-blocks, -f)  unlimited
data seg size           (kbytes, -d)  33554432
stack size              (kbytes, -s)  524288
core file size      (512-blocks, -c)  unlimited
max memory size         (kbytes, -m)  unlimited
locked memory           (kbytes, -l)  unlimited
max user processes              (-u)  89999
open files                      (-n)  1024
virtual mem size        (kbytes, -v)  unlimited
swap limit              (kbytes, -w)  unlimited
socket buffer size       (bytes, -b)  unlimited
pseudo-terminals                (-p)  unlimited
kqueues                         (-k)  unlimited
umtx shared locks               (-o)  unlimited
--End resource limits--
=======================<phase: check-sanity   >============================
===>  License BSD3CLAUSE accepted by the user
===========================================================================
=======================<phase: pkg-depends    >============================
===>   py27-Scrapy-1.6.0 depends on file: /usr/local/sbin/pkg - not found
===>   Installing existing package /packages/All/pkg-1.10.5_5.txz
[112amd64-default-build-as-user-job-24] Installing pkg-1.10.5_5...
[112amd64-default-build-as-user-job-24] Extracting pkg-1.10.5_5: .......... done
===>   py27-Scrapy-1.6.0 depends on file: /usr/local/sbin/pkg - found
===>   Returning to build of py27-Scrapy-1.6.0
===========================================================================
=======================<phase: fetch-depends  >============================
===========================================================================
=======================<phase: fetch          >============================
===>  License BSD3CLAUSE accepted by the user
===> Fetching all distfiles required by py27-Scrapy-1.6.0 for building
===========================================================================
=======================<phase: checksum       >============================
===>  License BSD3CLAUSE accepted by the user
===> Fetching all distfiles required by py27-Scrapy-1.6.0 for building
=> SHA256 Checksum OK for Scrapy-1.6.0.tar.gz.
===========================================================================
=======================<phase: extract-depends>============================
===========================================================================
=======================<phase: extract        >============================
===>  License BSD3CLAUSE accepted by the user
===> Fetching all distfiles required by py27-Scrapy-1.6.0 for building
===>  Extracting for py27-Scrapy-1.6.0
=> SHA256 Checksum OK for Scrapy-1.6.0.tar.gz.
===========================================================================
=======================<phase: patch-depends  >============================
===========================================================================
=======================<phase: patch          >============================
===>  Patching for py27-Scrapy-1.6.0
===========================================================================
=======================<phase: build-depends  >============================
===>   py27-Scrapy-1.6.0 depends on package: py27-setuptools>0 - not found
===>   Installing existing package /packages/All/py27-setuptools-40.8.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-setuptools-40.8.0...
[112amd64-default-build-as-user-job-24] `-- Installing python27-2.7.16...
[112amd64-default-build-as-user-job-24] |   `-- Installing gettext-runtime-0.19.8.1_2...
[112amd64-default-build-as-user-job-24] |   | `-- Installing indexinfo-0.3.1...
[112amd64-default-build-as-user-job-24] |   | `-- Extracting indexinfo-0.3.1: .... done
[112amd64-default-build-as-user-job-24] |   `-- Extracting gettext-runtime-0.19.8.1_2: .......... done
[112amd64-default-build-as-user-job-24] |   `-- Installing libffi-3.2.1_3...
[112amd64-default-build-as-user-job-24] |   `-- Extracting libffi-3.2.1_3: .......... done
[112amd64-default-build-as-user-job-24] |   `-- Installing readline-7.0.5...
[112amd64-default-build-as-user-job-24] |   `-- Extracting readline-7.0.5: .......... done
[112amd64-default-build-as-user-job-24] `-- Extracting python27-2.7.16: .......... done
[112amd64-default-build-as-user-job-24] Extracting py27-setuptools-40.8.0: .......... done
Message from python27-2.7.16:
<snip>
copying scrapy/utils/signal.py -> build/lib/scrapy/utils
copying scrapy/utils/benchserver.py -> build/lib/scrapy/utils
copying scrapy/utils/trackref.py -> build/lib/scrapy/utils
copying scrapy/utils/test.py -> build/lib/scrapy/utils
copying scrapy/utils/misc.py -> build/lib/scrapy/utils
copying scrapy/utils/spider.py -> build/lib/scrapy/utils
copying scrapy/utils/request.py -> build/lib/scrapy/utils
copying scrapy/utils/url.py -> build/lib/scrapy/utils
copying scrapy/utils/boto.py -> build/lib/scrapy/utils
copying scrapy/utils/datatypes.py -> build/lib/scrapy/utils
copying scrapy/utils/ossignal.py -> build/lib/scrapy/utils
copying scrapy/utils/template.py -> build/lib/scrapy/utils
copying scrapy/utils/job.py -> build/lib/scrapy/utils
copying scrapy/utils/deprecate.py -> build/lib/scrapy/utils
copying scrapy/utils/conf.py -> build/lib/scrapy/utils
creating build/lib/scrapy/contracts
copying scrapy/contracts/default.py -> build/lib/scrapy/contracts
copying scrapy/contracts/__init__.py -> build/lib/scrapy/contracts
creating build/lib/scrapy/xlib
copying scrapy/xlib/__init__.py -> build/lib/scrapy/xlib
copying scrapy/xlib/tx.py -> build/lib/scrapy/xlib
copying scrapy/xlib/pydispatch.py -> build/lib/scrapy/xlib
creating build/lib/scrapy/linkextractors
copying scrapy/linkextractors/htmlparser.py -> build/lib/scrapy/linkextractors
copying scrapy/linkextractors/__init__.py -> build/lib/scrapy/linkextractors
copying scrapy/linkextractors/sgml.py -> build/lib/scrapy/linkextractors
copying scrapy/linkextractors/lxmlhtml.py -> build/lib/scrapy/linkextractors
copying scrapy/linkextractors/regex.py -> build/lib/scrapy/linkextractors
creating build/lib/scrapy/pipelines
copying scrapy/pipelines/files.py -> build/lib/scrapy/pipelines
copying scrapy/pipelines/__init__.py -> build/lib/scrapy/pipelines
copying scrapy/pipelines/media.py -> build/lib/scrapy/pipelines
copying scrapy/pipelines/images.py -> build/lib/scrapy/pipelines
creating build/lib/scrapy/selector
copying scrapy/selector/__init__.py -> build/lib/scrapy/selector
copying scrapy/selector/unified.py -> build/lib/scrapy/selector
copying scrapy/selector/lxmlsel.py -> build/lib/scrapy/selector
copying scrapy/selector/csstranslator.py -> build/lib/scrapy/selector
creating build/lib/scrapy/core/downloader
copying scrapy/core/downloader/__init__.py -> build/lib/scrapy/core/downloader
copying scrapy/core/downloader/webclient.py -> build/lib/scrapy/core/downloader
copying scrapy/core/downloader/contextfactory.py -> build/lib/scrapy/core/downloader
copying scrapy/core/downloader/middleware.py -> build/lib/scrapy/core/downloader
copying scrapy/core/downloader/tls.py -> build/lib/scrapy/core/downloader
creating build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/__init__.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/http11.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/ftp.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/http.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/datauri.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/s3.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/http10.py -> build/lib/scrapy/core/downloader/handlers
copying scrapy/core/downloader/handlers/file.py -> build/lib/scrapy/core/downloader/handlers
creating build/lib/scrapy/http/request
copying scrapy/http/request/rpc.py -> build/lib/scrapy/http/request
copying scrapy/http/request/__init__.py -> build/lib/scrapy/http/request
copying scrapy/http/request/form.py -> build/lib/scrapy/http/request
creating build/lib/scrapy/http/response
copying scrapy/http/response/xml.py -> build/lib/scrapy/http/response
copying scrapy/http/response/__init__.py -> build/lib/scrapy/http/response
copying scrapy/http/response/html.py -> build/lib/scrapy/http/response
copying scrapy/http/response/text.py -> build/lib/scrapy/http/response
running egg_info
writing requirements to Scrapy.egg-info/requires.txt
writing Scrapy.egg-info/PKG-INFO
writing top-level names to Scrapy.egg-info/top_level.txt
writing dependency_links to Scrapy.egg-info/dependency_links.txt
writing entry points to Scrapy.egg-info/entry_points.txt
reading manifest file 'Scrapy.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
warning: no files found matching 'license.txt' under directory 'scrapy'
no previously-included directories found matching 'docs/build'
warning: no files found matching '*' under directory 'bin'
warning: no previously-included files matching '__pycache__' found anywhere in distribution
warning: no previously-included files matching '*.py[cod]' found anywhere in distribution
writing manifest file 'Scrapy.egg-info/SOURCES.txt'
copying scrapy/VERSION -> build/lib/scrapy
copying scrapy/mime.types -> build/lib/scrapy
creating build/lib/scrapy/templates
creating build/lib/scrapy/templates/project
copying scrapy/templates/project/scrapy.cfg -> build/lib/scrapy/templates/project
creating build/lib/scrapy/templates/project/module
copying scrapy/templates/project/module/__init__.py -> build/lib/scrapy/templates/project/module
copying scrapy/templates/project/module/items.py.tmpl -> build/lib/scrapy/templates/project/module
copying scrapy/templates/project/module/middlewares.py.tmpl -> build/lib/scrapy/templates/project/module
copying scrapy/templates/project/module/pipelines.py.tmpl -> build/lib/scrapy/templates/project/module
copying scrapy/templates/project/module/settings.py.tmpl -> build/lib/scrapy/templates/project/module
creating build/lib/scrapy/templates/project/module/spiders
copying scrapy/templates/project/module/spiders/__init__.py -> build/lib/scrapy/templates/project/module/spiders
creating build/lib/scrapy/templates/spiders
copying scrapy/templates/spiders/basic.tmpl -> build/lib/scrapy/templates/spiders
copying scrapy/templates/spiders/crawl.tmpl -> build/lib/scrapy/templates/spiders
copying scrapy/templates/spiders/csvfeed.tmpl -> build/lib/scrapy/templates/spiders
copying scrapy/templates/spiders/xmlfeed.tmpl -> build/lib/scrapy/templates/spiders
===========================================================================
=======================<phase: run-depends    >============================
===>   py27-Scrapy-1.6.0 depends on package: py27-twisted>=13.1.0 - not found
===>   Installing existing package /packages/All/py27-twisted-17.9.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-twisted-17.9.0...
[112amd64-default-build-as-user-job-24] `-- Installing py27-Automat-0.7.0...
[112amd64-default-build-as-user-job-24] |   `-- Installing py27-attrs-18.2.0...
[112amd64-default-build-as-user-job-24] |   `-- Extracting py27-attrs-18.2.0: .......... done
[112amd64-default-build-as-user-job-24] |   `-- Installing py27-six-1.12.0...
[112amd64-default-build-as-user-job-24] |   `-- Extracting py27-six-1.12.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Extracting py27-Automat-0.7.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-constantly-15.1.0...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-constantly-15.1.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-hyperlink-18.0.0...
[112amd64-default-build-as-user-job-24] |   `-- Installing py27-idna-2.8...
[112amd64-default-build-as-user-job-24] |   `-- Extracting py27-idna-2.8: .......... done
[112amd64-default-build-as-user-job-24] `-- Extracting py27-hyperlink-18.0.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-incremental-17.5.0...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-incremental-17.5.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-zope.interface-4.6.0...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-zope.interface-4.6.0: .......... done
[112amd64-default-build-as-user-job-24] Extracting py27-twisted-17.9.0: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-twisted>=13.1.0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-lxml>0 - not found
===>   Installing existing package /packages/All/py27-lxml-4.3.1.txz
[112amd64-default-build-as-user-job-24] Installing py27-lxml-4.3.1...
[112amd64-default-build-as-user-job-24] `-- Installing libxml2-2.9.8...
[112amd64-default-build-as-user-job-24] `-- Extracting libxml2-2.9.8: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing libxslt-1.1.32...
[112amd64-default-build-as-user-job-24] |   `-- Installing libgcrypt-1.8.4_1...
[112amd64-default-build-as-user-job-24] |   | `-- Installing libgpg-error-1.35...
[112amd64-default-build-as-user-job-24] |   | `-- Extracting libgpg-error-1.35: .......... done
[112amd64-default-build-as-user-job-24] |   `-- Extracting libgcrypt-1.8.4_1: .......... done
[112amd64-default-build-as-user-job-24] `-- Extracting libxslt-1.1.32: .......... done
[112amd64-default-build-as-user-job-24] Extracting py27-lxml-4.3.1: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-lxml>0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-sqlite3>0 - not found
===>   Installing existing package /packages/All/py27-sqlite3-2.7.16_7.txz
[112amd64-default-build-as-user-job-24] Installing py27-sqlite3-2.7.16_7...
[112amd64-default-build-as-user-job-24] `-- Installing sqlite3-3.27.1...
[112amd64-default-build-as-user-job-24] `-- Extracting sqlite3-3.27.1: .......... done
[112amd64-default-build-as-user-job-24] Extracting py27-sqlite3-2.7.16_7: ........ done
===>   py27-Scrapy-1.6.0 depends on package: py27-sqlite3>0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-w3lib>=1.17.0 - not found
===>   Installing existing package /packages/All/py27-w3lib-1.20.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-w3lib-1.20.0...
[112amd64-default-build-as-user-job-24] Extracting py27-w3lib-1.20.0: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-w3lib>=1.17.0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-cssselect>=0.9 - not found
===>   Installing existing package /packages/All/py27-cssselect-0.9.1_2.txz
[112amd64-default-build-as-user-job-24] Installing py27-cssselect-0.9.1_2...
[112amd64-default-build-as-user-job-24] Extracting py27-cssselect-0.9.1_2: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-cssselect>=0.9 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-queuelib>0 - not found
===>   Installing existing package /packages/All/py27-queuelib-1.5.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-queuelib-1.5.0...
[112amd64-default-build-as-user-job-24] Extracting py27-queuelib-1.5.0: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-queuelib>0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-pydispatcher>=2.0.5 - not found
===>   Installing existing package /packages/All/py27-pydispatcher-2.0.5.txz
[112amd64-default-build-as-user-job-24] Installing py27-pydispatcher-2.0.5...
[112amd64-default-build-as-user-job-24] Extracting py27-pydispatcher-2.0.5: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-pydispatcher>=2.0.5 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-service_identity>0 - not found
===>   Installing existing package /packages/All/py27-service_identity-17.0.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-service_identity-17.0.0...
[112amd64-default-build-as-user-job-24] `-- Installing py27-characteristic-14.3.0...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-characteristic-14.3.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-openssl-18.0.0...
[112amd64-default-build-as-user-job-24] |   `-- Installing py27-cryptography-2.3...
[112amd64-default-build-as-user-job-24] |   | `-- Installing py27-asn1crypto-0.22.0...
[112amd64-default-build-as-user-job-24] |   | `-- Extracting py27-asn1crypto-0.22.0: .......... done
[112amd64-default-build-as-user-job-24] |   | `-- Installing py27-cffi-1.12.2...
[112amd64-default-build-as-user-job-24] |   |   `-- Installing py27-pycparser-2.18...
[112amd64-default-build-as-user-job-24] |   |   `-- Extracting py27-pycparser-2.18: .......... done
[112amd64-default-build-as-user-job-24] |   | `-- Extracting py27-cffi-1.12.2: .......... done
[112amd64-default-build-as-user-job-24] |   | `-- Installing py27-enum34-1.1.6...
[112amd64-default-build-as-user-job-24] |   | `-- Extracting py27-enum34-1.1.6: .......... done
[112amd64-default-build-as-user-job-24] |   | `-- Installing py27-ipaddress-1.0.22...
[112amd64-default-build-as-user-job-24] |   | `-- Extracting py27-ipaddress-1.0.22: .......... done
[112amd64-default-build-as-user-job-24] |   `-- Extracting py27-cryptography-2.3: .......... done
[112amd64-default-build-as-user-job-24] `-- Extracting py27-openssl-18.0.0: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-pyasn1-0.4.5...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-pyasn1-0.4.5: .......... done
[112amd64-default-build-as-user-job-24] `-- Installing py27-pyasn1-modules-0.2.3...
[112amd64-default-build-as-user-job-24] `-- Extracting py27-pyasn1-modules-0.2.3: .......... done
[112amd64-default-build-as-user-job-24] Extracting py27-service_identity-17.0.0: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-service_identity>0 - found
===>   Returning to build of py27-Scrapy-1.6.0
===>   py27-Scrapy-1.6.0 depends on package: py27-six>=1.5.2 - found
===>   py27-Scrapy-1.6.0 depends on package: py27-parsel>=1.5 - not found
===>   Installing existing package /packages/All/py27-parsel-1.2.0.txz
[112amd64-default-build-as-user-job-24] Installing py27-parsel-1.2.0...
[112amd64-default-build-as-user-job-24] Extracting py27-parsel-1.2.0: .......... done
===>   py27-Scrapy-1.6.0 depends on package: py27-parsel>=1.5 - not found
*** Error code 1

Stop.
make: stopped in /usr/ports/www/py-scrapy



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?201903170616.x2H6GO0A090446>