=>> Building misc/py-llama-cpp-python build started at Sat Jan 24 10:20:48 EST 2026 port directory: /usr/ports/misc/py-llama-cpp-python package name: py311-llama-cpp-python-0.3.16 building for: FreeBSD pkg.DaveG.ca 13.5-RELEASE-p7 FreeBSD 13.5-RELEASE-p7 1305000 amd64 maintained by: yuri@FreeBSD.org Makefile datestamp: -rw-r--r-- 1 root wheel 1740 Oct 28 16:11 /usr/ports/misc/py-llama-cpp-python/Makefile Ports top last git commit: 076a295eb33c Ports top unclean checkout: no Port dir last git commit: b6660c1d142a Port dir unclean checkout: no Poudriere version: poudriere-git-3.4.2 Host OSVERSION: 1500068 Jail OSVERSION: 1305000 Job Id: 11 ---Begin Environment--- SHELL=/bin/csh OSVERSION=1305000 UNAME_v=FreeBSD 13.5-RELEASE-p7 1305000 UNAME_r=13.5-RELEASE-p7 BLOCKSIZE=K MAIL=/var/mail/root MM_CHARSET=UTF-8 LANG=C.UTF-8 STATUS=1 HOME=/root PATH=/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin MAKE_OBJDIR_CHECK_WRITABLE=0 LOCALBASE=/usr/local USER=root POUDRIERE_NAME=poudriere-git LIBEXECPREFIX=/usr/local/libexec/poudriere POUDRIERE_VERSION=3.4.2 MASTERMNT=/usr/local/poudriere/data/.m/freebsd_13_5_x64-HEAD-ALL/ref LC_COLLATE=C POUDRIERE_BUILD_TYPE=bulk PACKAGE_BUILDING=yes SAVED_TERM=tmux-256color OUTPUT_REDIRECTED_STDERR=4 OUTPUT_REDIRECTED=1 PWD=/usr/local/poudriere/data/.m/freebsd_13_5_x64-HEAD-ALL/11/.p OUTPUT_REDIRECTED_STDOUT=3 P_PORTS_FEATURES=FLAVORS SUBPACKAGES SELECTED_OPTIONS MASTERNAME=freebsd_13_5_x64-HEAD-ALL SCRIPTPREFIX=/usr/local/share/poudriere SCRIPTNAME=bulk.sh OLDPWD=/usr/local/poudriere/data/.m/freebsd_13_5_x64-HEAD-ALL/ref/.p/pool POUDRIERE_PKGNAME=poudriere-git-3.4.2 SCRIPTPATH=/usr/local/share/poudriere/bulk.sh POUDRIEREPATH=/usr/local/bin/poudriere ---End Environment--- ---Begin Poudriere Port Flags/Env--- PORT_FLAGS= PKGENV= FLAVOR=py311 MAKE_ARGS= FLAVOR=py311 ---End Poudriere Port Flags/Env--- ---Begin OPTIONS List--- ---End OPTIONS List--- --MAINTAINER-- yuri@FreeBSD.org --End MAINTAINER-- --CONFIGURE_ARGS-- --End CONFIGURE_ARGS-- --CONFIGURE_ENV-- PYTHON="/usr/local/bin/python3.11" XDG_DATA_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.cache HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig SHELL=/bin/sh CONFIG_SHELL=/bin/sh --End CONFIGURE_ENV-- --MAKE_ENV-- XDG_DATA_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CONFIG_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 XDG_CACHE_HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.cache HOME=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311 TMPDIR="/tmp" PATH=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/local/sbin:/usr/local/bin:/root/bin PKG_CONFIG_LIBDIR=/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/.pkgconfig:/usr/local/libdata/pkgconfig:/usr/local/share/pkgconfig:/usr/libdata/pkgconfig MK_DEBUG_FILES=no MK_KERNEL_SYMBOLS=no SHELL=/bin/sh NO_LINT=YES PREFIX=/usr/local LOCALBASE=/usr/local CC="cc" CFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " CPP="cpp" CPPFLAGS="" LDFLAGS=" " LIBS="" CXX="c++" CXXFLAGS="-O2 -pipe -fstack-protector-strong -fno-strict-aliasing " BSD_INSTALL_PROGRAM="install -s -m 555" BSD_INSTALL_LIB="install -s -m 0644" BSD_INSTALL_SCRIPT="install -m 555" BSD_INSTALL_DATA="install -m 0644" BSD_INSTALL_MAN="install -m 444" --End MAKE_ENV-- --PLIST_SUB-- PYTHON_INCLUDEDIR=include/python3.11 PYTHON_LIBDIR=lib/python3.11 PYTHON_PLATFORM=freebsd13 PYTHON_SITELIBDIR=lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_BASESUFFIX=311 PYTHON_TAG=.cpython-311 PYTHON_SOABI=.cpython-311 PYTHON_VER=3.11 PYTHON_BASEVER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" OSREL=13.5 PREFIX=%D LOCALBASE=/usr/local RESETPREFIX=/usr/local LIB32DIR=lib DOCSDIR="share/doc/llama-cpp-python" EXAMPLESDIR="share/examples/llama-cpp-python" DATADIR="share/llama-cpp-python" WWWDIR="www/llama-cpp-python" ETCDIR="etc/llama-cpp-python" --End PLIST_SUB-- --SUB_LIST-- PYTHON_INCLUDEDIR=/usr/local/include/python3.11 PYTHON_LIBDIR=/usr/local/lib/python3.11 PYTHON_PLATFORM=freebsd13 PYTHON_SITELIBDIR=/usr/local/lib/python3.11/site-packages PYTHON_SUFFIX=311 PYTHON_BASESUFFIX=311 PYTHON_TAG=.cpython-311 PYTHON_SOABI=.cpython-311 PYTHON_VER=3.11 PYTHON_BASEVER=3.11 PYTHON_VERSION=python3.11 PYTHON2="@comment " PYTHON3="" PREFIX=/usr/local LOCALBASE=/usr/local DATADIR=/usr/local/share/llama-cpp-python DOCSDIR=/usr/local/share/doc/llama-cpp-python EXAMPLESDIR=/usr/local/share/examples/llama-cpp-python WWWDIR=/usr/local/www/llama-cpp-python ETCDIR=/usr/local/etc/llama-cpp-python --End SUB_LIST-- ---Begin make.conf--- DEFAULT_VERSIONS+=pgsql=15 DEFAULT_VERSIONS+=php=82 DISABLE_LICENSES=yes SELECTED_OPTIONS+=OPENBLAS ARCHDEF=AMD64K10h64SSE3 USE_PACKAGE_DEPENDS=yes BATCH=yes WRKDIRPREFIX=/wrkdirs PORTSDIR=/usr/ports PACKAGES=/packages DISTDIR=/distfiles FORCE_PACKAGE=yes PACKAGE_BUILDING=yes PACKAGE_BUILDING_FLAVORS=yes #### #### DEFAULT_VERSIONS+=pgsql=15 DEFAULT_VERSIONS+=php=82 DISABLE_LICENSES=yes SELECTED_OPTIONS+=OPENBLAS ARCHDEF=AMD64K10h64SSE3 WITH_CCACHE_BUILD=yes CCACHE_DIR=/root/.ccache #### Misc Poudriere #### .include "/etc/make.conf.ports_env" GID=0 UID=0 ---End make.conf--- --Resource limits-- cpu time (seconds, -t) unlimited file size (512-blocks, -f) unlimited data seg size (kbytes, -d) 33554432 stack size (kbytes, -s) 524288 core file size (512-blocks, -c) unlimited max memory size (kbytes, -m) unlimited locked memory (kbytes, -l) unlimited max user processes (-u) 89999 open files (-n) 8192 virtual mem size (kbytes, -v) unlimited swap limit (kbytes, -w) unlimited socket buffer size (bytes, -b) unlimited pseudo-terminals (-p) unlimited kqueues (-k) unlimited umtx shared locks (-o) unlimited --End resource limits-- =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/sbin/pkg - not found ===> Installing existing package /packages/All/pkg-2.5.1.pkg [pkg.DaveG.ca] Installing pkg-2.5.1... [pkg.DaveG.ca] Extracting pkg-2.5.1: .......... done ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/sbin/pkg - found ===> Returning to build of py311-llama-cpp-python-0.3.16 =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.16 for building =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.16 for building => SHA256 Checksum OK for abetlen-llama-cpp-python-v0.3.16_GH0.tar.gz. => SHA256 Checksum OK for ggerganov-llama.cpp-b6598_GH0.tar.gz. =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Fetching all distfiles required by py311-llama-cpp-python-0.3.16 for building ===> Extracting for py311-llama-cpp-python-0.3.16 => SHA256 Checksum OK for abetlen-llama-cpp-python-v0.3.16_GH0.tar.gz. => SHA256 Checksum OK for ggerganov-llama.cpp-b6598_GH0.tar.gz. =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Patching for py311-llama-cpp-python-0.3.16 ===> Applying FreeBSD patches for py311-llama-cpp-python-0.3.16 from /usr/ports/misc/py-llama-cpp-python/files =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-scikit-build-core>=0.9.2 - not found ===> Installing existing package /packages/All/py311-scikit-build-core-0.11.6.pkg [pkg.DaveG.ca] Installing py311-scikit-build-core-0.11.6... [pkg.DaveG.ca] `-- Installing py311-packaging-25.0... [pkg.DaveG.ca] | `-- Installing python311-3.11.14... [pkg.DaveG.ca] | | `-- Installing gettext-runtime-0.26... [pkg.DaveG.ca] | | `-- Installing indexinfo-0.3.1_1... [pkg.DaveG.ca] | | `-- Extracting indexinfo-0.3.1_1: . done [pkg.DaveG.ca] | | `-- Extracting gettext-runtime-0.26: .......... done [pkg.DaveG.ca] | | `-- Installing libffi-3.5.1... [pkg.DaveG.ca] | | `-- Extracting libffi-3.5.1: .......... done [pkg.DaveG.ca] | | `-- Installing mpdecimal-4.0.1... [pkg.DaveG.ca] | | `-- Extracting mpdecimal-4.0.1: .......... done [pkg.DaveG.ca] | | `-- Installing readline-8.3.3... [pkg.DaveG.ca] | | `-- Extracting readline-8.3.3: .......... done [pkg.DaveG.ca] | `-- Extracting python311-3.11.14: .......... done [pkg.DaveG.ca] `-- Extracting py311-packaging-25.0: .......... done [pkg.DaveG.ca] `-- Installing py311-pathspec-1.0.3... [pkg.DaveG.ca] `-- Extracting py311-pathspec-1.0.3: .......... done [pkg.DaveG.ca] Extracting py311-scikit-build-core-0.11.6: .......... done ===== Message from python311-3.11.14: -- Note that some standard Python modules are provided as separate ports as they require additional dependencies. They are available as: py311-gdbm databases/py-gdbm@py311 py311-sqlite3 databases/py-sqlite3@py311 py311-tkinter x11-toolkits/py-tkinter@py311 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-scikit-build-core>=0.9.2 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on executable: cmake - not found ===> Installing existing package /packages/All/cmake-core-3.31.10.pkg [pkg.DaveG.ca] Installing cmake-core-3.31.10... [pkg.DaveG.ca] `-- Installing expat-2.7.3... [pkg.DaveG.ca] `-- Extracting expat-2.7.3: .......... done [pkg.DaveG.ca] `-- Installing jsoncpp-1.9.6_1... [pkg.DaveG.ca] `-- Extracting jsoncpp-1.9.6_1: .......... done [pkg.DaveG.ca] `-- Installing libidn2-2.3.8... [pkg.DaveG.ca] | `-- Installing libunistring-1.4.1... [pkg.DaveG.ca] | `-- Extracting libunistring-1.4.1: .......... done [pkg.DaveG.ca] `-- Extracting libidn2-2.3.8: .......... done [pkg.DaveG.ca] `-- Installing libuv-1.51.0... [pkg.DaveG.ca] `-- Extracting libuv-1.51.0: .......... done [pkg.DaveG.ca] `-- Installing rhash-1.4.6... [pkg.DaveG.ca] `-- Extracting rhash-1.4.6: .......... done [pkg.DaveG.ca] Extracting cmake-core-3.31.10: .......... done ===> py311-llama-cpp-python-0.3.16 depends on executable: cmake - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/bin/python3.11 - found ===> py311-llama-cpp-python-0.3.16 depends on package: py311-build>=0 - not found ===> Installing existing package /packages/All/py311-build-1.3.0.pkg [pkg.DaveG.ca] Installing py311-build-1.3.0... [pkg.DaveG.ca] `-- Installing py311-pyproject-hooks-1.2.0... [pkg.DaveG.ca] `-- Extracting py311-pyproject-hooks-1.2.0: .......... done [pkg.DaveG.ca] Extracting py311-build-1.3.0: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-build>=0 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-installer>=0 - not found ===> Installing existing package /packages/All/py311-installer-0.7.0.pkg [pkg.DaveG.ca] Installing py311-installer-0.7.0... [pkg.DaveG.ca] Extracting py311-installer-0.7.0: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-installer>=0 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/bin/ccache - not found ===> Installing existing package /packages/All/ccache-3.7.12_8.pkg [pkg.DaveG.ca] Installing ccache-3.7.12_8... [pkg.DaveG.ca] Extracting ccache-3.7.12_8: ......... done Create compiler links... create symlink for cc create symlink for cc (world) create symlink for c++ create symlink for c++ (world) create symlink for CC create symlink for CC (world) create symlink for clang create symlink for clang (world) create symlink for clang++ create symlink for clang++ (world) ===== Message from ccache-3.7.12_8: -- NOTE: Please read /usr/local/share/doc/ccache/ccache-howto-freebsd.txt for information on using ccache with FreeBSD ports and src. ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/bin/ccache - found ===> Returning to build of py311-llama-cpp-python-0.3.16 =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.16 depends on shared library: libvulkan.so - not found ===> Installing existing package /packages/All/vulkan-loader-1.4.336.pkg [pkg.DaveG.ca] Installing vulkan-loader-1.4.336... [pkg.DaveG.ca] `-- Installing libX11-1.8.12,1... [pkg.DaveG.ca] | `-- Installing libxcb-1.17.0... [pkg.DaveG.ca] | | `-- Installing libXau-1.0.12... [pkg.DaveG.ca] | | `-- Extracting libXau-1.0.12: .......... done [pkg.DaveG.ca] | | `-- Installing libXdmcp-1.1.5... [pkg.DaveG.ca] | | `-- Installing xorgproto-2024.1... [pkg.DaveG.ca] | | `-- Extracting xorgproto-2024.1: .......... done [pkg.DaveG.ca] | | `-- Extracting libXdmcp-1.1.5: ...... done [pkg.DaveG.ca] | `-- Extracting libxcb-1.17.0: .......... done [pkg.DaveG.ca] `-- Extracting libX11-1.8.12,1: .......... done [pkg.DaveG.ca] `-- Installing libXrandr-1.5.4... [pkg.DaveG.ca] | `-- Installing libXext-1.3.6,1... [pkg.DaveG.ca] | `-- Extracting libXext-1.3.6,1: .......... done [pkg.DaveG.ca] | `-- Installing libXrender-0.9.12... [pkg.DaveG.ca] | `-- Extracting libXrender-0.9.12: ....... done [pkg.DaveG.ca] `-- Extracting libXrandr-1.5.4: .......... done [pkg.DaveG.ca] `-- Installing wayland-1.24.0_2... [pkg.DaveG.ca] | `-- Installing libepoll-shim-0.0.20240608... [pkg.DaveG.ca] | `-- Extracting libepoll-shim-0.0.20240608: .......... done [pkg.DaveG.ca] | `-- Installing libxml2-2.14.6... [pkg.DaveG.ca] | `-- Extracting libxml2-2.14.6: .......... done [pkg.DaveG.ca] `-- Extracting wayland-1.24.0_2: .......... done [pkg.DaveG.ca] Extracting vulkan-loader-1.4.336: ....... done ===== Message from wayland-1.24.0_2: -- Wayland requires XDG_RUNTIME_DIR to be defined to a path that will contain "wayland-%d" unix(4) sockets. This is usually handled by consolekit2 (via ck-launch-session) or pam_xdg (via login). ===> py311-llama-cpp-python-0.3.16 depends on shared library: libvulkan.so - found (/usr/local/lib/libvulkan.so) ===> Returning to build of py311-llama-cpp-python-0.3.16 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Configuring for py311-llama-cpp-python-0.3.16 =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Building for py311-llama-cpp-python-0.3.16 * Getting build dependencies for wheel... * Building wheel... *** scikit-build-core 0.11.6 using CMake 3.31.10 (wheel) *** Configuring CMake... loading initial cache file /tmp/tmpui5d1wk8/build/CMakeInit.txt -- The C compiler identification is Clang 19.1.7 -- The CXX compiler identification is Clang 19.1.7 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/local/libexec/ccache/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/local/libexec/ccache/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done CMAKE_BUILD_TYPE=Release -- Could NOT find Git (missing: GIT_EXECUTABLE) CMake Warning at vendor/llama.cpp/cmake/build-info.cmake:14 (message): Git not found. Build info will not be accurate. Call Stack (most recent call first): vendor/llama.cpp/CMakeLists.txt:98 (include) -- The ASM compiler identification is Clang with GNU-like command-line -- Found assembler: /usr/local/libexec/ccache/cc -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed -- Check if compiler accepts -pthread -- Check if compiler accepts -pthread - yes -- Found Threads: TRUE -- ccache found, compilation results will be cached. Disable with GGML_CCACHE=OFF. -- CMAKE_SYSTEM_PROCESSOR: amd64 -- GGML_SYSTEM_ARCH: x86 -- Including CPU backend -- Found OpenMP_C: -fopenmp=libomp (found version "5.1") -- Found OpenMP_CXX: -fopenmp=libomp (found version "5.1") -- Found OpenMP: TRUE (found version "5.1") -- x86 detected -- Adding CPU backend variant ggml-cpu: -march=native -- ggml version: 0.9.0-dev -- ggml commit: unknown -- Looking for pthread_create in pthreads -- Looking for pthread_create in pthreads - not found -- Looking for pthread_create in pthread -- Looking for pthread_create in pthread - found CMake Warning at vendor/llama.cpp/common/CMakeLists.txt:32 (message): Git repository not found; to enable automatic generation of build info, make sure Git is installed and the project is a Git repository. -- Found OpenSSL: /usr/lib/libcrypto.so (found version "1.1.1w") -- Performing Test OPENSSL_VERSION_SUPPORTED -- Performing Test OPENSSL_VERSION_SUPPORTED - Failed CMake Warning (dev) at CMakeLists.txt:13 (install): Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:108 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:21 (install): Target llama has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:108 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:13 (install): Target ggml has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:109 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:21 (install): Target ggml has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:109 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:13 (install): Target mtmd has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:162 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. CMake Warning (dev) at CMakeLists.txt:21 (install): Target mtmd has PUBLIC_HEADER files but no PUBLIC_HEADER DESTINATION. Call Stack (most recent call first): CMakeLists.txt:162 (llama_cpp_python_install_target) This warning is for project developers. Use -Wno-dev to suppress it. -- Configuring done (19.6s) -- Generating done (0.1s) -- Build files have been written to: /tmp/tmpui5d1wk8/build *** Building project with Unix Makefiles... Change Dir: '/tmp/tmpui5d1wk8/build' Run Build Command(s): /usr/local/bin/cmake -E env VERBOSE=1 /usr/bin/make -f Makefile /usr/local/bin/cmake -S/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 -B/tmp/tmpui5d1wk8/build --check-build-system CMakeFiles/Makefile.cmake 0 /usr/local/bin/cmake -E cmake_progress_start /tmp/tmpui5d1wk8/build/CMakeFiles /tmp/tmpui5d1wk8/build//CMakeFiles/progress.marks /usr/bin/make -f CMakeFiles/Makefile2 all /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/build [ 1%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.c.o -MF CMakeFiles/ggml-base.dir/ggml.c.o.d -o CMakeFiles/ggml-base.dir/ggml.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml.c [ 2%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml.cpp.o -MF CMakeFiles/ggml-base.dir/ggml.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml.cpp [ 3%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-alloc.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-alloc.c.o -MF CMakeFiles/ggml-base.dir/ggml-alloc.c.o.d -o CMakeFiles/ggml-base.dir/ggml-alloc.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-alloc.c [ 4%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-backend.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-backend.cpp.o -MF CMakeFiles/ggml-base.dir/ggml-backend.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml-backend.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-backend.cpp [ 5%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-opt.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-opt.cpp.o -MF CMakeFiles/ggml-base.dir/ggml-opt.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml-opt.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-opt.cpp [ 7%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-threading.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-threading.cpp.o -MF CMakeFiles/ggml-base.dir/ggml-threading.cpp.o.d -o CMakeFiles/ggml-base.dir/ggml-threading.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-threading.cpp [ 8%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/ggml-quants.c.o -MF CMakeFiles/ggml-base.dir/ggml-quants.c.o.d -o CMakeFiles/ggml-base.dir/ggml-quants.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-quants.c [ 9%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/gguf.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BUILD -DGGML_COMMIT=\"unknown\" -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_VERSION=\"0.9.0-dev\" -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_base_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-base.dir/gguf.cpp.o -MF CMakeFiles/ggml-base.dir/gguf.cpp.o.d -o CMakeFiles/ggml-base.dir/gguf.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/gguf.cpp [ 10%] Linking CXX shared library ../../../../bin/libggml-base.so cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/ggml-base.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/ggml-base.dir/link.d -shared -Wl,-soname,libggml-base.so -o ../../../../bin/libggml-base.so "CMakeFiles/ggml-base.dir/ggml.c.o" "CMakeFiles/ggml-base.dir/ggml.cpp.o" "CMakeFiles/ggml-base.dir/ggml-alloc.c.o" "CMakeFiles/ggml-base.dir/ggml-backend.cpp.o" "CMakeFiles/ggml-base.dir/ggml-opt.cpp.o" "CMakeFiles/ggml-base.dir/ggml-threading.cpp.o" "CMakeFiles/ggml-base.dir/ggml-quants.c.o" "CMakeFiles/ggml-base.dir/gguf.cpp.o" -Wl,-rpath,"\$ORIGIN" -lm -pthread [ 10%] Built target ggml-base /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/build [ 11%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.c [ 13%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.cpp [ 14%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/repack.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/repack.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/repack.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/repack.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/repack.cpp [ 15%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/hbm.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/hbm.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/hbm.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/hbm.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/hbm.cpp [ 16%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/quants.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/quants.c.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/quants.c.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/quants.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/quants.c [ 17%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/traits.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/traits.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/traits.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/traits.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/traits.cpp [ 19%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/amx.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/amx.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/amx.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/amx.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/amx/amx.cpp [ 20%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp [ 21%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/binary-ops.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/binary-ops.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/binary-ops.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/binary-ops.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/binary-ops.cpp [ 22%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/unary-ops.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/unary-ops.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/unary-ops.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/unary-ops.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/unary-ops.cpp [ 23%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/vec.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/vec.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/vec.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/vec.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/vec.cpp [ 25%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ops.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/ops.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/ops.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/ops.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/ops.cpp [ 26%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/llamafile/sgemm.cpp [ 27%] Building C object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/quants.c.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/cc -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu11 -fPIC -Wshadow -Wstrict-prototypes -Wpointer-arith -Wmissing-prototypes -Werror=implicit-int -Werror=implicit-function-declaration -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wdouble-promotion -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/quants.c.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/quants.c.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/quants.c.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/arch/x86/quants.c [ 28%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/repack.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_BUILD -DGGML_BACKEND_SHARED -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU_REPACK -DGGML_USE_LLAMAFILE -DGGML_USE_OPENMP -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_cpu_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/repack.cpp.o -MF CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/repack.cpp.o.d -o CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/repack.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-cpu/arch/x86/repack.cpp [ 29%] Linking CXX shared library ../../../../bin/libggml-cpu.so cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/ggml-cpu.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/ggml-cpu.dir/link.d -shared -Wl,-soname,libggml-cpu.so -o ../../../../bin/libggml-cpu.so "CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.c.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/ggml-cpu.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/repack.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/hbm.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/quants.c.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/traits.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/amx.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/binary-ops.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/unary-ops.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/vec.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/ops.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/llamafile/sgemm.cpp.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/quants.c.o" "CMakeFiles/ggml-cpu.dir/ggml-cpu/arch/x86/repack.cpp.o" -Wl,-rpath,"\$ORIGIN" ../../../../bin/libggml-base.so /usr/lib/libomp.so [ 29%] Built target ggml-cpu /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/build.make vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/build [ 30%] Building CXX object vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/ggml-backend-reg.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_BUILD -DGGML_SCHED_MAX_COPIES=4 -DGGML_SHARED -DGGML_USE_CPU -D_XOPEN_SOURCE=600 -D__BSD_VISIBLE -Dggml_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/ggml/src/CMakeFiles/ggml.dir/ggml-backend-reg.cpp.o -MF CMakeFiles/ggml.dir/ggml-backend-reg.cpp.o.d -o CMakeFiles/ggml.dir/ggml-backend-reg.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/ggml-backend-reg.cpp [ 32%] Linking CXX shared library ../../../../bin/libggml.so cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/ggml/src && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/ggml.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/ggml.dir/link.d -shared -Wl,-soname,libggml.so -o ../../../../bin/libggml.so "CMakeFiles/ggml.dir/ggml-backend-reg.cpp.o" -Wl,-rpath,"\$ORIGIN" ../../../../bin/libggml-cpu.so ../../../../bin/libggml-base.so [ 32%] Built target ggml /usr/bin/make -f vendor/llama.cpp/src/CMakeFiles/llama.dir/build.make vendor/llama.cpp/src/CMakeFiles/llama.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src/CMakeFiles/llama.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/src/CMakeFiles/llama.dir/build.make vendor/llama.cpp/src/CMakeFiles/llama.dir/build [ 33%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama.cpp.o -MF CMakeFiles/llama.dir/llama.cpp.o.d -o CMakeFiles/llama.dir/llama.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama.cpp [ 34%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-adapter.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-adapter.cpp.o -MF CMakeFiles/llama.dir/llama-adapter.cpp.o.d -o CMakeFiles/llama.dir/llama-adapter.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-adapter.cpp [ 35%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-arch.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-arch.cpp.o -MF CMakeFiles/llama.dir/llama-arch.cpp.o.d -o CMakeFiles/llama.dir/llama-arch.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-arch.cpp [ 36%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-batch.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-batch.cpp.o -MF CMakeFiles/llama.dir/llama-batch.cpp.o.d -o CMakeFiles/llama.dir/llama-batch.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-batch.cpp [ 38%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-chat.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-chat.cpp.o -MF CMakeFiles/llama.dir/llama-chat.cpp.o.d -o CMakeFiles/llama.dir/llama-chat.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-chat.cpp [ 39%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-context.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-context.cpp.o -MF CMakeFiles/llama.dir/llama-context.cpp.o.d -o CMakeFiles/llama.dir/llama-context.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-context.cpp [ 40%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-cparams.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-cparams.cpp.o -MF CMakeFiles/llama.dir/llama-cparams.cpp.o.d -o CMakeFiles/llama.dir/llama-cparams.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-cparams.cpp [ 41%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-grammar.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-grammar.cpp.o -MF CMakeFiles/llama.dir/llama-grammar.cpp.o.d -o CMakeFiles/llama.dir/llama-grammar.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-grammar.cpp [ 42%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-graph.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-graph.cpp.o -MF CMakeFiles/llama.dir/llama-graph.cpp.o.d -o CMakeFiles/llama.dir/llama-graph.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-graph.cpp [ 44%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-hparams.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-hparams.cpp.o -MF CMakeFiles/llama.dir/llama-hparams.cpp.o.d -o CMakeFiles/llama.dir/llama-hparams.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-hparams.cpp [ 45%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-impl.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-impl.cpp.o -MF CMakeFiles/llama.dir/llama-impl.cpp.o.d -o CMakeFiles/llama.dir/llama-impl.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-impl.cpp [ 46%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-io.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-io.cpp.o -MF CMakeFiles/llama.dir/llama-io.cpp.o.d -o CMakeFiles/llama.dir/llama-io.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-io.cpp [ 47%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-kv-cache.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-kv-cache.cpp.o -MF CMakeFiles/llama.dir/llama-kv-cache.cpp.o.d -o CMakeFiles/llama.dir/llama-kv-cache.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-kv-cache.cpp [ 48%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-kv-cache-iswa.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-kv-cache-iswa.cpp.o -MF CMakeFiles/llama.dir/llama-kv-cache-iswa.cpp.o.d -o CMakeFiles/llama.dir/llama-kv-cache-iswa.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-kv-cache-iswa.cpp [ 50%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory.cpp.o -MF CMakeFiles/llama.dir/llama-memory.cpp.o.d -o CMakeFiles/llama.dir/llama-memory.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-memory.cpp [ 51%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory-hybrid.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory-hybrid.cpp.o -MF CMakeFiles/llama.dir/llama-memory-hybrid.cpp.o.d -o CMakeFiles/llama.dir/llama-memory-hybrid.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-memory-hybrid.cpp [ 52%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory-recurrent.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-memory-recurrent.cpp.o -MF CMakeFiles/llama.dir/llama-memory-recurrent.cpp.o.d -o CMakeFiles/llama.dir/llama-memory-recurrent.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-memory-recurrent.cpp [ 53%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-mmap.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-mmap.cpp.o -MF CMakeFiles/llama.dir/llama-mmap.cpp.o.d -o CMakeFiles/llama.dir/llama-mmap.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-mmap.cpp /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-mmap.cpp:488:45: warning: comparison of integers of different signs: 'rlim_t' (aka 'long') and 'size_t' (aka 'unsigned long') [-Wsign-compare] 488 | if (suggest && (lock_limit.rlim_max > lock_limit.rlim_cur + size)) { | ~~~~~~~~~~~~~~~~~~~ ^ ~~~~~~~~~~~~~~~~~~~~~~~~~~ 1 warning generated. [ 54%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model-loader.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model-loader.cpp.o -MF CMakeFiles/llama.dir/llama-model-loader.cpp.o.d -o CMakeFiles/llama.dir/llama-model-loader.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-model-loader.cpp [ 55%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model-saver.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model-saver.cpp.o -MF CMakeFiles/llama.dir/llama-model-saver.cpp.o.d -o CMakeFiles/llama.dir/llama-model-saver.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-model-saver.cpp [ 57%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-model.cpp.o -MF CMakeFiles/llama.dir/llama-model.cpp.o.d -o CMakeFiles/llama.dir/llama-model.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-model.cpp [ 58%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-quant.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-quant.cpp.o -MF CMakeFiles/llama.dir/llama-quant.cpp.o.d -o CMakeFiles/llama.dir/llama-quant.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-quant.cpp [ 59%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-sampling.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-sampling.cpp.o -MF CMakeFiles/llama.dir/llama-sampling.cpp.o.d -o CMakeFiles/llama.dir/llama-sampling.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-sampling.cpp [ 60%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-vocab.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/llama-vocab.cpp.o -MF CMakeFiles/llama.dir/llama-vocab.cpp.o.d -o CMakeFiles/llama.dir/llama-vocab.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/llama-vocab.cpp [ 61%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/unicode-data.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/unicode-data.cpp.o -MF CMakeFiles/llama.dir/unicode-data.cpp.o.d -o CMakeFiles/llama.dir/unicode-data.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/unicode-data.cpp [ 63%] Building CXX object vendor/llama.cpp/src/CMakeFiles/llama.dir/unicode.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dllama_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/src/CMakeFiles/llama.dir/unicode.cpp.o -MF CMakeFiles/llama.dir/unicode.cpp.o.d -o CMakeFiles/llama.dir/unicode.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/unicode.cpp [ 64%] Linking CXX shared library ../../../bin/libllama.so cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/src && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama.dir/link.d -shared -Wl,-soname,libllama.so -o ../../../bin/libllama.so CMakeFiles/llama.dir/llama.cpp.o "CMakeFiles/llama.dir/llama-adapter.cpp.o" "CMakeFiles/llama.dir/llama-arch.cpp.o" "CMakeFiles/llama.dir/llama-batch.cpp.o" "CMakeFiles/llama.dir/llama-chat.cpp.o" "CMakeFiles/llama.dir/llama-context.cpp.o" "CMakeFiles/llama.dir/llama-cparams.cpp.o" "CMakeFiles/llama.dir/llama-grammar.cpp.o" "CMakeFiles/llama.dir/llama-graph.cpp.o" "CMakeFiles/llama.dir/llama-hparams.cpp.o" "CMakeFiles/llama.dir/llama-impl.cpp.o" "CMakeFiles/llama.dir/llama-io.cpp.o" "CMakeFiles/llama.dir/llama-kv-cache.cpp.o" "CMakeFiles/llama.dir/llama-kv-cache-iswa.cpp.o" "CMakeFiles/llama.dir/llama-memory.cpp.o" "CMakeFiles/llama.dir/llama-memory-hybrid.cpp.o" "CMakeFiles/llama.dir/llama-memory-recurrent.cpp.o" "CMakeFiles/llama.dir/llama-mmap.cpp.o" "CMakeFiles/llama.dir/llama-model-loader.cpp.o" "CMakeFiles/llama.dir/llama-model-saver.cpp.o" "CMakeFiles/llama.dir/llama-model.cpp.o" "CMakeFiles/llama.dir/llama-quant.cpp.o" "CMakeFiles/llama.dir/llama-sampling.cpp.o" "CMakeFiles/llama.dir/llama-vocab.cpp.o" "CMakeFiles/llama.dir/unicode-data.cpp.o" CMakeFiles/llama.dir/unicode.cpp.o -Wl,-rpath,"\$ORIGIN" ../../../bin/libggml.so ../../../bin/libggml-cpu.so ../../../bin/libggml-base.so [ 64%] Built target llama /usr/bin/make -f vendor/llama.cpp/common/CMakeFiles/build_info.dir/build.make vendor/llama.cpp/common/CMakeFiles/build_info.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common/CMakeFiles/build_info.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/common/CMakeFiles/build_info.dir/build.make vendor/llama.cpp/common/CMakeFiles/build_info.dir/build [ 65%] Building CXX object vendor/llama.cpp/common/CMakeFiles/build_info.dir/build-info.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -MD -MT vendor/llama.cpp/common/CMakeFiles/build_info.dir/build-info.cpp.o -MF CMakeFiles/build_info.dir/build-info.cpp.o.d -o CMakeFiles/build_info.dir/build-info.cpp.o -c /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common/build-info.cpp [ 65%] Built target build_info /usr/bin/make -f vendor/llama.cpp/common/CMakeFiles/common.dir/build.make vendor/llama.cpp/common/CMakeFiles/common.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common/CMakeFiles/common.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/common/CMakeFiles/common.dir/build.make vendor/llama.cpp/common/CMakeFiles/common.dir/build [ 66%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/arg.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/arg.cpp.o -MF CMakeFiles/common.dir/arg.cpp.o.d -o CMakeFiles/common.dir/arg.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/arg.cpp [ 67%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/chat-parser.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/chat-parser.cpp.o -MF CMakeFiles/common.dir/chat-parser.cpp.o.d -o CMakeFiles/common.dir/chat-parser.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/chat-parser.cpp [ 69%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/chat.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/chat.cpp.o -MF CMakeFiles/common.dir/chat.cpp.o.d -o CMakeFiles/common.dir/chat.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/chat.cpp [ 70%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/common.cpp.o -MF CMakeFiles/common.dir/common.cpp.o.d -o CMakeFiles/common.dir/common.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/common.cpp [ 71%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/console.cpp.o -MF CMakeFiles/common.dir/console.cpp.o.d -o CMakeFiles/common.dir/console.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/console.cpp [ 72%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/json-partial.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/json-partial.cpp.o -MF CMakeFiles/common.dir/json-partial.cpp.o.d -o CMakeFiles/common.dir/json-partial.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/json-partial.cpp [ 73%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/json-schema-to-grammar.cpp.o -MF CMakeFiles/common.dir/json-schema-to-grammar.cpp.o.d -o CMakeFiles/common.dir/json-schema-to-grammar.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/json-schema-to-grammar.cpp [ 75%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/llguidance.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/llguidance.cpp.o -MF CMakeFiles/common.dir/llguidance.cpp.o.d -o CMakeFiles/common.dir/llguidance.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/llguidance.cpp [ 76%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/log.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/log.cpp.o -MF CMakeFiles/common.dir/log.cpp.o.d -o CMakeFiles/common.dir/log.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/log.cpp [ 77%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/ngram-cache.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/ngram-cache.cpp.o -MF CMakeFiles/common.dir/ngram-cache.cpp.o.d -o CMakeFiles/common.dir/ngram-cache.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/ngram-cache.cpp [ 78%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/regex-partial.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/regex-partial.cpp.o -MF CMakeFiles/common.dir/regex-partial.cpp.o.d -o CMakeFiles/common.dir/regex-partial.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/regex-partial.cpp [ 79%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/sampling.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/sampling.cpp.o -MF CMakeFiles/common.dir/sampling.cpp.o.d -o CMakeFiles/common.dir/sampling.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/sampling.cpp [ 80%] Building CXX object vendor/llama.cpp/common/CMakeFiles/common.dir/speculative.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -pthread -MD -MT vendor/llama.cpp/common/CMakeFiles/common.dir/speculative.cpp.o -MF CMakeFiles/common.dir/speculative.cpp.o.d -o CMakeFiles/common.dir/speculative.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/speculative.cpp [ 82%] Linking CXX static library libcommon.a cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && /usr/local/bin/cmake -P CMakeFiles/common.dir/cmake_clean_target.cmake cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/common && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/common.dir/link.txt --verbose=1 /usr/bin/llvm-ar qc libcommon.a CMakeFiles/common.dir/arg.cpp.o "CMakeFiles/common.dir/chat-parser.cpp.o" CMakeFiles/common.dir/chat.cpp.o CMakeFiles/common.dir/common.cpp.o CMakeFiles/common.dir/console.cpp.o "CMakeFiles/common.dir/json-partial.cpp.o" "CMakeFiles/common.dir/json-schema-to-grammar.cpp.o" CMakeFiles/common.dir/llguidance.cpp.o CMakeFiles/common.dir/log.cpp.o "CMakeFiles/common.dir/ngram-cache.cpp.o" "CMakeFiles/common.dir/regex-partial.cpp.o" CMakeFiles/common.dir/sampling.cpp.o CMakeFiles/common.dir/speculative.cpp.o "CMakeFiles/build_info.dir/build-info.cpp.o" /usr/bin/llvm-ranlib libcommon.a [ 82%] Built target common /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/build [ 83%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dmtmd_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wno-cast-qual -pthread -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd.cpp.o -MF CMakeFiles/mtmd.dir/mtmd.cpp.o.d -o CMakeFiles/mtmd.dir/mtmd.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/mtmd.cpp [ 84%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd-audio.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dmtmd_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wno-cast-qual -pthread -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd-audio.cpp.o -MF CMakeFiles/mtmd.dir/mtmd-audio.cpp.o.d -o CMakeFiles/mtmd.dir/mtmd-audio.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/mtmd-audio.cpp [ 85%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/clip.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dmtmd_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wno-cast-qual -pthread -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/clip.cpp.o -MF CMakeFiles/mtmd.dir/clip.cpp.o.d -o CMakeFiles/mtmd.dir/clip.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/clip.cpp [ 86%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd-helper.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_BUILD -DLLAMA_SHARED -Dmtmd_EXPORTS -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../.. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/../../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -fPIC -Wno-cast-qual -pthread -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/mtmd.dir/mtmd-helper.cpp.o -MF CMakeFiles/mtmd.dir/mtmd-helper.cpp.o.d -o CMakeFiles/mtmd.dir/mtmd-helper.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/mtmd-helper.cpp [ 88%] Linking CXX shared library libmtmd.so cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/mtmd.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -fPIC -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/mtmd.dir/link.d -shared -Wl,-soname,libmtmd.so -o libmtmd.so CMakeFiles/mtmd.dir/mtmd.cpp.o "CMakeFiles/mtmd.dir/mtmd-audio.cpp.o" CMakeFiles/mtmd.dir/clip.cpp.o "CMakeFiles/mtmd.dir/mtmd-helper.cpp.o" -Wl,-rpath,"\$ORIGIN" ../../../../bin/libllama.so ../../../../bin/libggml.so ../../../../bin/libggml-cpu.so ../../../../bin/libggml-base.so -lpthread [ 88%] Built target mtmd /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/build [ 89%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/deprecation-warning.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-llava-cli.dir/deprecation-warning.cpp.o -MF CMakeFiles/llama-llava-cli.dir/deprecation-warning.cpp.o.d -o CMakeFiles/llama-llava-cli.dir/deprecation-warning.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/deprecation-warning.cpp [ 90%] Linking CXX executable llama-llava-cli cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama-llava-cli.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama-llava-cli.dir/link.d "CMakeFiles/llama-llava-cli.dir/deprecation-warning.cpp.o" -o llama-llava-cli [ 90%] Built target llama-llava-cli /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/build [ 91%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/deprecation-warning.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-gemma3-cli.dir/deprecation-warning.cpp.o -MF CMakeFiles/llama-gemma3-cli.dir/deprecation-warning.cpp.o.d -o CMakeFiles/llama-gemma3-cli.dir/deprecation-warning.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/deprecation-warning.cpp [ 92%] Linking CXX executable llama-gemma3-cli cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama-gemma3-cli.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama-gemma3-cli.dir/link.d "CMakeFiles/llama-gemma3-cli.dir/deprecation-warning.cpp.o" -o llama-gemma3-cli [ 92%] Built target llama-gemma3-cli /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/build [ 94%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/deprecation-warning.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-minicpmv-cli.dir/deprecation-warning.cpp.o -MF CMakeFiles/llama-minicpmv-cli.dir/deprecation-warning.cpp.o.d -o CMakeFiles/llama-minicpmv-cli.dir/deprecation-warning.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/deprecation-warning.cpp [ 95%] Linking CXX executable llama-minicpmv-cli cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama-minicpmv-cli.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama-minicpmv-cli.dir/link.d "CMakeFiles/llama-minicpmv-cli.dir/deprecation-warning.cpp.o" -o llama-minicpmv-cli [ 95%] Built target llama-minicpmv-cli /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/build [ 96%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/deprecation-warning.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-qwen2vl-cli.dir/deprecation-warning.cpp.o -MF CMakeFiles/llama-qwen2vl-cli.dir/deprecation-warning.cpp.o.d -o CMakeFiles/llama-qwen2vl-cli.dir/deprecation-warning.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/deprecation-warning.cpp [ 97%] Linking CXX executable llama-qwen2vl-cli cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama-qwen2vl-cli.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama-qwen2vl-cli.dir/link.d "CMakeFiles/llama-qwen2vl-cli.dir/deprecation-warning.cpp.o" -o llama-qwen2vl-cli [ 97%] Built target llama-qwen2vl-cli /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/depend cd /tmp/tmpui5d1wk8/build && /usr/local/bin/cmake -E cmake_depends "Unix Makefiles" /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16 /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/DependInfo.cmake "--color=" /usr/bin/make -f vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/build.make vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/build [ 98%] Building CXX object vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/mtmd-cli.cpp.o cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && ccache /usr/local/libexec/ccache/c++ -DGGML_BACKEND_SHARED -DGGML_SHARED -DGGML_USE_CPU -DLLAMA_SHARED -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/common/../vendor -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/src/../include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/. -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/include -I/wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/ggml/include -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -pthread -MD -MT vendor/llama.cpp/tools/mtmd/CMakeFiles/llama-mtmd-cli.dir/mtmd-cli.cpp.o -MF CMakeFiles/llama-mtmd-cli.dir/mtmd-cli.cpp.o.d -o CMakeFiles/llama-mtmd-cli.dir/mtmd-cli.cpp.o -c /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/vendor/llama.cpp/tools/mtmd/mtmd-cli.cpp [100%] Linking CXX executable llama-mtmd-cli cd /tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd && /usr/local/bin/cmake -E cmake_link_script CMakeFiles/llama-mtmd-cli.dir/link.txt --verbose=1 /usr/local/libexec/ccache/c++ -O2 -pipe -fstack-protector-strong -fno-strict-aliasing -O3 -DNDEBUG -Xlinker --dependency-file=CMakeFiles/llama-mtmd-cli.dir/link.d "CMakeFiles/llama-mtmd-cli.dir/mtmd-cli.cpp.o" -o llama-mtmd-cli -Wl,-rpath,/tmp/tmpui5d1wk8/build/vendor/llama.cpp/tools/mtmd:/tmp/tmpui5d1wk8/build/bin ../../common/libcommon.a libmtmd.so ../../../../bin/libllama.so ../../../../bin/libggml.so ../../../../bin/libggml-cpu.so ../../../../bin/libggml-base.so -lpthread [100%] Built target llama-mtmd-cli /usr/local/bin/cmake -E cmake_progress_start /tmp/tmpui5d1wk8/build/CMakeFiles 0 *** Installing project into wheel... -- Install configuration: "Release" -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/libggml-cpu.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/libggml.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-cpu.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-alloc.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-backend.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-blas.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-cann.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-cpp.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-cuda.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-opt.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-metal.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-rpc.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-sycl.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-vulkan.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/ggml-webgpu.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/gguf.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/libggml-base.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/cmake/ggml/ggml-config.cmake -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/cmake/ggml/ggml-version.cmake -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/libllama.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/llama.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/llama-cpp.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/cmake/llama/llama-config.cmake -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/cmake/llama/llama-version.cmake -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/bin/convert_hf_to_gguf.py -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/pkgconfig/llama.pc -- Installing: /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/llama_cpp/lib/libllama.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/llama_cpp/lib/libllama.so -- Installing: /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/llama_cpp/lib/libggml.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/llama_cpp/lib/libggml.so -- Installing: /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/llama_cpp/lib/libggml-base.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/llama_cpp/lib/libggml-base.so -- Installing: /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/llama_cpp/lib/libggml-cpu.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/llama_cpp/lib/libggml-cpu.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/lib/libmtmd.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/mtmd.h -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/include/mtmd-helper.h -- Installing: /wrkdirs/usr/ports/misc/py-llama-cpp-python/work-py311/llama-cpp-python-0.3.16/llama_cpp/lib/libmtmd.so -- Installing: /tmp/tmpui5d1wk8/wheel/platlib/llama_cpp/lib/libmtmd.so *** Making wheel... *** Created llama_cpp_python-0.3.16-cp311-cp311-freebsd_13_5_release_p7_amd64.whl Successfully built llama_cpp_python-0.3.16-cp311-cp311-freebsd_13_5_release_p7_amd64.whl =========================================================================== =================================================== ===== env: USE_PACKAGE_DEPENDS_ONLY=1 USER=root UID=0 GID=0 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-diskcache>=5.6.1 - not found ===> Installing existing package /packages/All/py311-diskcache-5.6.3_2.pkg [pkg.DaveG.ca] Installing py311-diskcache-5.6.3_2... [pkg.DaveG.ca] `-- Installing py311-django42-4.2.27... [pkg.DaveG.ca] | `-- Installing py311-asgiref-3.11.0... [pkg.DaveG.ca] | `-- Extracting py311-asgiref-3.11.0: .......... done [pkg.DaveG.ca] | `-- Installing py311-sqlite3-3.11.14_10... [pkg.DaveG.ca] | | `-- Installing sqlite3-3.50.4_2,1... [pkg.DaveG.ca] | | `-- Installing libedit-3.1.20251016_1,1... [pkg.DaveG.ca] | | `-- Extracting libedit-3.1.20251016_1,1: .......... done [pkg.DaveG.ca] | | `-- Extracting sqlite3-3.50.4_2,1: .......... done [pkg.DaveG.ca] | `-- Extracting py311-sqlite3-3.11.14_10: ...... done [pkg.DaveG.ca] | `-- Installing py311-sqlparse-0.5.5... [pkg.DaveG.ca] | `-- Extracting py311-sqlparse-0.5.5: .......... done [pkg.DaveG.ca] `-- Extracting py311-django42-4.2.27: .......... done [pkg.DaveG.ca] Extracting py311-diskcache-5.6.3_2: .......... done ===== Message from py311-django42-4.2.27: -- ===> NOTICE: This port is deprecated; you may wish to reconsider installing it: Django 4.2 is now in Extended Support, receives security fixes only and will reach End-Of-Life in April 2026. Upgrade to a later Django version. See https://www.djangoproject.com/download/. It is scheduled to be removed on or after 2026-05-31. ===> py311-llama-cpp-python-0.3.16 depends on package: py311-diskcache>=5.6.1 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-Jinja2>=2.11.3 - not found ===> Installing existing package /packages/All/py311-Jinja2-3.1.6.pkg [pkg.DaveG.ca] Installing py311-Jinja2-3.1.6... [pkg.DaveG.ca] `-- Installing py311-Babel-2.17.0_1... [pkg.DaveG.ca] `-- Extracting py311-Babel-2.17.0_1: .......... done [pkg.DaveG.ca] `-- Installing py311-markupsafe-3.0.3... [pkg.DaveG.ca] `-- Extracting py311-markupsafe-3.0.3: .......... done [pkg.DaveG.ca] Extracting py311-Jinja2-3.1.6: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-Jinja2>=2.11.3 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-numpy>=1.16,1<1.27,1 - not found ===> Installing existing package /packages/All/py311-numpy-1.26.4_12,1.pkg [pkg.DaveG.ca] Installing py311-numpy-1.26.4_12,1... [pkg.DaveG.ca] `-- Installing gcc14-14.2.0_4... [pkg.DaveG.ca] | `-- Installing binutils-2.44,1... [pkg.DaveG.ca] | | `-- Installing zstd-1.5.7_1... [pkg.DaveG.ca] | | `-- Extracting zstd-1.5.7_1: .......... done [pkg.DaveG.ca] | `-- Extracting binutils-2.44,1: .......... done [pkg.DaveG.ca] | `-- Installing gmp-6.3.0... [pkg.DaveG.ca] | `-- Extracting gmp-6.3.0: .......... done [pkg.DaveG.ca] | `-- Installing mpc-1.3.1_1... [pkg.DaveG.ca] | | `-- Installing mpfr-4.2.2,1... [pkg.DaveG.ca] | | `-- Extracting mpfr-4.2.2,1: .......... done [pkg.DaveG.ca] | `-- Extracting mpc-1.3.1_1: ...... done [pkg.DaveG.ca] `-- Extracting gcc14-14.2.0_4: .......... done create symlink for gcc14 create symlink for gcc14 (world) create symlink for g++14 create symlink for g++14 (world) create symlink for cpp14 create symlink for cpp14 (world) [pkg.DaveG.ca] `-- Installing openblas-0.3.30,2... [pkg.DaveG.ca] `-- Extracting openblas-0.3.30,2: .......... done [pkg.DaveG.ca] `-- Installing suitesparse-umfpack-6.3.7... [pkg.DaveG.ca] | `-- Installing suitesparse-amd-3.3.4... [pkg.DaveG.ca] | | `-- Installing suitesparse-config-7.12.1... [pkg.DaveG.ca] | | `-- Extracting suitesparse-config-7.12.1: .......... done [pkg.DaveG.ca] | `-- Extracting suitesparse-amd-3.3.4: .......... done [pkg.DaveG.ca] | `-- Installing suitesparse-camd-3.3.5... [pkg.DaveG.ca] | `-- Extracting suitesparse-camd-3.3.5: .......... done [pkg.DaveG.ca] | `-- Installing suitesparse-ccolamd-3.3.5... [pkg.DaveG.ca] | `-- Extracting suitesparse-ccolamd-3.3.5: .......... done [pkg.DaveG.ca] | `-- Installing suitesparse-cholmod-5.3.4... [pkg.DaveG.ca] | | `-- Installing suitesparse-colamd-3.3.5... [pkg.DaveG.ca] | | `-- Extracting suitesparse-colamd-3.3.5: .......... done [pkg.DaveG.ca] | `-- Extracting suitesparse-cholmod-5.3.4: .......... done [pkg.DaveG.ca] `-- Extracting suitesparse-umfpack-6.3.7: .......... done [pkg.DaveG.ca] Extracting py311-numpy-1.26.4_12,1: .......... done ===== Message from gcc14-14.2.0_4: -- To ensure binaries built with this toolchain find appropriate versions of the necessary run-time libraries, you may want to link using -Wl,-rpath=/usr/local/lib/gcc14 For ports leveraging USE_GCC, USES=compiler, or USES=fortran this happens transparently. ===> py311-llama-cpp-python-0.3.16 depends on package: py311-numpy>=1.16,1<1.27,1 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-typing-extensions>=4.5.0 - not found ===> Installing existing package /packages/All/py311-typing-extensions-4.15.0.pkg [pkg.DaveG.ca] Installing py311-typing-extensions-4.15.0... [pkg.DaveG.ca] Extracting py311-typing-extensions-4.15.0: ....... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-typing-extensions>=4.5.0 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-fastapi>=0.100.0 - not found ===> Installing existing package /packages/All/py311-fastapi-0.128.0.pkg [pkg.DaveG.ca] Installing py311-fastapi-0.128.0... [pkg.DaveG.ca] `-- Installing py311-annotated-doc-0.0.4... [pkg.DaveG.ca] `-- Extracting py311-annotated-doc-0.0.4: .......... done [pkg.DaveG.ca] `-- Installing py311-pydantic2-2.12.5... [pkg.DaveG.ca] | `-- Installing py311-annotated-types-0.7.0... [pkg.DaveG.ca] | `-- Extracting py311-annotated-types-0.7.0: .......... done [pkg.DaveG.ca] | `-- Installing py311-pydantic-core-2.41.5_2... [pkg.DaveG.ca] | `-- Extracting py311-pydantic-core-2.41.5_2: .......... done [pkg.DaveG.ca] | `-- Installing py311-typing-inspection-0.4.2... [pkg.DaveG.ca] | `-- Extracting py311-typing-inspection-0.4.2: .......... done [pkg.DaveG.ca] `-- Extracting py311-pydantic2-2.12.5: .......... done [pkg.DaveG.ca] `-- Installing py311-starlette-0.50.0... [pkg.DaveG.ca] | `-- Installing py311-anyio-4.12.0... [pkg.DaveG.ca] | | `-- Installing py311-idna-3.11... [pkg.DaveG.ca] | | `-- Extracting py311-idna-3.11: .......... done [pkg.DaveG.ca] | | `-- Installing py311-truststore-0.10.4... [pkg.DaveG.ca] | | `-- Extracting py311-truststore-0.10.4: .......... done [pkg.DaveG.ca] | `-- Extracting py311-anyio-4.12.0: .......... done [pkg.DaveG.ca] | `-- Installing py311-httpx-0.28.1_1... [pkg.DaveG.ca] | | `-- Installing py311-certifi-2025.11.12... [pkg.DaveG.ca] | | `-- Extracting py311-certifi-2025.11.12: .......... done [pkg.DaveG.ca] | | `-- Installing py311-h2-4.1.0_1... [pkg.DaveG.ca] | | `-- Installing py311-hpack-4.0.0_1... [pkg.DaveG.ca] | | `-- Extracting py311-hpack-4.0.0_1: .......... done [pkg.DaveG.ca] | | `-- Installing py311-hyperframe-6.0.0_1... [pkg.DaveG.ca] | | `-- Extracting py311-hyperframe-6.0.0_1: .......... done [pkg.DaveG.ca] | | `-- Extracting py311-h2-4.1.0_1: .......... done [pkg.DaveG.ca] | | `-- Installing py311-httpcore-1.0.9... [pkg.DaveG.ca] | | `-- Installing py311-h11-0.16.0... [pkg.DaveG.ca] | | `-- Extracting py311-h11-0.16.0: .......... done [pkg.DaveG.ca] | | `-- Extracting py311-httpcore-1.0.9: .......... done [pkg.DaveG.ca] | | `-- Installing py311-socksio-1.0.0_1... [pkg.DaveG.ca] | | `-- Extracting py311-socksio-1.0.0_1: .......... done [pkg.DaveG.ca] | `-- Extracting py311-httpx-0.28.1_1: .......... done [pkg.DaveG.ca] | `-- Installing py311-itsdangerous-2.2.0... [pkg.DaveG.ca] | `-- Extracting py311-itsdangerous-2.2.0: .......... done [pkg.DaveG.ca] | `-- Installing py311-python-multipart-0.0.21... [pkg.DaveG.ca] | `-- Extracting py311-python-multipart-0.0.21: .......... done [pkg.DaveG.ca] | `-- Installing py311-pyyaml-6.0.3... [pkg.DaveG.ca] | | `-- Installing libyaml-0.2.5... [pkg.DaveG.ca] | | `-- Extracting libyaml-0.2.5: ...... done [pkg.DaveG.ca] | `-- Extracting py311-pyyaml-6.0.3: .......... done [pkg.DaveG.ca] `-- Extracting py311-starlette-0.50.0: .......... done [pkg.DaveG.ca] Extracting py311-fastapi-0.128.0: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-fastapi>=0.100.0 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-pydantic-settings>=2.0.1 - not found ===> Installing existing package /packages/All/py311-pydantic-settings-2.6.0.pkg [pkg.DaveG.ca] Installing py311-pydantic-settings-2.6.0... [pkg.DaveG.ca] `-- Installing py311-python-dotenv-1.2.1... [pkg.DaveG.ca] | `-- Installing py311-click-8.1.7_1... [pkg.DaveG.ca] | `-- Extracting py311-click-8.1.7_1: .......... done [pkg.DaveG.ca] `-- Extracting py311-python-dotenv-1.2.1: .......... done [pkg.DaveG.ca] Extracting py311-pydantic-settings-2.6.0: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-pydantic-settings>=2.0.1 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-sse-starlette>=1.6.1 - not found ===> Installing existing package /packages/All/py311-sse-starlette-3.1.2.pkg [pkg.DaveG.ca] Installing py311-sse-starlette-3.1.2... [pkg.DaveG.ca] `-- Installing py311-uvicorn-0.40.0... [pkg.DaveG.ca] `-- Extracting py311-uvicorn-0.40.0: .......... done [pkg.DaveG.ca] Extracting py311-sse-starlette-3.1.2: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-sse-starlette>=1.6.1 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-starlette-context>=0.3.6 - not found ===> Installing existing package /packages/All/py311-starlette-context-0.4.0.pkg [pkg.DaveG.ca] Installing py311-starlette-context-0.4.0... [pkg.DaveG.ca] Extracting py311-starlette-context-0.4.0: .......... done ===> py311-llama-cpp-python-0.3.16 depends on package: py311-starlette-context>=0.3.6 - found ===> Returning to build of py311-llama-cpp-python-0.3.16 ===> py311-llama-cpp-python-0.3.16 depends on package: py311-uvicorn>=0.22.0 - found ===> py311-llama-cpp-python-0.3.16 depends on package: py311-pyyaml>=5.1 - found ===> py311-llama-cpp-python-0.3.16 depends on file: /usr/local/bin/python3.11 - found =========================================================================== =================================================== ===== env: NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Staging for py311-llama-cpp-python-0.3.16 ===> Generating temporary packing list ====> Compressing man pages (compress-man) =========================================================================== =================================================== ===== env: 'PKG_NOTES=build_timestamp ports_top_git_hash ports_top_checkout_unclean port_git_hash port_checkout_unclean built_by' 'PKG_NOTE_build_timestamp=2026-01-24T15:20:48+0000' 'PKG_NOTE_ports_top_git_hash=076a295eb33c' 'PKG_NOTE_ports_top_checkout_unclean=no' 'PKG_NOTE_port_git_hash=b6660c1d142a' 'PKG_NOTE_port_checkout_unclean=no' 'PKG_NOTE_built_by=poudriere-git-3.4.2' NO_DEPENDS=yes USER=root UID=0 GID=0 ===> Building packages for py311-llama-cpp-python-0.3.16 ===> Building py311-llama-cpp-python-0.3.16 =========================================================================== =>> Cleaning up wrkdir ===> Cleaning for py311-llama-cpp-python-0.3.16 build of misc/py-llama-cpp-python@py311 | py311-llama-cpp-python-0.3.16 ended at Sat Jan 24 10:33:02 EST 2026 build time: 00:12:17