1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
| ...
Building wheels for collected packages: llama-cpp-python, peft
Building wheel for llama-cpp-python (pyproject.toml) ... error
error: subprocess-exited-with-error
× Building wheel for llama-cpp-python (pyproject.toml) did not run successfully.
│ exit code: 1
╰─> [131 lines of output]
...
-- Configuring done (0.3s)
-- Generating done (0.0s)
-- Build files have been written to: /tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55/_skbuild/linux-x86_64-3.9/cmake-build
[1/2] Generating /tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55/vendor/llama.cpp/libllama.so
FAILED: /tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55/vendor/llama.cpp/libllama.so
cd /tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55/vendor/llama.cpp && make libllama.so
I llama.cpp build info:
I UNAME_S: Linux
I UNAME_P: x86_64
I UNAME_M: x86_64
I CFLAGS: -I. -O3 -std=c11 -fPIC -DNDEBUG -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -march=native -mtune=native
I CXXFLAGS: -I. -I./examples -O3 -std=c++11 -fPIC -DNDEBUG -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native
I LDFLAGS:
I CC: cc (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
I CXX: g++ (Ubuntu 7.5.0-3ubuntu1~18.04) 7.5.0
g++ -I. -I./examples -O3 -std=c++11 -fPIC -DNDEBUG -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wno-multichar -pthread -march=native -mtune=native -c llama.cpp -o llama.o
llama.cpp: In function ‘size_t llama_set_state_data(llama_context*, const uint8_t*)’:
llama.cpp:2610:36: warning: cast from type ‘const uint8_t* {aka const unsigned char*}’ to type ‘void*’ casts away qualifiers [-Wcast-qual]
kin3d->data = (void *) in;
^~
llama.cpp:2614:36: warning: cast from type ‘const uint8_t* {aka const unsigned char*}’ to type ‘void*’ casts away qualifiers [-Wcast-qual]
vin3d->data = (void *) in;
^~
cc -I. -O3 -std=c11 -fPIC -DNDEBUG -Wall -Wextra -Wpedantic -Wcast-qual -Wdouble-promotion -Wshadow -Wstrict-prototypes -Wpointer-arith -pthread -march=native -mtune=native -c ggml.c -o ggml.o
ggml.c: In function ‘ggml_vec_dot_q4_2_q8_0’:
ggml.c:3253:40: warning: implicit declaration of function ‘_mm256_set_m128’; did you mean ‘_mm256_set_epi8’? [-Wimplicit-function-declaration]
const __m256 d = _mm256_mul_ps(_mm256_set_m128(d1, d0), _mm256_broadcast_ss(&y[i].d));
^~~~~~~~~~~~~~~
_mm256_set_epi8
ggml.c:3253:40: error: incompatible type for argument 1 of ‘_mm256_mul_ps’
In file included from /usr/lib/gcc/x86_64-linux-gnu/7/include/immintrin.h:41:0,
from ggml.c:189:
/usr/lib/gcc/x86_64-linux-gnu/7/include/avxintrin.h:318:1: note: expected ‘__m256 {aka __vector(8) float}’ but argument is of type ‘int’
_mm256_mul_ps (__m256 __A, __m256 __B)
^~~~~~~~~~~~~
ggml.c:3257:22: warning: implicit declaration of function ‘_mm256_set_m128i’; did you mean ‘_mm256_set_epi8’? [-Wimplicit-function-declaration]
__m256i bx = _mm256_set_m128i(bx1, bx0);
^~~~~~~~~~~~~~~~
_mm256_set_epi8
ggml.c:3257:22: error: incompatible types when initializing type ‘__m256i {aka __vector(4) long long int}’ using type ‘int’
Makefile:185: recipe for target 'ggml.o' failed
make: *** [ggml.o] Error 1
ninja: build stopped: subcommand failed.
Traceback (most recent call last):
File "/tmp/pip-build-env-cm_vdfny/overlay/lib/python3.9/site-packages/skbuild/setuptools_wrap.py", line 674, in setup
cmkr.make(make_args, install_target=cmake_install_target, env=env)
File "/tmp/pip-build-env-cm_vdfny/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 696, in make
self.make_impl(clargs=clargs, config=config, source_dir=source_dir, install_target=install_target, env=env)
File "/tmp/pip-build-env-cm_vdfny/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 741, in make_impl
raise SKBuildError(msg)
An error occurred while building with CMake.
Command:
/tmp/pip-build-env-cm_vdfny/overlay/lib/python3.9/site-packages/cmake/data/bin/cmake --build . --target install --config Release --
Install target:
install
Source directory:
/tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55
Working directory:
/tmp/pip-install-rzg2ayje/llama-cpp-python_ae749d81f3f74840a75abef221510a55/_skbuild/linux-x86_64-3.9/cmake-build
Please check the install target is valid and see CMake's output for more information.
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
ERROR: Failed building wheel for llama-cpp-python
Building wheel for peft (pyproject.toml) ... done
Created wheel for peft: filename=peft-0.4.0.dev0-py3-none-any.whl size=56306 sha256=aeadc3499eecab0d78df3bb60230221046062dd49030b88f0f36d154a1b01fcd
Stored in directory: /tmp/pip-ephem-wheel-cache-x1u8mjti/wheels/45/06/33/0048c03714539b315d99beef9ea6b6dad0fd5750105e221583
Successfully built peft
Failed to build llama-cpp-python
ERROR: Could not build wheels for llama-cpp-python, which is required to install pyproject.toml-based projects
|