2023-09-16 14:57:49 -04:00
|
|
|
cmake_minimum_required(VERSION 3.21)
|
2023-03-23 13:54:14 -04:00
|
|
|
|
|
|
|
|
project(llama_cpp)
|
|
|
|
|
|
2023-09-12 18:29:18 -04:00
|
|
|
option(LLAMA_BUILD "Build llama.cpp shared library and install alongside python package" ON)
|
2023-11-28 04:55:21 -05:00
|
|
|
option(LLAVA_BUILD "Build llava shared library and install alongside python package" ON)
|
2023-04-25 01:36:37 -04:00
|
|
|
|
2024-07-01 23:11:49 -04:00
|
|
|
function(llama_cpp_python_install_target target)
|
2024-11-15 19:31:12 -05:00
|
|
|
if(NOT TARGET ${target})
|
|
|
|
|
return()
|
|
|
|
|
endif()
|
|
|
|
|
|
2024-07-01 23:11:49 -04:00
|
|
|
install(
|
|
|
|
|
TARGETS ${target}
|
|
|
|
|
LIBRARY DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
RUNTIME DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
ARCHIVE DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
FRAMEWORK DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
RESOURCE DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
install(
|
|
|
|
|
TARGETS ${target}
|
|
|
|
|
LIBRARY DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
RUNTIME DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
ARCHIVE DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
FRAMEWORK DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
RESOURCE DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
set_target_properties(${target} PROPERTIES
|
|
|
|
|
INSTALL_RPATH "$ORIGIN"
|
|
|
|
|
BUILD_WITH_INSTALL_RPATH TRUE
|
|
|
|
|
)
|
2024-07-01 23:17:02 -04:00
|
|
|
if(UNIX)
|
|
|
|
|
if(APPLE)
|
|
|
|
|
set_target_properties(${target} PROPERTIES
|
|
|
|
|
INSTALL_RPATH "@loader_path"
|
|
|
|
|
BUILD_WITH_INSTALL_RPATH TRUE
|
|
|
|
|
)
|
|
|
|
|
else()
|
|
|
|
|
set_target_properties(${target} PROPERTIES
|
|
|
|
|
INSTALL_RPATH "$ORIGIN"
|
|
|
|
|
BUILD_WITH_INSTALL_RPATH TRUE
|
|
|
|
|
)
|
|
|
|
|
endif()
|
|
|
|
|
endif()
|
2024-07-01 23:11:49 -04:00
|
|
|
endfunction()
|
2024-07-01 22:03:19 -04:00
|
|
|
|
2023-09-12 18:29:18 -04:00
|
|
|
if (LLAMA_BUILD)
|
2023-04-08 02:39:17 -04:00
|
|
|
set(BUILD_SHARED_LIBS "On")
|
2023-11-10 16:32:36 -05:00
|
|
|
|
2024-07-01 21:28:11 -04:00
|
|
|
set(CMAKE_SKIP_BUILD_RPATH FALSE)
|
|
|
|
|
|
|
|
|
|
# When building, don't use the install RPATH already
|
|
|
|
|
# (but later on when installing)
|
|
|
|
|
set(CMAKE_BUILD_WITH_INSTALL_RPATH FALSE)
|
|
|
|
|
|
|
|
|
|
# Add the automatically determined parts of the RPATH
|
|
|
|
|
# which point to directories outside the build tree to the install RPATH
|
|
|
|
|
set(CMAKE_INSTALL_RPATH_USE_LINK_PATH TRUE)
|
2024-07-01 23:03:26 -04:00
|
|
|
set(CMAKE_SKIP_RPATH FALSE)
|
2024-07-01 21:28:11 -04:00
|
|
|
|
2024-10-22 09:33:48 -04:00
|
|
|
# Enable building of the common library
|
|
|
|
|
set(LLAMA_BUILD_COMMON ON CACHE BOOL "Build llama.cpp common library" FORCE)
|
|
|
|
|
|
2024-12-06 04:14:59 -05:00
|
|
|
# Architecture detection and settings for Apple platforms
|
2024-04-06 01:36:53 -04:00
|
|
|
if (APPLE)
|
2024-12-06 04:14:59 -05:00
|
|
|
# Get the target architecture
|
|
|
|
|
execute_process(
|
|
|
|
|
COMMAND uname -m
|
|
|
|
|
OUTPUT_VARIABLE HOST_ARCH
|
|
|
|
|
OUTPUT_STRIP_TRAILING_WHITESPACE
|
|
|
|
|
)
|
|
|
|
|
|
|
|
|
|
# If CMAKE_OSX_ARCHITECTURES is not set, use the host architecture
|
|
|
|
|
if(NOT CMAKE_OSX_ARCHITECTURES)
|
|
|
|
|
set(CMAKE_OSX_ARCHITECTURES ${HOST_ARCH} CACHE STRING "Build architecture for macOS" FORCE)
|
|
|
|
|
endif()
|
|
|
|
|
|
|
|
|
|
message(STATUS "Host architecture: ${HOST_ARCH}")
|
|
|
|
|
message(STATUS "Target architecture: ${CMAKE_OSX_ARCHITECTURES}")
|
|
|
|
|
|
|
|
|
|
# Configure based on target architecture
|
|
|
|
|
if(CMAKE_OSX_ARCHITECTURES STREQUAL "x86_64")
|
|
|
|
|
# Intel Mac settings
|
|
|
|
|
set(GGML_AVX "OFF" CACHE BOOL "ggml: enable AVX" FORCE)
|
|
|
|
|
set(GGML_AVX2 "OFF" CACHE BOOL "ggml: enable AVX2" FORCE)
|
|
|
|
|
set(GGML_FMA "OFF" CACHE BOOL "ggml: enable FMA" FORCE)
|
|
|
|
|
set(GGML_F16C "OFF" CACHE BOOL "ggml: enable F16C" FORCE)
|
|
|
|
|
endif()
|
|
|
|
|
|
|
|
|
|
# Metal settings (enable for both architectures)
|
|
|
|
|
set(GGML_METAL "ON" CACHE BOOL "ggml: enable Metal" FORCE)
|
|
|
|
|
set(GGML_METAL_EMBED_LIBRARY "ON" CACHE BOOL "ggml: embed metal library" FORCE)
|
2024-03-20 17:46:09 +01:00
|
|
|
endif()
|
|
|
|
|
|
2023-09-12 18:29:18 -04:00
|
|
|
add_subdirectory(vendor/llama.cpp)
|
2024-07-01 23:11:49 -04:00
|
|
|
llama_cpp_python_install_target(llama)
|
|
|
|
|
llama_cpp_python_install_target(ggml)
|
2024-11-15 18:00:30 -05:00
|
|
|
|
2024-11-15 19:31:12 -05:00
|
|
|
llama_cpp_python_install_target(ggml-base)
|
2024-11-15 18:00:30 -05:00
|
|
|
|
2024-11-15 19:31:12 -05:00
|
|
|
llama_cpp_python_install_target(ggml-amx)
|
|
|
|
|
llama_cpp_python_install_target(ggml-blas)
|
|
|
|
|
llama_cpp_python_install_target(ggml-can)
|
|
|
|
|
llama_cpp_python_install_target(ggml-cpu)
|
|
|
|
|
llama_cpp_python_install_target(ggml-cuda)
|
|
|
|
|
llama_cpp_python_install_target(ggml-hip)
|
|
|
|
|
llama_cpp_python_install_target(ggml-kompute)
|
|
|
|
|
llama_cpp_python_install_target(ggml-metal)
|
|
|
|
|
llama_cpp_python_install_target(ggml-musa)
|
|
|
|
|
llama_cpp_python_install_target(ggml-rpc)
|
|
|
|
|
llama_cpp_python_install_target(ggml-sycl)
|
|
|
|
|
llama_cpp_python_install_target(ggml-vulkan)
|
2024-11-15 18:00:30 -05:00
|
|
|
|
2023-09-16 14:57:49 -04:00
|
|
|
# Workaround for Windows + CUDA https://github.com/abetlen/llama-cpp-python/issues/563
|
2024-07-01 23:28:19 -04:00
|
|
|
if (WIN32)
|
2024-06-03 22:42:34 -06:00
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:llama>
|
2024-07-01 23:28:19 -04:00
|
|
|
DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:llama>
|
|
|
|
|
DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:ggml>
|
|
|
|
|
DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:ggml>
|
|
|
|
|
DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
2024-06-03 22:42:34 -06:00
|
|
|
)
|
|
|
|
|
endif()
|
2023-11-10 16:32:36 -05:00
|
|
|
|
2023-11-28 04:55:21 -05:00
|
|
|
if (LLAVA_BUILD)
|
2024-05-10 10:18:47 -04:00
|
|
|
if (LLAMA_CUBLAS OR LLAMA_CUDA)
|
2024-02-11 12:53:59 -06:00
|
|
|
add_compile_definitions(GGML_USE_CUBLAS)
|
2024-05-10 10:18:47 -04:00
|
|
|
add_compile_definitions(GGML_USE_CUDA)
|
2024-02-11 12:53:59 -06:00
|
|
|
endif()
|
|
|
|
|
|
|
|
|
|
if (LLAMA_METAL)
|
|
|
|
|
add_compile_definitions(GGML_USE_METAL)
|
|
|
|
|
endif()
|
|
|
|
|
|
2023-11-28 04:55:21 -05:00
|
|
|
# Building llava
|
|
|
|
|
add_subdirectory(vendor/llama.cpp/examples/llava)
|
|
|
|
|
set_target_properties(llava_shared PROPERTIES OUTPUT_NAME "llava")
|
2024-12-06 04:14:59 -05:00
|
|
|
|
2023-11-28 04:55:21 -05:00
|
|
|
if (WIN32)
|
|
|
|
|
set_target_properties(llava_shared PROPERTIES CUDA_ARCHITECTURES OFF)
|
|
|
|
|
endif()
|
2024-07-01 23:11:49 -04:00
|
|
|
llama_cpp_python_install_target(llava_shared)
|
2024-07-01 23:28:19 -04:00
|
|
|
if (WIN32)
|
|
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:llava_shared>
|
|
|
|
|
DESTINATION ${CMAKE_CURRENT_SOURCE_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
install(
|
|
|
|
|
FILES $<TARGET_RUNTIME_DLLS:llava_shared>
|
|
|
|
|
DESTINATION ${SKBUILD_PLATLIB_DIR}/llama_cpp/lib
|
|
|
|
|
)
|
|
|
|
|
endif()
|
2024-10-22 09:33:48 -04:00
|
|
|
|
|
|
|
|
# Fix for llava build: Add include directory for llama.h
|
|
|
|
|
# Move these commands after the add_subdirectory call
|
|
|
|
|
target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
|
|
|
|
|
target_include_directories(llava PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
|
|
|
|
|
|
|
|
|
|
if (BUILD_SHARED_LIBS)
|
|
|
|
|
target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
|
|
|
|
|
target_include_directories(llava_shared PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/ggml/include)
|
|
|
|
|
endif()
|
|
|
|
|
|
|
|
|
|
target_include_directories(llama-llava-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
|
|
|
|
|
target_include_directories(llama-minicpmv-cli PUBLIC ${CMAKE_CURRENT_SOURCE_DIR}/vendor/llama.cpp/include)
|
2023-11-10 16:32:36 -05:00
|
|
|
endif()
|
2024-07-09 00:29:29 -04:00
|
|
|
endif()
|