The misc/llama.cpp port
llama-cpp-0.0.4589 – LLM inference system (cvsweb github mirror)
Description
Inference of Meta's LLaMA model (and others) in pure C/C++ with minimal setup and state-of-the-art performance on a wide range of hardwareWWW: https://github.com/ggerganov/llama.cpp
Maintainer
The OpenBSD ports mailing-list
Categories
Build dependencies
Files
- /usr/local/bin/convert_hf_to_gguf.py
- /usr/local/bin/llama-batched
- /usr/local/bin/llama-batched-bench
- /usr/local/bin/llama-bench
- /usr/local/bin/llama-cli
- /usr/local/bin/llama-convert-llama2c-to-ggml
- /usr/local/bin/llama-cvector-generator
- /usr/local/bin/llama-embedding
- /usr/local/bin/llama-eval-callback
- /usr/local/bin/llama-export-lora
- /usr/local/bin/llama-gbnf-validator
- /usr/local/bin/llama-gen-docs
- /usr/local/bin/llama-gguf
- /usr/local/bin/llama-gguf-hash
- /usr/local/bin/llama-gguf-split
- /usr/local/bin/llama-gritlm
- /usr/local/bin/llama-imatrix
- /usr/local/bin/llama-infill
- /usr/local/bin/llama-llava-cli
- /usr/local/bin/llama-lookahead
- /usr/local/bin/llama-lookup
- /usr/local/bin/llama-lookup-create
- /usr/local/bin/llama-lookup-merge
- /usr/local/bin/llama-lookup-stats
- /usr/local/bin/llama-minicpmv-cli
- /usr/local/bin/llama-parallel
- /usr/local/bin/llama-passkey
- /usr/local/bin/llama-perplexity
- /usr/local/bin/llama-quantize
- /usr/local/bin/llama-quantize-stats
- /usr/local/bin/llama-qwen2vl-cli
- /usr/local/bin/llama-retrieval
- /usr/local/bin/llama-run
- /usr/local/bin/llama-save-load-state
- /usr/local/bin/llama-server
- /usr/local/bin/llama-simple
- /usr/local/bin/llama-simple-chat
- /usr/local/bin/llama-speculative
- /usr/local/bin/llama-speculative-simple
- /usr/local/bin/llama-tokenize
- /usr/local/bin/llama-tts
- /usr/local/bin/test-arg-parser
- /usr/local/bin/test-autorelease
- /usr/local/bin/test-backend-ops
- /usr/local/bin/test-barrier
- /usr/local/bin/test-chat-template
- /usr/local/bin/test-gguf
- /usr/local/bin/test-grammar-integration
- /usr/local/bin/test-grammar-parser
- /usr/local/bin/test-json-schema-to-grammar
- /usr/local/bin/test-llama-grammar
- /usr/local/bin/test-log
- /usr/local/bin/test-model-load-cancel
- /usr/local/bin/test-quantize-fns
- /usr/local/bin/test-quantize-perf
- /usr/local/bin/test-rope
- /usr/local/bin/test-sampling
- /usr/local/bin/test-tokenizer-0
- /usr/local/bin/test-tokenizer-1-bpe
- /usr/local/bin/test-tokenizer-1-spm
- /usr/local/include/ggml-alloc.h
- /usr/local/include/ggml-backend.h
- /usr/local/include/ggml-blas.h
- /usr/local/include/ggml-cann.h
- /usr/local/include/ggml-cpu.h
- /usr/local/include/ggml-cuda.h
- /usr/local/include/ggml-kompute.h
- /usr/local/include/ggml-metal.h
- /usr/local/include/ggml-opt.h
- /usr/local/include/ggml-rpc.h
- /usr/local/include/ggml-sycl.h
- /usr/local/include/ggml-vulkan.h
- /usr/local/include/ggml.h
- /usr/local/include/gguf.h
- /usr/local/include/llama-cpp.h
- /usr/local/include/llama.h
- /usr/local/lib/cmake/
- /usr/local/lib/cmake/ggml/
- /usr/local/lib/cmake/ggml/ggml-config.cmake
- /usr/local/lib/cmake/ggml/ggml-version.cmake
- /usr/local/lib/cmake/llama/
- /usr/local/lib/cmake/llama/llama-config.cmake
- /usr/local/lib/cmake/llama/llama-version.cmake
- /usr/local/lib/libggml-base.so.0.0
- /usr/local/lib/libggml-cpu.so.0.0
- /usr/local/lib/libggml.so.0.0
- /usr/local/lib/libllama.so.0.0
- /usr/local/lib/libllava_shared.so.0.0
- /usr/local/lib/pkgconfig/llama.pc