Skip to content

kjh2159/llama.cpp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4,567 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

llama.cpp

llama

License: MIT Release Server

Manifesto / ggml / ops

LLM inference in C/C++

Recent API changes

Hot topics


Quick start (llama.cpp)

The quick start guide of the basic llama.cpp follows the original repository. The main focus of IGNITE for on-device inference is based on llama-cli, guided by the llama-completion.

Quick start (IGNITE)

Model download

python downloader.py

Through this file, you can download models, which are pre-selected to evaluate themselves on IGNITE. If there is no preferred model, you can download and run your own gguf models also.

Build (on-device)

cd scripts && sh build-android.sh && cd ..

Run (on-device)

chmod +x scripts-termux/run.sh
su -c "sh scripts-termux/run.sh"

Build (Linux)

cd scripts && sh build.sh && cd ..

Run (Linux)

./build/bin/ignite \
    -m models/qwen-1.5-0.5b-chat-q4k.gguf \
    -cnv \
    --temp 0 \
    --top-k 1 \
    --threads 1 \
    --output-path outputs/hotpot_0_0.csv \
    --json-path dataset/hotpot_qa_30.json

This will be filled up. Please wait.

About

No description, website, or topics provided.

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors