Flutter plugin wrapping llama.cpp to enable fully offline, on-device chat inference powering Maath:ai.
The Maathai_llama plugin is a Flutter plugin that integrates llama.cpp via native bindings to run large language models entirely on-device. Maath:ai uses this plugin to deliver private, offline conversations with no server dependency.
These steps mirror the repository instructions.
git clone https://github.com/kipkirui/Maathai_llama
cd Maathai_llama
git submodule update --init --recursive
cd maathai_llamma
flutter pub get
cd example
flutter run --device-id <device>
See the repo for updates and platform specifics.
Inference is performed locally; prompts and responses remain on the device. This aligns with Maath:ai’s edge privacy model.