Skip to content

Support for local LLMs #3

@iwr-redmond

Description

@iwr-redmond

It would be helpful to add a package like xllamacpp to facilitate local VLM inference rather than being reliant on the Google API.

The xllamacpp package supports Vulkan and MPS inference as well as CUDA. Sample inference code can be reviewed here.

Metadata

Metadata

Assignees

Labels

No labels
No labels

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions