Skip to content

avble/av_llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

25 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Overview

  • Gen AI stuff

Tech-stack

  • A lightweight OpenAI API compatible server: av_connect http server in C++
  • Text-generation: llama.cp
  • Web UI: Provide a simple web UI interface to explore/experiment (borrowed from @llama.cpp project)

A snapshot

demo-1

Quick started

Obtain the latest container from docker hub

** Currently, the docker build is quite outdated build **

docker image pull harryavble/av_llm

Access to Web interface at http://127.0.0.1:8080

Supported model

Download model and run

docker run -p 8080:8080 -v $your_host_model_folder:/work/model av_llm ./av_llm -m /work/model/$your_model_file

Compile and run

$ cmake -B build && cmake --build build
$ build/av_llm -m <path to gguf file>

UI

Should work with below UI

Future work

  • Support more LLM models
  • Support more OpenAI API server
  • Support more application

Note

This is demonstration version, some issues or error checking is not fully validated.
Contact me via avble.harry dot gmail.com if any

Reference

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published