Thanks to visit codestin.com
Credit goes to github.com

Skip to content

aifoundry-org/olmo-llamafile

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Description

Allows to convert gguf-ed LMMs to llamafile and upload them to public HF repo You can do that from your local machine and from GitHub Actions

Status

  • Convert gguf model to llamafile and upload to huggingface using docker
  • GitHub actions for convert gguf model to llamafile and upload to huggingface
  • GitHub actions for convert raw model to llamafile and upload to huggingface

Up and running

Preparation

  1. Create HF repo
  2. Create Access Key with write permission and save it somewhere

Upload from local machine

  1. Create .env file and set there corresponding vars:
cp .env.example .env
  1. Up container
docker compose up -d olmo
  1. Copy gguf-ed model
docker cp OLMo-1.7-7B-hf.Q8_0.gguf llfiler-olmo-1:/app/
  1. Connect to container shell
docker exec -it llfiler-olmo-1 /bin/bash
  1. Convert ggufed model to llamafile:
llamafile-0.8.6/bin/llamafile-convert OLMo-1.7-7B-hf.Q8_0.gguf
  1. Upload to HF:
huggingface-cli upload "$HF_REPO" "$HF_REPO_FILE"

Upload from repository

  1. Copy .github/workflows/main.yml workflow to your repo
  2. Add secret HF_TOKEN to your repo secrets
  3. Input HF_REPO, HF_REPO_FILE, REMOTE_GGUF_MODEL, LLAMAFILE_RELEASE on workflow start

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published