-
Notifications
You must be signed in to change notification settings - Fork 20
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Signed-off-by: Eric Curtin <ecurtin@redhat.com>
- Loading branch information
1 parent
8df8552
commit e1502a8
Showing
17 changed files
with
242 additions
and
759 deletions.
There are no files selected for viewing
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,3 +1,81 @@ | ||
# Project archived | ||
# Project undergoing rewrite | ||
|
||
New version at https://github.com/containers/ramalama | ||
The project is undergoing a complete rewrite in python3, the dependancy on containers will be removed. It will run on macOS and Linux native at a minimum and will support pulling, running and serving models from the following types of repos: huggingface, oci, ollama. | ||
|
||
You can still play away with the POC though :) | ||
|
||
# ramalama | ||
|
||
The goal of ramalama is to make AI even more boring. | ||
|
||
## Install | ||
|
||
Install ramalama by running this one-liner: | ||
|
||
``` | ||
curl -fsSL https://raw.githubusercontent.com/containers/ramalama/main/install.sh | sudo bash | ||
``` | ||
|
||
## Usage | ||
|
||
### Pulling Models | ||
|
||
You can pull a model using the `pull` command. By default, it pulls from the ollama registry. | ||
|
||
``` | ||
$ ramalama pull granite-code | ||
``` | ||
|
||
## Diagram | ||
|
||
``` | ||
+----------------+ | ||
| | | ||
| ramalama run | | ||
| | | ||
+-------+--------+ | ||
| | ||
v | ||
+----------------+ +-----------------------+ +------------------+ | ||
| | | Pull runtime layer | | Pull model layer | | ||
| Auto-detect +--->| for llama.cpp +--->| i.e. granite | | ||
| hardware type | | (CPU, Vulkan, AMD, | | | | ||
| | | Nvidia, Intel, | +------------------+ | ||
+----------------+ | Apple Silicon, etc.) | | Repo options: | | ||
+-----------------------+ +-+-------+------+-+ | ||
| | | | ||
v v v | ||
+---------+ +------+ +----------+ | ||
| Hugging | | quay | | Ollama | | ||
| Face | | | | Registry | | ||
+-------+-+ +---+--+ +-+--------+ | ||
| | | | ||
v v v | ||
+------------------+ | ||
| Start container | | ||
| with llama.cpp | | ||
| and granite | | ||
| model | | ||
+------------------+ | ||
``` | ||
|
||
## In development | ||
|
||
Regard this alpha, everything is under development, so expect breaking changes, luckily it's easy to reset everything and re-install: | ||
|
||
``` | ||
rm -rf /var/lib/ramalama # only required if running as root user | ||
rm -rf $HOME/.local/share/ramalama | ||
``` | ||
|
||
and install again. | ||
|
||
## Credit where credit is due | ||
|
||
For the vast majority of AI/LLM software we use, under the covers the heavy lifting is being done by: | ||
|
||
https://github.com/ggerganov/llama.cpp | ||
|
||
so if you like this tool, give llama.cpp repo a :star:, and hey, give us a :star: too while you are at it. | ||
|
||
![image](https://github.com/user-attachments/assets/d7a91662-5903-4117-ad41-2b193a852ea1) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file was deleted.
Oops, something went wrong.
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.