mirror of
https://github.com/mudler/LocalAI.git
synced 2024-06-07 19:40:48 +00:00
c5c77d2b0d
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
247 lines
6.5 KiB
Markdown
247 lines
6.5 KiB
Markdown
|
|
+++
|
|
disableToc = false
|
|
title = "Mods"
|
|
weight = 2
|
|
+++
|
|
|
|
<p>
|
|
<img src="https://github.com/charmbracelet/mods/assets/25087/5442bf46-b908-47af-bf4e-60f7c38951c4" width="630" alt="Mods product art and type treatment"/>
|
|
<br>
|
|
</p>
|
|
|
|
AI for the command line, built for pipelines.
|
|
|
|
<p><img src="https://vhs.charm.sh/vhs-5Uyj0U6Hlqi1LVIIRyYKM5.gif" width="900" alt="a GIF of mods running"></p>
|
|
|
|
LLM based AI is really good at interpreting the output of commands and
|
|
returning the results in CLI friendly text formats like Markdown. Mods is a
|
|
simple tool that makes it super easy to use AI on the command line and in your
|
|
pipelines. Mods works with [OpenAI](https://platform.openai.com/account/api-keys)
|
|
and [LocalAI](https://github.com/go-skynet/LocalAI)
|
|
|
|
To get started, [install Mods](#installation) and check out some of the
|
|
examples below. Since Mods has built-in Markdown formatting, you may also want
|
|
to grab [Glow](https://github.com/charmbracelet/glow) to give the output some
|
|
_pizzazz_.
|
|
|
|
Github Link - https://github.com/charmbracelet/mods
|
|
|
|
## What Can It Do?
|
|
|
|
Mods works by reading standard in and prefacing it with a prompt supplied in
|
|
the `mods` arguments. It sends the input text to an LLM and prints out the
|
|
result, optionally asking the LLM to format the response as Markdown. This
|
|
gives you a way to "question" the output of a command. Mods will also work on
|
|
standard in or an argument supplied prompt individually.
|
|
|
|
## Installation
|
|
|
|
Mods works with OpenAI compatible endpoints. By default, Mods is configured to
|
|
support OpenAI's official API and a LocalAI installation running on port 8080.
|
|
You can configure additional endpoints in your settings file by running
|
|
`mods --settings`.
|
|
|
|
### LocalAI
|
|
|
|
LocalAI allows you to run a multitude of models locally. Mods works with the
|
|
GPT4ALL-J model as setup in [this tutorial](https://github.com/go-skynet/LocalAI#example-use-gpt4all-j-model).
|
|
You can define more LocalAI models and endpoints with `mods --settings`.
|
|
|
|
### Install Mods
|
|
|
|
```bash
|
|
# macOS or Linux
|
|
brew install charmbracelet/tap/mods
|
|
|
|
# Arch Linux (btw)
|
|
yay -S mods
|
|
|
|
# Debian/Ubuntu
|
|
sudo mkdir -p /etc/apt/keyrings
|
|
curl -fsSL https://repo.charm.sh/apt/gpg.key | sudo gpg --dearmor -o /etc/apt/keyrings/charm.gpg
|
|
echo "deb [signed-by=/etc/apt/keyrings/charm.gpg] https://repo.charm.sh/apt/ * *" | sudo tee /etc/apt/sources.list.d/charm.list
|
|
sudo apt update && sudo apt install mods
|
|
|
|
# Fedora/RHEL
|
|
echo '[charm]
|
|
name=Charm
|
|
baseurl=https://repo.charm.sh/yum/
|
|
enabled=1
|
|
gpgcheck=1
|
|
gpgkey=https://repo.charm.sh/yum/gpg.key' | sudo tee /etc/yum.repos.d/charm.repo
|
|
sudo yum install mods
|
|
```
|
|
|
|
Or, download it:
|
|
|
|
- [Packages][releases] are available in Debian and RPM formats
|
|
- [Binaries][releases] are available for Linux, macOS, and Windows
|
|
|
|
[releases]: https://github.com/charmbracelet/mods/releases
|
|
|
|
Or, just install it with `go`:
|
|
|
|
```sh
|
|
go install github.com/charmbracelet/mods@latest
|
|
```
|
|
|
|
## Saving conversations
|
|
|
|
Conversations save automatically. They are identified by their latest prompt.
|
|
Similar to Git, conversations have a SHA-1 identifier and a title. Conversations
|
|
can be updated, maintaining their SHA-1 identifier but changing their title.
|
|
|
|
<p><img src="https://vhs.charm.sh/vhs-6MMscpZwgzohYYMfTrHErF.gif" width="900" alt="a GIF listing and showing saved conversations."></p>
|
|
|
|
## Settings
|
|
|
|
`--settings`
|
|
|
|
Mods lets you tune your query with a variety of settings. You can configure
|
|
Mods with `mods --settings` or pass the settings as environment variables
|
|
and flags.
|
|
|
|
#### Model
|
|
|
|
`-m`, `--model`, `MODS_MODEL`
|
|
|
|
Mods uses `gpt-4` with OpenAI by default but you can specify any model as long
|
|
as your account has access to it or you have installed locally with LocalAI.
|
|
|
|
You can add new models to the settings with `mods --settings`.
|
|
You can also specify a model and an API endpoint with `-m` and `-a`
|
|
to use models not in the settings file.
|
|
|
|
#### Title
|
|
|
|
`-t`, `--title`
|
|
|
|
Set a custom save title for the conversation.
|
|
|
|
#### Continue last
|
|
|
|
`-C`, `--continue-last`
|
|
|
|
Continues the previous conversation.
|
|
|
|
#### Continue
|
|
|
|
`-c`, `--continue`
|
|
|
|
Continue from the last response or a given title or SHA1.
|
|
|
|
#### List
|
|
|
|
`-l`, `--list`
|
|
|
|
Lists all saved conversations.
|
|
|
|
#### Show
|
|
|
|
`-s`, `--show`
|
|
|
|
Show the saved conversation the given title or SHA1.
|
|
|
|
#### Delete
|
|
|
|
`--delete`
|
|
|
|
Deletes the saved conversation with the given title or SHA1.
|
|
|
|
#### Format As Markdown
|
|
|
|
`-f`, `--format`, `MODS_FORMAT`
|
|
|
|
Ask the LLM to format the response as markdown. You can edit the text passed to
|
|
the LLM with `mods --settings` then changing the `format-text` value.
|
|
|
|
#### Raw
|
|
|
|
`-r`, `--raw`, `MODS_RAW`
|
|
|
|
Print the raw response without syntax highlighting, even when connect to a TTY.
|
|
|
|
#### Max Tokens
|
|
|
|
`--max-tokens`, `MODS_MAX_TOKENS`
|
|
|
|
Max tokens tells the LLM to respond in less than this number of tokens. LLMs
|
|
are better at longer responses so values larger than 256 tend to work best.
|
|
|
|
#### Temperature
|
|
|
|
`--temp`, `MODS_TEMP`
|
|
|
|
Sampling temperature is a number between 0.0 and 2.0 and determines how
|
|
confident the model is in its choices. Higher values make the output more
|
|
random and lower values make it more deterministic.
|
|
|
|
#### TopP
|
|
|
|
`--topp`, `MODS_TOPP`
|
|
|
|
Top P is an alternative to sampling temperature. It's a number between 0.0 and
|
|
2.0 with smaller numbers narrowing the domain from which the model will create
|
|
its response.
|
|
|
|
#### No Limit
|
|
|
|
`--no-limit`, `MODS_NO_LIMIT`
|
|
|
|
By default Mods attempts to size the input to the maximum size the allowed by
|
|
the model. You can potentially squeeze a few more tokens into the input by
|
|
setting this but also risk getting a max token exceeded error from the OpenAI API.
|
|
|
|
#### Include Prompt
|
|
|
|
`-P`, `--prompt`, `MODS_INCLUDE_PROMPT`
|
|
|
|
Include prompt will preface the response with the entire prompt, both standard
|
|
in and the prompt supplied by the arguments.
|
|
|
|
#### Include Prompt Args
|
|
|
|
`-p`, `--prompt-args`, `MODS_INCLUDE_PROMPT_ARGS`
|
|
|
|
Include prompt args will include _only_ the prompt supplied by the arguments.
|
|
This can be useful if your standard in content is long and you just a want a
|
|
summary before the response.
|
|
|
|
#### Max Retries
|
|
|
|
`--max-retries`, `MODS_MAX_RETRIES`
|
|
|
|
The maximum number of retries to failed API calls. The retries happen with an
|
|
exponential backoff.
|
|
|
|
#### Fanciness
|
|
|
|
`--fanciness`, `MODS_FANCINESS`
|
|
|
|
Your desired level of fanciness.
|
|
|
|
#### Quiet
|
|
|
|
`-q`, `--quiet`, `MODS_QUIET`
|
|
|
|
Output nothing to standard err.
|
|
|
|
#### Reset Settings
|
|
|
|
`--reset-settings`
|
|
|
|
Backup your old settings file and reset everything to the defaults.
|
|
|
|
#### No Cache
|
|
|
|
`--no-cache`, `MODS_NO_CACHE`
|
|
|
|
Disables conversation saving.
|
|
|
|
#### HTTP Proxy
|
|
|
|
`-x`, `--http-proxy`, `MODS_HTTP_PROXY`
|
|
|
|
Use the HTTP proxy to the connect the API endpoints.
|