mirror of
https://github.com/mudler/LocalAI.git
synced 2024-06-07 19:40:48 +00:00
chore: update README to include new Helm values (#369)
This commit is contained in:
parent
04797a80e1
commit
76c561a908
76
README.md
76
README.md
@ -608,47 +608,69 @@ It should work, however you need to make sure you give enough resources to the c
|
|||||||
LocalAI can be installed inside Kubernetes with helm.
|
LocalAI can be installed inside Kubernetes with helm.
|
||||||
|
|
||||||
<details>
|
<details>
|
||||||
|
By default, the helm chart will install LocalAI instance using the ggml-gpt4all-j model without persistent storage.
|
||||||
|
|
||||||
1. Add the helm repo
|
1. Add the helm repo
|
||||||
```bash
|
```bash
|
||||||
helm repo add go-skynet https://go-skynet.github.io/helm-charts/
|
helm repo add go-skynet https://go-skynet.github.io/helm-charts/
|
||||||
```
|
```
|
||||||
1. Create a values files with your settings:
|
2. Install the helm chart:
|
||||||
```bash
|
```bash
|
||||||
cat <<EOF > values.yaml
|
helm repo update
|
||||||
|
helm install local-ai go-skynet/local-ai -f values.yaml
|
||||||
|
```
|
||||||
|
> **Note:** For further configuration options, see the [helm chart repository on GitHub](https://github.com/go-skynet/helm-charts).
|
||||||
|
### Example values
|
||||||
|
Deploy a single LocalAI pod with 6GB of persistent storage serving up a `ggml-gpt4all-j` model with custom prompt.
|
||||||
|
```yaml
|
||||||
|
### values.yaml
|
||||||
|
|
||||||
deployment:
|
deployment:
|
||||||
image: quay.io/go-skynet/local-ai:latest
|
# Adjust the number of threads and context size for model inference
|
||||||
env:
|
env:
|
||||||
threads: 4
|
threads: 14
|
||||||
contextSize: 1024
|
contextSize: 512
|
||||||
modelsPath: "/models"
|
|
||||||
# Optionally create a PVC, mount the PV to the LocalAI Deployment,
|
# Set the pod requests/limits
|
||||||
# and download a model to prepopulate the models directory
|
resources:
|
||||||
modelsVolume:
|
limits:
|
||||||
enabled: true
|
cpu: 4000m
|
||||||
url: "https://gpt4all.io/models/ggml-gpt4all-j.bin"
|
memory: 7000Mi
|
||||||
pvc:
|
requests:
|
||||||
|
cpu: 100m
|
||||||
|
memory: 6000Mi
|
||||||
|
|
||||||
|
# Add a custom prompt template for the ggml-gpt4all-j model
|
||||||
|
promptTemplates:
|
||||||
|
# The name of the model this template belongs to
|
||||||
|
ggml-gpt4all-j.bin.tmpl: |
|
||||||
|
This is my custom prompt template...
|
||||||
|
### Prompt:
|
||||||
|
{{.Input}}
|
||||||
|
### Response:
|
||||||
|
|
||||||
|
# Model configuration
|
||||||
|
models:
|
||||||
|
# Don't re-download models on pod creation
|
||||||
|
forceDownload: false
|
||||||
|
|
||||||
|
# List of models to download and serve
|
||||||
|
list:
|
||||||
|
- url: "https://gpt4all.io/models/ggml-gpt4all-j.bin"
|
||||||
|
# Optional basic HTTP authentication
|
||||||
|
basicAuth: base64EncodedCredentials
|
||||||
|
|
||||||
|
# Enable 6Gb of persistent storage models and prompt templates
|
||||||
|
persistence:
|
||||||
|
enabled: true
|
||||||
size: 6Gi
|
size: 6Gi
|
||||||
accessModes:
|
|
||||||
- ReadWriteOnce
|
|
||||||
auth:
|
|
||||||
# Optional value for HTTP basic access authentication header
|
|
||||||
basic: "" # 'username:password' base64 encoded
|
|
||||||
service:
|
service:
|
||||||
type: ClusterIP
|
type: ClusterIP
|
||||||
annotations: {}
|
annotations: {}
|
||||||
# If using an AWS load balancer, you'll need to override the default 60s load balancer idle timeout
|
# If using an AWS load balancer, you'll need to override the default 60s load balancer idle timeout
|
||||||
# service.beta.kubernetes.io/aws-load-balancer-connection-idle-timeout: "1200"
|
# service.beta.kubernetes.io/aws-load-balancer-connection-idle-timeout: "1200"
|
||||||
EOF
|
|
||||||
```
|
```
|
||||||
3. Install the helm chart:
|
|
||||||
```bash
|
|
||||||
helm repo update
|
|
||||||
helm install local-ai go-skynet/local-ai -f values.yaml
|
|
||||||
```
|
|
||||||
|
|
||||||
Check out also the [helm chart repository on GitHub](https://github.com/go-skynet/helm-charts).
|
|
||||||
|
|
||||||
</details>
|
</details>
|
||||||
|
|
||||||
## Supported OpenAI API endpoints
|
## Supported OpenAI API endpoints
|
||||||
|
Loading…
Reference in New Issue
Block a user