LocalAI/pkg/model/loader.go

167 lines
4.0 KiB
Go
Raw Normal View History

2023-04-11 22:02:39 +00:00
package model
import (
2023-04-08 08:46:51 +00:00
"bytes"
"context"
"fmt"
2023-04-10 10:02:40 +00:00
"io/ioutil"
"os"
"path/filepath"
2023-04-10 10:02:40 +00:00
"strings"
"sync"
2023-04-08 08:46:51 +00:00
"text/template"
"github.com/go-skynet/LocalAI/pkg/grpc"
process "github.com/mudler/go-processmanager"
"github.com/rs/zerolog/log"
)
type ModelLoader struct {
ModelPath string
mu sync.Mutex
2023-05-10 13:20:21 +00:00
// TODO: this needs generics
models map[string]*grpc.Client
grpcProcesses map[string]*process.Process
2023-05-11 14:34:16 +00:00
promptsTemplates map[string]*template.Template
}
func NewModelLoader(modelPath string) *ModelLoader {
return &ModelLoader{
2023-05-11 14:34:16 +00:00
ModelPath: modelPath,
models: make(map[string]*grpc.Client),
2023-05-11 14:34:16 +00:00
promptsTemplates: make(map[string]*template.Template),
grpcProcesses: make(map[string]*process.Process),
}
}
2023-04-20 16:33:02 +00:00
func (ml *ModelLoader) ExistsInModelPath(s string) bool {
_, err := os.Stat(filepath.Join(ml.ModelPath, s))
2023-04-20 16:33:02 +00:00
return err == nil
}
2023-04-10 10:02:40 +00:00
func (ml *ModelLoader) ListModels() ([]string, error) {
files, err := ioutil.ReadDir(ml.ModelPath)
2023-04-10 10:02:40 +00:00
if err != nil {
return []string{}, err
}
models := []string{}
for _, file := range files {
2023-04-20 16:33:02 +00:00
// Skip templates, YAML and .keep files
if strings.HasSuffix(file.Name(), ".tmpl") || strings.HasSuffix(file.Name(), ".keep") || strings.HasSuffix(file.Name(), ".yaml") || strings.HasSuffix(file.Name(), ".yml") {
continue
2023-04-10 10:02:40 +00:00
}
2023-04-20 16:33:02 +00:00
models = append(models, file.Name())
2023-04-10 10:02:40 +00:00
}
return models, nil
}
2023-04-08 08:46:51 +00:00
func (ml *ModelLoader) TemplatePrefix(modelName string, in interface{}) (string, error) {
ml.mu.Lock()
defer ml.mu.Unlock()
m, ok := ml.promptsTemplates[modelName]
if !ok {
modelFile := filepath.Join(ml.ModelPath, modelName)
if err := ml.loadTemplateIfExists(modelName, modelFile); err != nil {
return "", err
}
t, exists := ml.promptsTemplates[modelName]
if exists {
m = t
}
}
if m == nil {
return "", fmt.Errorf("failed loading any template")
2023-04-08 08:46:51 +00:00
}
var buf bytes.Buffer
if err := m.Execute(&buf, in); err != nil {
return "", err
}
return buf.String(), nil
}
2023-04-20 16:33:02 +00:00
func (ml *ModelLoader) loadTemplateIfExists(modelName, modelFile string) error {
// Check if the template was already loaded
if _, ok := ml.promptsTemplates[modelName]; ok {
return nil
}
// Check if the model path exists
// skip any error here - we run anyway if a template does not exist
2023-04-20 16:33:02 +00:00
modelTemplateFile := fmt.Sprintf("%s.tmpl", modelName)
if !ml.ExistsInModelPath(modelTemplateFile) {
return nil
}
dat, err := os.ReadFile(filepath.Join(ml.ModelPath, modelTemplateFile))
if err != nil {
return err
}
// Parse the template
tmpl, err := template.New("prompt").Parse(string(dat))
if err != nil {
return err
}
ml.promptsTemplates[modelName] = tmpl
return nil
}
func (ml *ModelLoader) LoadModel(modelName string, loader func(string) (*grpc.Client, error)) (*grpc.Client, error) {
ml.mu.Lock()
defer ml.mu.Unlock()
// Check if we already have a loaded model
if model := ml.checkIsLoaded(modelName); model != nil {
2023-04-20 16:33:02 +00:00
log.Debug().Msgf("Model already loaded in memory: %s", modelName)
return model, nil
}
2023-04-20 16:33:02 +00:00
// Load the model and keep it in memory for later use
modelFile := filepath.Join(ml.ModelPath, modelName)
2023-04-20 16:33:02 +00:00
log.Debug().Msgf("Loading model in memory from file: %s", modelFile)
2023-05-11 14:34:16 +00:00
model, err := loader(modelFile)
if err != nil {
return nil, err
}
// If there is a prompt template, load it
2023-04-20 16:33:02 +00:00
if err := ml.loadTemplateIfExists(modelName, modelFile); err != nil {
return nil, err
2023-04-08 08:46:51 +00:00
}
2023-04-20 16:33:02 +00:00
ml.models[modelName] = model
2023-05-11 14:34:16 +00:00
return model, nil
}
func (ml *ModelLoader) checkIsLoaded(s string) *grpc.Client {
if m, ok := ml.models[s]; ok {
log.Debug().Msgf("Model already loaded in memory: %s", s)
if !m.HealthCheck(context.Background()) {
log.Debug().Msgf("GRPC Model not responding", s)
if !ml.grpcProcesses[s].IsAlive() {
log.Debug().Msgf("GRPC Process is not responding", s)
// stop and delete the process, this forces to re-load the model and re-create again the service
ml.grpcProcesses[s].Stop()
delete(ml.grpcProcesses, s)
delete(ml.models, s)
return nil
}
}
return m
}
return nil
}