2023-04-07 09:30:59 +00:00
|
|
|
package main
|
|
|
|
|
|
|
|
import (
|
2023-04-08 08:46:51 +00:00
|
|
|
"bytes"
|
2023-04-07 09:30:59 +00:00
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"path/filepath"
|
|
|
|
"sync"
|
2023-04-08 08:46:51 +00:00
|
|
|
"text/template"
|
2023-04-07 09:30:59 +00:00
|
|
|
|
|
|
|
llama "github.com/go-skynet/go-llama.cpp"
|
|
|
|
)
|
|
|
|
|
|
|
|
type ModelLoader struct {
|
2023-04-08 08:46:51 +00:00
|
|
|
modelPath string
|
|
|
|
mu sync.Mutex
|
|
|
|
models map[string]*llama.LLama
|
|
|
|
promptsTemplates map[string]*template.Template
|
2023-04-07 09:30:59 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func NewModelLoader(modelPath string) *ModelLoader {
|
2023-04-08 08:46:51 +00:00
|
|
|
return &ModelLoader{modelPath: modelPath, models: make(map[string]*llama.LLama), promptsTemplates: make(map[string]*template.Template)}
|
2023-04-07 09:30:59 +00:00
|
|
|
}
|
|
|
|
|
2023-04-08 08:46:51 +00:00
|
|
|
func (ml *ModelLoader) TemplatePrefix(modelName string, in interface{}) (string, error) {
|
|
|
|
ml.mu.Lock()
|
|
|
|
defer ml.mu.Unlock()
|
|
|
|
|
|
|
|
m, ok := ml.promptsTemplates[modelName]
|
|
|
|
if !ok {
|
|
|
|
// try to find a s.bin
|
|
|
|
modelBin := fmt.Sprintf("%s.bin", modelName)
|
|
|
|
m, ok = ml.promptsTemplates[modelBin]
|
|
|
|
if !ok {
|
|
|
|
return "", fmt.Errorf("no prompt template available")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var buf bytes.Buffer
|
|
|
|
|
|
|
|
if err := m.Execute(&buf, in); err != nil {
|
|
|
|
return "", err
|
|
|
|
}
|
|
|
|
return buf.String(), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (ml *ModelLoader) LoadModel(modelName string, opts ...llama.ModelOption) (*llama.LLama, error) {
|
2023-04-07 09:30:59 +00:00
|
|
|
ml.mu.Lock()
|
|
|
|
defer ml.mu.Unlock()
|
|
|
|
|
|
|
|
// Check if we already have a loaded model
|
2023-04-08 08:46:51 +00:00
|
|
|
modelFile := filepath.Join(ml.modelPath, modelName)
|
2023-04-07 09:30:59 +00:00
|
|
|
|
|
|
|
if m, ok := ml.models[modelFile]; ok {
|
|
|
|
return m, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Check if the model path exists
|
|
|
|
if _, err := os.Stat(modelFile); os.IsNotExist(err) {
|
|
|
|
// try to find a s.bin
|
|
|
|
modelBin := fmt.Sprintf("%s.bin", modelFile)
|
|
|
|
if _, err := os.Stat(modelBin); os.IsNotExist(err) {
|
|
|
|
return nil, err
|
|
|
|
} else {
|
|
|
|
modelFile = modelBin
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Load the model and keep it in memory for later use
|
|
|
|
model, err := llama.New(modelFile, opts...)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
2023-04-08 08:46:51 +00:00
|
|
|
// If there is a prompt template, load it
|
|
|
|
|
|
|
|
modelTemplateFile := fmt.Sprintf("%s.tmpl", modelFile)
|
|
|
|
// Check if the model path exists
|
|
|
|
if _, err := os.Stat(modelTemplateFile); err == nil {
|
|
|
|
dat, err := os.ReadFile(modelTemplateFile)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Parse the template
|
|
|
|
tmpl, err := template.New("prompt").Parse(string(dat))
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
ml.promptsTemplates[modelFile] = tmpl
|
|
|
|
}
|
|
|
|
|
2023-04-07 09:30:59 +00:00
|
|
|
ml.models[modelFile] = model
|
|
|
|
return model, err
|
|
|
|
}
|