2023-07-14 23:19:43 +00:00
|
|
|
package bloomz
|
|
|
|
|
|
|
|
// This is a wrapper to statisfy the GRPC service interface
|
|
|
|
// It is meant to be used by the main executable that is the server for the specific backend type (falcon, gpt3, etc)
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
|
|
|
|
"github.com/go-skynet/LocalAI/pkg/grpc/base"
|
|
|
|
pb "github.com/go-skynet/LocalAI/pkg/grpc/proto"
|
2023-08-18 19:23:14 +00:00
|
|
|
"github.com/rs/zerolog/log"
|
2023-07-14 23:19:43 +00:00
|
|
|
|
|
|
|
"github.com/go-skynet/bloomz.cpp"
|
|
|
|
)
|
|
|
|
|
|
|
|
type LLM struct {
|
|
|
|
base.Base
|
|
|
|
|
|
|
|
bloomz *bloomz.Bloomz
|
|
|
|
}
|
|
|
|
|
|
|
|
func (llm *LLM) Load(opts *pb.ModelOptions) error {
|
2023-08-18 19:23:14 +00:00
|
|
|
if llm.Base.State != pb.StatusResponse_UNINITIALIZED {
|
|
|
|
log.Warn().Msgf("bloomz backend loading %s while already in state %s!", opts.Model, llm.Base.State.String())
|
|
|
|
}
|
|
|
|
|
|
|
|
llm.Base.Lock()
|
|
|
|
defer llm.Base.Unlock()
|
2023-08-07 20:39:10 +00:00
|
|
|
model, err := bloomz.New(opts.ModelFile)
|
2023-07-14 23:19:43 +00:00
|
|
|
llm.bloomz = model
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
func buildPredictOptions(opts *pb.PredictOptions) []bloomz.PredictOption {
|
|
|
|
predictOptions := []bloomz.PredictOption{
|
|
|
|
bloomz.SetTemperature(float64(opts.Temperature)),
|
|
|
|
bloomz.SetTopP(float64(opts.TopP)),
|
|
|
|
bloomz.SetTopK(int(opts.TopK)),
|
|
|
|
bloomz.SetTokens(int(opts.Tokens)),
|
|
|
|
bloomz.SetThreads(int(opts.Threads)),
|
|
|
|
}
|
|
|
|
|
|
|
|
if opts.Seed != 0 {
|
|
|
|
predictOptions = append(predictOptions, bloomz.SetSeed(int(opts.Seed)))
|
|
|
|
}
|
|
|
|
|
|
|
|
return predictOptions
|
|
|
|
}
|
|
|
|
|
|
|
|
func (llm *LLM) Predict(opts *pb.PredictOptions) (string, error) {
|
2023-08-18 19:23:14 +00:00
|
|
|
llm.Base.Lock()
|
|
|
|
defer llm.Base.Unlock()
|
|
|
|
|
2023-07-14 23:19:43 +00:00
|
|
|
return llm.bloomz.Predict(opts.Prompt, buildPredictOptions(opts)...)
|
|
|
|
}
|
|
|
|
|
|
|
|
// fallback to Predict
|
|
|
|
func (llm *LLM) PredictStream(opts *pb.PredictOptions, results chan string) error {
|
2023-08-18 19:23:14 +00:00
|
|
|
llm.Base.Lock()
|
|
|
|
|
2023-07-14 23:19:43 +00:00
|
|
|
go func() {
|
|
|
|
res, err := llm.bloomz.Predict(opts.Prompt, buildPredictOptions(opts)...)
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
fmt.Println("err: ", err)
|
|
|
|
}
|
|
|
|
results <- res
|
|
|
|
close(results)
|
2023-08-18 19:23:14 +00:00
|
|
|
llm.Base.Unlock()
|
2023-07-14 23:19:43 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|