Ollama
All about Ollama.
package main
import (
"context"
"fmt"
"log"
"github.com/hupe1980/golc/integration/ollama"
"github.com/hupe1980/golc/model"
"github.com/hupe1980/golc/model/llm"
"github.com/hupe1980/golc/prompt"
)
// Start ollama
// docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
// docker exec -it ollama ollama run llama2
func main() {
client := ollama.New("http://localhost:11434")
llm, err := llm.NewOllama(client, func(o *llm.OllamaOptions) {
o.ModelName = "llama2"
})
if err != nil {
log.Fatal(err)
}
res, err := model.GeneratePrompt(context.Background(), llm, prompt.StringPromptValue("Hello llama2!"))
if err != nil {
log.Fatal(err)
}
fmt.Println(res.Generations[0].Text)
}
Last modified December 21, 2023: Refactor amazon bedrock support (120f6a1)