mirror of
https://github.com/nomic-ai/gpt4all
synced 2024-11-06 09:20:33 +00:00
d3c9c47bd2
* WIP * Fix includes * Try to fix linking issues * Refinements * allow to load MPT and llama models too * cleanup, add example, add README
128 lines
2.6 KiB
Go
128 lines
2.6 KiB
Go
package gpt4all
|
|
|
|
type PredictOptions struct {
|
|
ContextSize, RepeatLastN, Tokens, TopK, Batch int
|
|
TopP, Temperature, ContextErase, RepeatPenalty float64
|
|
}
|
|
|
|
type PredictOption func(p *PredictOptions)
|
|
|
|
var DefaultOptions PredictOptions = PredictOptions{
|
|
Tokens: 200,
|
|
TopK: 10,
|
|
TopP: 0.90,
|
|
Temperature: 0.96,
|
|
Batch: 1,
|
|
ContextErase: 0.55,
|
|
ContextSize: 1024,
|
|
RepeatLastN: 10,
|
|
RepeatPenalty: 1.2,
|
|
}
|
|
|
|
var DefaultModelOptions ModelOptions = ModelOptions{
|
|
Threads: 4,
|
|
ModelType: GPTJType,
|
|
}
|
|
|
|
type ModelOptions struct {
|
|
Threads int
|
|
ModelType ModelType
|
|
}
|
|
type ModelOption func(p *ModelOptions)
|
|
|
|
type ModelType int
|
|
|
|
const (
|
|
LLaMAType ModelType = 0
|
|
GPTJType ModelType = iota
|
|
MPTType ModelType = iota
|
|
)
|
|
|
|
// SetTokens sets the number of tokens to generate.
|
|
func SetTokens(tokens int) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.Tokens = tokens
|
|
}
|
|
}
|
|
|
|
// SetTopK sets the value for top-K sampling.
|
|
func SetTopK(topk int) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.TopK = topk
|
|
}
|
|
}
|
|
|
|
// SetTopP sets the value for nucleus sampling.
|
|
func SetTopP(topp float64) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.TopP = topp
|
|
}
|
|
}
|
|
|
|
// SetRepeatPenalty sets the repeat penalty.
|
|
func SetRepeatPenalty(ce float64) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.RepeatPenalty = ce
|
|
}
|
|
}
|
|
|
|
// SetRepeatLastN sets the RepeatLastN.
|
|
func SetRepeatLastN(ce int) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.RepeatLastN = ce
|
|
}
|
|
}
|
|
|
|
// SetContextErase sets the context erase %.
|
|
func SetContextErase(ce float64) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.ContextErase = ce
|
|
}
|
|
}
|
|
|
|
// SetTemperature sets the temperature value for text generation.
|
|
func SetTemperature(temp float64) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.Temperature = temp
|
|
}
|
|
}
|
|
|
|
// SetBatch sets the batch size.
|
|
func SetBatch(size int) PredictOption {
|
|
return func(p *PredictOptions) {
|
|
p.Batch = size
|
|
}
|
|
}
|
|
|
|
// Create a new PredictOptions object with the given options.
|
|
func NewPredictOptions(opts ...PredictOption) PredictOptions {
|
|
p := DefaultOptions
|
|
for _, opt := range opts {
|
|
opt(&p)
|
|
}
|
|
return p
|
|
}
|
|
|
|
// SetThreads sets the number of threads to use for text generation.
|
|
func SetThreads(c int) ModelOption {
|
|
return func(p *ModelOptions) {
|
|
p.Threads = c
|
|
}
|
|
}
|
|
|
|
// SetModelType sets the model type.
|
|
func SetModelType(c ModelType) ModelOption {
|
|
return func(p *ModelOptions) {
|
|
p.ModelType = c
|
|
}
|
|
}
|
|
|
|
// Create a new PredictOptions object with the given options.
|
|
func NewModelOptions(opts ...ModelOption) ModelOptions {
|
|
p := DefaultModelOptions
|
|
for _, opt := range opts {
|
|
opt(&p)
|
|
}
|
|
return p
|
|
}
|