2023-03-27 18:27:33 +08:00
|
|
|
package utils
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"github.com/BurntSushi/toml"
|
|
|
|
"net/http"
|
|
|
|
logger2 "openai/logger"
|
|
|
|
"openai/types"
|
|
|
|
"os"
|
|
|
|
)
|
|
|
|
|
|
|
|
func NewDefaultConfig() *types.Config {
|
|
|
|
return &types.Config{
|
2023-03-28 10:17:36 +08:00
|
|
|
Listen: "0.0.0.0:5678",
|
|
|
|
ProxyURL: make([]string, 0),
|
|
|
|
EnableAuth: true,
|
|
|
|
AccessKey: "yangjian102621@gmail.com",
|
2023-03-27 18:27:33 +08:00
|
|
|
|
|
|
|
Session: types.Session{
|
|
|
|
SecretKey: RandString(64),
|
|
|
|
Name: "CHAT_SESSION_ID",
|
|
|
|
Domain: "",
|
|
|
|
Path: "/",
|
|
|
|
MaxAge: 86400,
|
|
|
|
Secure: true,
|
|
|
|
HttpOnly: false,
|
|
|
|
SameSite: http.SameSiteLaxMode,
|
|
|
|
},
|
|
|
|
Chat: types.Chat{
|
|
|
|
ApiURL: "https://api.openai.com/v1/chat/completions",
|
|
|
|
ApiKeys: []string{""},
|
|
|
|
Model: "gpt-3.5-turbo",
|
|
|
|
MaxTokens: 1024,
|
|
|
|
Temperature: 0.9,
|
|
|
|
EnableContext: true,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
var logger = logger2.GetLogger()
|
|
|
|
|
|
|
|
func LoadConfig(configFile string) (*types.Config, error) {
|
|
|
|
var config *types.Config
|
|
|
|
_, err := os.Stat(configFile)
|
|
|
|
if err != nil {
|
|
|
|
logger.Errorf("Error open config file: %s", err.Error())
|
|
|
|
config = NewDefaultConfig()
|
|
|
|
// save config
|
|
|
|
err := SaveConfig(config, configFile)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return config, nil
|
|
|
|
}
|
|
|
|
_, err = toml.DecodeFile(configFile, &config)
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return config, err
|
|
|
|
}
|
|
|
|
|
|
|
|
func SaveConfig(config *types.Config, configFile string) error {
|
|
|
|
buf := new(bytes.Buffer)
|
|
|
|
encoder := toml.NewEncoder(buf)
|
|
|
|
if err := encoder.Encode(&config); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
return os.WriteFile(configFile, buf.Bytes(), 0644)
|
|
|
|
}
|