mirror of
https://github.com/casdoor/casdoor.git
synced 2025-05-23 02:35:49 +08:00
117 lines
2.6 KiB
Go
117 lines
2.6 KiB
Go
// Copyright 2023 The Casdoor Authors. All Rights Reserved.
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package ai
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/sashabaranov/go-openai"
|
|
)
|
|
|
|
func queryAnswer(authToken string, question string, timeout int) (string, error) {
|
|
//fmt.Printf("Question: %s\n", question)
|
|
|
|
client := getProxyClientFromToken(authToken)
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Duration(2+timeout*2)*time.Second)
|
|
defer cancel()
|
|
|
|
resp, err := client.CreateChatCompletion(
|
|
ctx,
|
|
openai.ChatCompletionRequest{
|
|
Model: openai.GPT3Dot5Turbo,
|
|
Messages: []openai.ChatCompletionMessage{
|
|
{
|
|
Role: openai.ChatMessageRoleUser,
|
|
Content: question,
|
|
},
|
|
},
|
|
},
|
|
)
|
|
if err != nil {
|
|
//fmt.Printf("%s\n", err.Error())
|
|
return "", err
|
|
}
|
|
|
|
res := resp.Choices[0].Message.Content
|
|
res = strings.Trim(res, "\n")
|
|
//fmt.Printf("Answer: %s\n\n", res)
|
|
return res, nil
|
|
}
|
|
|
|
func QueryAnswerSafe(authToken string, question string) string {
|
|
var res string
|
|
var err error
|
|
for i := 0; i < 10; i++ {
|
|
res, err = queryAnswer(authToken, question, i)
|
|
if err != nil {
|
|
if i > 0 {
|
|
fmt.Printf("\tFailed (%d): %s\n", i+1, err.Error())
|
|
}
|
|
} else {
|
|
break
|
|
}
|
|
}
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
return res
|
|
}
|
|
|
|
func QueryAnswerStream(authToken string, question string, writer io.Writer, builder *strings.Builder) error {
|
|
client := getProxyClientFromToken(authToken)
|
|
|
|
ctx := context.Background()
|
|
|
|
respStream, err := client.CreateCompletionStream(
|
|
ctx,
|
|
openai.CompletionRequest{
|
|
Model: openai.GPT3TextDavinci003,
|
|
Prompt: question,
|
|
MaxTokens: 50,
|
|
Stream: true,
|
|
},
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
defer respStream.Close()
|
|
|
|
for {
|
|
completion, streamErr := respStream.Recv()
|
|
if streamErr != nil {
|
|
if streamErr == io.EOF {
|
|
break
|
|
}
|
|
return streamErr
|
|
}
|
|
|
|
// Write the streamed data as Server-Sent Events
|
|
if _, err := fmt.Fprintf(writer, "data: %s\n\n", completion.Choices[0].Text); err != nil {
|
|
return err
|
|
}
|
|
|
|
// Append the response to the strings.Builder
|
|
builder.WriteString(completion.Choices[0].Text)
|
|
}
|
|
|
|
return nil
|
|
}
|