mirror of
https://github.com/CherryHQ/cherry-studio.git
synced 2025-12-27 12:51:26 +08:00
feat: Update message handling to include optional uiMessages parameter
- Modify BaseParams type to make uiMessages optional. - Refactor message preparation in HomeWindow and ActionUtils to handle modelMessages and uiMessages separately. - Ensure compatibility with updated message structures in fetchChatCompletion calls.
This commit is contained in:
parent
10301a1f79
commit
f8a22b5e55
@ -1307,7 +1307,7 @@ type BaseParams = {
|
||||
options?: FetchChatCompletionOptions
|
||||
onChunkReceived: (chunk: Chunk) => void
|
||||
topicId?: string // 添加 topicId 参数
|
||||
uiMessages: Message[]
|
||||
uiMessages?: Message[]
|
||||
}
|
||||
|
||||
type MessagesParams = BaseParams & {
|
||||
|
||||
@ -266,13 +266,17 @@ const HomeWindow: FC<{ draggable?: boolean }> = ({ draggable = true }) => {
|
||||
newAssistant.webSearchProviderId = undefined
|
||||
newAssistant.mcpServers = undefined
|
||||
newAssistant.knowledge_bases = undefined
|
||||
const llmMessages = await ConversationService.prepareMessagesForModel(messagesForContext, newAssistant)
|
||||
const { modelMessages, uiMessages } = await ConversationService.prepareMessagesForModel(
|
||||
messagesForContext,
|
||||
newAssistant
|
||||
)
|
||||
|
||||
await fetchChatCompletion({
|
||||
messages: llmMessages,
|
||||
messages: modelMessages,
|
||||
assistant: newAssistant,
|
||||
options: {},
|
||||
topicId,
|
||||
uiMessages: uiMessages,
|
||||
onChunkReceived: (chunk: Chunk) => {
|
||||
switch (chunk.type) {
|
||||
case ChunkType.THINKING_START:
|
||||
|
||||
@ -64,12 +64,13 @@ export const processMessages = async (
|
||||
newAssistant.webSearchProviderId = undefined
|
||||
newAssistant.mcpServers = undefined
|
||||
newAssistant.knowledge_bases = undefined
|
||||
const llmMessages = await ConversationService.prepareMessagesForModel([userMessage], newAssistant)
|
||||
const { modelMessages, uiMessages } = await ConversationService.prepareMessagesForModel([userMessage], newAssistant)
|
||||
|
||||
await fetchChatCompletion({
|
||||
messages: llmMessages,
|
||||
messages: modelMessages,
|
||||
assistant: newAssistant,
|
||||
options: {},
|
||||
uiMessages: uiMessages,
|
||||
onChunkReceived: (chunk: Chunk) => {
|
||||
if (finished) {
|
||||
return
|
||||
|
||||
Loading…
Reference in New Issue
Block a user