Merge pull request #3205 from H0llyW00dzZ/summarizelogic
Refactor Summarize Logic
This commit is contained in:
commit
e38b527ac2
|
@ -557,6 +557,7 @@ export const useChatStore = createPersistStore(
|
|||
messages: topicMessages,
|
||||
config: {
|
||||
model: getSummarizeModel(session.mask.modelConfig.model),
|
||||
stream: false,
|
||||
},
|
||||
onFinish(message) {
|
||||
get().updateCurrentSession(
|
||||
|
@ -600,6 +601,10 @@ export const useChatStore = createPersistStore(
|
|||
historyMsgLength > modelConfig.compressMessageLengthThreshold &&
|
||||
modelConfig.sendMemory
|
||||
) {
|
||||
/** Destruct max_tokens while summarizing
|
||||
* this param is just shit
|
||||
**/
|
||||
const { max_tokens, ...modelcfg } = modelConfig;
|
||||
api.llm.chat({
|
||||
messages: toBeSummarizedMsgs.concat(
|
||||
createMessage({
|
||||
|
@ -609,7 +614,7 @@ export const useChatStore = createPersistStore(
|
|||
}),
|
||||
),
|
||||
config: {
|
||||
...modelConfig,
|
||||
...modelcfg,
|
||||
stream: true,
|
||||
model: getSummarizeModel(session.mask.modelConfig.model),
|
||||
},
|
||||
|
|
Loading…
Reference in New Issue