@@ -16,13 +16,13 @@ export const createClient = () => {
16
16
17
17
const nodeToStep = ( node : string ) => {
18
18
switch ( node ) {
19
- case "route_at_start_node " :
19
+ case "analyze_and_route_query " :
20
20
return 0 ;
21
- case "generate_questions " :
21
+ case "create_research_plan " :
22
22
return 1 ;
23
- case "research_node " :
23
+ case "conduct_research " :
24
24
return 2 ;
25
- case "generate " :
25
+ case "respond " :
26
26
return 3 ;
27
27
default :
28
28
return 0 ;
@@ -86,17 +86,17 @@ export function useGraph() {
86
86
} ;
87
87
88
88
const getOrCreateThread = async ( ) => {
89
- const assistantIdCookie = getCookie ( "oc_assistant_id " ) ;
89
+ const assistantIdCookie = getCookie ( "clc_py_assistant_id " ) ;
90
90
if ( assistantIdCookie ) {
91
91
setAssistantId ( assistantIdCookie ) ;
92
92
return ;
93
93
}
94
94
const client = createClient ( ) ;
95
95
const assistant = await client . assistants . create ( {
96
- graphId : "agent " ,
96
+ graphId : "chat " ,
97
97
} ) ;
98
98
setAssistantId ( assistant . assistant_id ) ;
99
- setCookie ( "oc_assistant_id " , assistant . assistant_id ) ;
99
+ setCookie ( "clc_py_assistant_id " , assistant . assistant_id ) ;
100
100
} ;
101
101
102
102
const streamMessage = async ( params : GraphInput ) => {
@@ -198,10 +198,10 @@ export function useGraph() {
198
198
const node = chunk ?. data ?. metadata ?. langgraph_node ;
199
199
if (
200
200
[
201
- "route_at_start_node " ,
202
- "generate_questions " ,
203
- "research_node " ,
204
- "generate " ,
201
+ "analyze_and_route_query " ,
202
+ "create_research_plan " ,
203
+ "conduct_research " ,
204
+ "respond " ,
205
205
] . includes ( node )
206
206
) {
207
207
setMessages ( ( prevMessages ) => {
@@ -235,7 +235,7 @@ export function useGraph() {
235
235
} ) ;
236
236
}
237
237
238
- if ( node === "generate " ) {
238
+ if ( node === "respond " ) {
239
239
setMessages ( ( prevMessages ) => {
240
240
const selectedDocumentsAIMessage = new AIMessage ( {
241
241
content : "" ,
@@ -254,7 +254,7 @@ export function useGraph() {
254
254
}
255
255
256
256
if ( chunk . data . event === "on_chat_model_stream" ) {
257
- if ( chunk . data . metadata . langgraph_node === "route_at_start_node " ) {
257
+ if ( chunk . data . metadata . langgraph_node === "analyze_and_route_query " ) {
258
258
const message = chunk . data . data . chunk ;
259
259
const toolCallChunk = message . tool_call_chunks ?. [ 0 ] ;
260
260
fullRoutingStr += toolCallChunk ?. args || "" ;
@@ -302,7 +302,7 @@ export function useGraph() {
302
302
}
303
303
}
304
304
305
- if ( chunk . data . metadata . langgraph_node === "general " ) {
305
+ if ( chunk . data . metadata . langgraph_node === "respond_to_general_query " ) {
306
306
const message = chunk . data . data . chunk ;
307
307
setMessages ( ( prevMessages ) => {
308
308
const existingMessageIndex = prevMessages . findIndex (
@@ -329,7 +329,7 @@ export function useGraph() {
329
329
} ) ;
330
330
}
331
331
332
- if ( chunk . data . metadata . langgraph_node === "generate_questions " ) {
332
+ if ( chunk . data . metadata . langgraph_node === "create_research_plan " ) {
333
333
const message = chunk . data . data . chunk ;
334
334
generatingQuestionsMessageId = message . id ;
335
335
const toolCallChunk = message . tool_call_chunks ?. [ 0 ] ;
@@ -405,7 +405,7 @@ export function useGraph() {
405
405
}
406
406
}
407
407
408
- if ( chunk . data . metadata . langgraph_node === "generate " ) {
408
+ if ( chunk . data . metadata . langgraph_node === "respond " ) {
409
409
const message = chunk . data . data . chunk ;
410
410
setMessages ( ( prevMessages ) => {
411
411
const existingMessageIndex = prevMessages . findIndex (
@@ -444,10 +444,10 @@ export function useGraph() {
444
444
445
445
if ( chunk . data . event === "on_chain_end" ) {
446
446
if (
447
- chunk . data . metadata . langgraph_node === "research_node " &&
447
+ chunk . data . metadata . langgraph_node === "conduct_research " &&
448
448
chunk . data . data ?. output &&
449
449
typeof chunk . data . data . output === "object" &&
450
- "sub_question " in chunk . data . data . output
450
+ "question " in chunk . data . data . output
451
451
) {
452
452
setMessages ( ( prevMessages ) => {
453
453
const foundIndex = prevMessages . findIndex (
@@ -466,9 +466,7 @@ export function useGraph() {
466
466
if ( toolCall . name === "generating_questions" ) {
467
467
const updatedQuestions = toolCall . args . questions . map (
468
468
( q : any ) => {
469
- if (
470
- q . question === chunk . data . data . output . sub_question
471
- ) {
469
+ if ( q . question === chunk . data . data . output . question ) {
472
470
return {
473
471
...q ,
474
472
queries : chunk . data . data . output . queries ,
@@ -509,7 +507,7 @@ export function useGraph() {
509
507
}
510
508
511
509
if (
512
- [ "generate " , "general " , "more_info " ] . includes (
510
+ [ "respond " , "respond_to_general_query " , "ask_for_more_info " ] . includes (
513
511
chunk ?. data ?. metadata ?. langgraph_node ,
514
512
)
515
513
) {
@@ -544,14 +542,16 @@ export function useGraph() {
544
542
} ) ;
545
543
}
546
544
547
- if ( chunk . data . metadata . langgraph_node === "generate " ) {
545
+ if ( chunk . data . metadata . langgraph_node === "respond " ) {
548
546
const inputDocuments = chunk . data . data . input . documents ;
549
- const message = chunk . data . data . output . messages ;
547
+ const message = chunk . data . data . output . messages [ 0 ] ;
548
+ console . log ( "setting messages respond" , chunk ) ;
550
549
setMessages ( ( prevMessages ) => {
551
550
const existingMessageIndex = prevMessages . findIndex (
552
551
( pMsg ) => pMsg . id === message . id ,
553
552
) ;
554
553
if ( existingMessageIndex !== - 1 ) {
554
+ console . log ( "message.content" , message . content ) ;
555
555
const newMessageWithLinks = new AIMessage ( {
556
556
...message ,
557
557
content : addDocumentLinks ( message . content , inputDocuments ) ,
0 commit comments