Spaces:
Paused
Paused
matt HOFFNER
commited on
Commit
·
eef247f
1
Parent(s):
644a95f
SSE get request
Browse files- src/app/search/web/page.jsx +28 -39
- src/pages/api/llm.js +10 -8
src/app/search/web/page.jsx
CHANGED
@@ -1,55 +1,44 @@
|
|
1 |
-
"use client"
|
2 |
import { useEffect, useState } from "react";
|
3 |
|
4 |
export default function WebSearchPage({ searchParams }) {
|
5 |
const [aiResponse, setAiResponse] = useState("");
|
|
|
6 |
|
7 |
useEffect(() => {
|
8 |
if (!searchParams || !searchParams.searchTerm) return;
|
9 |
|
10 |
const { searchTerm, start = "1" } = searchParams;
|
11 |
const url = new URL('/api/llm', window.location.origin);
|
|
|
12 |
|
13 |
-
fetch
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
20 |
-
startIndex: start
|
21 |
-
})
|
22 |
-
})
|
23 |
-
.then(response => {
|
24 |
-
if (!response.ok) {
|
25 |
-
throw new Error("HTTP error " + response.status);
|
26 |
-
}
|
27 |
-
// Create a reader to stream the response body
|
28 |
-
const reader = response.body.getReader();
|
29 |
-
|
30 |
-
// Read and process the response body chunks
|
31 |
-
return new ReadableStream({
|
32 |
-
start(controller) {
|
33 |
-
function push() {
|
34 |
-
reader.read().then(({ done, value }) => {
|
35 |
-
if (done) {
|
36 |
-
// Close the stream when done
|
37 |
-
controller.close();
|
38 |
-
return;
|
39 |
-
}
|
40 |
-
// Decode the response chunk and append it to the existing response
|
41 |
-
setAiResponse(prev => prev + new TextDecoder().decode(value));
|
42 |
-
// Get the next chunk
|
43 |
-
push();
|
44 |
-
});
|
45 |
-
}
|
46 |
-
push();
|
47 |
-
}
|
48 |
-
});
|
49 |
-
})
|
50 |
-
.catch(console.error);
|
51 |
}, [searchParams]);
|
52 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
console.log(aiResponse);
|
54 |
|
55 |
return <>{aiResponse}</>;
|
|
|
|
|
1 |
import { useEffect, useState } from "react";
|
2 |
|
3 |
export default function WebSearchPage({ searchParams }) {
|
4 |
const [aiResponse, setAiResponse] = useState("");
|
5 |
+
const [eventSource, setEventSource] = useState(null);
|
6 |
|
7 |
useEffect(() => {
|
8 |
if (!searchParams || !searchParams.searchTerm) return;
|
9 |
|
10 |
const { searchTerm, start = "1" } = searchParams;
|
11 |
const url = new URL('/api/llm', window.location.origin);
|
12 |
+
url.search = new URLSearchParams({ question: searchTerm, startIndex: start });
|
13 |
|
14 |
+
// No need to make a fetch request. Directly open the EventSource connection.
|
15 |
+
const es = new EventSource(url);
|
16 |
+
setEventSource(es);
|
17 |
+
|
18 |
+
return () => {
|
19 |
+
if (es) es.close(); // Close the EventSource when the component is unmounted.
|
20 |
+
};
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
}, [searchParams]);
|
22 |
|
23 |
+
// Add event listener for the EventSource
|
24 |
+
useEffect(() => {
|
25 |
+
if (eventSource) {
|
26 |
+
eventSource.onmessage = (event) => {
|
27 |
+
setAiResponse(prev => prev + event.data);
|
28 |
+
};
|
29 |
+
|
30 |
+
eventSource.onerror = (event) => {
|
31 |
+
console.error("EventSource failed:", event);
|
32 |
+
};
|
33 |
+
}
|
34 |
+
|
35 |
+
return () => {
|
36 |
+
if (eventSource) {
|
37 |
+
eventSource.close();
|
38 |
+
}
|
39 |
+
};
|
40 |
+
}, [eventSource]);
|
41 |
+
|
42 |
console.log(aiResponse);
|
43 |
|
44 |
return <>{aiResponse}</>;
|
src/pages/api/llm.js
CHANGED
@@ -2,10 +2,15 @@ import { Configuration, OpenAIApi } from "openai";
|
|
2 |
import { GoogleCustomSearch } from "openai-function-calling-tools";
|
3 |
|
4 |
export default function handler(req, res) {
|
|
|
|
|
|
|
|
|
|
|
5 |
const QUESTION = req.body.question;
|
6 |
|
7 |
if (!QUESTION) {
|
8 |
-
res.status(400).send({ error: 'Question is missing in request' });
|
9 |
return;
|
10 |
}
|
11 |
|
@@ -15,10 +20,6 @@ export default function handler(req, res) {
|
|
15 |
const openai = new OpenAIApi(configuration);
|
16 |
|
17 |
const messages = [
|
18 |
-
{
|
19 |
-
role: "system",
|
20 |
-
content: "You are a helpful assistant.",
|
21 |
-
},
|
22 |
{
|
23 |
role: "user",
|
24 |
content: QUESTION,
|
@@ -38,8 +39,9 @@ export default function handler(req, res) {
|
|
38 |
const response = await openai.createChatCompletion({
|
39 |
model: "gpt-3.5-turbo-0613",
|
40 |
messages,
|
41 |
-
functions: [
|
42 |
temperature: 0,
|
|
|
43 |
});
|
44 |
|
45 |
return response;
|
@@ -63,7 +65,7 @@ export default function handler(req, res) {
|
|
63 |
const args = response.data.choices[0].message.function_call.arguments;
|
64 |
|
65 |
const fn = functions[fnName];
|
66 |
-
const result = await fn.
|
67 |
|
68 |
messages.push({
|
69 |
role: "assistant",
|
@@ -87,4 +89,4 @@ export default function handler(req, res) {
|
|
87 |
console.error(err);
|
88 |
res.status(500).send({ error: 'Internal Server Error' });
|
89 |
});
|
90 |
-
}
|
|
|
2 |
import { GoogleCustomSearch } from "openai-function-calling-tools";
|
3 |
|
4 |
export default function handler(req, res) {
|
5 |
+
if (req.method !== 'POST') {
|
6 |
+
res.status(405).send({ error: 'Method Not Allowed', method: req.method });
|
7 |
+
return;
|
8 |
+
}
|
9 |
+
|
10 |
const QUESTION = req.body.question;
|
11 |
|
12 |
if (!QUESTION) {
|
13 |
+
res.status(400).send({ error: 'Question is missing in request body' });
|
14 |
return;
|
15 |
}
|
16 |
|
|
|
20 |
const openai = new OpenAIApi(configuration);
|
21 |
|
22 |
const messages = [
|
|
|
|
|
|
|
|
|
23 |
{
|
24 |
role: "user",
|
25 |
content: QUESTION,
|
|
|
39 |
const response = await openai.createChatCompletion({
|
40 |
model: "gpt-3.5-turbo-0613",
|
41 |
messages,
|
42 |
+
functions: [googleCustomSearchSchema],
|
43 |
temperature: 0,
|
44 |
+
stream: true
|
45 |
});
|
46 |
|
47 |
return response;
|
|
|
65 |
const args = response.data.choices[0].message.function_call.arguments;
|
66 |
|
67 |
const fn = functions[fnName];
|
68 |
+
const result = await fn(...Object.values(JSON.parse(args)));
|
69 |
|
70 |
messages.push({
|
71 |
role: "assistant",
|
|
|
89 |
console.error(err);
|
90 |
res.status(500).send({ error: 'Internal Server Error' });
|
91 |
});
|
92 |
+
}
|