mirror of
https://github.com/multipleof4/sune.git
synced 2026-01-14 00:27:56 +00:00
Update sw.js
This commit is contained in:
345
public/sw.js
345
public/sw.js
@@ -1,245 +1,178 @@
|
|||||||
// sw.js — debug-friendly service worker
|
// /sw.js
|
||||||
// Replace your existing worker with this (or merge the relevant parts).
|
// Enhanced: tracks whether streaming fetches are being teed and reports status on ping.
|
||||||
importScripts('https://cdn.jsdelivr.net/npm/localforage@1.10.0/dist/localforage.min.js');
|
// Drop this at the root: /sw.js
|
||||||
|
|
||||||
const THREADS_KEY = 'threads_v1';
|
const TARGET_SUBSTRING = 'openrouter.ai/api/v1/chat/completions'; // adjust if different
|
||||||
const TARGET_SUBSTRING = 'openrouter.ai/api/v1/chat/completions'; // keep your real target here
|
const STATE_TTL_MS = 24 * 60 * 60 * 1000;
|
||||||
const LOG = (...a) => { console.log('[sw-debug]', ...a) };
|
|
||||||
|
|
||||||
const gid = () => Math.random().toString(36).slice(2,9) + '-' + Date.now().toString(36);
|
const state = {
|
||||||
|
totalIntercepted: 0,
|
||||||
|
activeStreams: {}, // streamId -> { url, startedAt, bytes, lastProgressAt, status }
|
||||||
|
lastStreamSummary: null // summary of last finished stream
|
||||||
|
};
|
||||||
|
|
||||||
// ---------- small idb helpers using localforage ----------
|
const gid = () => Math.random().toString(36).slice(2, 9) + '-' + Date.now().toString(36);
|
||||||
async function readThreads() {
|
|
||||||
try { const v = await localforage.getItem(THREADS_KEY); return Array.isArray(v) ? v : []; }
|
|
||||||
catch (e) { LOG('readThreads err', e); return []; }
|
|
||||||
}
|
|
||||||
async function writeThreads(arr) {
|
|
||||||
try { await localforage.setItem(THREADS_KEY, arr); }
|
|
||||||
catch (e) { LOG('writeThreads err', e); throw e; }
|
|
||||||
}
|
|
||||||
function pickThread(threads) {
|
|
||||||
if (!threads || !threads.length) return null;
|
|
||||||
// prefer newest updatedAt
|
|
||||||
threads.sort((a,b) => (b.updatedAt||0) - (a.updatedAt||0));
|
|
||||||
return threads[0];
|
|
||||||
}
|
|
||||||
async function upsertStreamMessage(streamId, text, meta = {}) {
|
|
||||||
const threads = await readThreads();
|
|
||||||
let th = pickThread(threads);
|
|
||||||
const now = Date.now();
|
|
||||||
if (!th) {
|
|
||||||
th = { id: 'sw-' + gid(), title: 'Missed while backgrounded', pinned: false, updatedAt: now, messages: [] };
|
|
||||||
threads.unshift(th);
|
|
||||||
}
|
|
||||||
let msgIndex = -1;
|
|
||||||
for (let i = th.messages.length - 1; i >= 0; i--) {
|
|
||||||
if (th.messages[i] && th.messages[i].sw_streamId === streamId) { msgIndex = i; break; }
|
|
||||||
}
|
|
||||||
const contentParts = [{ type: 'text', text }];
|
|
||||||
if (msgIndex >= 0) {
|
|
||||||
const ex = th.messages[msgIndex];
|
|
||||||
ex.content = text;
|
|
||||||
ex.contentParts = contentParts;
|
|
||||||
ex.updatedAt = now;
|
|
||||||
ex._sw_lastSave = now;
|
|
||||||
ex._sw_meta = Object.assign({}, ex._sw_meta || {}, meta);
|
|
||||||
} else {
|
|
||||||
th.messages.push({
|
|
||||||
id: 'swmsg-' + gid(),
|
|
||||||
role: 'assistant',
|
|
||||||
content: text,
|
|
||||||
contentParts,
|
|
||||||
kind: 'assistant',
|
|
||||||
sw_saved: true,
|
|
||||||
sw_streamId: streamId,
|
|
||||||
createdAt: now,
|
|
||||||
updatedAt: now,
|
|
||||||
_sw_meta: meta
|
|
||||||
});
|
|
||||||
}
|
|
||||||
th.updatedAt = now;
|
|
||||||
await writeThreads(threads);
|
|
||||||
return { threadId: th.id };
|
|
||||||
}
|
|
||||||
async function finalizeStream(streamId, meta = {}) {
|
|
||||||
const threads = await readThreads();
|
|
||||||
const th = pickThread(threads);
|
|
||||||
if (!th) return;
|
|
||||||
for (let i = th.messages.length - 1; i >= 0; i--) {
|
|
||||||
const m = th.messages[i];
|
|
||||||
if (m && m.sw_streamId === streamId) {
|
|
||||||
m._sw_meta = Object.assign({}, m._sw_meta || {}, meta, { completeAt: Date.now() });
|
|
||||||
m.updatedAt = Date.now();
|
|
||||||
th.updatedAt = Date.now();
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
await writeThreads(threads);
|
|
||||||
// notify clients
|
|
||||||
const clientsList = await self.clients.matchAll({ includeUncontrolled: true, type: 'window' });
|
|
||||||
for (const c of clientsList) {
|
|
||||||
try { c.postMessage({ type: 'stream-saved', streamId, meta }); } catch(e) {}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// ---------- helpers ----------
|
async function broadcast(msg) {
|
||||||
async function listSwStreams() {
|
|
||||||
const threads = await readThreads();
|
|
||||||
const found = [];
|
|
||||||
for (const t of (threads || [])) {
|
|
||||||
for (const m of (t.messages || [])) {
|
|
||||||
if (m && m.sw_streamId) found.push({
|
|
||||||
threadId: t.id,
|
|
||||||
threadTitle: t.title,
|
|
||||||
messageId: m.id,
|
|
||||||
sw_streamId: m.sw_streamId,
|
|
||||||
snippet: (m.content||'').slice(0,200),
|
|
||||||
updatedAt: m.updatedAt
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return found;
|
|
||||||
}
|
|
||||||
async function notifyClients(msg) {
|
|
||||||
try {
|
try {
|
||||||
const list = await self.clients.matchAll({ includeUncontrolled: true, type: 'window' });
|
const clientsList = await self.clients.matchAll({ includeUncontrolled: true, type: 'window' });
|
||||||
for (const c of list) {
|
for (const c of clientsList) {
|
||||||
try { c.postMessage(msg); } catch(e) {}
|
try { c.postMessage(msg); } catch (e) { /* ignore client errors */ }
|
||||||
}
|
}
|
||||||
} catch(e) { LOG('notifyClients err', e); }
|
} catch(e) { /* ignore */ }
|
||||||
}
|
}
|
||||||
|
|
||||||
// ---------- install/activate ----------
|
function cleanupState() {
|
||||||
self.addEventListener('install', e => { self.skipWaiting(); });
|
const now = Date.now();
|
||||||
self.addEventListener('activate', e => { e.waitUntil(self.clients.claim()); });
|
for (const k of Object.keys(state.activeStreams)) {
|
||||||
|
if ((now - (state.activeStreams[k].lastProgressAt || state.activeStreams[k].startedAt)) > STATE_TTL_MS) {
|
||||||
|
delete state.activeStreams[k];
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// install/activate
|
||||||
|
self.addEventListener('install', (e) => { self.skipWaiting(); });
|
||||||
|
self.addEventListener('activate', (e) => { e.waitUntil(self.clients.claim()); });
|
||||||
|
|
||||||
|
// fetch: attempt to tee target streaming responses and track progress
|
||||||
|
self.addEventListener('fetch', (event) => {
|
||||||
|
try {
|
||||||
|
const url = String(event.request.url || '');
|
||||||
|
if (!url.includes(TARGET_SUBSTRING)) return; // not our target
|
||||||
|
|
||||||
// ---------- fetch: keep existing behavior for your real target, plus a debug probe ----------
|
|
||||||
self.addEventListener('fetch', event => {
|
|
||||||
const reqUrl = event.request.url || '';
|
|
||||||
// debug probe route — the sune will call /__sw_tee_test to force a tee test:
|
|
||||||
if (new URL(reqUrl).pathname === '/__sw_tee_test') {
|
|
||||||
event.respondWith((async () => {
|
event.respondWith((async () => {
|
||||||
const probeId = new URL(reqUrl).searchParams.get('probeId') || gid();
|
const upstream = await fetch(event.request);
|
||||||
LOG('Received tee-test probe', probeId);
|
|
||||||
|
|
||||||
// create a simulated streaming source inside the SW
|
// if there's no body (or not a readable stream), just forward
|
||||||
const encoder = new TextEncoder();
|
if (!upstream || !upstream.body) return upstream;
|
||||||
const stream = new ReadableStream({
|
|
||||||
start(ctrl) {
|
|
||||||
let count = 0;
|
|
||||||
const id = setInterval(() => {
|
|
||||||
count++;
|
|
||||||
const chunk = `probe(${probeId}) chunk ${count}\n`;
|
|
||||||
ctrl.enqueue(encoder.encode(chunk));
|
|
||||||
// after 6 chunks stop
|
|
||||||
if (count >= 6) { clearInterval(id); ctrl.close(); }
|
|
||||||
}, 300);
|
|
||||||
}
|
|
||||||
});
|
|
||||||
|
|
||||||
// tee the generated stream: one branch to client, one branch for SW consumption/save
|
// create a stream id and register active stream
|
||||||
const [clientBranch, swBranch] = stream.tee();
|
const streamId = 'sw-' + gid();
|
||||||
|
const meta = { url, startedAt: Date.now(), bytes: 0, lastProgressAt: Date.now(), status: 'started' };
|
||||||
|
state.totalIntercepted = (state.totalIntercepted || 0) + 1;
|
||||||
|
state.activeStreams[streamId] = meta;
|
||||||
|
// notify clients
|
||||||
|
broadcast({ type: 'sw-intercept-start', streamId, meta });
|
||||||
|
|
||||||
// in SW consume swBranch and write to localforage
|
// tee the body: one goes to client, one we consume in SW
|
||||||
(async () => {
|
const [clientStream, swStream] = upstream.body.tee();
|
||||||
|
|
||||||
|
const savePromise = (async () => {
|
||||||
try {
|
try {
|
||||||
const reader = swBranch.getReader();
|
const reader = swStream.getReader();
|
||||||
const dec = new TextDecoder('utf-8');
|
const decoder = new TextDecoder('utf-8');
|
||||||
let collected = '';
|
let decodedSoFar = '';
|
||||||
let bytes = 0;
|
let chunkCount = 0;
|
||||||
|
let lastBroadcastMs = 0;
|
||||||
|
const BROADCAST_THROTTLE_MS = 800;
|
||||||
|
const BROADCAST_BYTES = 16 * 1024; // also broadcast every ~16KB
|
||||||
|
|
||||||
while (true) {
|
while (true) {
|
||||||
const { value, done } = await reader.read();
|
const { value, done } = await reader.read();
|
||||||
if (done) break;
|
if (done) break;
|
||||||
const chunkText = dec.decode(value, { stream: true });
|
chunkCount++;
|
||||||
collected += chunkText;
|
// count bytes
|
||||||
bytes += (value && value.byteLength) ? value.byteLength : chunkText.length;
|
const bytes = value ? value.byteLength || 0 : 0;
|
||||||
// periodic save: write the collected so far into threads_v1 under probe stream id
|
meta.bytes += bytes;
|
||||||
await upsertStreamMessage('probe-' + probeId, collected, { probeId, bytesSoFar: bytes });
|
meta.lastProgressAt = Date.now();
|
||||||
// notify clients incrementally
|
|
||||||
await notifyClients({ type: 'tee-probe-chunk', probeId, bytes, snippet: chunkText.slice(0,200) });
|
// append decoded snippet for quick preview
|
||||||
|
try { decodedSoFar += decoder.decode(value, { stream: true }); } catch (e) { /* ignore decode */ }
|
||||||
|
|
||||||
|
// occasional broadcasts (throttle)
|
||||||
|
const now = Date.now();
|
||||||
|
if (now - lastBroadcastMs > BROADCAST_THROTTLE_MS || meta.bytes >= (meta._lastBroadcastBytes || 0) + BROADCAST_BYTES) {
|
||||||
|
meta._lastBroadcastBytes = meta.bytes;
|
||||||
|
lastBroadcastMs = now;
|
||||||
|
broadcast({
|
||||||
|
type: 'sw-intercept-progress',
|
||||||
|
streamId,
|
||||||
|
meta: { bytes: meta.bytes, lastProgressAt: meta.lastProgressAt, snippet: decodedSoFar.slice(-1024) }
|
||||||
|
});
|
||||||
|
}
|
||||||
}
|
}
|
||||||
await finalizeStream('probe-' + probeId, { totalBytes: bytes, probeId });
|
|
||||||
LOG('tee-probe: save complete', probeId, 'bytes', bytes);
|
// finalize
|
||||||
await notifyClients({ type: 'tee-probe-complete', probeId, totalBytes: bytes });
|
meta.status = 'finished';
|
||||||
|
meta.endedAt = Date.now();
|
||||||
|
state.lastStreamSummary = {
|
||||||
|
streamId, url, startedAt: meta.startedAt, endedAt: meta.endedAt, totalBytes: meta.bytes
|
||||||
|
};
|
||||||
|
// remove from active
|
||||||
|
delete state.activeStreams[streamId];
|
||||||
|
broadcast({ type: 'sw-intercept-end', streamId, meta: { totalBytes: meta.bytes, endedAt: meta.endedAt } });
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
LOG('tee-probe save error', err);
|
meta.status = 'error';
|
||||||
await notifyClients({ type: 'tee-probe-error', probeId, error: String(err) });
|
meta.error = String(err && err.message ? err.message : err);
|
||||||
|
meta.lastProgressAt = Date.now();
|
||||||
|
delete state.activeStreams[streamId];
|
||||||
|
broadcast({ type: 'sw-intercept-error', streamId, meta: { error: meta.error } });
|
||||||
|
console.error('sw: stream save error', err);
|
||||||
}
|
}
|
||||||
})();
|
})();
|
||||||
|
|
||||||
// return the client branch as a streaming response
|
// keep the SW alive while we process the swStream
|
||||||
return new Response(clientBranch, { status: 200, headers: { 'Content-Type': 'text/plain; charset=utf-8' } });
|
event.waitUntil(savePromise);
|
||||||
|
|
||||||
|
// forward the response to the page using the clientStream
|
||||||
|
return new Response(clientStream, {
|
||||||
|
status: upstream.status,
|
||||||
|
statusText: upstream.statusText,
|
||||||
|
headers: upstream.headers
|
||||||
|
});
|
||||||
})());
|
})());
|
||||||
return;
|
} catch (e) {
|
||||||
|
console.error('sw: fetch handler error', e);
|
||||||
|
} finally {
|
||||||
|
cleanupState();
|
||||||
}
|
}
|
||||||
|
|
||||||
// your production streaming intercept — minimal example:
|
|
||||||
if (reqUrl.includes(TARGET_SUBSTRING)) {
|
|
||||||
// let your existing logic run — simple pass-through or tee logic you already had
|
|
||||||
// For demonstration, do a straight throughfetch (or optionally tee/save like earlier example)
|
|
||||||
event.respondWith(fetch(event.request));
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
|
|
||||||
// not handled by SW: let it go
|
|
||||||
});
|
});
|
||||||
|
|
||||||
// ---------- message handling: richer PING/DIAG ----------
|
// message handler: support PING (simple) and PING_STATUS (detailed)
|
||||||
self.addEventListener('message', event => {
|
self.addEventListener('message', (event) => {
|
||||||
const data = event.data || {};
|
const data = event.data || {};
|
||||||
try {
|
try {
|
||||||
if (data && data.type === 'PING') {
|
if (data && data.type === 'PING') {
|
||||||
// reply with diagnostic info (prefer port)
|
// original ping behavior: support MessageChannel
|
||||||
(async () => {
|
if (event.ports && event.ports[0]) {
|
||||||
const streams = await listSwStreams();
|
event.ports[0].postMessage({ type: 'PONG', ts: Date.now(), ok: true });
|
||||||
const info = {
|
} else if (event.source && typeof event.source.postMessage === 'function') {
|
||||||
type: 'PONG',
|
try { event.source.postMessage({ type: 'PONG', ts: Date.now(), ok: true }); } catch(e) {}
|
||||||
ts: Date.now(),
|
} else {
|
||||||
ok: true,
|
broadcast({ type: 'PONG', ts: Date.now(), ok: true });
|
||||||
canTeeProbe: true,
|
}
|
||||||
savedStreamCount: streams.length,
|
|
||||||
lastSaved: streams[0] || null
|
|
||||||
};
|
|
||||||
if (event.ports && event.ports[0]) {
|
|
||||||
event.ports[0].postMessage(info);
|
|
||||||
} else if (event.source && typeof event.source.postMessage === 'function') {
|
|
||||||
event.source.postMessage(info);
|
|
||||||
} else {
|
|
||||||
await notifyClients(info);
|
|
||||||
}
|
|
||||||
})();
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (data && data.type === 'list-sw-streams') {
|
if (data && data.type === 'PING_STATUS') {
|
||||||
(async () => {
|
// return current SW status: activeStreams summary + lastStreamSummary + totalIntercepted
|
||||||
const streams = await listSwStreams();
|
const reply = {
|
||||||
const payload = { type: 'sw-streams-list', streams };
|
type: 'PONG_STATUS',
|
||||||
if (event.ports && event.ports[0]) event.ports[0].postMessage(payload);
|
ts: Date.now(),
|
||||||
else if (event.source && typeof event.source.postMessage === 'function') event.source.postMessage(payload);
|
totalIntercepted: state.totalIntercepted || 0,
|
||||||
else await notifyClients(payload);
|
activeStreams: Object.entries(state.activeStreams).map(([id, m]) => ({ streamId: id, url: m.url, bytes: m.bytes, status: m.status, startedAt: m.startedAt })),
|
||||||
})();
|
lastStreamSummary: state.lastStreamSummary || null
|
||||||
|
};
|
||||||
|
// reply on MessageChannel port if present, else try source, else broadcast
|
||||||
|
if (event.ports && event.ports[0]) {
|
||||||
|
event.ports[0].postMessage(reply);
|
||||||
|
} else if (event.source && typeof event.source.postMessage === 'function') {
|
||||||
|
try { event.source.postMessage(reply); } catch(e) {}
|
||||||
|
} else {
|
||||||
|
broadcast(reply);
|
||||||
|
}
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// allow direct command to run a tee probe from client (alternative path)
|
// support a request for the sw to return its current state (no port)
|
||||||
if (data && data.type === 'run-tee-probe' && data.probeId) {
|
if (data && data.type === 'GET_STATE') {
|
||||||
// reply that probe will be run — the fetch handler will simulate the stream
|
const snapshot = { totalIntercepted: state.totalIntercepted || 0, activeCount: Object.keys(state.activeStreams).length, last: state.lastStreamSummary || null };
|
||||||
(async () => {
|
if (event.ports && event.ports[0]) event.ports[0].postMessage({ type:'STATE', snapshot });
|
||||||
try {
|
else if (event.source && event.source.postMessage) event.source.postMessage({ type:'STATE', snapshot });
|
||||||
const probeUrl = '/__sw_tee_test?probeId=' + encodeURIComponent(data.probeId);
|
|
||||||
// perform fetch from SW to itself to cause internal handling (not required — client can call the endpoint)
|
|
||||||
// but we simply reply OK and let the client hit the endpoint so the client receives the response body
|
|
||||||
if (event.ports && event.ports[0]) event.ports[0].postMessage({ type:'run-tee-probe-ok', probeId: data.probeId, probeUrl });
|
|
||||||
else if (event.source && typeof event.source.postMessage === 'function') event.source.postMessage({ type:'run-tee-probe-ok', probeId: data.probeId, probeUrl });
|
|
||||||
} catch (e) {
|
|
||||||
if (event.ports && event.ports[0]) event.ports[0].postMessage({ type:'run-tee-probe-error', error: String(e) });
|
|
||||||
}
|
|
||||||
})();
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
} catch (e) {
|
||||||
} catch (err) {
|
console.error('sw: message handler error', e);
|
||||||
LOG('message handler error', err);
|
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
|||||||
Reference in New Issue
Block a user