bostr/bouncer.js

470 lines
17 KiB
JavaScript
Raw Permalink Normal View History

"use strict";
const { version } = require("./package.json");
const WebSocket = require("ws");
const querystring = require("querystring");
const { validateEvent, nip19, matchFilters, mergeFilters, getFilterLimit } = require("nostr-tools");
const auth = require("./auth.js");
const nip42 = require("./nip42.js");
let { relays, allowed_publishers, approved_publishers, blocked_publishers, log_about_relays, authorized_keys, private_keys, reconnect_time, wait_eose, pause_on_limit, max_eose_score, broadcast_ratelimit, upstream_ratelimit_expiration, max_client_subs, idle_sessions, cache_relays, noscraper } = require(process.env.BOSTR_CONFIG_PATH || "./config");
log_about_relays = process.env.LOG_ABOUT_RELAYS || log_about_relays;
authorized_keys = authorized_keys?.map(i => i.startsWith("npub") ? nip19.decode(i).data : i);
allowed_publishers = allowed_publishers?.map(i => i.startsWith("npub") ? nip19.decode(i).data : i);
blocked_publishers = blocked_publishers?.map(i => i.startsWith("npub") ? nip19.decode(i).data : i);
// CL MaxEoseScore: Set <max_eose_score> as 0 if configured relays is under of the expected number from <max_eose_score>
if (relays.length < max_eose_score) max_eose_score = 0;
// The following warning will be removed in the next 2 stable release
if (approved_publishers?.length) {
allowed_publishers = approved_publishers?.map(i => i.startsWith("npub") ? nip19.decode(i).data : i);
console.warn(process.pid, "[config]");
console.warn(process.pid, "[config]", "!!! Attention !!!");
console.warn(process.pid, "[config]", "approved_publishers is deprecated. rename as allowed_publishers");
console.warn(process.pid, "[config]");
}
const csess = new Map(); // this is used for relays.
const userRelays = new Map(); // per ID contains Set() of <WebSocket>
const idleSess = new Set();
let stats = {
_global: {
raw_rx: 0,
rx: 0,
tx: 0,
f: 0
}
};
// CL - User socket
function handleConnection(ws, req, onClose) {
let query = querystring.parse(req.url.slice(2));
let authKey = null;
let authorized = true;
let sessStarted = false;
let lastEvent = Date.now();
ws.ip = req.headers["x-forwarded-for"]?.split(",")[0] || req.socket.address()?.address;
ws.subs = new Map(); // contains filter submitted by clients. per subID
ws.pause_subs = new Set(); // pause subscriptions from receiving events after reached over <filter.limit> until all relays send EOSE. per subID
ws.events = new Map(); // only to prevent the retransmit of the same event. per subID
ws.my_events = new Set(); // for event retransmitting.
ws.pendingEOSE = new Map(); // each contain subID
ws.reconnectTimeout = new Set(); // relays timeout() before reconnection. Only use after client disconnected.
ws.subalias = new Map();
ws.fakesubalias = new Map();
ws.mergedFilters = new Map();
ws.pubkey = null;
ws.rejectKinds = query.reject?.split(",").map(_ => parseInt(_));
ws.acceptKinds = query.accept?.split(",").map(_ => parseInt(_));
ws.forcedLimit = parseInt(query.limit);
ws.accurateMode = parseInt(query.accurate);
ws.saveMode = parseInt(query.save);
if (noscraper || authorized_keys?.length) {
authKey = Date.now() + Math.random().toString(36);
authorized = false;
ws.send(JSON.stringify(["AUTH", authKey]));
} else if (Object.keys(private_keys).length) {
// If there is no whitelist, Then we ask to client what is their public key.
// We will enable NIP-42 function for this session if user pubkey was available & valid in <private_keys>.
// There is no need to limit this session. We only ask who is this user.
// If it was the users listed at <private_keys> in config.js, Then the user could use NIP-42 protected relays.
authKey = Date.now() + Math.random().toString(36);
ws.send(JSON.stringify(["AUTH", authKey]));
}
console.log(process.pid, `->- ${ws.ip} connected [${req.headers["user-agent"] || ""}]`);
ws.on("message", data => {
try {
data = JSON.parse(data);
} catch {
return ws.send(
JSON.stringify(["NOTICE", "error: bad JSON."])
)
}
switch (data[0]) {
case "EVENT":
if (!validateEvent(data[1])) return ws.send(JSON.stringify(["NOTICE", "error: invalid event"]));
if (data[1].kind == 22242) return ws.send(JSON.stringify(["OK", data[1].id, false, "rejected: kind 22242"]));
if (blocked_publishers?.includes(data[1].pubkey)) return ws.send(JSON.stringify(["OK", data[1].id, false, "blocked: this event author is blacklisted."]));
if (!authorized) {
ws.send(JSON.stringify(["OK", data[1].id, false, "auth-required: authentication is required to perform this action."]));
return ws.send(JSON.stringify(["AUTH", authKey]));
}
if (
allowed_publishers?.length &&
!allowed_publishers?.includes(data[1].pubkey)
) return ws.send(JSON.stringify(["OK", data[1]?.id, false, "rejected: unauthorized"]));
if (broadcast_ratelimit && (broadcast_ratelimit > (Date.now() - lastEvent))) {
lastEvent = Date.now();
return ws.send(JSON.stringify(["OK", data[1]?.id, false, "rate-limited: request too fast."]));
}
lastEvent = Date.now();
ws.my_events.add(data[1]);
if (!sessStarted) {
console.log(process.pid, `>>>`, `${ws.ip} executed ${data[0]} command for the first. Initializing session`);
getIdleSess(ws);
sessStarted = true;
}
bc(data, ws.id);
ws.send(JSON.stringify(["OK", data[1]?.id, true, ""]));
break;
case "REQ": {
if (data.length < 3) return ws.send(JSON.stringify(["NOTICE", "error: bad request."]));
if (typeof(data[1]) !== "string") return ws.send(JSON.stringify(["NOTICE", "error: expected subID a string. but got the otherwise."]));
if (typeof(data[2]) !== "object") return ws.send(JSON.stringify(["CLOSED", data[1], "error: expected filter to be obj, instead gives the otherwise."]));
if (!authorized) {
ws.send(JSON.stringify(["CLOSED", data[1], "auth-required: authentication is required to perform this action."]));
return ws.send(JSON.stringify(["AUTH", authKey]));
}
if ((max_client_subs !== -1) && (ws.subs.size > max_client_subs)) return ws.send(JSON.stringify(["CLOSED", data[1], "rate-limited: too many subscriptions."]));
const origID = data[1];
if (ws.subs.has(data[1])) {
const faked = ws.fakesubalias.get(origID);
ws.subs.delete(origID);
ws.events.delete(origID);
ws.pendingEOSE.delete(origID);
ws.pause_subs.delete(origID);
ws.fakesubalias.delete(origID);
ws.subalias.delete(faked);
ws.mergedFilters.delete(origID);
bc(["CLOSE", faked], ws.id);
};
const faked = Date.now() + Math.random().toString(36);
let filters = data.slice(2);
let filter = mergeFilters(...filters);
for (const fn in filters) {
if (!Array.isArray(filters[fn].kinds)) {
filters[fn].kinds = ws.acceptKinds;
continue;
} else {
filters[fn].kinds = filters[fn].kinds?.filter(kind => {
if (ws.rejectKinds && ws.rejectKinds.includes(kind)) return false;
if (ws.acceptKinds && !ws.acceptKinds.includes(kind)) return false;
return true;
});
}
if (filters[fn].limit > ws.forcedLimit)
filters[fn].limit = ws.forcedLimit;
}
if (!sessStarted) {
console.log(process.pid, `>>>`, `${ws.ip} executed ${data[0]} command for the first. Initializing session`);
getIdleSess(ws);
sessStarted = true;
}
ws.subs.set(origID, filters);
ws.events.set(origID, new Set());
ws.pause_subs.delete(origID);
ws.subalias.set(faked, origID);
ws.fakesubalias.set(origID, faked);
if (!filter.since) filter.since = Math.floor(Date.now() / 1000); // Will not impact everything. Only used for handling passing pause_on_limit (or save mode)
ws.mergedFilters.set(origID, filter);
data[1] = faked;
bc(data, ws.id);
if (filter.limit < 1) return ws.send(JSON.stringify(["EOSE", origID]));
ws.pendingEOSE.set(origID, 0);
break;
}
case "CLOSE":
if (!authorized) return;
if (typeof(data[1]) !== "string") return ws.send(JSON.stringify(["NOTICE", "error: bad request."]));
if (!ws.subs.has(data[1])) return ws.send(JSON.stringify(["CLOSED", data[1], "error: this sub is not opened."]));
const origID = data[1];
const faked = ws.fakesubalias.get(origID);
ws.subs.delete(origID);
ws.events.delete(origID);
ws.pendingEOSE.delete(origID);
ws.pause_subs.delete(origID);
ws.fakesubalias.delete(origID);
ws.subalias.delete(faked);
ws.mergedFilters.delete(origID);
data[1] = faked;
bc(data, ws.id);
ws.send(JSON.stringify(["CLOSED", origID, ""]));
break;
case "AUTH":
if (authorized) return;
if (auth(authKey, data[1], ws, req)) {
ws.pubkey = data[1].pubkey;
console.log(process.pid, "---", ws.ip, "successfully authorized as", ws.pubkey, private_keys[ws.pubkey] ? "(admin)" : "(user)");
if (authorized) return;
authorized = true;
lastEvent = Date.now();
}
break;
default:
ws.send(JSON.stringify(["NOTICE", `error: unrecognized command: ${data[0]}`]));
break;
}
});
ws.on('error', console.error);
ws.on('close', _ => {
onClose();
console.log(process.pid, "---", `${ws.ip} disconnected`);
if (!sessStarted) return;
for (const sock of userRelays.get(ws.id)) {
sock.terminate();
}
userRelays.delete(ws.id);
csess.delete(ws.id);
});
}
// WS - New session for client $id
// mostly for new idle session.
function newsess() {
const id = Date.now() + "_" + process.pid + "_" + Math.random();
userRelays.set(id, new Set());
csess.set(id, null);
idleSess.add(id);
if (cache_relays) {
for (const url of cache_relays) {
newConn(url, id);
}
}
for (const url of relays) {
newConn(url, id);
}
}
// WS - Broadcast message to every existing sockets
function bc(msg, id, toCacheOnly) {
for (const relay of userRelays.get(id)) {
if (relay.readyState !== 1) continue;
if (toCacheOnly && !relay.isCache) continue;
// skip the ratelimit after <config.upstream_ratelimit_expiration>
if ((upstream_ratelimit_expiration) > (Date.now() - relay.ratelimit)) continue;
relay.send(JSON.stringify(msg));
}
}
function getIdleSess(ws) {
ws.id = idleSess.values().next().value;
idleSess.delete(ws.id);
csess.set(ws.id, ws);
if (log_about_relays) console.log(process.pid, "---", ws.ip, "is now using session", ws.id);
newsess();
}
// WS - Sessions
function newConn(addr, id, reconn_t = 0) {
if (!csess.has(id)) return;
if (!stats[addr]) stats[addr] = { raw_rx: 0, rx: 0, tx: 0, f: 0 };
const relay = new WebSocket(addr, {
headers: {
"User-Agent": `Bostr ${version}; The nostr relay bouncer; https://github.com/Yonle/bostr`,
},
noDelay: true,
allowSynchronousEvents: true
});
relay.isCache = cache_relays?.includes(addr);
relay.ratelimit = 0;
relay.on('open', _ => {
if (!csess.has(id)) return relay.terminate();
const client = csess.get(id);
reconn_t = 0;
if (log_about_relays) console.log(process.pid, "---", id, `${addr} is connected`);
if (!client) return;
for (const i of client.my_events) {
relay.send(JSON.stringify(["EVENT", i]));
}
for (const i of client.subs) {
relay.send(JSON.stringify(["REQ", client.fakesubalias.get(i[0]), ...i[1]]));
}
});
relay.on('message', data => {
if (!csess.has(id)) return relay.terminate();
try {
data = JSON.parse(data);
} catch (error) {
return;
}
const client = csess.get(id);
if (!client) return;
switch (data[0]) {
case "EVENT": {
stats._global.raw_rx++;
stats[addr].raw_rx++;
if (data.length < 3 || typeof(data[1]) !== "string" || typeof(data[2]) !== "object") return;
if (!client.subalias.has(data[1])) return;
data[1] = client.subalias.get(data[1]);
if (client.events.get(data[1]).has(data[2]?.id)) return; // No need to transmit once it has been transmitted before.
if (!relay.isCache) bc(["EVENT", data[2]], id, true); // store to cache relay
const filter = client.mergedFilters.get(data[1]);
if (client.pause_subs.has(data[1]) && (filter.since > data[2].created_at) && !relay.isCache) return;
if (client.rejectKinds && client.rejectKinds.includes(data[2]?.id)) return;
const filters = client.subs.get(data[1]);
if (!matchFilters(filters, data[2])) return;
const NotInSearchQuery = "search" in filter && !data[2]?.content?.toLowerCase().includes(filter.search.toLowerCase());
if (NotInSearchQuery) return;
client.events.get(data[1]).add(data[2]?.id);
client.send(JSON.stringify(data));
stats._global.rx++;
stats[addr].rx++;
// Now count for REQ limit requested by client.
// If it's at the limit, Send EOSE to client and delete pendingEOSE of subID
// Skip if EOSE has been omitted
if (!client.pendingEOSE.has(data[1]) || client.pause_subs.has(data[1])) return;
const limit = getFilterLimit(filter);
if (limit === Infinity) return;
if (client.events.get(data[1]).size >= limit) {
// Once reached to <filter.limit>, send EOSE to client.
client.send(JSON.stringify(["EOSE", data[1]]));
if (!client.accurateMode && (client.saveMode || pause_on_limit)) {
client.pause_subs.add(data[1]);
} else {
client.pendingEOSE.delete(data[1]);
}
}
break;
}
case "EOSE":
if (!client.subalias.has(data[1])) return;
data[1] = client.subalias.get(data[1]);
if (!client.pendingEOSE.has(data[1])) return;
client.pendingEOSE.set(data[1], client.pendingEOSE.get(data[1]) + 1);
if (log_about_relays) console.log(process.pid, "---", id, `got EOSE from ${addr} for ${data[1]}. There are ${client.pendingEOSE.get(data[1])} EOSE received out of ${userRelays.get(id).size} connected relays.`);
if (wait_eose && ((client.pendingEOSE.get(data[1]) < max_eose_score) || (client.pendingEOSE.get(data[1]) < userRelays.get(id).size))) return;
client.pendingEOSE.delete(data[1]);
if (client.pause_subs.has(data[1])) {
client.pause_subs.delete(data[1]);
} else {
client.send(JSON.stringify(data));
}
break;
case "AUTH":
if (!private_keys || typeof(data[1]) !== "string" || !client.pubkey) return;
nip42(relay, client.pubkey, private_keys[client.pubkey], data[1]);
break;
case "NOTICE":
if (typeof(data[1]) !== "string") return;
if (data[1].startsWith("rate-limited")) relay.ratelimit = Date.now();
if (log_about_relays) console.log(process.pid, id, addr, data[0], data[1]);
stats._global.f++
stats[addr].f++
break;
case "CLOSED":
if ((typeof(data[1]) !== "string") || (typeof(data[2]) !== "string")) return;
if (data[2].startsWith("rate-limited")) relay.ratelimit = Date.now();
if (log_about_relays) console.log(process.pid, id, addr, data[0], data[1], data[2]);
if (data[2].length) {
stats._global.f++;
stats[addr].f++;
}
if (client.pendingEOSE.has(data[1])) client.pendingEOSE.set(data[1], client.pendingEOSE.get(data[1]) + 1);
break;
case "OK":
if ((typeof(data[1]) !== "string") || (typeof(data[2]) !== "boolean") || (typeof(data[3]) !== "string")) return;
if (data[3].startsWith("rate-limited")) relay.ratelimit = Date.now();
if (log_about_relays) console.log(process.pid, id, addr, data[0], data[1], data[2], data[3]);
switch (data[2]) {
case true:
stats._global.tx++;
stats[addr].tx++;
case false:
stats._global.f++
stats[addr].f++
}
break;
}
});
relay.on('error', _ => {
if (log_about_relays) console.error(process.pid, "-!-", id, addr, _.toString())
});
relay.on('close', _ => {
if (!userRelays.has(id)) return;
userRelays.get(id).delete(relay); // Remove this socket session from <client.relays> list
if (log_about_relays) console.log(process.pid, "-!-", id, "Disconnected from", addr);
reconn_t += reconnect_time || 5000
setTimeout(_ => {
newConn(addr, id, reconn_t);
}, reconn_t);
stats._global.f++
stats[addr].f++
});
relay.on('unexpected-response', (req, res) => {
if (!userRelays.has(id)) return;
userRelays.get(id).delete(relay);
if (res.statusCode >= 500) return relay.emit("close", null);
relays = relays.filter(_ => _ != addr);
console.log(process.pid, "-!-", `${addr} give status code ${res.statusCode}. Not (re)connect with new session again.`);
stats._global.f++
stats[addr].f++
});
userRelays.get(id).add(relay); // Add this socket session to <client.relays>
}
for (let i = 1; i <= (idle_sessions || 1); i++) {
newsess();
}
function getStat(n) {
if (!n) return stats;
return stats[n];
}
module.exports = {
handleConnection,
getStat
}