From afa87af01c79a9baa539f2992d32154d2a4739bd Mon Sep 17 00:00:00 2001 From: Adam Mathes Date: Sat, 14 Feb 2026 14:46:37 -0800 Subject: task: delete vanilla js prototype\n\n- Removed vanilla/ directory and web/dist/vanilla directory\n- Updated Makefile, Dockerfile, and CI workflow to remove vanilla references\n- Cleaned up web/web.go to remove vanilla embed and routes\n- Verified build and tests pass\n\nCloses NK-2tcnmq --- .../css-tree/lib/lexer/prepare-tokens.js | 50 ---------------------- 1 file changed, 50 deletions(-) delete mode 100644 vanilla/node_modules/css-tree/lib/lexer/prepare-tokens.js (limited to 'vanilla/node_modules/css-tree/lib/lexer/prepare-tokens.js') diff --git a/vanilla/node_modules/css-tree/lib/lexer/prepare-tokens.js b/vanilla/node_modules/css-tree/lib/lexer/prepare-tokens.js deleted file mode 100644 index 4243fa8..0000000 --- a/vanilla/node_modules/css-tree/lib/lexer/prepare-tokens.js +++ /dev/null @@ -1,50 +0,0 @@ -import { tokenize } from '../tokenizer/index.js'; - -const astToTokens = { - decorator(handlers) { - const tokens = []; - let curNode = null; - - return { - ...handlers, - node(node) { - const tmp = curNode; - curNode = node; - handlers.node.call(this, node); - curNode = tmp; - }, - emit(value, type, auto) { - tokens.push({ - type, - value, - node: auto ? null : curNode - }); - }, - result() { - return tokens; - } - }; - } -}; - -function stringToTokens(str) { - const tokens = []; - - tokenize(str, (type, start, end) => - tokens.push({ - type, - value: str.slice(start, end), - node: null - }) - ); - - return tokens; -} - -export default function(value, syntax) { - if (typeof value === 'string') { - return stringToTokens(value); - } - - return syntax.generate(value, astToTokens); -}; -- cgit v1.2.3