forked from FINAKON/HelpProject
1. Initial Commit - a boiler plate code and POC to realize the concept of context sensitive help 2. Frontend code written in ReactJS 3. Backend code written in Java, Spring Boot Framework 4. Frontend Start: pre-requisites : node, npm npm run dev ==> to start the frontend vite server 5. Backend Start: pre-requisites : java, mvn mvn spring-boot:run ==> to start the backend server 6. Visit http://localhost:5173/ for basic demo of help, press F1 in textboxes 7. Visit http://localhost:5173/editor and enter "admin123" to add/modify texts. Happy Coding !!! Thank you, Bhargava.
192 lines
5.4 KiB
JavaScript
192 lines
5.4 KiB
JavaScript
"use strict";
|
|
|
|
Object.defineProperty(exports, "__esModule", {
|
|
value: true
|
|
});
|
|
exports.TokenMap = void 0;
|
|
var _t = require("@babel/types");
|
|
const {
|
|
traverseFast,
|
|
VISITOR_KEYS
|
|
} = _t;
|
|
class TokenMap {
|
|
constructor(ast, tokens, source) {
|
|
this._tokens = void 0;
|
|
this._source = void 0;
|
|
this._nodesToTokenIndexes = new Map();
|
|
this._nodesOccurrencesCountCache = new Map();
|
|
this._tokensCache = new Map();
|
|
this._tokens = tokens;
|
|
this._source = source;
|
|
traverseFast(ast, node => {
|
|
const indexes = this._getTokensIndexesOfNode(node);
|
|
if (indexes.length > 0) this._nodesToTokenIndexes.set(node, indexes);
|
|
});
|
|
this._tokensCache = null;
|
|
}
|
|
has(node) {
|
|
return this._nodesToTokenIndexes.has(node);
|
|
}
|
|
getIndexes(node) {
|
|
return this._nodesToTokenIndexes.get(node);
|
|
}
|
|
find(node, condition) {
|
|
const indexes = this._nodesToTokenIndexes.get(node);
|
|
if (indexes) {
|
|
for (let k = 0; k < indexes.length; k++) {
|
|
const index = indexes[k];
|
|
const tok = this._tokens[index];
|
|
if (condition(tok, index)) return tok;
|
|
}
|
|
}
|
|
return null;
|
|
}
|
|
findLastIndex(node, condition) {
|
|
const indexes = this._nodesToTokenIndexes.get(node);
|
|
if (indexes) {
|
|
for (let k = indexes.length - 1; k >= 0; k--) {
|
|
const index = indexes[k];
|
|
const tok = this._tokens[index];
|
|
if (condition(tok, index)) return index;
|
|
}
|
|
}
|
|
return -1;
|
|
}
|
|
findMatching(node, test, occurrenceCount = 0) {
|
|
const indexes = this._nodesToTokenIndexes.get(node);
|
|
if (indexes) {
|
|
let i = 0;
|
|
const count = occurrenceCount;
|
|
if (count > 1) {
|
|
const cache = this._nodesOccurrencesCountCache.get(node);
|
|
if (cache && cache.test === test && cache.count < count) {
|
|
i = cache.i + 1;
|
|
occurrenceCount -= cache.count + 1;
|
|
}
|
|
}
|
|
for (; i < indexes.length; i++) {
|
|
const tok = this._tokens[indexes[i]];
|
|
if (this.matchesOriginal(tok, test)) {
|
|
if (occurrenceCount === 0) {
|
|
if (count > 0) {
|
|
this._nodesOccurrencesCountCache.set(node, {
|
|
test,
|
|
count,
|
|
i
|
|
});
|
|
}
|
|
return tok;
|
|
}
|
|
occurrenceCount--;
|
|
}
|
|
}
|
|
}
|
|
return null;
|
|
}
|
|
matchesOriginal(token, test) {
|
|
if (token.end - token.start !== test.length) return false;
|
|
if (token.value != null) return token.value === test;
|
|
return this._source.startsWith(test, token.start);
|
|
}
|
|
startMatches(node, test) {
|
|
const indexes = this._nodesToTokenIndexes.get(node);
|
|
if (!indexes) return false;
|
|
const tok = this._tokens[indexes[0]];
|
|
if (tok.start !== node.start) return false;
|
|
return this.matchesOriginal(tok, test);
|
|
}
|
|
endMatches(node, test) {
|
|
const indexes = this._nodesToTokenIndexes.get(node);
|
|
if (!indexes) return false;
|
|
const tok = this._tokens[indexes[indexes.length - 1]];
|
|
if (tok.end !== node.end) return false;
|
|
return this.matchesOriginal(tok, test);
|
|
}
|
|
_getTokensIndexesOfNode(node) {
|
|
if (node.start == null || node.end == null) return [];
|
|
const {
|
|
first,
|
|
last
|
|
} = this._findTokensOfNode(node, 0, this._tokens.length - 1);
|
|
let low = first;
|
|
const children = childrenIterator(node);
|
|
if ((node.type === "ExportNamedDeclaration" || node.type === "ExportDefaultDeclaration") && node.declaration && node.declaration.type === "ClassDeclaration") {
|
|
children.next();
|
|
}
|
|
const indexes = [];
|
|
for (const child of children) {
|
|
if (child == null) continue;
|
|
if (child.start == null || child.end == null) continue;
|
|
const childTok = this._findTokensOfNode(child, low, last);
|
|
const high = childTok.first;
|
|
for (let k = low; k < high; k++) indexes.push(k);
|
|
low = childTok.last + 1;
|
|
}
|
|
for (let k = low; k <= last; k++) indexes.push(k);
|
|
return indexes;
|
|
}
|
|
_findTokensOfNode(node, low, high) {
|
|
const cached = this._tokensCache.get(node);
|
|
if (cached) return cached;
|
|
const first = this._findFirstTokenOfNode(node.start, low, high);
|
|
const last = this._findLastTokenOfNode(node.end, first, high);
|
|
this._tokensCache.set(node, {
|
|
first,
|
|
last
|
|
});
|
|
return {
|
|
first,
|
|
last
|
|
};
|
|
}
|
|
_findFirstTokenOfNode(start, low, high) {
|
|
while (low <= high) {
|
|
const mid = high + low >> 1;
|
|
if (start < this._tokens[mid].start) {
|
|
high = mid - 1;
|
|
} else if (start > this._tokens[mid].start) {
|
|
low = mid + 1;
|
|
} else {
|
|
return mid;
|
|
}
|
|
}
|
|
return low;
|
|
}
|
|
_findLastTokenOfNode(end, low, high) {
|
|
while (low <= high) {
|
|
const mid = high + low >> 1;
|
|
if (end < this._tokens[mid].end) {
|
|
high = mid - 1;
|
|
} else if (end > this._tokens[mid].end) {
|
|
low = mid + 1;
|
|
} else {
|
|
return mid;
|
|
}
|
|
}
|
|
return high;
|
|
}
|
|
}
|
|
exports.TokenMap = TokenMap;
|
|
function* childrenIterator(node) {
|
|
if (node.type === "TemplateLiteral") {
|
|
yield node.quasis[0];
|
|
for (let i = 1; i < node.quasis.length; i++) {
|
|
yield node.expressions[i - 1];
|
|
yield node.quasis[i];
|
|
}
|
|
return;
|
|
}
|
|
const keys = VISITOR_KEYS[node.type];
|
|
for (const key of keys) {
|
|
const child = node[key];
|
|
if (!child) continue;
|
|
if (Array.isArray(child)) {
|
|
yield* child;
|
|
} else {
|
|
yield child;
|
|
}
|
|
}
|
|
}
|
|
|
|
//# sourceMappingURL=token-map.js.map
|