Spaces:
Sleeping
Sleeping
; | |
Object.defineProperty(exports, "__esModule", { | |
value: true | |
}); | |
exports.TokenMap = void 0; | |
var _t = require("@babel/types"); | |
const { | |
traverseFast, | |
VISITOR_KEYS | |
} = _t; | |
class TokenMap { | |
constructor(ast, tokens, source) { | |
this._tokens = void 0; | |
this._source = void 0; | |
this._nodesToTokenIndexes = new Map(); | |
this._nodesOccurrencesCountCache = new Map(); | |
this._tokensCache = new Map(); | |
this._tokens = tokens; | |
this._source = source; | |
traverseFast(ast, node => { | |
const indexes = this._getTokensIndexesOfNode(node); | |
if (indexes.length > 0) this._nodesToTokenIndexes.set(node, indexes); | |
}); | |
this._tokensCache = null; | |
} | |
has(node) { | |
return this._nodesToTokenIndexes.has(node); | |
} | |
getIndexes(node) { | |
return this._nodesToTokenIndexes.get(node); | |
} | |
find(node, condition) { | |
const indexes = this._nodesToTokenIndexes.get(node); | |
if (indexes) { | |
for (let k = 0; k < indexes.length; k++) { | |
const index = indexes[k]; | |
const tok = this._tokens[index]; | |
if (condition(tok, index)) return tok; | |
} | |
} | |
return null; | |
} | |
findLastIndex(node, condition) { | |
const indexes = this._nodesToTokenIndexes.get(node); | |
if (indexes) { | |
for (let k = indexes.length - 1; k >= 0; k--) { | |
const index = indexes[k]; | |
const tok = this._tokens[index]; | |
if (condition(tok, index)) return index; | |
} | |
} | |
return -1; | |
} | |
findMatching(node, test, occurrenceCount = 0) { | |
const indexes = this._nodesToTokenIndexes.get(node); | |
if (indexes) { | |
let i = 0; | |
const count = occurrenceCount; | |
if (count > 1) { | |
const cache = this._nodesOccurrencesCountCache.get(node); | |
if (cache && cache.test === test && cache.count < count) { | |
i = cache.i + 1; | |
occurrenceCount -= cache.count + 1; | |
} | |
} | |
for (; i < indexes.length; i++) { | |
const tok = this._tokens[indexes[i]]; | |
if (this.matchesOriginal(tok, test)) { | |
if (occurrenceCount === 0) { | |
if (count > 0) { | |
this._nodesOccurrencesCountCache.set(node, { | |
test, | |
count, | |
i | |
}); | |
} | |
return tok; | |
} | |
occurrenceCount--; | |
} | |
} | |
} | |
return null; | |
} | |
matchesOriginal(token, test) { | |
if (token.end - token.start !== test.length) return false; | |
if (token.value != null) return token.value === test; | |
return this._source.startsWith(test, token.start); | |
} | |
startMatches(node, test) { | |
const indexes = this._nodesToTokenIndexes.get(node); | |
if (!indexes) return false; | |
const tok = this._tokens[indexes[0]]; | |
if (tok.start !== node.start) return false; | |
return this.matchesOriginal(tok, test); | |
} | |
endMatches(node, test) { | |
const indexes = this._nodesToTokenIndexes.get(node); | |
if (!indexes) return false; | |
const tok = this._tokens[indexes[indexes.length - 1]]; | |
if (tok.end !== node.end) return false; | |
return this.matchesOriginal(tok, test); | |
} | |
_getTokensIndexesOfNode(node) { | |
if (node.start == null || node.end == null) return []; | |
const { | |
first, | |
last | |
} = this._findTokensOfNode(node, 0, this._tokens.length - 1); | |
let low = first; | |
const children = childrenIterator(node); | |
if ((node.type === "ExportNamedDeclaration" || node.type === "ExportDefaultDeclaration") && node.declaration && node.declaration.type === "ClassDeclaration") { | |
children.next(); | |
} | |
const indexes = []; | |
for (const child of children) { | |
if (child == null) continue; | |
if (child.start == null || child.end == null) continue; | |
const childTok = this._findTokensOfNode(child, low, last); | |
const high = childTok.first; | |
for (let k = low; k < high; k++) indexes.push(k); | |
low = childTok.last + 1; | |
} | |
for (let k = low; k <= last; k++) indexes.push(k); | |
return indexes; | |
} | |
_findTokensOfNode(node, low, high) { | |
const cached = this._tokensCache.get(node); | |
if (cached) return cached; | |
const first = this._findFirstTokenOfNode(node.start, low, high); | |
const last = this._findLastTokenOfNode(node.end, first, high); | |
this._tokensCache.set(node, { | |
first, | |
last | |
}); | |
return { | |
first, | |
last | |
}; | |
} | |
_findFirstTokenOfNode(start, low, high) { | |
while (low <= high) { | |
const mid = high + low >> 1; | |
if (start < this._tokens[mid].start) { | |
high = mid - 1; | |
} else if (start > this._tokens[mid].start) { | |
low = mid + 1; | |
} else { | |
return mid; | |
} | |
} | |
return low; | |
} | |
_findLastTokenOfNode(end, low, high) { | |
while (low <= high) { | |
const mid = high + low >> 1; | |
if (end < this._tokens[mid].end) { | |
high = mid - 1; | |
} else if (end > this._tokens[mid].end) { | |
low = mid + 1; | |
} else { | |
return mid; | |
} | |
} | |
return high; | |
} | |
} | |
exports.TokenMap = TokenMap; | |
function* childrenIterator(node) { | |
if (node.type === "TemplateLiteral") { | |
yield node.quasis[0]; | |
for (let i = 1; i < node.quasis.length; i++) { | |
yield node.expressions[i - 1]; | |
yield node.quasis[i]; | |
} | |
return; | |
} | |
const keys = VISITOR_KEYS[node.type]; | |
for (const key of keys) { | |
const child = node[key]; | |
if (!child) continue; | |
if (Array.isArray(child)) { | |
yield* child; | |
} else { | |
yield child; | |
} | |
} | |
} | |
//# sourceMappingURL=token-map.js.map | |