Jump To …

rewriter.coffee

#
this.exports: this unless process?
#

Tokens that must be balanced.

BALANCED_PAIRS: [['(', ')'], ['[', ']'], ['{', '}'], ['INDENT', 'OUTDENT'],
  ['PARAM_START', 'PARAM_END'], ['CALL_START', 'CALL_END'],
  ['INDEX_START', 'INDEX_END'], ['SOAKED_INDEX_START', 'SOAKED_INDEX_END']]
#

Tokens that signal the start of a balanced pair.

EXPRESSION_START: pair[0] for pair in BALANCED_PAIRS
#

Tokens that signal the end of a balanced pair.

EXPRESSION_TAIL: pair[1] for pair in BALANCED_PAIRS
#

Tokens that indicate the close of a clause of an expression.

EXPRESSION_CLOSE: ['CATCH', 'WHEN', 'ELSE', 'FINALLY'].concat(EXPRESSION_TAIL)
#

Tokens pairs that, in immediate succession, indicate an implicit call.

IMPLICIT_FUNC: ['IDENTIFIER', 'SUPER', ')', 'CALL_END', ']', 'INDEX_END']
IMPLICIT_BLOCK:['->', '=>', '{', '[', ',']
IMPLICIT_END:  ['IF', 'UNLESS', 'FOR', 'WHILE', 'TERMINATOR', 'INDENT', 'OUTDENT']
IMPLICIT_CALL: ['IDENTIFIER', 'NUMBER', 'STRING', 'JS', 'REGEX', 'NEW', 'PARAM_START',
                 'TRY', 'DELETE', 'TYPEOF', 'SWITCH',
                 'TRUE', 'FALSE', 'YES', 'NO', 'ON', 'OFF', '!', '!!', 'NOT',
                 '@', '->', '=>', '[', '(', '{']
#

The inverse mappings of token pairs we're trying to fix up.

INVERSES: {}
for pair in BALANCED_PAIRS
  INVERSES[pair[0]]: pair[1]
  INVERSES[pair[1]]: pair[0]
#

Single-line flavors of block expressions that have unclosed endings. The grammar can't disambiguate them, so we insert the implicit indentation.

SINGLE_LINERS: ['ELSE', "->", "=>", 'TRY', 'FINALLY', 'THEN']
SINGLE_CLOSERS: ['TERMINATOR', 'CATCH', 'FINALLY', 'ELSE', 'OUTDENT', 'LEADING_WHEN']
#

In order to keep the grammar simple, the stream of tokens that the Lexer emits is rewritten by the Rewriter, smoothing out ambiguities, mis-nested indentation, and single-line flavors of expressions.

exports.Rewriter: class Rewriter
#

Rewrite the token stream in multiple passes, one logical filter at a time. This could certainly be changed into a single pass through the stream, with a big ol' efficient switch, but it's much nicer like this.

  rewrite: (tokens) ->
    @tokens: tokens
    @adjust_comments()
    @remove_leading_newlines()
    @remove_mid_expression_newlines()
    @move_commas_outside_outdents()
    @close_open_calls_and_indexes()
    @add_implicit_indentation()
    @add_implicit_parentheses()
    @ensure_balance(BALANCED_PAIRS)
    @rewrite_closing_parens()
    @tokens
#

Rewrite the token stream, looking one token ahead and behind. Allow the return value of the block to tell us how many tokens to move forwards (or backwards) in the stream, to make sure we don't miss anything as the stream changes length under our feet.

  scan_tokens: (block) ->
    i: 0
    while true
      break unless @tokens[i]
      move: block(@tokens[i - 1], @tokens[i], @tokens[i + 1], i)
      i += move
    true
#

Massage newlines and indentations so that comments don't have to be correctly indented, or appear on their own line.

  adjust_comments: ->
    @scan_tokens (prev, token, post, i) =>
      return 1 unless token[0] is 'COMMENT'
      after:  @tokens[i + 2]
      if after and after[0] is 'INDENT'
        @tokens.splice(i + 2, 1)
        @tokens.splice(i, 0, after)
        return 1
      else if prev and prev[0] isnt 'TERMINATOR' and prev[0] isnt 'INDENT' and prev[0] isnt 'OUTDENT'
        @tokens.splice(i, 0, ['TERMINATOR', "\n", prev[2]])
        return 2
      else
        return 1
#

Leading newlines would introduce an ambiguity in the grammar, so we dispatch them here.

  remove_leading_newlines: ->
    @tokens.shift() if @tokens[0][0] is 'TERMINATOR'
#

Some blocks occur in the middle of expressions -- when we're expecting this, remove their trailing newlines.

  remove_mid_expression_newlines: ->
    @scan_tokens (prev, token, post, i) =>
      return 1 unless post and EXPRESSION_CLOSE.indexOf(post[0]) >= 0 and token[0] is 'TERMINATOR'
      @tokens.splice(i, 1)
      return 0
#

Make sure that we don't accidentally break trailing commas, which need to go on the outside of expression closers.

  move_commas_outside_outdents: ->
    @scan_tokens (prev, token, post, i) =>
      @tokens.splice(i, 1, token) if token[0] is 'OUTDENT' and prev[0] is ','
      return 1
#

We've tagged the opening parenthesis of a method call, and the opening bracket of an indexing operation. Match them with their close.

  close_open_calls_and_indexes: ->
    parens:   [0]
    brackets: [0]
    @scan_tokens (prev, token, post, i) =>
      switch token[0]
        when 'CALL_START'  then parens.push(0)
        when 'INDEX_START' then brackets.push(0)
        when '('           then parens[parens.length - 1] += 1
        when '['           then brackets[brackets.length - 1] += 1
        when ')'
          if parens[parens.length - 1] is 0
            parens.pop()
            token[0]: 'CALL_END'
          else
            parens[parens.length - 1] -= 1
        when ']'
          if brackets[brackets.length - 1] == 0
            brackets.pop()
            token[0]: 'INDEX_END'
          else
            brackets[brackets.length - 1] -= 1
      return 1
#

Methods may be optionally called without parentheses, for simple cases. Insert the implicit parentheses here, so that the parser doesn't have to deal with them.

  add_implicit_parentheses: ->
    stack: [0]
    @scan_tokens (prev, token, post, i) =>
      tag: token[0]
      stack.push(0) if tag is 'INDENT'
      if tag is 'OUTDENT'
        last: stack.pop()
        stack[stack.length - 1] += last
      if IMPLICIT_END.indexOf(tag) >= 0 or !post?
        return 1 if tag is 'INDENT' and prev and IMPLICIT_BLOCK.indexOf(prev[0]) >= 0
        if stack[stack.length - 1] > 0 or tag is 'INDENT'
          idx: if tag is 'OUTDENT' then i +1 else i
          stack_pointer: if tag is 'INDENT' then 2 else 1
          for tmp in [0...stack[stack.length - stack_pointer]]
            @tokens.splice(idx, 0, ['CALL_END', ')', token[2]])
          size: stack[stack.length - stack_pointer] + 1
          stack[stack.length - stack_pointer]: 0
          return size
      return 1 unless prev and IMPLICIT_FUNC.indexOf(prev[0]) >= 0 and IMPLICIT_CALL.indexOf(tag) >= 0
      @tokens.splice(i, 0, ['CALL_START', '(', token[2]])
      stack[stack.length - 1] += 1
      return 2
#

Because our grammar is LALR(1), it can't handle some single-line expressions that lack ending delimiters. Use the lexer to add the implicit blocks, so it doesn't need to. ')' can close a single-line block, but we need to make sure it's balanced.

  add_implicit_indentation: ->
    @scan_tokens (prev, token, post, i) =>
      return 1 unless SINGLE_LINERS.indexOf(token[0]) >= 0 and post[0] isnt 'INDENT' and
        not (token[0] is 'ELSE' and post[0] is 'IF')
      starter: token[0]
      @tokens.splice(i + 1, 0, ['INDENT', 2, token[2]])
      idx: i + 1
      parens: 0
      while true
        idx += 1
        tok: @tokens[idx]
        pre: @tokens[idx - 1]
        if (not tok or
            (SINGLE_CLOSERS.indexOf(tok[0]) >= 0 and tok[1] isnt ';') or
            (tok[0] is ')' && parens is 0)) and
            not (starter is 'ELSE' and tok[0] is 'ELSE')
          insertion: if pre[0] is "," then idx - 1 else idx
          @tokens.splice(insertion, 0, ['OUTDENT', 2, token[2]])
          break
        parens += 1 if tok[0] is '('
        parens -= 1 if tok[0] is ')'
      return 1 unless token[0] is 'THEN'
      @tokens.splice(i, 1)
      return 0
#

Ensure that all listed pairs of tokens are correctly balanced throughout the course of the token stream.

  ensure_balance: (pairs) ->
    levels: {}
    @scan_tokens (prev, token, post, i) =>
      for pair in pairs
        [open, close]: pair
        levels[open] ||= 0
        levels[open] += 1 if token[0] is open
        levels[open] -= 1 if token[0] is close
        throw new Error("too many ${token[1]} on line ${token[2] + 1}") if levels[open] < 0
      return 1
    unclosed: key for key, value of levels when value > 0
    throw new Error("unclosed ${unclosed[0]}") if unclosed.length
#

We'd like to support syntax like this: el.click((event) -> el.hide()) In order to accomplish this, move outdents that follow closing parens inwards, safely. The steps to accomplish this are:

  1. Check that all paired tokens are balanced and in order.
  2. Rewrite the stream with a stack: if you see an '(' or INDENT, add it to the stack. If you see an ')' or OUTDENT, pop the stack and replace it with the inverse of what we've just popped.
  3. Keep track of "debt" for tokens that we fake, to make sure we end up balanced in the end.
  rewrite_closing_parens: ->
    stack: []
    debt:  {}
    (debt[key]: 0) for key, val of INVERSES
    @scan_tokens (prev, token, post, i) =>
      tag: token[0]
      inv: INVERSES[token[0]]
#

Push openers onto the stack.

      if EXPRESSION_START.indexOf(tag) >= 0
        stack.push(token)
        return 1
#

The end of an expression, check stack and debt for a pair.

      else if EXPRESSION_TAIL.indexOf(tag) >= 0
#

If the tag is already in our debt, swallow it.

        if debt[inv] > 0
          debt[inv] -= 1
          @tokens.splice(i, 1)
          return 0
        else
#

Pop the stack of open delimiters.

          match: stack.pop()
          mtag:  match[0]
#

Continue onwards if it's the expected tag.

          if tag is INVERSES[mtag]
            return 1
          else
#

Unexpected close, insert correct close, adding to the debt.

            debt[mtag] += 1
            val: if mtag is 'INDENT' then match[1] else INVERSES[mtag]
            @tokens.splice(i, 0, [INVERSES[mtag], val])
            return 1
      else
        return 1