Python tokenize.ERRORTOKEN Examples
The following are 5
code examples of tokenize.ERRORTOKEN().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
tokenize
, or try the search function
.
Example #1
Source File: raw.py From linter-pylama with MIT License | 6 votes |
def _get_all_tokens(line, lines): '''Starting from *line*, generate the necessary tokens which represent the shortest tokenization possible. This is done by catching :exc:`tokenize.TokenError` when a multi-line string or statement is encountered. :returns: tokens, lines ''' buffer = line used_lines = [line] while True: try: tokens = _generate(buffer) except tokenize.TokenError: # A multi-line string or statement has been encountered: # start adding lines and stop when tokenize stops complaining pass else: if not any(t[0] == tokenize.ERRORTOKEN for t in tokens): return tokens, used_lines # Add another line next_line = next(lines) buffer = buffer + '\n' + next_line used_lines.append(next_line)
Example #2
Source File: evaluate.py From python_autocomplete with MIT License | 6 votes |
def __get_tokens(it): tokens: List[tokenize.TokenInfo] = [] try: for t in it: if t.type in tokenizer.SKIP_TOKENS: continue if t.type == tokenize.NEWLINE and t.string == '': continue if t.type == tokenize.DEDENT: continue if t.type == tokenize.ERRORTOKEN: continue tokens.append(t) except tokenize.TokenError as e: if not e.args[0].startswith('EOF in'): print(e) except IndentationError as e: print(e) return tokens
Example #3
Source File: tokens.py From vnpy_crypto with MIT License | 5 votes |
def python_tokenize(code): # Since formulas can only contain Python expressions, and Python # expressions cannot meaningfully contain newlines, we'll just remove all # the newlines up front to avoid any complications: code = code.replace("\n", " ").strip() it = tokenize.generate_tokens(StringIO(code).readline) try: for (pytype, string, (_, start), (_, end), code) in it: if pytype == tokenize.ENDMARKER: break origin = Origin(code, start, end) assert pytype not in (tokenize.NL, tokenize.NEWLINE) if pytype == tokenize.ERRORTOKEN: raise PatsyError("error tokenizing input " "(maybe an unclosed string?)", origin) if pytype == tokenize.COMMENT: raise PatsyError("comments are not allowed", origin) yield (pytype, string, origin) else: # pragma: no cover raise ValueError("stream ended without ENDMARKER?!?") except tokenize.TokenError as e: # TokenError is raised iff the tokenizer thinks that there is # some sort of multi-line construct in progress (e.g., an # unclosed parentheses, which in Python lets a virtual line # continue past the end of the physical line), and it hits the # end of the source text. We have our own error handling for # such cases, so just treat this as an end-of-stream. # # Just in case someone adds some other error case: assert e.args[0].startswith("EOF in multi-line") return
Example #4
Source File: config_parser.py From gin-config with Apache License 2.0 | 5 votes |
def _advance_one_token(self): self._current_token = ConfigParser.Token(*next(self._token_generator)) # Certain symbols (e.g., "$") cause ERRORTOKENs on all preceding space # characters. Find the first non-space or non-ERRORTOKEN token. while (self._current_token.kind == tokenize.ERRORTOKEN and self._current_token.value in ' \t'): self._current_token = ConfigParser.Token(*next(self._token_generator))
Example #5
Source File: tokens.py From Splunking-Crime with GNU Affero General Public License v3.0 | 5 votes |
def python_tokenize(code): # Since formulas can only contain Python expressions, and Python # expressions cannot meaningfully contain newlines, we'll just remove all # the newlines up front to avoid any complications: code = code.replace("\n", " ").strip() it = tokenize.generate_tokens(StringIO(code).readline) try: for (pytype, string, (_, start), (_, end), code) in it: if pytype == tokenize.ENDMARKER: break origin = Origin(code, start, end) assert pytype not in (tokenize.NL, tokenize.NEWLINE) if pytype == tokenize.ERRORTOKEN: raise PatsyError("error tokenizing input " "(maybe an unclosed string?)", origin) if pytype == tokenize.COMMENT: raise PatsyError("comments are not allowed", origin) yield (pytype, string, origin) else: # pragma: no cover raise ValueError("stream ended without ENDMARKER?!?") except tokenize.TokenError as e: # TokenError is raised iff the tokenizer thinks that there is # some sort of multi-line construct in progress (e.g., an # unclosed parentheses, which in Python lets a virtual line # continue past the end of the physical line), and it hits the # end of the source text. We have our own error handling for # such cases, so just treat this as an end-of-stream. # # Just in case someone adds some other error case: assert e.args[0].startswith("EOF in multi-line") return