mirror of
https://github.com/fish-shell/fish-shell
synced 2024-11-10 07:04:29 +00:00
557d8b0334
Keywords and options recently got dedicated highlighting roles inb3626d48e
(Highlight keywords differently, 2021-02-04) and711796ad1
(Highlight options differently, 2021-10-19) but still default to "command" and "parameter", respectively. The dedicated roles were not colored by our CSS theme, which makes a "test -f foo.txt" look weird: - "test" is dark blue (since it's a command) - "foo.txt" is light blue (since it's a parameter) - "-f" is black (weird!) The CSS theme doesn't support configuration, so the dedicated highlighting roles should always default to their fallback options. Make it so.
130 lines
4.1 KiB
Python
130 lines
4.1 KiB
Python
# This is a plugin for pygments that shells out to fish_indent.
|
|
|
|
# Example of how to use this:
|
|
# env PATH="/dir/containing/fish/indent/:$PATH" pygmentize -f terminal256 -l /path/to/fish_indent_lexer.py:FishIndentLexer -x ~/test.fish
|
|
|
|
from pygments.lexer import Lexer
|
|
from pygments.token import (
|
|
Name,
|
|
Comment,
|
|
String,
|
|
Error,
|
|
Number,
|
|
Operator,
|
|
Other,
|
|
Generic,
|
|
Whitespace,
|
|
String,
|
|
Text,
|
|
Punctuation,
|
|
)
|
|
import re
|
|
import subprocess
|
|
|
|
# A fallback token type.
|
|
DEFAULT = Text
|
|
|
|
# Mapping from fish token types to Pygments types.
|
|
ROLE_TO_TOKEN = {
|
|
"normal": Name.Variable,
|
|
"error": Generic.Error,
|
|
"command": Name.Function,
|
|
"keyword": Name.Function,
|
|
"statement_terminator": Punctuation,
|
|
"param": Name.Constant,
|
|
"option": Name.Constant,
|
|
"comment": Comment,
|
|
"match": DEFAULT,
|
|
"search_match": DEFAULT,
|
|
"operat": Operator,
|
|
"escape": String.Escape,
|
|
"quote": String.Single, # note, may be changed to double dynamically
|
|
"redirection": Punctuation, # ?
|
|
"autosuggestion": Other, # in practice won't be generated
|
|
"selection": DEFAULT,
|
|
"pager_progress": DEFAULT,
|
|
"pager_background": DEFAULT,
|
|
"pager_prefix": DEFAULT,
|
|
"pager_completion": DEFAULT,
|
|
"pager_description": DEFAULT,
|
|
"pager_secondary_background": DEFAULT,
|
|
"pager_secondary_prefix": DEFAULT,
|
|
"pager_secondary_completion": DEFAULT,
|
|
"pager_secondary_description": DEFAULT,
|
|
"pager_selected_background": DEFAULT,
|
|
"pager_selected_prefix": DEFAULT,
|
|
"pager_selected_completion": DEFAULT,
|
|
"pager_selected_description": DEFAULT,
|
|
}
|
|
|
|
|
|
def token_for_text_and_role(text, role):
|
|
"""Return the pygments token for some input text and a fish role
|
|
|
|
This applies any special cases of ROLE_TO_TOKEN.
|
|
"""
|
|
if text.isspace():
|
|
# Here fish will return 'normal' or 'statement_terminator' for newline.
|
|
return Text.Whitespace
|
|
elif role == "quote":
|
|
# Check for single or double.
|
|
return String.Single if text.startswith("'") else String.Double
|
|
else:
|
|
return ROLE_TO_TOKEN[role]
|
|
|
|
|
|
def tokenize_fish_command(code, offset):
|
|
"""Tokenize some fish code, offset in a parent string, by shelling
|
|
out to fish_indent.
|
|
|
|
fish_indent will output a list of csv lines: start,end,type.
|
|
|
|
This function returns a list of (start, tok, value) tuples, as
|
|
Pygments expects.
|
|
"""
|
|
proc = subprocess.Popen(
|
|
["fish_indent", "--pygments"],
|
|
stdin=subprocess.PIPE,
|
|
stdout=subprocess.PIPE,
|
|
universal_newlines=False,
|
|
)
|
|
stdout, _ = proc.communicate(code.encode("utf-8"))
|
|
result = []
|
|
for line in stdout.decode("utf-8").splitlines():
|
|
start, end, role = line.split(",")
|
|
start, end = int(start), int(end)
|
|
value = code[start:end]
|
|
tok = token_for_text_and_role(value, role)
|
|
result.append((start + offset, tok, value))
|
|
return result
|
|
|
|
|
|
class FishIndentLexer(Lexer):
|
|
name = "FishIndentLexer"
|
|
aliases = ["fish", "fish-docs-samples"]
|
|
filenames = ["*.fish"]
|
|
|
|
def get_tokens_unprocessed(self, input_text):
|
|
"""Return a list of (start, tok, value) tuples.
|
|
|
|
start is the index into the string
|
|
tok is the token type (as above)
|
|
value is the string contents of the token
|
|
"""
|
|
result = []
|
|
if not any(s.startswith(">") for s in input_text.splitlines()):
|
|
# No prompt, just tokenize everything.
|
|
result = tokenize_fish_command(input_text, 0)
|
|
else:
|
|
# We have a prompt line.
|
|
# Use a regexp because it will maintain string indexes for us.
|
|
regex = re.compile(r"^(>_?\s*)?(.*\n?)", re.MULTILINE)
|
|
for m in regex.finditer(input_text):
|
|
if m.group(1):
|
|
# Prompt line; highlight via fish syntax.
|
|
result.append((m.start(1), Generic.Prompt, m.group(1)))
|
|
result.extend(tokenize_fish_command(m.group(2), m.start(2)))
|
|
else:
|
|
# Non-prompt line representing output from a command.
|
|
result.append((m.start(2), Generic.Output, m.group(2)))
|
|
return result
|