Mercurial > yakumo_izuru > aya
diff vendor/github.com/alecthomas/chroma/v2/lexers/embedded/coq.xml @ 66:787b5ee0289d draft
Use vendored modules
Signed-off-by: Izuru Yakumo <yakumo.izuru@chaotic.ninja>
author | yakumo.izuru |
---|---|
date | Sun, 23 Jul 2023 13:18:53 +0000 |
parents | |
children |
line wrap: on
line diff
--- /dev/null Thu Jan 01 00:00:00 1970 +0000 +++ b/vendor/github.com/alecthomas/chroma/v2/lexers/embedded/coq.xml Sun Jul 23 13:18:53 2023 +0000 @@ -0,0 +1,136 @@ +<lexer> + <config> + <name>Coq</name> + <alias>coq</alias> + <filename>*.v</filename> + <mime_type>text/x-coq</mime_type> + </config> + <rules> + <state name="string"> + <rule pattern="[^"]+"> + <token type="LiteralStringDouble"/> + </rule> + <rule pattern=""""> + <token type="LiteralStringDouble"/> + </rule> + <rule pattern="""> + <token type="LiteralStringDouble"/> + <pop depth="1"/> + </rule> + </state> + <state name="dotted"> + <rule pattern="\s+"> + <token type="Text"/> + </rule> + <rule pattern="\."> + <token type="Punctuation"/> + </rule> + <rule pattern="[A-Z][\w\']*(?=\s*\.)"> + <token type="NameNamespace"/> + </rule> + <rule pattern="[A-Z][\w\']*"> + <token type="NameClass"/> + <pop depth="1"/> + </rule> + <rule pattern="[a-z][a-z0-9_\']*"> + <token type="Name"/> + <pop depth="1"/> + </rule> + <rule> + <pop depth="1"/> + </rule> + </state> + <state name="root"> + <rule pattern="\s+"> + <token type="Text"/> + </rule> + <rule pattern="false|true|\(\)|\[\]"> + <token type="NameBuiltinPseudo"/> + </rule> + <rule pattern="\(\*"> + <token type="Comment"/> + <push state="comment"/> + </rule> + <rule pattern="\b(Projections|Monomorphic|Polymorphic|Proposition|CoInductive|Hypothesis|CoFixpoint|Contextual|Definition|Parameters|Hypotheses|Structure|Inductive|Corollary|Implicits|Parameter|Variables|Arguments|Canonical|Printing|Coercion|Reserved|Universe|Notation|Instance|Fixpoint|Variable|Morphism|Relation|Existing|Implicit|Example|Theorem|Delimit|Defined|Rewrite|outside|Require|Resolve|Section|Context|Prenex|Strict|Module|Import|Export|Global|inside|Remark|Tactic|Search|Record|Scope|Unset|Check|Local|Close|Class|Graph|Proof|Lemma|Print|Axiom|Show|Goal|Open|Fact|Hint|Bind|Ltac|Save|View|Let|Set|All|End|Qed)\b"> + <token type="KeywordNamespace"/> + </rule> + <rule pattern="\b(exists2|nosimpl|struct|exists|return|forall|match|cofix|then|with|else|for|fix|let|fun|end|is|of|if|in|as)\b"> + <token type="Keyword"/> + </rule> + <rule pattern="\b(Type|Prop)\b"> + <token type="KeywordType"/> + </rule> + <rule pattern="\b(native_compute|setoid_rewrite|etransitivity|econstructor|transitivity|autorewrite|constructor|cutrewrite|vm_compute|bool_congr|generalize|inversion|induction|injection|nat_congr|intuition|destruct|suffices|erewrite|symmetry|nat_norm|replace|rewrite|compute|pattern|trivial|without|assert|unfold|change|eapply|intros|unlock|revert|rename|refine|eauto|tauto|after|right|congr|split|field|simpl|intro|clear|apply|using|subst|case|left|suff|loss|wlog|have|fold|ring|move|lazy|elim|pose|auto|red|cbv|hnf|cut|set)\b"> + <token type="Keyword"/> + </rule> + <rule pattern="\b(contradiction|discriminate|reflexivity|assumption|congruence|romega|omega|exact|solve|tauto|done|by)\b"> + <token type="KeywordPseudo"/> + </rule> + <rule pattern="\b(repeat|first|idtac|last|try|do)\b"> + <token type="KeywordReserved"/> + </rule> + <rule pattern="\b([A-Z][\w\']*)"> + <token type="Name"/> + </rule> + <rule pattern="(λ|Π|\|\}|\{\||\\/|/\\|=>|~|\}|\|]|\||\{<|\{|`|_|]|\[\||\[>|\[<|\[|\?\?|\?|>\}|>]|>|=|<->|<-|<|;;|;|:>|:=|::|:|\.\.|\.|->|-\.|-|,|\+|\*|\)|\(|&&|&|#|!=)"> + <token type="Operator"/> + </rule> + <rule pattern="([=<>@^|&+\*/$%-]|[!?~])?[!$%&*+\./:<=>?@^|~-]"> + <token type="Operator"/> + </rule> + <rule pattern="\b(unit|nat|bool|string|ascii|list)\b"> + <token type="KeywordType"/> + </rule> + <rule pattern="[^\W\d][\w']*"> + <token type="Name"/> + </rule> + <rule pattern="\d[\d_]*"> + <token type="LiteralNumberInteger"/> + </rule> + <rule pattern="0[xX][\da-fA-F][\da-fA-F_]*"> + <token type="LiteralNumberHex"/> + </rule> + <rule pattern="0[oO][0-7][0-7_]*"> + <token type="LiteralNumberOct"/> + </rule> + <rule pattern="0[bB][01][01_]*"> + <token type="LiteralNumberBin"/> + </rule> + <rule pattern="-?\d[\d_]*(.[\d_]*)?([eE][+\-]?\d[\d_]*)"> + <token type="LiteralNumberFloat"/> + </rule> + <rule pattern="'(?:(\\[\\\"'ntbr ])|(\\[0-9]{3})|(\\x[0-9a-fA-F]{2}))'"> + <token type="LiteralStringChar"/> + </rule> + <rule pattern="'.'"> + <token type="LiteralStringChar"/> + </rule> + <rule pattern="'"> + <token type="Keyword"/> + </rule> + <rule pattern="""> + <token type="LiteralStringDouble"/> + <push state="string"/> + </rule> + <rule pattern="[~?][a-z][\w\']*:"> + <token type="Name"/> + </rule> + </state> + <state name="comment"> + <rule pattern="[^(*)]+"> + <token type="Comment"/> + </rule> + <rule pattern="\(\*"> + <token type="Comment"/> + <push/> + </rule> + <rule pattern="\*\)"> + <token type="Comment"/> + <pop depth="1"/> + </rule> + <rule pattern="[(*)]"> + <token type="Comment"/> + </rule> + </state> + </rules> +</lexer> \ No newline at end of file