forked from Isabelle_DOF/Isabelle_DOF
571 lines
20 KiB
Standard ML
571 lines
20 KiB
Standard ML
(* Title: Pure/Thy/thy_output.ML
|
|
Author: Makarius
|
|
|
|
Theory document output.
|
|
*)
|
|
|
|
signature THY_OUTPUT =
|
|
sig
|
|
val output_document: Proof.context -> {markdown: bool} -> Input.source -> Latex.text list
|
|
val check_comments: Proof.context -> Symbol_Pos.T list -> unit
|
|
val output_token: Proof.context -> Token.T -> Latex.text list
|
|
val output_source: Proof.context -> string -> Latex.text list
|
|
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state}
|
|
val present_thy: Options.T -> theory -> segment list -> Latex.text list
|
|
val set_meta_args_parser : (theory -> string parser) -> unit
|
|
val pretty_term: Proof.context -> term -> Pretty.T
|
|
val pretty_thm: Proof.context -> thm -> Pretty.T
|
|
val lines: Latex.text list -> Latex.text list
|
|
val items: Latex.text list -> Latex.text list
|
|
val isabelle: Proof.context -> Latex.text list -> Latex.text
|
|
val isabelle_typewriter: Proof.context -> Latex.text list -> Latex.text
|
|
val typewriter: Proof.context -> string -> Latex.text
|
|
val verbatim: Proof.context -> string -> Latex.text
|
|
val source: Proof.context -> Token.src -> Latex.text
|
|
val pretty: Proof.context -> Pretty.T -> Latex.text
|
|
val pretty_source: Proof.context -> Token.src -> Pretty.T -> Latex.text
|
|
val pretty_items: Proof.context -> Pretty.T list -> Latex.text
|
|
val pretty_items_source: Proof.context -> Token.src -> Pretty.T list -> Latex.text
|
|
val antiquotation_pretty:
|
|
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
|
val antiquotation_pretty_source:
|
|
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
|
val antiquotation_raw:
|
|
binding -> 'a context_parser -> (Proof.context -> 'a -> Latex.text) -> theory -> theory
|
|
val antiquotation_verbatim:
|
|
binding -> 'a context_parser -> (Proof.context -> 'a -> string) -> theory -> theory
|
|
end;
|
|
|
|
structure Thy_Output: THY_OUTPUT =
|
|
struct
|
|
|
|
(* output document source *)
|
|
|
|
val output_symbols = single o Latex.symbols_output;
|
|
|
|
fun output_comment ctxt (kind, syms) =
|
|
(case kind of
|
|
Comment.Comment =>
|
|
Input.cartouche_content syms
|
|
|> output_document (ctxt |> Config.put Document_Antiquotation.thy_output_display false)
|
|
{markdown = false}
|
|
|> Latex.enclose_body "%\n\\isamarkupcmt{" "%\n}"
|
|
| Comment.Cancel =>
|
|
Symbol_Pos.cartouche_content syms
|
|
|> output_symbols
|
|
|> Latex.enclose_body "%\n\\isamarkupcancel{" "}"
|
|
| Comment.Latex =>
|
|
[Latex.symbols (Symbol_Pos.cartouche_content syms)])
|
|
and output_comment_document ctxt (comment, syms) =
|
|
(case comment of
|
|
SOME kind => output_comment ctxt (kind, syms)
|
|
| NONE => [Latex.symbols syms])
|
|
and output_document_text ctxt syms =
|
|
Comment.read_body syms |> maps (output_comment_document ctxt)
|
|
and output_document ctxt {markdown} source =
|
|
let
|
|
val pos = Input.pos_of source;
|
|
val syms = Input.source_explode source;
|
|
|
|
val output_antiquotes =
|
|
maps (Document_Antiquotation.evaluate (output_document_text ctxt) ctxt);
|
|
|
|
fun output_line line =
|
|
(if Markdown.line_is_item line then [Latex.string "\\item "] else []) @
|
|
output_antiquotes (Markdown.line_content line);
|
|
|
|
fun output_block (Markdown.Par lines) =
|
|
Latex.block (separate (Latex.string "\n") (map (Latex.block o output_line) lines))
|
|
| output_block (Markdown.List {kind, body, ...}) =
|
|
Latex.environment_block (Markdown.print_kind kind) (output_blocks body)
|
|
and output_blocks blocks = separate (Latex.string "\n\n") (map output_block blocks);
|
|
in
|
|
if Toplevel.is_skipped_proof (Toplevel.presentation_state ctxt) then []
|
|
else if markdown andalso exists (Markdown.is_control o Symbol_Pos.symbol) syms
|
|
then
|
|
let
|
|
val ants = Antiquote.parse_comments pos syms;
|
|
val reports = Antiquote.antiq_reports ants;
|
|
val blocks = Markdown.read_antiquotes ants;
|
|
val _ = Context_Position.reports ctxt (reports @ Markdown.reports blocks);
|
|
in output_blocks blocks end
|
|
else
|
|
let
|
|
val ants = Antiquote.parse_comments pos (trim (Symbol.is_blank o Symbol_Pos.symbol) syms);
|
|
val reports = Antiquote.antiq_reports ants;
|
|
val _ = Context_Position.reports ctxt (reports @ Markdown.text_reports ants);
|
|
in output_antiquotes ants end
|
|
end;
|
|
|
|
|
|
(* output tokens with formal comments *)
|
|
|
|
local
|
|
|
|
val output_symbols_antiq =
|
|
(fn Antiquote.Text syms => output_symbols syms
|
|
| Antiquote.Control {name = (name, _), body, ...} =>
|
|
Latex.string (Latex.output_symbols [Symbol.encode (Symbol.Control name)]) ::
|
|
output_symbols body
|
|
| Antiquote.Antiq {body, ...} =>
|
|
Latex.enclose_body "%\n\\isaantiq\n" "{}%\n\\endisaantiq\n" (output_symbols body));
|
|
|
|
fun output_comment_symbols ctxt {antiq} (comment, syms) =
|
|
(case (comment, antiq) of
|
|
(NONE, false) => output_symbols syms
|
|
| (NONE, true) =>
|
|
Antiquote.parse_comments (#1 (Symbol_Pos.range syms)) syms
|
|
|> maps output_symbols_antiq
|
|
| (SOME comment, _) => output_comment ctxt (comment, syms));
|
|
|
|
fun output_body ctxt antiq bg en syms =
|
|
Comment.read_body syms
|
|
|> maps (output_comment_symbols ctxt {antiq = antiq})
|
|
|> Latex.enclose_body bg en;
|
|
|
|
in
|
|
|
|
fun output_token ctxt tok =
|
|
let
|
|
fun output antiq bg en =
|
|
output_body ctxt antiq bg en (Input.source_explode (Token.input_of tok));
|
|
in
|
|
(case Token.kind_of tok of
|
|
Token.Comment NONE => []
|
|
| Token.Command => output false "\\isacommand{" "}"
|
|
| Token.Keyword =>
|
|
if Symbol.is_ascii_identifier (Token.content_of tok)
|
|
then output false "\\isakeyword{" "}"
|
|
else output false "" ""
|
|
| Token.String => output false "{\\isachardoublequoteopen}" "{\\isachardoublequoteclose}"
|
|
| Token.Alt_String => output false "{\\isacharbackquoteopen}" "{\\isacharbackquoteclose}"
|
|
| Token.Verbatim => output true "{\\isacharverbatimopen}" "{\\isacharverbatimclose}"
|
|
| Token.Cartouche => output false "{\\isacartoucheopen}" "{\\isacartoucheclose}"
|
|
| _ => output false "" "")
|
|
end handle ERROR msg => error (msg ^ Position.here (Token.pos_of tok));
|
|
|
|
fun output_source ctxt s =
|
|
output_body ctxt false "" "" (Symbol_Pos.explode (s, Position.none));
|
|
|
|
fun check_comments ctxt =
|
|
Comment.read_body #> List.app (fn (comment, syms) =>
|
|
let
|
|
val pos = #1 (Symbol_Pos.range syms);
|
|
val _ =
|
|
comment |> Option.app (fn kind =>
|
|
Context_Position.reports ctxt (map (pair pos) (Markup.cartouche :: Comment.markups kind)));
|
|
val _ = output_comment_symbols ctxt {antiq = false} (comment, syms);
|
|
in if comment = SOME Comment.Comment then check_comments ctxt syms else () end);
|
|
|
|
end;
|
|
|
|
|
|
|
|
(** present theory source **)
|
|
|
|
(*NB: arranging white space around command spans is a black art*)
|
|
|
|
val is_white = Token.is_space orf Token.is_informal_comment;
|
|
val is_black = not o is_white;
|
|
|
|
val is_white_comment = Token.is_informal_comment;
|
|
val is_black_comment = Token.is_formal_comment;
|
|
|
|
|
|
(* presentation tokens *)
|
|
|
|
datatype token =
|
|
Ignore_Token
|
|
| Basic_Token of Token.T
|
|
| Markup_Token of string * string * Input.source
|
|
| Markup_Env_Token of string * string * Input.source
|
|
| Raw_Token of Input.source;
|
|
|
|
fun basic_token pred (Basic_Token tok) = pred tok
|
|
| basic_token _ _ = false;
|
|
|
|
val white_token = basic_token is_white;
|
|
val white_comment_token = basic_token is_white_comment;
|
|
val blank_token = basic_token Token.is_blank;
|
|
val newline_token = basic_token Token.is_newline;
|
|
|
|
fun present_token ctxt tok =
|
|
(case tok of
|
|
Ignore_Token => []
|
|
| Basic_Token tok => output_token ctxt tok
|
|
| Markup_Token (cmd, meta_args, source) =>
|
|
Latex.enclose_body ("%\n\\isamarkup" ^ cmd (* ^ meta_args *) ^ "{") "%\n}\n"
|
|
(output_document ctxt {markdown = false} source)
|
|
| Markup_Env_Token (cmd, meta_args, source) =>
|
|
[Latex.environment_block
|
|
("isamarkup" ^ cmd (* ^ meta_args*))
|
|
(Latex.string meta_args :: output_document ctxt {markdown = true} source)]
|
|
| Raw_Token source =>
|
|
Latex.string "%\n" :: output_document ctxt {markdown = true} source @ [Latex.string "\n"]);
|
|
|
|
|
|
(* command spans *)
|
|
|
|
type command = string * Position.T * string list; (*name, position, tags*)
|
|
type source = (token * (string * int)) list; (*token, markup flag, meta-comment depth*)
|
|
|
|
datatype span = Span of command * (source * source * source * source) * bool;
|
|
|
|
fun make_span cmd src =
|
|
let
|
|
fun chop_newline (tok :: toks) =
|
|
if newline_token (fst tok) then ([tok], toks, true)
|
|
else ([], tok :: toks, false)
|
|
| chop_newline [] = ([], [], false);
|
|
val (((src_prefix, src_main), src_suffix1), (src_suffix2, src_appendix, newline)) =
|
|
src
|
|
|> chop_prefix (white_token o fst)
|
|
||>> chop_suffix (white_token o fst)
|
|
||>> chop_prefix (white_comment_token o fst)
|
|
||> chop_newline;
|
|
in Span (cmd, (src_prefix, src_main, src_suffix1 @ src_suffix2, src_appendix), newline) end;
|
|
|
|
|
|
(* present spans *)
|
|
|
|
local
|
|
|
|
fun err_bad_nesting pos =
|
|
error ("Bad nesting of commands in presentation" ^ pos);
|
|
|
|
fun edge which f (x: string option, y) =
|
|
if x = y then I
|
|
else (case which (x, y) of NONE => I | SOME txt => cons (Latex.string (f txt)));
|
|
|
|
val begin_tag = edge #2 Latex.begin_tag;
|
|
val end_tag = edge #1 Latex.end_tag;
|
|
fun open_delim delim e = edge #2 Latex.begin_delim e #> delim #> edge #2 Latex.end_delim e;
|
|
fun close_delim delim e = edge #1 Latex.begin_delim e #> delim #> edge #1 Latex.end_delim e;
|
|
|
|
fun read_tag s =
|
|
(case space_explode "%" s of
|
|
["", b] => (SOME b, NONE)
|
|
| [a, b] => (NONE, SOME (a, b))
|
|
| _ => error ("Bad document_tags specification: " ^ quote s));
|
|
|
|
in
|
|
|
|
fun make_command_tag options keywords =
|
|
let
|
|
val document_tags =
|
|
map read_tag (space_explode "," (Options.string options \<^system_option>\<open>document_tags\<close>));
|
|
val document_tags_default = map_filter #1 document_tags;
|
|
val document_tags_command = map_filter #2 document_tags;
|
|
in
|
|
fn {cmd_name, cmd_tags, tag, active_tag} => fn state => fn state' =>
|
|
let
|
|
val tag' = try hd (fold (update (op =)) cmd_tags (the_list tag));
|
|
|
|
val keyword_tags =
|
|
if cmd_name = "end" andalso Toplevel.is_end_theory state' then ["theory"]
|
|
else Keyword.command_tags keywords cmd_name;
|
|
val command_tags =
|
|
the_list (AList.lookup (op =) document_tags_command cmd_name) @
|
|
keyword_tags @ document_tags_default;
|
|
|
|
val active_tag' =
|
|
if is_some tag' then tag'
|
|
else
|
|
(case command_tags of
|
|
default_tag :: _ => SOME default_tag
|
|
| [] =>
|
|
if Keyword.is_vacuous keywords cmd_name andalso Toplevel.is_proof state
|
|
then active_tag
|
|
else NONE);
|
|
in {tag' = tag', active_tag' = active_tag'} end
|
|
end;
|
|
|
|
fun present_span thy command_tag span state state'
|
|
(tag_stack, active_tag, newline, latex, present_cont) =
|
|
let
|
|
val ctxt' =
|
|
Toplevel.presentation_context state'
|
|
handle Toplevel.UNDEF => Proof_Context.get_global thy Context.PureN;
|
|
val present = fold (fn (tok, (flag, 0)) =>
|
|
fold cons (present_token ctxt' tok)
|
|
#> cons (Latex.string flag)
|
|
| _ => I);
|
|
|
|
val Span ((cmd_name, cmd_pos, cmd_tags), srcs, span_newline) = span;
|
|
|
|
val (tag, tags) = tag_stack;
|
|
val {tag', active_tag'} =
|
|
command_tag {cmd_name = cmd_name, cmd_tags = cmd_tags, tag = tag, active_tag = active_tag}
|
|
state state';
|
|
val edge = (active_tag, active_tag');
|
|
|
|
val nesting = Toplevel.level state' - Toplevel.level state;
|
|
|
|
val newline' =
|
|
if is_none active_tag' then span_newline else newline;
|
|
|
|
val tag_stack' =
|
|
if nesting = 0 andalso not (Toplevel.is_proof state) then tag_stack
|
|
else if nesting >= 0 then (tag', replicate nesting tag @ tags)
|
|
else
|
|
(case drop (~ nesting - 1) tags of
|
|
tg :: tgs => (tg, tgs)
|
|
| [] => err_bad_nesting (Position.here cmd_pos));
|
|
|
|
val latex' =
|
|
latex
|
|
|> end_tag edge
|
|
|> close_delim (fst present_cont) edge
|
|
|> snd present_cont
|
|
|> open_delim (present (#1 srcs)) edge
|
|
|> begin_tag edge
|
|
|> present (#2 srcs);
|
|
val present_cont' =
|
|
if newline then (present (#3 srcs), present (#4 srcs))
|
|
else (I, present (#3 srcs) #> present (#4 srcs));
|
|
in (tag_stack', active_tag', newline', latex', present_cont') end;
|
|
|
|
fun present_trailer ((_, tags), active_tag, _, latex, present_cont) =
|
|
if not (null tags) then err_bad_nesting " at end of theory"
|
|
else
|
|
latex
|
|
|> end_tag (active_tag, NONE)
|
|
|> close_delim (fst present_cont) (active_tag, NONE)
|
|
|> snd present_cont;
|
|
|
|
end;
|
|
|
|
|
|
(* present_thy *)
|
|
|
|
local
|
|
|
|
val markup_true = "\\isamarkuptrue%\n";
|
|
val markup_false = "\\isamarkupfalse%\n";
|
|
|
|
val space_proper =
|
|
Scan.one Token.is_blank -- Scan.many is_white_comment -- Scan.one is_black;
|
|
|
|
val is_improper = not o (is_black orf Token.is_begin_ignore orf Token.is_end_ignore);
|
|
val improper = Scan.many is_improper;
|
|
val improper_end = Scan.repeat (Scan.unless space_proper (Scan.one is_improper));
|
|
val blank_end = Scan.repeat (Scan.unless space_proper (Scan.one Token.is_blank));
|
|
|
|
val opt_newline = Scan.option (Scan.one Token.is_newline);
|
|
|
|
val ignore =
|
|
Scan.depend (fn d => opt_newline |-- Scan.one Token.is_begin_ignore
|
|
>> pair (d + 1)) ||
|
|
Scan.depend (fn d => Scan.one Token.is_end_ignore --|
|
|
(if d = 0 then Scan.fail_with (K (fn () => "Bad nesting of meta-comments")) else opt_newline)
|
|
>> pair (d - 1));
|
|
|
|
val tag = (improper -- Parse.$$$ "%" -- improper) |-- Parse.!!! (Parse.tag_name --| blank_end);
|
|
|
|
val locale =
|
|
Scan.option ((Parse.$$$ "(" -- improper -- Parse.$$$ "in") |--
|
|
Parse.!!! (improper |-- Parse.name --| (improper -- Parse.$$$ ")")));
|
|
|
|
(*
|
|
val meta_args_parser_hook = Synchronized.var "meta args parser hook"
|
|
((fn thy => fn s => ("",s)): theory -> string parser);
|
|
*)
|
|
val meta_args_parser_hook = Unsynchronized.ref
|
|
((fn thy => fn s => ("",s)): theory -> string parser);
|
|
|
|
|
|
in
|
|
|
|
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state};
|
|
|
|
fun present_thy options thy (segments: segment list) =
|
|
let
|
|
|
|
val keywords = Thy_Header.get_keywords thy;
|
|
|
|
|
|
(* tokens *)
|
|
|
|
val ignored = Scan.state --| ignore
|
|
>> (fn d => (NONE, (Ignore_Token, ("", d))));
|
|
|
|
fun markup pred mk flag = Scan.peek (fn d =>
|
|
improper |--
|
|
Parse.position (Scan.one (fn tok =>
|
|
Token.is_command tok andalso pred keywords (Token.content_of tok))) --
|
|
Scan.repeat tag --
|
|
(improper |--
|
|
(Parse.!!!!
|
|
( (!meta_args_parser_hook thy)
|
|
-- ( (improper -- locale -- improper)
|
|
|-- (Parse.document_source))
|
|
--| improper_end)))
|
|
>> (fn (((tok, pos'), tags), (meta_args, source)) =>
|
|
let val name = Token.content_of tok
|
|
in (SOME (name, pos', tags), (mk (name, meta_args, source), (flag, d))) end));
|
|
|
|
val command = Scan.peek (fn d =>
|
|
Scan.optional (Scan.one Token.is_command_modifier ::: improper) [] --
|
|
Scan.one Token.is_command -- Scan.repeat tag
|
|
>> (fn ((cmd_mod, cmd), tags) =>
|
|
map (fn tok => (NONE, (Basic_Token tok, ("", d)))) cmd_mod @
|
|
[(SOME (Token.content_of cmd, Token.pos_of cmd, tags),
|
|
(Basic_Token cmd, (markup_false, d)))]));
|
|
|
|
val cmt = Scan.peek (fn d =>
|
|
Scan.one is_black_comment >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
|
|
|
val other = Scan.peek (fn d =>
|
|
Parse.not_eof >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
|
|
|
val tokens =
|
|
(ignored ||
|
|
markup Keyword.is_document_heading Markup_Token markup_true ||
|
|
markup Keyword.is_document_body Markup_Env_Token markup_true ||
|
|
markup Keyword.is_document_raw (Raw_Token o #3) "") >> single ||
|
|
command ||
|
|
(cmt || other) >> single;
|
|
|
|
|
|
(* spans *)
|
|
|
|
val is_eof = fn (_, (Basic_Token x, _)) => Token.is_eof x | _ => false;
|
|
val stopper = Scan.stopper (K (NONE, (Basic_Token Token.eof, ("", 0)))) is_eof;
|
|
|
|
val cmd = Scan.one (is_some o fst);
|
|
val non_cmd = Scan.one (is_none o fst andf not o is_eof) >> #2;
|
|
|
|
val white_comments = Scan.many (white_comment_token o fst o snd);
|
|
val blank = Scan.one (blank_token o fst o snd);
|
|
val newline = Scan.one (newline_token o fst o snd);
|
|
val before_cmd =
|
|
Scan.option (newline -- white_comments) --
|
|
Scan.option (newline -- white_comments) --
|
|
Scan.option (blank -- white_comments) -- cmd;
|
|
|
|
val span =
|
|
Scan.repeat non_cmd -- cmd --
|
|
Scan.repeat (Scan.unless before_cmd non_cmd) --
|
|
Scan.option (newline >> (single o snd))
|
|
>> (fn (((toks1, (cmd, tok2)), toks3), tok4) =>
|
|
make_span (the cmd) (toks1 @ (tok2 :: (toks3 @ the_default [] tok4))));
|
|
|
|
val spans = segments
|
|
|> maps (Command_Span.content o #span)
|
|
|> drop_suffix Token.is_space
|
|
|> Source.of_list
|
|
|> Source.source' 0 Token.stopper (Scan.error (Scan.bulk tokens >> flat))
|
|
|> Source.source stopper (Scan.error (Scan.bulk span))
|
|
|> Source.exhaust;
|
|
|
|
val command_results =
|
|
segments |> map_filter (fn {command, state, ...} =>
|
|
if Toplevel.is_ignored command then NONE else SOME (command, state));
|
|
|
|
|
|
(* present commands *)
|
|
|
|
val command_tag = make_command_tag options keywords;
|
|
|
|
fun present_command tr span st st' =
|
|
Toplevel.setmp_thread_position tr (present_span thy command_tag span st st');
|
|
|
|
fun present _ [] = I
|
|
| present st ((span, (tr, st')) :: rest) = present_command tr span st st'
|
|
#> present st' rest;
|
|
|
|
in
|
|
if length command_results = length spans then
|
|
((NONE, []), NONE, true, [], (I, I))
|
|
|> present Toplevel.toplevel (spans ~~ command_results)
|
|
|> present_trailer
|
|
|> rev
|
|
else error "Messed-up outer syntax for presentation"
|
|
end;
|
|
|
|
(*
|
|
fun set_meta_args_parser f = let val _ = writeln "Meta-args parser set to new value"
|
|
in (Unsynchronized.assign meta_args_parser_hook f) end
|
|
*)
|
|
fun set_meta_args_parser f = let val _ = writeln "Meta-args parser set to new value"
|
|
in ( meta_args_parser_hook := f) end
|
|
|
|
|
|
end;
|
|
|
|
|
|
|
|
(** standard output operations **)
|
|
|
|
(* pretty printing *)
|
|
|
|
fun pretty_term ctxt t =
|
|
Syntax.pretty_term (Variable.auto_fixes t ctxt) t;
|
|
|
|
fun pretty_thm ctxt = pretty_term ctxt o Thm.full_prop_of;
|
|
|
|
|
|
(* default output *)
|
|
|
|
val lines = separate (Latex.string "\\isanewline%\n");
|
|
val items = separate (Latex.string "\\isasep\\isanewline%\n");
|
|
|
|
fun isabelle ctxt body =
|
|
if Config.get ctxt Document_Antiquotation.thy_output_display
|
|
then Latex.environment_block "isabelle" body
|
|
else Latex.block (Latex.enclose_body "\\isa{" "}" body);
|
|
|
|
fun isabelle_typewriter ctxt body =
|
|
if Config.get ctxt Document_Antiquotation.thy_output_display
|
|
then Latex.environment_block "isabellett" body
|
|
else Latex.block (Latex.enclose_body "\\isatt{" "}" body);
|
|
|
|
fun typewriter ctxt s =
|
|
isabelle_typewriter ctxt [Latex.string (Latex.output_ascii s)];
|
|
|
|
fun verbatim ctxt =
|
|
if Config.get ctxt Document_Antiquotation.thy_output_display
|
|
then Document_Antiquotation.indent_lines ctxt #> typewriter ctxt
|
|
else split_lines #> map (typewriter ctxt) #> lines #> Latex.block;
|
|
|
|
fun source ctxt =
|
|
Token.args_of_src
|
|
#> map (Token.unparse #> Document_Antiquotation.prepare_lines ctxt)
|
|
#> space_implode " "
|
|
#> output_source ctxt
|
|
#> isabelle ctxt;
|
|
|
|
fun pretty ctxt =
|
|
Document_Antiquotation.output ctxt #> Latex.string #> single #> isabelle ctxt;
|
|
|
|
fun pretty_source ctxt src prt =
|
|
if Config.get ctxt Document_Antiquotation.thy_output_source
|
|
then source ctxt src else pretty ctxt prt;
|
|
|
|
fun pretty_items ctxt =
|
|
map (Document_Antiquotation.output ctxt #> Latex.string) #> items #> isabelle ctxt;
|
|
|
|
fun pretty_items_source ctxt src prts =
|
|
if Config.get ctxt Document_Antiquotation.thy_output_source
|
|
then source ctxt src else pretty_items ctxt prts;
|
|
|
|
|
|
(* antiquotation variants *)
|
|
|
|
fun antiquotation_pretty name scan f =
|
|
Document_Antiquotation.setup name scan
|
|
(fn {context = ctxt, argument = x, ...} => pretty ctxt (f ctxt x));
|
|
|
|
fun antiquotation_pretty_source name scan f =
|
|
Document_Antiquotation.setup name scan
|
|
(fn {context = ctxt, source = src, argument = x} => pretty_source ctxt src (f ctxt x));
|
|
|
|
fun antiquotation_raw name scan f =
|
|
Document_Antiquotation.setup name scan
|
|
(fn {context = ctxt, argument = x, ...} => f ctxt x);
|
|
|
|
fun antiquotation_verbatim name scan f =
|
|
antiquotation_raw name scan (fn ctxt => verbatim ctxt o f ctxt);
|
|
|
|
end;
|