some better trace infos over the LaTeX generator Bug
HOL-OCL/Isabelle_DOF/Isabelle2018 There was a failure building this commit
Details
HOL-OCL/Isabelle_DOF/Isabelle2018 There was a failure building this commit
Details
This commit is contained in:
parent
6b62e260cd
commit
ed1bef5cbf
29
Isa_DOF.thy
29
Isa_DOF.thy
|
@ -24,8 +24,8 @@ theory Isa_DOF (* Isabelle Document Ontology Framework *)
|
|||
keywords "+=" ":=" "accepts" "rejects"
|
||||
|
||||
and "title*" "subtitle*"
|
||||
"chapter*" "section*" "subsection*" "subsubsection*"
|
||||
"paragraph*" "subparagraph*"
|
||||
"chapter*" "section*" "subsection*" "subsubsection*"
|
||||
"paragraph*" "subparagraph*"
|
||||
"text*"
|
||||
"figure*"
|
||||
"side_by_side_figure*"
|
||||
|
@ -1537,7 +1537,7 @@ end
|
|||
end
|
||||
|
||||
val _ = Thy_Output.set_meta_args_parser
|
||||
(fn thy => let val _ = writeln "META_ARGS_PARSING"
|
||||
(fn thy => let val _ = error "META_ARGS_PARSING"
|
||||
in
|
||||
(Scan.optional ( ODL_Command_Parser.attributes
|
||||
>> ODL_LTX_Converter.meta_args_2_string thy) "")
|
||||
|
@ -1948,4 +1948,27 @@ end (* struct *)
|
|||
ML\<open>Pretty.text_fold; Pretty.unformatted_string_of\<close>
|
||||
ML\<open> (String.concatWith ","); Token.content_of\<close>
|
||||
ML\<open>open Theory\<close>
|
||||
|
||||
|
||||
ML\<open>
|
||||
val _ = Thy_Output.set_meta_args_parser
|
||||
(fn thy => (Scan.optional ( ODL_Command_Parser.attributes
|
||||
>> ODL_LTX_Converter.meta_args_2_string thy) ""))
|
||||
val _ = Thy_Output.set_meta_args_parser
|
||||
(fn thy => (Scan.optional ( ODL_Command_Parser.attributes
|
||||
>> ODL_LTX_Converter.meta_args_2_string thy) ""))
|
||||
\<close>
|
||||
ML\<open>
|
||||
Document.state;
|
||||
Session.get_keywords();
|
||||
Parse.command;
|
||||
Parse.tags;
|
||||
\<close>
|
||||
ML\<open>
|
||||
Outer_Syntax.print_commands @{theory};
|
||||
Outer_Syntax.parse_spans;
|
||||
Parse.!!!;
|
||||
|
||||
\<close>
|
||||
|
||||
end
|
||||
|
|
|
@ -9,11 +9,11 @@ text\<open>@{theory \<open>Draft.Conceptual\<close>} provides a monitor @{typ M}
|
|||
\<^theory_text>\<open>doc_class\<close>es @{typ M} is enabled for.\<close>
|
||||
open_monitor*[struct::M]
|
||||
|
||||
section*[a::A, x = "3"] \<open> Lorem ipsum dolor sit amet, ... \<close>
|
||||
sectionX[a::A, x = "3"] \<open> Lorem ipsum dolor sit amet, ... \<close>
|
||||
|
||||
text*[c1::C, x = "''beta''"] \<open> ... suspendisse non arcu malesuada mollis, nibh morbi, ... \<close>
|
||||
textX[c1::C, x = "''beta''"] \<open> ... suspendisse non arcu malesuada mollis, nibh morbi, ... \<close>
|
||||
|
||||
text*[d::D, a1 = "X3"] \<open> ... phasellus amet id massa nunc, pede suscipit repellendus,
|
||||
textX[d::D, a1 = "X3"] \<open> ... phasellus amet id massa nunc, pede suscipit repellendus,
|
||||
... @{C c1} @{thm "refl"}\<close>
|
||||
|
||||
|
||||
|
@ -21,9 +21,9 @@ update_instance*[d::D, a1 := X2]
|
|||
|
||||
text\<open> ... in ut tortor ... @{docitem \<open>a\<close>} ... @{A \<open>a\<close>}\<close>
|
||||
|
||||
text*[c2::C, x = "''delta''"] \<open> ... in ut tortor eleifend augue pretium consectetuer. \<close>
|
||||
textX[c2::C, x = "''delta''"] \<open> ... in ut tortor eleifend augue pretium consectetuer. \<close>
|
||||
|
||||
text*[f::F] \<open> Lectus accumsan velit ultrices, ... }\<close>
|
||||
textX[f::F] \<open> Lectus accumsan velit ultrices, ... }\<close>
|
||||
|
||||
theorem some_proof : "P" sorry
|
||||
|
||||
|
|
|
@ -47,10 +47,17 @@ doc_class M =
|
|||
|
||||
|
||||
text\<open>fdgh\<close>
|
||||
ML\<open>\<close>
|
||||
|
||||
ML\<open> Document.state();\<close>
|
||||
(*
|
||||
ML\<open> Session.get_keywords(); (* this looks to be really session global. *)
|
||||
Outer_Syntax.command; \<close>
|
||||
ML\<open> Thy_Header.get_keywords @{theory};(* this looks to be really theory global. *) \<close>
|
||||
*)
|
||||
|
||||
section* [ test :: A ] \<open> Test and Validation\<close>
|
||||
text\<open>Defining some document elements to be referenced in later on in another theory: \<close>
|
||||
text* [ sdf ] \<open> f @{thm refl}\<close>
|
||||
text* [ sdf ] \<open> f @{thm refl}\<close>
|
||||
text* [ sdfg ] \<open> fg @{thm refl}\<close>
|
||||
text* [ xxxy ] \<open> dd @{docitem \<open>sdfg\<close>} @{thm refl}\<close>
|
||||
|
||||
|
|
|
@ -0,0 +1,495 @@
|
|||
(* Title: Pure/Thy/thy_info.ML
|
||||
Author: Markus Wenzel, TU Muenchen
|
||||
|
||||
Global theory info database, with auto-loading according to theory and
|
||||
file dependencies.
|
||||
*)
|
||||
|
||||
signature THY_INFO =
|
||||
sig
|
||||
type presentation_context =
|
||||
{options: Options.T, file_pos: Position.T, adjust_pos: Position.T -> Position.T,
|
||||
segments: Thy_Output.segment list}
|
||||
val apply_presentation: presentation_context -> theory -> unit
|
||||
val add_presentation: (presentation_context -> theory -> unit) -> theory -> theory
|
||||
val get_names: unit -> string list
|
||||
val lookup_theory: string -> theory option
|
||||
val get_theory: string -> theory
|
||||
val master_directory: string -> Path.T
|
||||
val remove_thy: string -> unit
|
||||
type context =
|
||||
{options: Options.T,
|
||||
symbols: HTML.symbols,
|
||||
bibtex_entries: string list,
|
||||
last_timing: Toplevel.transition -> Time.time}
|
||||
val use_theories: context -> string -> Path.T -> (string * Position.T) list -> unit
|
||||
val use_thy: string -> unit
|
||||
val script_thy: Position.T -> string -> theory -> theory
|
||||
val register_thy: theory -> unit
|
||||
val finish: unit -> unit
|
||||
end;
|
||||
|
||||
structure Thy_Info: THY_INFO =
|
||||
struct
|
||||
|
||||
(** presentation of consolidated theory **)
|
||||
|
||||
type presentation_context =
|
||||
{options: Options.T, file_pos: Position.T, adjust_pos: Position.T -> Position.T,
|
||||
segments: Thy_Output.segment list};
|
||||
|
||||
structure Presentation = Theory_Data
|
||||
(
|
||||
type T = ((presentation_context -> theory -> unit) * stamp) list;
|
||||
val empty = [];
|
||||
val extend = I;
|
||||
fun merge data : T = Library.merge (eq_snd op =) data;
|
||||
);
|
||||
|
||||
fun apply_presentation (context: presentation_context) thy =
|
||||
ignore (Presentation.get thy |> Par_List.map (fn (f, _) => f context thy));
|
||||
|
||||
fun add_presentation f = Presentation.map (cons (f, stamp ()));
|
||||
|
||||
val _ =
|
||||
Theory.setup (add_presentation (fn {options, file_pos, segments, ...} => fn thy =>
|
||||
if exists (Toplevel.is_skipped_proof o #state) segments then ()
|
||||
else
|
||||
let
|
||||
val body = Thy_Output.present_thy options thy segments;
|
||||
val option = Present.document_option options;
|
||||
in
|
||||
if #disabled option then ()
|
||||
else
|
||||
let
|
||||
val latex = Latex.isabelle_body (Context.theory_name thy) body;
|
||||
val output = [Latex.output_text latex, Latex.output_positions file_pos latex];
|
||||
val _ =
|
||||
if Options.bool options "export_document"
|
||||
then Export.export thy "document.tex" output else ();
|
||||
val _ = if #enabled option then Present.theory_output thy output else ();
|
||||
in () end
|
||||
end));
|
||||
|
||||
|
||||
|
||||
(** thy database **)
|
||||
|
||||
(* messages *)
|
||||
|
||||
val show_path = space_implode " via " o map quote;
|
||||
|
||||
fun cycle_msg names = "Cyclic dependency of " ^ show_path names;
|
||||
|
||||
|
||||
(* derived graph operations *)
|
||||
|
||||
fun add_deps name parents G = String_Graph.add_deps_acyclic (name, parents) G
|
||||
handle String_Graph.CYCLES namess => error (cat_lines (map cycle_msg namess));
|
||||
|
||||
fun new_entry name parents entry =
|
||||
String_Graph.new_node (name, entry) #> add_deps name parents;
|
||||
|
||||
|
||||
(* global thys *)
|
||||
|
||||
type deps =
|
||||
{master: (Path.T * SHA1.digest), (*master dependencies for thy file*)
|
||||
imports: (string * Position.T) list}; (*source specification of imports (partially qualified)*)
|
||||
|
||||
fun make_deps master imports : deps = {master = master, imports = imports};
|
||||
|
||||
fun master_dir_deps (d: deps option) =
|
||||
the_default Path.current (Option.map (Path.dir o #1 o #master) d);
|
||||
|
||||
local
|
||||
val global_thys =
|
||||
Synchronized.var "Thy_Info.thys"
|
||||
(String_Graph.empty: (deps option * theory option) String_Graph.T);
|
||||
in
|
||||
fun get_thys () = Synchronized.value global_thys;
|
||||
fun change_thys f = Synchronized.change global_thys f;
|
||||
end;
|
||||
|
||||
fun get_names () = String_Graph.topological_order (get_thys ());
|
||||
|
||||
|
||||
(* access thy *)
|
||||
|
||||
fun lookup thys name = try (String_Graph.get_node thys) name;
|
||||
fun lookup_thy name = lookup (get_thys ()) name;
|
||||
|
||||
fun get thys name =
|
||||
(case lookup thys name of
|
||||
SOME thy => thy
|
||||
| NONE => error ("Theory loader: nothing known about theory " ^ quote name));
|
||||
|
||||
fun get_thy name = get (get_thys ()) name;
|
||||
|
||||
|
||||
(* access deps *)
|
||||
|
||||
val lookup_deps = Option.map #1 o lookup_thy;
|
||||
|
||||
val master_directory = master_dir_deps o #1 o get_thy;
|
||||
|
||||
|
||||
(* access theory *)
|
||||
|
||||
fun lookup_theory name =
|
||||
(case lookup_thy name of
|
||||
SOME (_, SOME theory) => SOME theory
|
||||
| _ => NONE);
|
||||
|
||||
fun get_theory name =
|
||||
(case lookup_theory name of
|
||||
SOME theory => theory
|
||||
| _ => error ("Theory loader: undefined entry for theory " ^ quote name));
|
||||
|
||||
val get_imports = Resources.imports_of o get_theory;
|
||||
|
||||
|
||||
|
||||
(** thy operations **)
|
||||
|
||||
(* remove *)
|
||||
|
||||
fun remove name thys =
|
||||
(case lookup thys name of
|
||||
NONE => thys
|
||||
| SOME (NONE, _) => error ("Cannot update finished theory " ^ quote name)
|
||||
| SOME _ =>
|
||||
let
|
||||
val succs = String_Graph.all_succs thys [name];
|
||||
val _ = writeln ("Theory loader: removing " ^ commas_quote succs);
|
||||
in fold String_Graph.del_node succs thys end);
|
||||
|
||||
val remove_thy = change_thys o remove;
|
||||
|
||||
|
||||
(* update *)
|
||||
|
||||
fun update deps theory thys =
|
||||
let
|
||||
val name = Context.theory_long_name theory;
|
||||
val parents = map Context.theory_long_name (Theory.parents_of theory);
|
||||
|
||||
val thys' = remove name thys;
|
||||
val _ = map (get thys') parents;
|
||||
in new_entry name parents (SOME deps, SOME theory) thys' end;
|
||||
|
||||
fun update_thy deps theory = change_thys (update deps theory);
|
||||
|
||||
|
||||
(* context *)
|
||||
|
||||
type context =
|
||||
{options: Options.T,
|
||||
symbols: HTML.symbols,
|
||||
bibtex_entries: string list,
|
||||
last_timing: Toplevel.transition -> Time.time};
|
||||
|
||||
fun default_context (): context =
|
||||
{options = Options.default (),
|
||||
symbols = HTML.no_symbols,
|
||||
bibtex_entries = [],
|
||||
last_timing = K Time.zeroTime};
|
||||
|
||||
|
||||
(* scheduling loader tasks *)
|
||||
|
||||
datatype result =
|
||||
Result of {theory: theory, exec_id: Document_ID.exec,
|
||||
present: unit -> unit, commit: unit -> unit, weight: int};
|
||||
|
||||
fun theory_result theory =
|
||||
Result {theory = theory, exec_id = Document_ID.none, present = I, commit = I, weight = 0};
|
||||
|
||||
fun result_theory (Result {theory, ...}) = theory;
|
||||
fun result_present (Result {present, ...}) = present;
|
||||
fun result_commit (Result {commit, ...}) = commit;
|
||||
fun result_ord (Result {weight = i, ...}, Result {weight = j, ...}) = int_ord (j, i);
|
||||
|
||||
fun join_theory (Result {theory, exec_id, ...}) =
|
||||
let
|
||||
val _ = Execution.join [exec_id];
|
||||
val res = Exn.capture Thm.consolidate_theory theory;
|
||||
val exns = maps Task_Queue.group_status (Execution.peek exec_id);
|
||||
in res :: map Exn.Exn exns end;
|
||||
|
||||
datatype task =
|
||||
Task of string list * (theory list -> result) |
|
||||
Finished of theory;
|
||||
|
||||
fun task_finished (Task _) = false
|
||||
| task_finished (Finished _) = true;
|
||||
|
||||
fun task_parents deps (parents: string list) = map (the o AList.lookup (op =) deps) parents;
|
||||
|
||||
val schedule_seq =
|
||||
String_Graph.schedule (fn deps => fn (_, task) =>
|
||||
(case task of
|
||||
Task (parents, body) =>
|
||||
let
|
||||
val result = body (task_parents deps parents);
|
||||
val _ = Par_Exn.release_all (join_theory result);
|
||||
val _ = result_present result ();
|
||||
val _ = result_commit result ();
|
||||
in result_theory result end
|
||||
| Finished thy => thy)) #> ignore;
|
||||
|
||||
val schedule_futures = Thread_Attributes.uninterruptible (fn _ => fn tasks =>
|
||||
let
|
||||
val futures = tasks
|
||||
|> String_Graph.schedule (fn deps => fn (name, task) =>
|
||||
(case task of
|
||||
Task (parents, body) =>
|
||||
(singleton o Future.forks)
|
||||
{name = "theory:" ^ name, group = NONE,
|
||||
deps = map (Future.task_of o #2) deps, pri = 0, interrupts = true}
|
||||
(fn () =>
|
||||
(case filter (not o can Future.join o #2) deps of
|
||||
[] => body (map (result_theory o Future.join) (task_parents deps parents))
|
||||
| bad =>
|
||||
error
|
||||
("Failed to load theory " ^ quote name ^
|
||||
" (unresolved " ^ commas_quote (map #1 bad) ^ ")")))
|
||||
| Finished theory => Future.value (theory_result theory)));
|
||||
|
||||
val results1 = futures
|
||||
|> maps (fn future =>
|
||||
(case Future.join_result future of
|
||||
Exn.Res result => join_theory result
|
||||
| Exn.Exn exn => [Exn.Exn exn]));
|
||||
|
||||
val results2 = futures
|
||||
|> map_filter (Exn.get_res o Future.join_result)
|
||||
|> sort result_ord
|
||||
|> Par_List.map (fn result => Exn.capture (result_present result) ());
|
||||
|
||||
(* FIXME more precise commit order (!?) *)
|
||||
val results3 = futures
|
||||
|> map (fn future => Exn.capture (fn () => result_commit (Future.join future) ()) ());
|
||||
|
||||
(* FIXME avoid global Execution.reset (!??) *)
|
||||
val results4 = map Exn.Exn (maps Task_Queue.group_status (Execution.reset ()));
|
||||
|
||||
val _ = Par_Exn.release_all (results1 @ results2 @ results3 @ results4);
|
||||
in () end);
|
||||
|
||||
|
||||
(* eval theory *)
|
||||
|
||||
fun excursion keywords master_dir last_timing init elements =
|
||||
let
|
||||
fun prepare_span st span =
|
||||
Command_Span.content span
|
||||
|> Command.read keywords (Command.read_thy st) master_dir init ([], ~1)
|
||||
|> (fn tr => Toplevel.put_timing (last_timing tr) tr);
|
||||
|
||||
fun element_result span_elem (st, _) =
|
||||
let
|
||||
val elem = Thy_Syntax.map_element (prepare_span st) span_elem;
|
||||
val (results, st') = Toplevel.element_result keywords elem st;
|
||||
val pos' = Toplevel.pos_of (Thy_Syntax.last_element elem);
|
||||
in (results, (st', pos')) end;
|
||||
|
||||
val (results, (end_state, end_pos)) =
|
||||
fold_map element_result elements (Toplevel.toplevel, Position.none);
|
||||
|
||||
val thy = Toplevel.end_theory end_pos end_state;
|
||||
in (results, thy) end;
|
||||
|
||||
fun eval_thy (context: context) update_time master_dir header text_pos text parents =
|
||||
let
|
||||
val {options, symbols, bibtex_entries, last_timing} = context;
|
||||
val (name, _) = #name header;
|
||||
val keywords =
|
||||
fold (curry Keyword.merge_keywords o Thy_Header.get_keywords) parents
|
||||
(Keyword.add_keywords (#keywords header) Keyword.empty_keywords);
|
||||
|
||||
val _ = writeln "eval_thy 1";
|
||||
|
||||
val spans = Outer_Syntax.parse_spans (Token.explode keywords text_pos text);
|
||||
val _ = writeln "eval_thy 2";
|
||||
val elements = Thy_Syntax.parse_elements keywords spans;
|
||||
val _ = writeln "eval_thy 3";
|
||||
|
||||
fun init () =
|
||||
Resources.begin_theory master_dir header parents
|
||||
|> Present.begin_theory bibtex_entries update_time
|
||||
(fn () => implode (map (HTML.present_span symbols keywords) spans));
|
||||
|
||||
val (results, thy) =
|
||||
cond_timeit true ("theory " ^ quote name)
|
||||
(fn () => excursion keywords master_dir last_timing init elements);
|
||||
|
||||
fun present () =
|
||||
let
|
||||
val _ = writeln "eval_thy 4 - present"
|
||||
val segments = (spans ~~ maps Toplevel.join_results results)
|
||||
|> map (fn (span, (tr, st')) => {span = span, command = tr, state = st'});
|
||||
|
||||
val X = if null segments then (writeln "eval_thy 5 - no segments";[])
|
||||
else List.concat(map (fn X => Command_Span.content (#span X)) segments)
|
||||
val Y = (String.concatWith "::") (map Token.content_of X)
|
||||
val _ = writeln("eval_thy 5 BEGIN\n"^Y^"eval_thy 6 END:"^Context.theory_name thy^"\n")
|
||||
|
||||
val context: presentation_context =
|
||||
{options = options, file_pos = text_pos, adjust_pos = I, segments = segments};
|
||||
in apply_presentation context thy end;
|
||||
in (thy, present, size text) end;
|
||||
|
||||
|
||||
(* require_thy -- checking database entries wrt. the file-system *)
|
||||
|
||||
local
|
||||
|
||||
fun required_by _ [] = ""
|
||||
| required_by s initiators = s ^ "(required by " ^ show_path (rev initiators) ^ ")";
|
||||
|
||||
fun load_thy context initiators update_time deps text (name, pos) keywords parents =
|
||||
let
|
||||
val _ = remove_thy name;
|
||||
val _ = writeln ("Loading theory " ^ quote name ^ required_by " " initiators);
|
||||
val _ = Output.try_protocol_message (Markup.loading_theory name) [];
|
||||
|
||||
val {master = (thy_path, _), imports} = deps;
|
||||
val dir = Path.dir thy_path;
|
||||
val header = Thy_Header.make (name, pos) imports keywords;
|
||||
|
||||
val _ = Position.reports (map #2 imports ~~ map Theory.get_markup parents);
|
||||
|
||||
val exec_id = Document_ID.make ();
|
||||
val _ =
|
||||
Execution.running Document_ID.none exec_id [] orelse
|
||||
raise Fail ("Failed to register execution: " ^ Document_ID.print exec_id);
|
||||
|
||||
val timing_start = Timing.start ();
|
||||
|
||||
val text_pos = Position.put_id (Document_ID.print exec_id) (Path.position thy_path);
|
||||
val (theory, present, weight) =
|
||||
eval_thy context update_time dir header text_pos text
|
||||
(if name = Context.PureN then [Context.the_global_context ()] else parents);
|
||||
|
||||
val timing_result = Timing.result timing_start;
|
||||
val timing_props = [Markup.theory_timing, (Markup.nameN, name)];
|
||||
val _ = Output.try_protocol_message (timing_props @ Markup.timing_properties timing_result) []
|
||||
|
||||
fun commit () = update_thy deps theory;
|
||||
in
|
||||
Result {theory = theory, exec_id = exec_id, present = present, commit = commit, weight = weight}
|
||||
end;
|
||||
|
||||
fun check_deps dir name =
|
||||
(case lookup_deps name of
|
||||
SOME NONE => (true, NONE, Position.none, get_imports name, [])
|
||||
| NONE =>
|
||||
let val {master, text, theory_pos, imports, keywords} = Resources.check_thy dir name
|
||||
in (false, SOME (make_deps master imports, text), theory_pos, imports, keywords) end
|
||||
| SOME (SOME {master, ...}) =>
|
||||
let
|
||||
val {master = master', text = text', theory_pos = theory_pos', imports = imports',
|
||||
keywords = keywords'} = Resources.check_thy dir name;
|
||||
val deps' = SOME (make_deps master' imports', text');
|
||||
val current =
|
||||
#2 master = #2 master' andalso
|
||||
(case lookup_theory name of
|
||||
NONE => false
|
||||
| SOME theory => Resources.loaded_files_current theory);
|
||||
in (current, deps', theory_pos', imports', keywords') end);
|
||||
|
||||
in
|
||||
|
||||
fun require_thys context initiators qualifier dir strs tasks =
|
||||
fold_map (require_thy context initiators qualifier dir) strs tasks |>> forall I
|
||||
and require_thy context initiators qualifier dir (s, require_pos) tasks =
|
||||
let
|
||||
val {master_dir, theory_name, ...} = Resources.import_name qualifier dir s;
|
||||
in
|
||||
(case try (String_Graph.get_node tasks) theory_name of
|
||||
SOME task => (task_finished task, tasks)
|
||||
| NONE =>
|
||||
let
|
||||
val _ = member (op =) initiators theory_name andalso error (cycle_msg initiators);
|
||||
|
||||
val (current, deps, theory_pos, imports, keywords) = check_deps master_dir theory_name
|
||||
handle ERROR msg =>
|
||||
cat_error msg
|
||||
("The error(s) above occurred for theory " ^ quote theory_name ^
|
||||
Position.here require_pos ^ required_by "\n" initiators);
|
||||
|
||||
val qualifier' = Resources.theory_qualifier theory_name;
|
||||
val dir' = Path.append dir (master_dir_deps (Option.map #1 deps));
|
||||
|
||||
val parents = map (#theory_name o Resources.import_name qualifier' dir' o #1) imports;
|
||||
val (parents_current, tasks') =
|
||||
require_thys context (theory_name :: initiators) qualifier' dir' imports tasks;
|
||||
|
||||
val all_current = current andalso parents_current;
|
||||
val task =
|
||||
if all_current then Finished (get_theory theory_name)
|
||||
else
|
||||
(case deps of
|
||||
NONE => raise Fail "Malformed deps"
|
||||
| SOME (dep, text) =>
|
||||
let
|
||||
val update_time = serial ();
|
||||
val load =
|
||||
load_thy context initiators update_time
|
||||
dep text (theory_name, theory_pos) keywords;
|
||||
in Task (parents, load) end);
|
||||
|
||||
val tasks'' = new_entry theory_name parents task tasks';
|
||||
in (all_current, tasks'') end)
|
||||
end;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* use theories *)
|
||||
|
||||
fun use_theories context qualifier master_dir imports =
|
||||
let val (_, tasks) = require_thys context [] qualifier master_dir imports String_Graph.empty
|
||||
in if Multithreading.max_threads () > 1 then schedule_futures tasks else schedule_seq tasks end;
|
||||
|
||||
fun use_thy name =
|
||||
use_theories (default_context ()) Resources.default_qualifier
|
||||
Path.current [(name, Position.none)];
|
||||
|
||||
|
||||
(* toplevel scripting -- without maintaining database *)
|
||||
|
||||
fun script_thy pos txt thy =
|
||||
let
|
||||
val trs =
|
||||
Outer_Syntax.parse thy pos txt
|
||||
|> map (Toplevel.modify_init (K thy));
|
||||
val end_pos = if null trs then pos else Toplevel.pos_of (List.last trs);
|
||||
val end_state = fold (Toplevel.command_exception true) trs Toplevel.toplevel;
|
||||
in Toplevel.end_theory end_pos end_state end;
|
||||
|
||||
|
||||
(* register theory *)
|
||||
|
||||
fun register_thy theory =
|
||||
let
|
||||
val name = Context.theory_long_name theory;
|
||||
val {master, ...} = Resources.check_thy (Resources.master_directory theory) name;
|
||||
val imports = Resources.imports_of theory;
|
||||
in
|
||||
change_thys (fn thys =>
|
||||
let
|
||||
val thys' = remove name thys;
|
||||
val _ = writeln ("Registering theory " ^ quote name);
|
||||
in update (make_deps master imports) theory thys' end)
|
||||
end;
|
||||
|
||||
|
||||
(* finish all theories *)
|
||||
|
||||
fun finish () = change_thys (String_Graph.map (fn _ => fn (_, entry) => (NONE, entry)));
|
||||
|
||||
end;
|
||||
|
||||
fun use_thy name = Runtime.toplevel_program (fn () => Thy_Info.use_thy name);
|
|
@ -0,0 +1,484 @@
|
|||
(* Title: Pure/Thy/thy_info.ML
|
||||
Author: Markus Wenzel, TU Muenchen
|
||||
|
||||
Global theory info database, with auto-loading according to theory and
|
||||
file dependencies.
|
||||
*)
|
||||
|
||||
signature THY_INFO =
|
||||
sig
|
||||
type presentation_context =
|
||||
{options: Options.T, file_pos: Position.T, adjust_pos: Position.T -> Position.T,
|
||||
segments: Thy_Output.segment list}
|
||||
val apply_presentation: presentation_context -> theory -> unit
|
||||
val add_presentation: (presentation_context -> theory -> unit) -> theory -> theory
|
||||
val get_names: unit -> string list
|
||||
val lookup_theory: string -> theory option
|
||||
val get_theory: string -> theory
|
||||
val master_directory: string -> Path.T
|
||||
val remove_thy: string -> unit
|
||||
type context =
|
||||
{options: Options.T,
|
||||
symbols: HTML.symbols,
|
||||
bibtex_entries: string list,
|
||||
last_timing: Toplevel.transition -> Time.time}
|
||||
val use_theories: context -> string -> Path.T -> (string * Position.T) list -> unit
|
||||
val use_thy: string -> unit
|
||||
val script_thy: Position.T -> string -> theory -> theory
|
||||
val register_thy: theory -> unit
|
||||
val finish: unit -> unit
|
||||
end;
|
||||
|
||||
structure Thy_Info: THY_INFO =
|
||||
struct
|
||||
|
||||
(** presentation of consolidated theory **)
|
||||
|
||||
type presentation_context =
|
||||
{options: Options.T, file_pos: Position.T, adjust_pos: Position.T -> Position.T,
|
||||
segments: Thy_Output.segment list};
|
||||
|
||||
structure Presentation = Theory_Data
|
||||
(
|
||||
type T = ((presentation_context -> theory -> unit) * stamp) list;
|
||||
val empty = [];
|
||||
val extend = I;
|
||||
fun merge data : T = Library.merge (eq_snd op =) data;
|
||||
);
|
||||
|
||||
fun apply_presentation (context: presentation_context) thy =
|
||||
ignore (Presentation.get thy |> Par_List.map (fn (f, _) => f context thy));
|
||||
|
||||
fun add_presentation f = Presentation.map (cons (f, stamp ()));
|
||||
|
||||
val _ =
|
||||
Theory.setup (add_presentation (fn {options, file_pos, segments, ...} => fn thy =>
|
||||
if exists (Toplevel.is_skipped_proof o #state) segments then ()
|
||||
else
|
||||
let
|
||||
val body = Thy_Output.present_thy options thy segments;
|
||||
val option = Present.document_option options;
|
||||
in
|
||||
if #disabled option then ()
|
||||
else
|
||||
let
|
||||
val latex = Latex.isabelle_body (Context.theory_name thy) body;
|
||||
val output = [Latex.output_text latex, Latex.output_positions file_pos latex];
|
||||
val _ =
|
||||
if Options.bool options "export_document"
|
||||
then Export.export thy "document.tex" output else ();
|
||||
val _ = if #enabled option then Present.theory_output thy output else ();
|
||||
in () end
|
||||
end));
|
||||
|
||||
|
||||
|
||||
(** thy database **)
|
||||
|
||||
(* messages *)
|
||||
|
||||
val show_path = space_implode " via " o map quote;
|
||||
|
||||
fun cycle_msg names = "Cyclic dependency of " ^ show_path names;
|
||||
|
||||
|
||||
(* derived graph operations *)
|
||||
|
||||
fun add_deps name parents G = String_Graph.add_deps_acyclic (name, parents) G
|
||||
handle String_Graph.CYCLES namess => error (cat_lines (map cycle_msg namess));
|
||||
|
||||
fun new_entry name parents entry =
|
||||
String_Graph.new_node (name, entry) #> add_deps name parents;
|
||||
|
||||
|
||||
(* global thys *)
|
||||
|
||||
type deps =
|
||||
{master: (Path.T * SHA1.digest), (*master dependencies for thy file*)
|
||||
imports: (string * Position.T) list}; (*source specification of imports (partially qualified)*)
|
||||
|
||||
fun make_deps master imports : deps = {master = master, imports = imports};
|
||||
|
||||
fun master_dir_deps (d: deps option) =
|
||||
the_default Path.current (Option.map (Path.dir o #1 o #master) d);
|
||||
|
||||
local
|
||||
val global_thys =
|
||||
Synchronized.var "Thy_Info.thys"
|
||||
(String_Graph.empty: (deps option * theory option) String_Graph.T);
|
||||
in
|
||||
fun get_thys () = Synchronized.value global_thys;
|
||||
fun change_thys f = Synchronized.change global_thys f;
|
||||
end;
|
||||
|
||||
fun get_names () = String_Graph.topological_order (get_thys ());
|
||||
|
||||
|
||||
(* access thy *)
|
||||
|
||||
fun lookup thys name = try (String_Graph.get_node thys) name;
|
||||
fun lookup_thy name = lookup (get_thys ()) name;
|
||||
|
||||
fun get thys name =
|
||||
(case lookup thys name of
|
||||
SOME thy => thy
|
||||
| NONE => error ("Theory loader: nothing known about theory " ^ quote name));
|
||||
|
||||
fun get_thy name = get (get_thys ()) name;
|
||||
|
||||
|
||||
(* access deps *)
|
||||
|
||||
val lookup_deps = Option.map #1 o lookup_thy;
|
||||
|
||||
val master_directory = master_dir_deps o #1 o get_thy;
|
||||
|
||||
|
||||
(* access theory *)
|
||||
|
||||
fun lookup_theory name =
|
||||
(case lookup_thy name of
|
||||
SOME (_, SOME theory) => SOME theory
|
||||
| _ => NONE);
|
||||
|
||||
fun get_theory name =
|
||||
(case lookup_theory name of
|
||||
SOME theory => theory
|
||||
| _ => error ("Theory loader: undefined entry for theory " ^ quote name));
|
||||
|
||||
val get_imports = Resources.imports_of o get_theory;
|
||||
|
||||
|
||||
|
||||
(** thy operations **)
|
||||
|
||||
(* remove *)
|
||||
|
||||
fun remove name thys =
|
||||
(case lookup thys name of
|
||||
NONE => thys
|
||||
| SOME (NONE, _) => error ("Cannot update finished theory " ^ quote name)
|
||||
| SOME _ =>
|
||||
let
|
||||
val succs = String_Graph.all_succs thys [name];
|
||||
val _ = writeln ("Theory loader: removing " ^ commas_quote succs);
|
||||
in fold String_Graph.del_node succs thys end);
|
||||
|
||||
val remove_thy = change_thys o remove;
|
||||
|
||||
|
||||
(* update *)
|
||||
|
||||
fun update deps theory thys =
|
||||
let
|
||||
val name = Context.theory_long_name theory;
|
||||
val parents = map Context.theory_long_name (Theory.parents_of theory);
|
||||
|
||||
val thys' = remove name thys;
|
||||
val _ = map (get thys') parents;
|
||||
in new_entry name parents (SOME deps, SOME theory) thys' end;
|
||||
|
||||
fun update_thy deps theory = change_thys (update deps theory);
|
||||
|
||||
|
||||
(* context *)
|
||||
|
||||
type context =
|
||||
{options: Options.T,
|
||||
symbols: HTML.symbols,
|
||||
bibtex_entries: string list,
|
||||
last_timing: Toplevel.transition -> Time.time};
|
||||
|
||||
fun default_context (): context =
|
||||
{options = Options.default (),
|
||||
symbols = HTML.no_symbols,
|
||||
bibtex_entries = [],
|
||||
last_timing = K Time.zeroTime};
|
||||
|
||||
|
||||
(* scheduling loader tasks *)
|
||||
|
||||
datatype result =
|
||||
Result of {theory: theory, exec_id: Document_ID.exec,
|
||||
present: unit -> unit, commit: unit -> unit, weight: int};
|
||||
|
||||
fun theory_result theory =
|
||||
Result {theory = theory, exec_id = Document_ID.none, present = I, commit = I, weight = 0};
|
||||
|
||||
fun result_theory (Result {theory, ...}) = theory;
|
||||
fun result_present (Result {present, ...}) = present;
|
||||
fun result_commit (Result {commit, ...}) = commit;
|
||||
fun result_ord (Result {weight = i, ...}, Result {weight = j, ...}) = int_ord (j, i);
|
||||
|
||||
fun join_theory (Result {theory, exec_id, ...}) =
|
||||
let
|
||||
val _ = Execution.join [exec_id];
|
||||
val res = Exn.capture Thm.consolidate_theory theory;
|
||||
val exns = maps Task_Queue.group_status (Execution.peek exec_id);
|
||||
in res :: map Exn.Exn exns end;
|
||||
|
||||
datatype task =
|
||||
Task of string list * (theory list -> result) |
|
||||
Finished of theory;
|
||||
|
||||
fun task_finished (Task _) = false
|
||||
| task_finished (Finished _) = true;
|
||||
|
||||
fun task_parents deps (parents: string list) = map (the o AList.lookup (op =) deps) parents;
|
||||
|
||||
val schedule_seq =
|
||||
String_Graph.schedule (fn deps => fn (_, task) =>
|
||||
(case task of
|
||||
Task (parents, body) =>
|
||||
let
|
||||
val result = body (task_parents deps parents);
|
||||
val _ = Par_Exn.release_all (join_theory result);
|
||||
val _ = result_present result ();
|
||||
val _ = result_commit result ();
|
||||
in result_theory result end
|
||||
| Finished thy => thy)) #> ignore;
|
||||
|
||||
val schedule_futures = Thread_Attributes.uninterruptible (fn _ => fn tasks =>
|
||||
let
|
||||
val futures = tasks
|
||||
|> String_Graph.schedule (fn deps => fn (name, task) =>
|
||||
(case task of
|
||||
Task (parents, body) =>
|
||||
(singleton o Future.forks)
|
||||
{name = "theory:" ^ name, group = NONE,
|
||||
deps = map (Future.task_of o #2) deps, pri = 0, interrupts = true}
|
||||
(fn () =>
|
||||
(case filter (not o can Future.join o #2) deps of
|
||||
[] => body (map (result_theory o Future.join) (task_parents deps parents))
|
||||
| bad =>
|
||||
error
|
||||
("Failed to load theory " ^ quote name ^
|
||||
" (unresolved " ^ commas_quote (map #1 bad) ^ ")")))
|
||||
| Finished theory => Future.value (theory_result theory)));
|
||||
|
||||
val results1 = futures
|
||||
|> maps (fn future =>
|
||||
(case Future.join_result future of
|
||||
Exn.Res result => join_theory result
|
||||
| Exn.Exn exn => [Exn.Exn exn]));
|
||||
|
||||
val results2 = futures
|
||||
|> map_filter (Exn.get_res o Future.join_result)
|
||||
|> sort result_ord
|
||||
|> Par_List.map (fn result => Exn.capture (result_present result) ());
|
||||
|
||||
(* FIXME more precise commit order (!?) *)
|
||||
val results3 = futures
|
||||
|> map (fn future => Exn.capture (fn () => result_commit (Future.join future) ()) ());
|
||||
|
||||
(* FIXME avoid global Execution.reset (!??) *)
|
||||
val results4 = map Exn.Exn (maps Task_Queue.group_status (Execution.reset ()));
|
||||
|
||||
val _ = Par_Exn.release_all (results1 @ results2 @ results3 @ results4);
|
||||
in () end);
|
||||
|
||||
|
||||
(* eval theory *)
|
||||
|
||||
fun excursion keywords master_dir last_timing init elements =
|
||||
let
|
||||
fun prepare_span st span =
|
||||
Command_Span.content span
|
||||
|> Command.read keywords (Command.read_thy st) master_dir init ([], ~1)
|
||||
|> (fn tr => Toplevel.put_timing (last_timing tr) tr);
|
||||
|
||||
fun element_result span_elem (st, _) =
|
||||
let
|
||||
val elem = Thy_Syntax.map_element (prepare_span st) span_elem;
|
||||
val (results, st') = Toplevel.element_result keywords elem st;
|
||||
val pos' = Toplevel.pos_of (Thy_Syntax.last_element elem);
|
||||
in (results, (st', pos')) end;
|
||||
|
||||
val (results, (end_state, end_pos)) =
|
||||
fold_map element_result elements (Toplevel.toplevel, Position.none);
|
||||
|
||||
val thy = Toplevel.end_theory end_pos end_state;
|
||||
in (results, thy) end;
|
||||
|
||||
fun eval_thy (context: context) update_time master_dir header text_pos text parents =
|
||||
let
|
||||
val {options, symbols, bibtex_entries, last_timing} = context;
|
||||
val (name, _) = #name header;
|
||||
val keywords =
|
||||
fold (curry Keyword.merge_keywords o Thy_Header.get_keywords) parents
|
||||
(Keyword.add_keywords (#keywords header) Keyword.empty_keywords);
|
||||
|
||||
val spans = Outer_Syntax.parse_spans (Token.explode keywords text_pos text);
|
||||
val elements = Thy_Syntax.parse_elements keywords spans;
|
||||
|
||||
fun init () =
|
||||
Resources.begin_theory master_dir header parents
|
||||
|> Present.begin_theory bibtex_entries update_time
|
||||
(fn () => implode (map (HTML.present_span symbols keywords) spans));
|
||||
|
||||
val (results, thy) =
|
||||
cond_timeit true ("theory " ^ quote name)
|
||||
(fn () => excursion keywords master_dir last_timing init elements);
|
||||
|
||||
fun present () =
|
||||
let
|
||||
val segments = (spans ~~ maps Toplevel.join_results results)
|
||||
|> map (fn (span, (tr, st')) => {span = span, command = tr, state = st'});
|
||||
val context: presentation_context =
|
||||
{options = options, file_pos = text_pos, adjust_pos = I, segments = segments};
|
||||
in apply_presentation context thy end;
|
||||
in (thy, present, size text) end;
|
||||
|
||||
|
||||
(* require_thy -- checking database entries wrt. the file-system *)
|
||||
|
||||
local
|
||||
|
||||
fun required_by _ [] = ""
|
||||
| required_by s initiators = s ^ "(required by " ^ show_path (rev initiators) ^ ")";
|
||||
|
||||
fun load_thy context initiators update_time deps text (name, pos) keywords parents =
|
||||
let
|
||||
val _ = remove_thy name;
|
||||
val _ = writeln ("Loading theory " ^ quote name ^ required_by " " initiators);
|
||||
val _ = Output.try_protocol_message (Markup.loading_theory name) [];
|
||||
|
||||
val {master = (thy_path, _), imports} = deps;
|
||||
val dir = Path.dir thy_path;
|
||||
val header = Thy_Header.make (name, pos) imports keywords;
|
||||
|
||||
val _ = Position.reports (map #2 imports ~~ map Theory.get_markup parents);
|
||||
|
||||
val exec_id = Document_ID.make ();
|
||||
val _ =
|
||||
Execution.running Document_ID.none exec_id [] orelse
|
||||
raise Fail ("Failed to register execution: " ^ Document_ID.print exec_id);
|
||||
|
||||
val timing_start = Timing.start ();
|
||||
|
||||
val text_pos = Position.put_id (Document_ID.print exec_id) (Path.position thy_path);
|
||||
val (theory, present, weight) =
|
||||
eval_thy context update_time dir header text_pos text
|
||||
(if name = Context.PureN then [Context.the_global_context ()] else parents);
|
||||
|
||||
val timing_result = Timing.result timing_start;
|
||||
val timing_props = [Markup.theory_timing, (Markup.nameN, name)];
|
||||
val _ = Output.try_protocol_message (timing_props @ Markup.timing_properties timing_result) []
|
||||
|
||||
fun commit () = update_thy deps theory;
|
||||
in
|
||||
Result {theory = theory, exec_id = exec_id, present = present, commit = commit, weight = weight}
|
||||
end;
|
||||
|
||||
fun check_deps dir name =
|
||||
(case lookup_deps name of
|
||||
SOME NONE => (true, NONE, Position.none, get_imports name, [])
|
||||
| NONE =>
|
||||
let val {master, text, theory_pos, imports, keywords} = Resources.check_thy dir name
|
||||
in (false, SOME (make_deps master imports, text), theory_pos, imports, keywords) end
|
||||
| SOME (SOME {master, ...}) =>
|
||||
let
|
||||
val {master = master', text = text', theory_pos = theory_pos', imports = imports',
|
||||
keywords = keywords'} = Resources.check_thy dir name;
|
||||
val deps' = SOME (make_deps master' imports', text');
|
||||
val current =
|
||||
#2 master = #2 master' andalso
|
||||
(case lookup_theory name of
|
||||
NONE => false
|
||||
| SOME theory => Resources.loaded_files_current theory);
|
||||
in (current, deps', theory_pos', imports', keywords') end);
|
||||
|
||||
in
|
||||
|
||||
fun require_thys context initiators qualifier dir strs tasks =
|
||||
fold_map (require_thy context initiators qualifier dir) strs tasks |>> forall I
|
||||
and require_thy context initiators qualifier dir (s, require_pos) tasks =
|
||||
let
|
||||
val {master_dir, theory_name, ...} = Resources.import_name qualifier dir s;
|
||||
in
|
||||
(case try (String_Graph.get_node tasks) theory_name of
|
||||
SOME task => (task_finished task, tasks)
|
||||
| NONE =>
|
||||
let
|
||||
val _ = member (op =) initiators theory_name andalso error (cycle_msg initiators);
|
||||
|
||||
val (current, deps, theory_pos, imports, keywords) = check_deps master_dir theory_name
|
||||
handle ERROR msg =>
|
||||
cat_error msg
|
||||
("The error(s) above occurred for theory " ^ quote theory_name ^
|
||||
Position.here require_pos ^ required_by "\n" initiators);
|
||||
|
||||
val qualifier' = Resources.theory_qualifier theory_name;
|
||||
val dir' = Path.append dir (master_dir_deps (Option.map #1 deps));
|
||||
|
||||
val parents = map (#theory_name o Resources.import_name qualifier' dir' o #1) imports;
|
||||
val (parents_current, tasks') =
|
||||
require_thys context (theory_name :: initiators) qualifier' dir' imports tasks;
|
||||
|
||||
val all_current = current andalso parents_current;
|
||||
val task =
|
||||
if all_current then Finished (get_theory theory_name)
|
||||
else
|
||||
(case deps of
|
||||
NONE => raise Fail "Malformed deps"
|
||||
| SOME (dep, text) =>
|
||||
let
|
||||
val update_time = serial ();
|
||||
val load =
|
||||
load_thy context initiators update_time
|
||||
dep text (theory_name, theory_pos) keywords;
|
||||
in Task (parents, load) end);
|
||||
|
||||
val tasks'' = new_entry theory_name parents task tasks';
|
||||
in (all_current, tasks'') end)
|
||||
end;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* use theories *)
|
||||
|
||||
fun use_theories context qualifier master_dir imports =
|
||||
let val (_, tasks) = require_thys context [] qualifier master_dir imports String_Graph.empty
|
||||
in if Multithreading.max_threads () > 1 then schedule_futures tasks else schedule_seq tasks end;
|
||||
|
||||
fun use_thy name =
|
||||
use_theories (default_context ()) Resources.default_qualifier
|
||||
Path.current [(name, Position.none)];
|
||||
|
||||
|
||||
(* toplevel scripting -- without maintaining database *)
|
||||
|
||||
fun script_thy pos txt thy =
|
||||
let
|
||||
val trs =
|
||||
Outer_Syntax.parse thy pos txt
|
||||
|> map (Toplevel.modify_init (K thy));
|
||||
val end_pos = if null trs then pos else Toplevel.pos_of (List.last trs);
|
||||
val end_state = fold (Toplevel.command_exception true) trs Toplevel.toplevel;
|
||||
in Toplevel.end_theory end_pos end_state end;
|
||||
|
||||
|
||||
(* register theory *)
|
||||
|
||||
fun register_thy theory =
|
||||
let
|
||||
val name = Context.theory_long_name theory;
|
||||
val {master, ...} = Resources.check_thy (Resources.master_directory theory) name;
|
||||
val imports = Resources.imports_of theory;
|
||||
in
|
||||
change_thys (fn thys =>
|
||||
let
|
||||
val thys' = remove name thys;
|
||||
val _ = writeln ("Registering theory " ^ quote name);
|
||||
in update (make_deps master imports) theory thys' end)
|
||||
end;
|
||||
|
||||
|
||||
(* finish all theories *)
|
||||
|
||||
fun finish () = change_thys (String_Graph.map (fn _ => fn (_, entry) => (NONE, entry)));
|
||||
|
||||
end;
|
||||
|
||||
fun use_thy name = Runtime.toplevel_program (fn () => Thy_Info.use_thy name);
|
|
@ -0,0 +1,686 @@
|
|||
(* Title: Pure/Thy/thy_output.ML
|
||||
Author: Markus Wenzel, TU Muenchen
|
||||
|
||||
Theory document output with antiquotations.
|
||||
*)
|
||||
|
||||
signature THY_OUTPUT =
|
||||
sig
|
||||
val display: bool Config.T
|
||||
val quotes: bool Config.T
|
||||
val margin: int Config.T
|
||||
val indent: int Config.T
|
||||
val source: bool Config.T
|
||||
val break: bool Config.T
|
||||
val modes: string Config.T
|
||||
val add_wrapper: ((unit -> string) -> unit -> string) -> Proof.context -> Proof.context
|
||||
val add_option: binding -> (string -> Proof.context -> Proof.context) -> theory -> theory
|
||||
val check_command: Proof.context -> xstring * Position.T -> string
|
||||
val check_option: Proof.context -> xstring * Position.T -> string
|
||||
val print_antiquotations: bool -> Proof.context -> unit
|
||||
val antiquotation: binding -> 'a context_parser ->
|
||||
({source: Token.src, state: Toplevel.state, context: Proof.context} -> 'a -> string) ->
|
||||
theory -> theory
|
||||
val boolean: string -> bool
|
||||
val integer: string -> int
|
||||
val eval_antiquote: Toplevel.state -> Antiquote.text_antiquote -> string
|
||||
val output_text: Toplevel.state -> {markdown: bool} -> Input.source -> string
|
||||
val present_thy: theory -> (Toplevel.transition * Toplevel.state) list -> Token.T list -> Buffer.T
|
||||
val set_meta_args_parser : (theory -> string parser) -> unit
|
||||
val pretty_text: Proof.context -> string -> Pretty.T
|
||||
val pretty_term: Proof.context -> term -> Pretty.T
|
||||
val pretty_thm: Proof.context -> thm -> Pretty.T
|
||||
val str_of_source: Token.src -> string
|
||||
val maybe_pretty_source: (Proof.context -> 'a -> Pretty.T) -> Proof.context ->
|
||||
Token.src -> 'a list -> Pretty.T list
|
||||
val string_of_margin: Proof.context -> Pretty.T -> string
|
||||
val output: Proof.context -> Pretty.T list -> string
|
||||
val verbatim_text: Proof.context -> string -> string
|
||||
val document_command: {markdown: bool} -> (xstring * Position.T) option * Input.source ->
|
||||
Toplevel.transition -> Toplevel.transition
|
||||
end;
|
||||
|
||||
structure Thy_Output: THY_OUTPUT =
|
||||
struct
|
||||
|
||||
(** options **)
|
||||
|
||||
val display = Attrib.setup_option_bool ("thy_output_display", \<^here>);
|
||||
val break = Attrib.setup_option_bool ("thy_output_break", \<^here>);
|
||||
val quotes = Attrib.setup_option_bool ("thy_output_quotes", \<^here>);
|
||||
val margin = Attrib.setup_option_int ("thy_output_margin", \<^here>);
|
||||
val indent = Attrib.setup_option_int ("thy_output_indent", \<^here>);
|
||||
val source = Attrib.setup_option_bool ("thy_output_source", \<^here>);
|
||||
val modes = Attrib.setup_option_string ("thy_output_modes", \<^here>);
|
||||
|
||||
|
||||
structure Wrappers = Proof_Data
|
||||
(
|
||||
type T = ((unit -> string) -> unit -> string) list;
|
||||
fun init _ = [];
|
||||
);
|
||||
|
||||
fun add_wrapper wrapper = Wrappers.map (cons wrapper);
|
||||
|
||||
val wrap = Wrappers.get #> fold (fn wrapper => fn f => wrapper f);
|
||||
|
||||
|
||||
|
||||
(** maintain global antiquotations **)
|
||||
|
||||
structure Antiquotations = Theory_Data
|
||||
(
|
||||
type T =
|
||||
(Token.src -> Toplevel.state -> Proof.context -> string) Name_Space.table *
|
||||
(string -> Proof.context -> Proof.context) Name_Space.table;
|
||||
val empty : T =
|
||||
(Name_Space.empty_table Markup.document_antiquotationN,
|
||||
Name_Space.empty_table Markup.document_antiquotation_optionN);
|
||||
val extend = I;
|
||||
fun merge ((commands1, options1), (commands2, options2)) : T =
|
||||
(Name_Space.merge_tables (commands1, commands2),
|
||||
Name_Space.merge_tables (options1, options2));
|
||||
);
|
||||
|
||||
val get_antiquotations = Antiquotations.get o Proof_Context.theory_of;
|
||||
|
||||
fun add_command name cmd thy = thy
|
||||
|> Antiquotations.map (apfst (Name_Space.define (Context.Theory thy) true (name, cmd) #> snd));
|
||||
|
||||
fun add_option name opt thy = thy
|
||||
|> Antiquotations.map (apsnd (Name_Space.define (Context.Theory thy) true (name, opt) #> snd));
|
||||
|
||||
fun check_command ctxt = #1 o Name_Space.check (Context.Proof ctxt) (#1 (get_antiquotations ctxt));
|
||||
|
||||
fun check_option ctxt = #1 o Name_Space.check (Context.Proof ctxt) (#2 (get_antiquotations ctxt));
|
||||
|
||||
fun command src state ctxt =
|
||||
let val (src', f) = Token.check_src ctxt (#1 o get_antiquotations) src
|
||||
in f src' state ctxt end;
|
||||
|
||||
fun option ((xname, pos), s) ctxt =
|
||||
let
|
||||
val (_, opt) =
|
||||
Name_Space.check (Context.Proof ctxt) (#2 (get_antiquotations ctxt)) (xname, pos);
|
||||
in opt s ctxt end;
|
||||
|
||||
fun print_antiquotations verbose ctxt =
|
||||
let
|
||||
val (commands, options) = get_antiquotations ctxt;
|
||||
val command_names = map #1 (Name_Space.markup_table verbose ctxt commands);
|
||||
val option_names = map #1 (Name_Space.markup_table verbose ctxt options);
|
||||
in
|
||||
[Pretty.big_list "document antiquotations:" (map Pretty.mark_str command_names),
|
||||
Pretty.big_list "document antiquotation options:" (map Pretty.mark_str option_names)]
|
||||
end |> Pretty.writeln_chunks;
|
||||
|
||||
fun antiquotation name scan body =
|
||||
add_command name
|
||||
(fn src => fn state => fn ctxt =>
|
||||
let val (x, ctxt') = Token.syntax scan src ctxt
|
||||
in body {source = src, state = state, context = ctxt'} x end);
|
||||
|
||||
|
||||
|
||||
(** syntax of antiquotations **)
|
||||
|
||||
(* option values *)
|
||||
|
||||
fun boolean "" = true
|
||||
| boolean "true" = true
|
||||
| boolean "false" = false
|
||||
| boolean s = error ("Bad boolean value: " ^ quote s);
|
||||
|
||||
fun integer s =
|
||||
let
|
||||
fun int ss =
|
||||
(case Library.read_int ss of (i, []) => i
|
||||
| _ => error ("Bad integer value: " ^ quote s));
|
||||
in (case Symbol.explode s of "-" :: ss => ~ (int ss) | ss => int ss) end;
|
||||
|
||||
|
||||
(* outer syntax *)
|
||||
|
||||
local
|
||||
|
||||
val property =
|
||||
Parse.position Parse.name -- Scan.optional (Parse.$$$ "=" |-- Parse.!!! Parse.name) "";
|
||||
|
||||
val properties =
|
||||
Scan.optional (Parse.$$$ "[" |-- Parse.!!! (Parse.enum "," property --| Parse.$$$ "]")) [];
|
||||
|
||||
in
|
||||
|
||||
val antiq =
|
||||
Parse.!!!
|
||||
(Parse.token Parse.liberal_name -- properties -- Parse.args --| Scan.ahead Parse.eof)
|
||||
>> (fn ((name, props), args) => (props, name :: args));
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* eval antiquote *)
|
||||
|
||||
local
|
||||
|
||||
fun eval_antiq state (opts, src) =
|
||||
let
|
||||
val preview_ctxt = fold option opts (Toplevel.presentation_context_of state);
|
||||
val print_ctxt = Context_Position.set_visible false preview_ctxt;
|
||||
|
||||
fun cmd ctxt = wrap ctxt (fn () => command src state ctxt) ();
|
||||
val _ = cmd preview_ctxt;
|
||||
val print_modes = space_explode "," (Config.get print_ctxt modes) @ [Latex.latexN];
|
||||
in Print_Mode.with_modes print_modes (fn () => cmd print_ctxt) () end;
|
||||
|
||||
in
|
||||
|
||||
fun eval_antiquote _ (Antiquote.Text ss) = Symbol_Pos.content ss
|
||||
| eval_antiquote state (Antiquote.Control {name, body, ...}) =
|
||||
eval_antiq state
|
||||
([], Token.make_src name (if null body then [] else [Token.read_cartouche body]))
|
||||
| eval_antiquote state (Antiquote.Antiq {range = (pos, _), body, ...}) =
|
||||
let
|
||||
val keywords =
|
||||
(case try Toplevel.presentation_context_of state of
|
||||
SOME ctxt => Thy_Header.get_keywords' ctxt
|
||||
| NONE =>
|
||||
error ("Unknown context -- cannot expand document antiquotations" ^
|
||||
Position.here pos));
|
||||
in eval_antiq state (Token.read_antiq keywords antiq (body, pos)) end;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* output text *)
|
||||
|
||||
fun output_text state {markdown} source =
|
||||
let
|
||||
val is_reported =
|
||||
(case try Toplevel.context_of state of
|
||||
SOME ctxt => Context_Position.is_visible ctxt
|
||||
| NONE => true);
|
||||
|
||||
val pos = Input.pos_of source;
|
||||
val syms = Input.source_explode source;
|
||||
|
||||
val _ =
|
||||
if is_reported then
|
||||
Position.report pos (Markup.language_document (Input.is_delimited source))
|
||||
else ();
|
||||
|
||||
val output_antiquotes = map (eval_antiquote state) #> implode;
|
||||
|
||||
fun output_line line =
|
||||
(if Markdown.line_is_item line then "\\item " else "") ^
|
||||
output_antiquotes (Markdown.line_content line);
|
||||
|
||||
fun output_blocks blocks = space_implode "\n\n" (map output_block blocks)
|
||||
and output_block (Markdown.Par lines) = cat_lines (map output_line lines)
|
||||
| output_block (Markdown.List {kind, body, ...}) =
|
||||
Latex.environment (Markdown.print_kind kind) (output_blocks body);
|
||||
in
|
||||
if Toplevel.is_skipped_proof state then ""
|
||||
else if markdown andalso exists (Markdown.is_control o Symbol_Pos.symbol) syms
|
||||
then
|
||||
let
|
||||
val ants = Antiquote.parse pos syms;
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val blocks = Markdown.read_antiquotes ants;
|
||||
val _ = if is_reported then Position.reports (reports @ Markdown.reports blocks) else ();
|
||||
in output_blocks blocks end
|
||||
else
|
||||
let
|
||||
val ants = Antiquote.parse pos (Symbol_Pos.trim_blanks syms);
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val _ = if is_reported then Position.reports (reports @ Markdown.text_reports ants) else ();
|
||||
in output_antiquotes ants end
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** present theory source **)
|
||||
|
||||
(*NB: arranging white space around command spans is a black art*)
|
||||
|
||||
(* presentation tokens *)
|
||||
|
||||
datatype token =
|
||||
No_Token
|
||||
| Basic_Token of Token.T
|
||||
| Markup_Token of string * string * Input.source
|
||||
| Markup_Env_Token of string * string * Input.source
|
||||
| Raw_Token of Input.source;
|
||||
|
||||
fun basic_token pred (Basic_Token tok) = pred tok
|
||||
| basic_token _ _ = false;
|
||||
|
||||
val improper_token = basic_token Token.is_improper;
|
||||
val comment_token = basic_token Token.is_comment;
|
||||
val blank_token = basic_token Token.is_blank;
|
||||
val newline_token = basic_token Token.is_newline;
|
||||
|
||||
|
||||
(* output token *)
|
||||
|
||||
fun output_token state tok =
|
||||
(case tok of
|
||||
No_Token => ""
|
||||
| Basic_Token tok => Latex.output_token tok
|
||||
| Markup_Token (cmd, meta_args, source) =>
|
||||
"%\n\\isamarkup" ^ cmd ^ meta_args ^ "{" ^ output_text state {markdown = false} source ^ "%\n}\n"
|
||||
| Markup_Env_Token (cmd, meta_args, source) =>
|
||||
Latex.environment ("isamarkup" ^ cmd)
|
||||
(meta_args ^ output_text state {markdown = true} source)
|
||||
| Raw_Token source =>
|
||||
"%\n" ^ output_text state {markdown = true} source ^ "\n");
|
||||
|
||||
|
||||
(* command spans *)
|
||||
|
||||
type command = string * Position.T * string list; (*name, position, tags*)
|
||||
type source = (token * (string * int)) list; (*token, markup flag, meta-comment depth*)
|
||||
|
||||
datatype span = Span of command * (source * source * source * source) * bool;
|
||||
|
||||
fun make_span cmd src =
|
||||
let
|
||||
fun take_newline (tok :: toks) =
|
||||
if newline_token (fst tok) then ([tok], toks, true)
|
||||
else ([], tok :: toks, false)
|
||||
| take_newline [] = ([], [], false);
|
||||
val (((src_prefix, src_main), src_suffix1), (src_suffix2, src_appendix, newline)) =
|
||||
src
|
||||
|> take_prefix (improper_token o fst)
|
||||
||>> take_suffix (improper_token o fst)
|
||||
||>> take_prefix (comment_token o fst)
|
||||
||> take_newline;
|
||||
in Span (cmd, (src_prefix, src_main, src_suffix1 @ src_suffix2, src_appendix), newline) end;
|
||||
|
||||
|
||||
(* present spans *)
|
||||
|
||||
local
|
||||
|
||||
fun err_bad_nesting pos =
|
||||
error ("Bad nesting of commands in presentation" ^ pos);
|
||||
|
||||
fun edge which f (x: string option, y) =
|
||||
if x = y then I
|
||||
else (case which (x, y) of NONE => I | SOME txt => Buffer.add (f txt));
|
||||
|
||||
val begin_tag = edge #2 Latex.begin_tag;
|
||||
val end_tag = edge #1 Latex.end_tag;
|
||||
fun open_delim delim e = edge #2 Latex.begin_delim e #> delim #> edge #2 Latex.end_delim e;
|
||||
fun close_delim delim e = edge #1 Latex.begin_delim e #> delim #> edge #1 Latex.end_delim e;
|
||||
|
||||
in
|
||||
|
||||
fun present_span keywords span state state' (tag_stack, active_tag, newline, buffer, present_cont) =
|
||||
let
|
||||
val present = fold (fn (tok, (flag, 0)) =>
|
||||
Buffer.add (output_token state' tok)
|
||||
#> Buffer.add flag
|
||||
| _ => I);
|
||||
|
||||
val Span ((cmd_name, cmd_pos, cmd_tags), srcs, span_newline) = span;
|
||||
|
||||
val (tag, tags) = tag_stack;
|
||||
val tag' = try hd (fold (update (op =)) cmd_tags (the_list tag));
|
||||
|
||||
val nesting = Toplevel.level state' - Toplevel.level state;
|
||||
|
||||
val active_tag' =
|
||||
if is_some tag' then tag'
|
||||
else if cmd_name = "end" andalso not (Toplevel.is_toplevel state') then NONE
|
||||
else
|
||||
(case Keyword.command_tags keywords cmd_name of
|
||||
default_tag :: _ => SOME default_tag
|
||||
| [] =>
|
||||
if Keyword.is_vacuous keywords cmd_name andalso Toplevel.is_proof state
|
||||
then active_tag
|
||||
else NONE);
|
||||
|
||||
val edge = (active_tag, active_tag');
|
||||
|
||||
val newline' =
|
||||
if is_none active_tag' then span_newline else newline;
|
||||
|
||||
val tag_stack' =
|
||||
if nesting = 0 andalso not (Toplevel.is_proof state) then tag_stack
|
||||
else if nesting >= 0 then (tag', replicate nesting tag @ tags)
|
||||
else
|
||||
(case drop (~ nesting - 1) tags of
|
||||
tg :: tgs => (tg, tgs)
|
||||
| [] => err_bad_nesting (Position.here cmd_pos));
|
||||
|
||||
val buffer' =
|
||||
buffer
|
||||
|> end_tag edge
|
||||
|> close_delim (fst present_cont) edge
|
||||
|> snd present_cont
|
||||
|> open_delim (present (#1 srcs)) edge
|
||||
|> begin_tag edge
|
||||
|> present (#2 srcs);
|
||||
val present_cont' =
|
||||
if newline then (present (#3 srcs), present (#4 srcs))
|
||||
else (I, present (#3 srcs) #> present (#4 srcs));
|
||||
in (tag_stack', active_tag', newline', buffer', present_cont') end;
|
||||
|
||||
fun present_trailer ((_, tags), active_tag, _, buffer, present_cont) =
|
||||
if not (null tags) then err_bad_nesting " at end of theory"
|
||||
else
|
||||
buffer
|
||||
|> end_tag (active_tag, NONE)
|
||||
|> close_delim (fst present_cont) (active_tag, NONE)
|
||||
|> snd present_cont;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* present_thy *)
|
||||
|
||||
local
|
||||
|
||||
val markup_true = "\\isamarkuptrue%\n";
|
||||
val markup_false = "\\isamarkupfalse%\n";
|
||||
|
||||
val space_proper =
|
||||
Scan.one Token.is_blank -- Scan.many Token.is_comment -- Scan.one Token.is_proper;
|
||||
|
||||
val is_improper = not o (Token.is_proper orf Token.is_begin_ignore orf Token.is_end_ignore);
|
||||
val improper = Scan.many is_improper;
|
||||
val improper_end = Scan.repeat (Scan.unless space_proper (Scan.one is_improper));
|
||||
val blank_end = Scan.repeat (Scan.unless space_proper (Scan.one Token.is_blank));
|
||||
|
||||
val opt_newline = Scan.option (Scan.one Token.is_newline);
|
||||
|
||||
val ignore =
|
||||
Scan.depend (fn d => opt_newline |-- Scan.one Token.is_begin_ignore
|
||||
>> pair (d + 1)) ||
|
||||
Scan.depend (fn d => Scan.one Token.is_end_ignore --|
|
||||
(if d = 0 then Scan.fail_with (K (fn () => "Bad nesting of meta-comments")) else opt_newline)
|
||||
>> pair (d - 1));
|
||||
|
||||
val tag = (improper -- Parse.$$$ "%" -- improper) |-- Parse.!!! (Parse.tag_name --| blank_end);
|
||||
|
||||
val locale =
|
||||
Scan.option ((Parse.$$$ "(" -- improper -- Parse.$$$ "in") |--
|
||||
Parse.!!! (improper |-- Parse.name --| (improper -- Parse.$$$ ")")));
|
||||
|
||||
val meta_args_parser_hook = Unsynchronized.ref((fn thy => fn s => ("",s)): theory -> string parser)
|
||||
|
||||
|
||||
in
|
||||
|
||||
fun present_thy thy command_results toks =
|
||||
let
|
||||
val keywords = Thy_Header.get_keywords thy;
|
||||
|
||||
(* tokens *)
|
||||
|
||||
val ignored = Scan.state --| ignore
|
||||
>> (fn d => (NONE, (No_Token, ("", d))));
|
||||
|
||||
fun markup pred mk flag = Scan.peek (fn d =>
|
||||
improper |--
|
||||
Parse.position (Scan.one (fn tok => Token.is_command tok andalso
|
||||
pred keywords (Token.content_of tok))) --
|
||||
Scan.repeat tag --
|
||||
(improper |--
|
||||
(Parse.!!!!
|
||||
( (!meta_args_parser_hook thy)
|
||||
-- ( (improper -- locale -- improper)
|
||||
|-- (Parse.document_source))
|
||||
--| improper_end)))
|
||||
>> (fn (((tok, pos'), tags), (meta_args,source)) =>
|
||||
let val name = Token.content_of tok
|
||||
in (SOME (name, pos', tags), (mk (name, meta_args, source), (flag, d))) end));
|
||||
|
||||
val command = Scan.peek (fn d =>
|
||||
Scan.optional (Scan.one Token.is_command_modifier ::: improper) [] --
|
||||
Scan.one Token.is_command -- Scan.repeat tag
|
||||
>> (fn ((cmd_mod, cmd), tags) =>
|
||||
map (fn tok => (NONE, (Basic_Token tok, ("", d)))) cmd_mod @
|
||||
[(SOME (Token.content_of cmd, Token.pos_of cmd, tags),
|
||||
(Basic_Token cmd, (markup_false, d)))]));
|
||||
|
||||
val cmt = Scan.peek (fn d =>
|
||||
(Parse.$$$ "--" || Parse.$$$ Symbol.comment) |--
|
||||
Parse.!!!! (improper |-- Parse.document_source) >>
|
||||
(fn source => (NONE, (Markup_Token ("cmt", "", source), ("", d)))));
|
||||
|
||||
val other = Scan.peek (fn d =>
|
||||
Parse.not_eof >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
||||
|
||||
val tokens =
|
||||
(ignored ||
|
||||
markup Keyword.is_document_heading Markup_Token markup_true ||
|
||||
markup Keyword.is_document_body Markup_Env_Token markup_true ||
|
||||
markup Keyword.is_document_raw (Raw_Token o #3) "") >> single ||
|
||||
command ||
|
||||
(cmt || other) >> single;
|
||||
|
||||
|
||||
(* spans *)
|
||||
|
||||
val is_eof = fn (_, (Basic_Token x, _)) => Token.is_eof x | _ => false;
|
||||
val stopper = Scan.stopper (K (NONE, (Basic_Token Token.eof, ("", 0)))) is_eof;
|
||||
|
||||
val cmd = Scan.one (is_some o fst);
|
||||
val non_cmd = Scan.one (is_none o fst andf not o is_eof) >> #2;
|
||||
|
||||
val comments = Scan.many (comment_token o fst o snd);
|
||||
val blank = Scan.one (blank_token o fst o snd);
|
||||
val newline = Scan.one (newline_token o fst o snd);
|
||||
val before_cmd =
|
||||
Scan.option (newline -- comments) --
|
||||
Scan.option (newline -- comments) --
|
||||
Scan.option (blank -- comments) -- cmd;
|
||||
|
||||
val span =
|
||||
Scan.repeat non_cmd -- cmd --
|
||||
Scan.repeat (Scan.unless before_cmd non_cmd) --
|
||||
Scan.option (newline >> (single o snd))
|
||||
>> (fn (((toks1, (cmd, tok2)), toks3), tok4) =>
|
||||
make_span (the cmd) (toks1 @ (tok2 :: (toks3 @ the_default [] tok4))));
|
||||
|
||||
val spans = toks
|
||||
|> take_suffix Token.is_space |> #1
|
||||
|> Source.of_list
|
||||
|> Source.source' 0 Token.stopper (Scan.error (Scan.bulk tokens >> flat))
|
||||
|> Source.source stopper (Scan.error (Scan.bulk span))
|
||||
|> Source.exhaust;
|
||||
|
||||
(* present commands *)
|
||||
|
||||
fun present_command tr span st st' =
|
||||
Toplevel.setmp_thread_position tr (present_span keywords span st st');
|
||||
|
||||
fun present _ [] = I
|
||||
| present st (((tr, st'), span) :: rest) = present_command tr span st st' #> present st' rest;
|
||||
in
|
||||
if length command_results = length spans then
|
||||
((NONE, []), NONE, true, Buffer.empty, (I, I))
|
||||
|> present Toplevel.toplevel (command_results ~~ spans)
|
||||
|> present_trailer
|
||||
else error "Messed-up outer syntax for presentation"
|
||||
end;
|
||||
|
||||
fun set_meta_args_parser f = (meta_args_parser_hook:= f)
|
||||
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** setup default output **)
|
||||
|
||||
(* options *)
|
||||
|
||||
val _ = Theory.setup
|
||||
(add_option @{binding show_types} (Config.put show_types o boolean) #>
|
||||
add_option @{binding show_sorts} (Config.put show_sorts o boolean) #>
|
||||
add_option @{binding show_structs} (Config.put show_structs o boolean) #>
|
||||
add_option @{binding show_question_marks} (Config.put show_question_marks o boolean) #>
|
||||
add_option @{binding show_abbrevs} (Config.put show_abbrevs o boolean) #>
|
||||
add_option @{binding names_long} (Config.put Name_Space.names_long o boolean) #>
|
||||
add_option @{binding names_short} (Config.put Name_Space.names_short o boolean) #>
|
||||
add_option @{binding names_unique} (Config.put Name_Space.names_unique o boolean) #>
|
||||
add_option @{binding eta_contract} (Config.put Syntax_Trans.eta_contract o boolean) #>
|
||||
add_option @{binding display} (Config.put display o boolean) #>
|
||||
add_option @{binding break} (Config.put break o boolean) #>
|
||||
add_option @{binding quotes} (Config.put quotes o boolean) #>
|
||||
add_option @{binding mode} (add_wrapper o Print_Mode.with_modes o single) #>
|
||||
add_option @{binding margin} (Config.put margin o integer) #>
|
||||
add_option @{binding indent} (Config.put indent o integer) #>
|
||||
add_option @{binding source} (Config.put source o boolean) #>
|
||||
add_option @{binding goals_limit} (Config.put Goal_Display.goals_limit o integer));
|
||||
|
||||
|
||||
(* basic pretty printing *)
|
||||
|
||||
fun perhaps_trim ctxt =
|
||||
not (Config.get ctxt display) ? Symbol.trim_blanks;
|
||||
|
||||
fun pretty_text ctxt =
|
||||
Pretty.chunks o map Pretty.str o map (perhaps_trim ctxt) o split_lines;
|
||||
|
||||
fun pretty_term ctxt t = Syntax.pretty_term (Variable.auto_fixes t ctxt) t;
|
||||
|
||||
fun pretty_thm ctxt = pretty_term ctxt o Thm.full_prop_of;
|
||||
|
||||
fun pretty_term_style ctxt (style, t) =
|
||||
pretty_term ctxt (style t);
|
||||
|
||||
fun pretty_thm_style ctxt (style, th) =
|
||||
pretty_term ctxt (style (Thm.full_prop_of th));
|
||||
|
||||
fun pretty_term_typ ctxt (style, t) =
|
||||
let val t' = style t
|
||||
in pretty_term ctxt (Type.constraint (Term.fastype_of t') t') end;
|
||||
|
||||
fun pretty_term_typeof ctxt (style, t) =
|
||||
Syntax.pretty_typ ctxt (Term.fastype_of (style t));
|
||||
|
||||
fun pretty_const ctxt c =
|
||||
let
|
||||
val t = Const (c, Consts.type_scheme (Proof_Context.consts_of ctxt) c)
|
||||
handle TYPE (msg, _, _) => error msg;
|
||||
val ([t'], _) = Variable.import_terms true [t] ctxt;
|
||||
in pretty_term ctxt t' end;
|
||||
|
||||
fun pretty_abbrev ctxt s =
|
||||
let
|
||||
val t = Syntax.read_term (Proof_Context.set_mode Proof_Context.mode_abbrev ctxt) s;
|
||||
fun err () = error ("Abbreviated constant expected: " ^ Syntax.string_of_term ctxt t);
|
||||
val (head, args) = Term.strip_comb t;
|
||||
val (c, T) = Term.dest_Const head handle TERM _ => err ();
|
||||
val (U, u) = Consts.the_abbreviation (Proof_Context.consts_of ctxt) c
|
||||
handle TYPE _ => err ();
|
||||
val t' = Term.betapplys (Envir.expand_atom T (U, u), args);
|
||||
val eq = Logic.mk_equals (t, t');
|
||||
val ctxt' = Variable.auto_fixes eq ctxt;
|
||||
in Proof_Context.pretty_term_abbrev ctxt' eq end;
|
||||
|
||||
fun pretty_locale ctxt (name, pos) =
|
||||
let
|
||||
val thy = Proof_Context.theory_of ctxt
|
||||
in (Pretty.str o Locale.extern thy o Locale.check thy) (name, pos) end;
|
||||
|
||||
fun pretty_class ctxt =
|
||||
Pretty.str o Proof_Context.extern_class ctxt o Proof_Context.read_class ctxt;
|
||||
|
||||
fun pretty_type ctxt s =
|
||||
let val Type (name, _) = Proof_Context.read_type_name {proper = true, strict = false} ctxt s
|
||||
in Pretty.str (Proof_Context.extern_type ctxt name) end;
|
||||
|
||||
fun pretty_prf full ctxt = Proof_Syntax.pretty_clean_proof_of ctxt full;
|
||||
|
||||
fun pretty_theory ctxt (name, pos) = (Theory.check ctxt (name, pos); Pretty.str name);
|
||||
|
||||
|
||||
(* default output *)
|
||||
|
||||
val str_of_source = space_implode " " o map Token.unparse o Token.args_of_src;
|
||||
|
||||
fun maybe_pretty_source pretty ctxt src xs =
|
||||
map (pretty ctxt) xs (*always pretty in order to exhibit errors!*)
|
||||
|> (if Config.get ctxt source then K [pretty_text ctxt (str_of_source src)] else I);
|
||||
|
||||
fun string_of_margin ctxt = Pretty.string_of_margin (Config.get ctxt margin);
|
||||
|
||||
fun output ctxt prts =
|
||||
prts
|
||||
|> Config.get ctxt quotes ? map Pretty.quote
|
||||
|> (if Config.get ctxt display then
|
||||
map (Pretty.indent (Config.get ctxt indent) #> string_of_margin ctxt #> Output.output)
|
||||
#> space_implode "\\isasep\\isanewline%\n"
|
||||
#> Latex.environment "isabelle"
|
||||
else
|
||||
map
|
||||
((if Config.get ctxt break then string_of_margin ctxt else Pretty.unformatted_string_of)
|
||||
#> Output.output)
|
||||
#> space_implode "\\isasep\\isanewline%\n"
|
||||
#> enclose "\\isa{" "}");
|
||||
|
||||
|
||||
(* verbatim text *)
|
||||
|
||||
fun verbatim_text ctxt =
|
||||
if Config.get ctxt display then
|
||||
split_lines #> map (prefix (Symbol.spaces (Config.get ctxt indent))) #> cat_lines #>
|
||||
Latex.output_ascii #> Latex.environment "isabellett"
|
||||
else
|
||||
split_lines #>
|
||||
map (Latex.output_ascii #> enclose "\\isatt{" "}") #>
|
||||
space_implode "\\isasep\\isanewline%\n";
|
||||
|
||||
|
||||
(* antiquotations for basic entities *)
|
||||
|
||||
local
|
||||
|
||||
fun basic_entities name scan pretty =
|
||||
antiquotation name scan (fn {source, context = ctxt, ...} =>
|
||||
output ctxt o maybe_pretty_source pretty ctxt source);
|
||||
|
||||
fun basic_entities_style name scan pretty =
|
||||
antiquotation name scan (fn {source, context = ctxt, ...} => fn (style, xs) =>
|
||||
output ctxt
|
||||
(maybe_pretty_source (fn ctxt => fn x => pretty ctxt (style, x)) ctxt source xs));
|
||||
|
||||
fun basic_entity name scan = basic_entities name (scan >> single);
|
||||
|
||||
in
|
||||
|
||||
val _ = Theory.setup
|
||||
(basic_entities_style @{binding thm} (Term_Style.parse -- Attrib.thms) pretty_thm_style #>
|
||||
basic_entity @{binding prop} (Term_Style.parse -- Args.prop) pretty_term_style #>
|
||||
basic_entity @{binding term} (Term_Style.parse -- Args.term) pretty_term_style #>
|
||||
basic_entity @{binding term_type} (Term_Style.parse -- Args.term) pretty_term_typ #>
|
||||
basic_entity @{binding typeof} (Term_Style.parse -- Args.term) pretty_term_typeof #>
|
||||
basic_entity @{binding const} (Args.const {proper = true, strict = false}) pretty_const #>
|
||||
basic_entity @{binding abbrev} (Scan.lift Args.embedded_inner_syntax) pretty_abbrev #>
|
||||
basic_entity @{binding typ} Args.typ_abbrev Syntax.pretty_typ #>
|
||||
basic_entity @{binding locale} (Scan.lift (Parse.position Args.name)) pretty_locale #>
|
||||
basic_entity @{binding class} (Scan.lift Args.embedded_inner_syntax) pretty_class #>
|
||||
basic_entity @{binding type} (Scan.lift Args.embedded) pretty_type #>
|
||||
basic_entities @{binding prf} Attrib.thms (pretty_prf false) #>
|
||||
basic_entities @{binding full_prf} Attrib.thms (pretty_prf true) #>
|
||||
basic_entity @{binding theory} (Scan.lift (Parse.position Args.name)) pretty_theory);
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(** document command **)
|
||||
|
||||
|
||||
fun document_command markdown (loc, txt) =
|
||||
Toplevel.keep (fn state =>
|
||||
(case loc of
|
||||
NONE => ignore (output_text state markdown txt)
|
||||
| SOME (_, pos) =>
|
||||
error ("Illegal target specification -- not a theory context" ^ Position.here pos))) o
|
||||
Toplevel.present_local_theory loc (fn state => ignore (output_text state markdown txt));
|
||||
|
||||
end;
|
|
@ -410,9 +410,13 @@ type segment = {span: Command_Span.span, command: Toplevel.transition, state: To
|
|||
|
||||
fun present_thy options thy (segments: segment list) =
|
||||
let
|
||||
val X = if null segments then Command_Span.content (#span(hd segments)) else (writeln "HUMPFLEPUMPF0";[])
|
||||
|
||||
val X = List.concat(map (fn X => Command_Span.content (#span X)) segments)
|
||||
val Y = (String.concatWith "::") (map Token.content_of X)
|
||||
val _ = writeln("HUMPFLEPUMF"^Y^Context.theory_name thy)
|
||||
val _ = writeln("present_thy BEGIN:"^Context.theory_name thy^"\n"
|
||||
^Y^
|
||||
"present_thy END:"^Context.theory_name thy^"\n")
|
||||
|
||||
val keywords = Thy_Header.get_keywords thy;
|
||||
|
||||
|
||||
|
@ -453,7 +457,8 @@ fun present_thy options thy (segments: segment list) =
|
|||
Scan.one Token.is_command -- Scan.repeat tag
|
||||
>> (fn ((cmd_mod, cmd), tags) =>
|
||||
map (fn tok => (NONE, (Basic_Token tok, ("", d)))) cmd_mod @
|
||||
[(SOME (let val _ = writeln("ZZZ::" ^ Token.content_of cmd)
|
||||
[(SOME (let val _ = writeln("present_thy ZZZ:"^ Context.theory_name thy
|
||||
^ Token.content_of cmd^"\n")
|
||||
in
|
||||
Token.content_of cmd
|
||||
end, Token.pos_of cmd, tags),
|
||||
|
@ -504,7 +509,7 @@ fun present_thy options thy (segments: segment list) =
|
|||
>> (fn (((toks1, (cmd, tok2)), toks3), tok4) =>
|
||||
make_span (the cmd) (toks1 @ (tok2 :: (toks3 @ the_default [] tok4))));
|
||||
|
||||
val _ = writeln("HUMPFLEPUMF'"^Y)
|
||||
val _ = writeln("HUMPFLEPUMF'")
|
||||
|
||||
val spans = segments
|
||||
|> maps (Command_Span.content o #span)
|
||||
|
@ -529,7 +534,7 @@ fun present_thy options thy (segments: segment list) =
|
|||
fun present _ [] = I
|
||||
| present st ((span, (tr, st')) :: rest) = present_command tr span st st' #> present st' rest;
|
||||
|
||||
val _ = writeln("HUMPFLEPUMF''"^Y^Context.theory_name thy)
|
||||
val _ = writeln("HUMPFLEPUMF''"^Context.theory_name thy)
|
||||
|
||||
in
|
||||
if length command_results = length spans then
|
||||
|
@ -540,7 +545,8 @@ fun present_thy options thy (segments: segment list) =
|
|||
else error "Messed-up outer syntax for presentation"
|
||||
end;
|
||||
|
||||
fun set_meta_args_parser f = (meta_args_parser_hook:= f)
|
||||
fun set_meta_args_parser f = let val _ = writeln "Meta-args parser set to new value"
|
||||
in (meta_args_parser_hook:= f) end
|
||||
|
||||
end;
|
||||
|
||||
|
|
|
@ -0,0 +1,555 @@
|
|||
(* Title: Pure/Thy/thy_output.ML
|
||||
Author: Makarius
|
||||
|
||||
Theory document output.
|
||||
*)
|
||||
|
||||
signature THY_OUTPUT =
|
||||
sig
|
||||
val output_document: Proof.context -> {markdown: bool} -> Input.source -> Latex.text list
|
||||
val check_comments: Proof.context -> Symbol_Pos.T list -> unit
|
||||
val output_token: Proof.context -> Token.T -> Latex.text list
|
||||
val output_source: Proof.context -> string -> Latex.text list
|
||||
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state}
|
||||
val present_thy: Options.T -> theory -> segment list -> Latex.text list
|
||||
val set_meta_args_parser : (theory -> string parser) -> unit
|
||||
val pretty_term: Proof.context -> term -> Pretty.T
|
||||
val pretty_thm: Proof.context -> thm -> Pretty.T
|
||||
val lines: Latex.text list -> Latex.text list
|
||||
val items: Latex.text list -> Latex.text list
|
||||
val isabelle: Proof.context -> Latex.text list -> Latex.text
|
||||
val isabelle_typewriter: Proof.context -> Latex.text list -> Latex.text
|
||||
val typewriter: Proof.context -> string -> Latex.text
|
||||
val verbatim: Proof.context -> string -> Latex.text
|
||||
val source: Proof.context -> Token.src -> Latex.text
|
||||
val pretty: Proof.context -> Pretty.T -> Latex.text
|
||||
val pretty_source: Proof.context -> Token.src -> Pretty.T -> Latex.text
|
||||
val pretty_items: Proof.context -> Pretty.T list -> Latex.text
|
||||
val pretty_items_source: Proof.context -> Token.src -> Pretty.T list -> Latex.text
|
||||
val antiquotation_pretty:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
||||
val antiquotation_pretty_source:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
||||
val antiquotation_raw:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Latex.text) -> theory -> theory
|
||||
val antiquotation_verbatim:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> string) -> theory -> theory
|
||||
end;
|
||||
|
||||
structure Thy_Output: THY_OUTPUT =
|
||||
struct
|
||||
|
||||
(* output document source *)
|
||||
|
||||
val output_symbols = single o Latex.symbols_output;
|
||||
|
||||
fun output_comment ctxt (kind, syms) =
|
||||
(case kind of
|
||||
Comment.Comment =>
|
||||
Input.cartouche_content syms
|
||||
|> output_document (ctxt |> Config.put Document_Antiquotation.thy_output_display false)
|
||||
{markdown = false}
|
||||
|> Latex.enclose_body "%\n\\isamarkupcmt{" "%\n}"
|
||||
| Comment.Cancel =>
|
||||
Symbol_Pos.cartouche_content syms
|
||||
|> output_symbols
|
||||
|> Latex.enclose_body "%\n\\isamarkupcancel{" "}"
|
||||
| Comment.Latex =>
|
||||
[Latex.symbols (Symbol_Pos.cartouche_content syms)])
|
||||
and output_comment_document ctxt (comment, syms) =
|
||||
(case comment of
|
||||
SOME kind => output_comment ctxt (kind, syms)
|
||||
| NONE => [Latex.symbols syms])
|
||||
and output_document_text ctxt syms =
|
||||
Comment.read_body syms |> maps (output_comment_document ctxt)
|
||||
and output_document ctxt {markdown} source =
|
||||
let
|
||||
val pos = Input.pos_of source;
|
||||
val syms = Input.source_explode source;
|
||||
|
||||
val output_antiquotes =
|
||||
maps (Document_Antiquotation.evaluate (output_document_text ctxt) ctxt);
|
||||
|
||||
fun output_line line =
|
||||
(if Markdown.line_is_item line then [Latex.string "\\item "] else []) @
|
||||
output_antiquotes (Markdown.line_content line);
|
||||
|
||||
fun output_block (Markdown.Par lines) =
|
||||
Latex.block (separate (Latex.string "\n") (map (Latex.block o output_line) lines))
|
||||
| output_block (Markdown.List {kind, body, ...}) =
|
||||
Latex.environment_block (Markdown.print_kind kind) (output_blocks body)
|
||||
and output_blocks blocks = separate (Latex.string "\n\n") (map output_block blocks);
|
||||
in
|
||||
if Toplevel.is_skipped_proof (Toplevel.presentation_state ctxt) then []
|
||||
else if markdown andalso exists (Markdown.is_control o Symbol_Pos.symbol) syms
|
||||
then
|
||||
let
|
||||
val ants = Antiquote.parse_comments pos syms;
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val blocks = Markdown.read_antiquotes ants;
|
||||
val _ = Context_Position.reports ctxt (reports @ Markdown.reports blocks);
|
||||
in output_blocks blocks end
|
||||
else
|
||||
let
|
||||
val ants = Antiquote.parse_comments pos (trim (Symbol.is_blank o Symbol_Pos.symbol) syms);
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val _ = Context_Position.reports ctxt (reports @ Markdown.text_reports ants);
|
||||
in output_antiquotes ants end
|
||||
end;
|
||||
|
||||
|
||||
(* output tokens with formal comments *)
|
||||
|
||||
local
|
||||
|
||||
val output_symbols_antiq =
|
||||
(fn Antiquote.Text syms => output_symbols syms
|
||||
| Antiquote.Control {name = (name, _), body, ...} =>
|
||||
Latex.string (Latex.output_symbols [Symbol.encode (Symbol.Control name)]) ::
|
||||
output_symbols body
|
||||
| Antiquote.Antiq {body, ...} =>
|
||||
Latex.enclose_body "%\n\\isaantiq\n" "{}%\n\\endisaantiq\n" (output_symbols body));
|
||||
|
||||
fun output_comment_symbols ctxt {antiq} (comment, syms) =
|
||||
(case (comment, antiq) of
|
||||
(NONE, false) => output_symbols syms
|
||||
| (NONE, true) =>
|
||||
Antiquote.parse_comments (#1 (Symbol_Pos.range syms)) syms
|
||||
|> maps output_symbols_antiq
|
||||
| (SOME comment, _) => output_comment ctxt (comment, syms));
|
||||
|
||||
fun output_body ctxt antiq bg en syms =
|
||||
Comment.read_body syms
|
||||
|> maps (output_comment_symbols ctxt {antiq = antiq})
|
||||
|> Latex.enclose_body bg en;
|
||||
|
||||
in
|
||||
|
||||
fun output_token ctxt tok =
|
||||
let
|
||||
fun output antiq bg en =
|
||||
output_body ctxt antiq bg en (Input.source_explode (Token.input_of tok));
|
||||
in
|
||||
(case Token.kind_of tok of
|
||||
Token.Comment NONE => []
|
||||
| Token.Command => output false "\\isacommand{" "}"
|
||||
| Token.Keyword =>
|
||||
if Symbol.is_ascii_identifier (Token.content_of tok)
|
||||
then output false "\\isakeyword{" "}"
|
||||
else output false "" ""
|
||||
| Token.String => output false "{\\isachardoublequoteopen}" "{\\isachardoublequoteclose}"
|
||||
| Token.Alt_String => output false "{\\isacharbackquoteopen}" "{\\isacharbackquoteclose}"
|
||||
| Token.Verbatim => output true "{\\isacharverbatimopen}" "{\\isacharverbatimclose}"
|
||||
| Token.Cartouche => output false "{\\isacartoucheopen}" "{\\isacartoucheclose}"
|
||||
| _ => output false "" "")
|
||||
end handle ERROR msg => error (msg ^ Position.here (Token.pos_of tok));
|
||||
|
||||
fun output_source ctxt s =
|
||||
output_body ctxt false "" "" (Symbol_Pos.explode (s, Position.none));
|
||||
|
||||
fun check_comments ctxt =
|
||||
Comment.read_body #> List.app (fn (comment, syms) =>
|
||||
let
|
||||
val pos = #1 (Symbol_Pos.range syms);
|
||||
val _ =
|
||||
comment |> Option.app (fn kind =>
|
||||
Context_Position.reports ctxt (map (pair pos) (Markup.cartouche :: Comment.markups kind)));
|
||||
val _ = output_comment_symbols ctxt {antiq = false} (comment, syms);
|
||||
in if comment = SOME Comment.Comment then check_comments ctxt syms else () end);
|
||||
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** present theory source **)
|
||||
|
||||
(*NB: arranging white space around command spans is a black art*)
|
||||
|
||||
val is_white = Token.is_space orf Token.is_informal_comment;
|
||||
val is_black = not o is_white;
|
||||
|
||||
val is_white_comment = Token.is_informal_comment;
|
||||
val is_black_comment = Token.is_formal_comment;
|
||||
|
||||
|
||||
(* presentation tokens *)
|
||||
|
||||
datatype token =
|
||||
Ignore_Token
|
||||
| Basic_Token of Token.T
|
||||
| Markup_Token of string * Input.source
|
||||
| Markup_Env_Token of string * Input.source
|
||||
| Raw_Token of Input.source;
|
||||
|
||||
fun basic_token pred (Basic_Token tok) = pred tok
|
||||
| basic_token _ _ = false;
|
||||
|
||||
val white_token = basic_token is_white;
|
||||
val white_comment_token = basic_token is_white_comment;
|
||||
val blank_token = basic_token Token.is_blank;
|
||||
val newline_token = basic_token Token.is_newline;
|
||||
|
||||
fun present_token ctxt tok =
|
||||
(case tok of
|
||||
Ignore_Token => []
|
||||
| Basic_Token tok => output_token ctxt tok
|
||||
| Markup_Token (cmd, source) =>
|
||||
Latex.enclose_body ("%\n\\isamarkup" ^ cmd ^ "{") "%\n}\n"
|
||||
(output_document ctxt {markdown = false} source)
|
||||
| Markup_Env_Token (cmd, source) =>
|
||||
[Latex.environment_block ("isamarkup" ^ cmd) (output_document ctxt {markdown = true} source)]
|
||||
| Raw_Token source =>
|
||||
Latex.string "%\n" :: output_document ctxt {markdown = true} source @ [Latex.string "\n"]);
|
||||
|
||||
|
||||
(* command spans *)
|
||||
|
||||
type command = string * Position.T * string list; (*name, position, tags*)
|
||||
type source = (token * (string * int)) list; (*token, markup flag, meta-comment depth*)
|
||||
|
||||
datatype span = Span of command * (source * source * source * source) * bool;
|
||||
|
||||
fun make_span cmd src =
|
||||
let
|
||||
fun chop_newline (tok :: toks) =
|
||||
if newline_token (fst tok) then ([tok], toks, true)
|
||||
else ([], tok :: toks, false)
|
||||
| chop_newline [] = ([], [], false);
|
||||
val (((src_prefix, src_main), src_suffix1), (src_suffix2, src_appendix, newline)) =
|
||||
src
|
||||
|> chop_prefix (white_token o fst)
|
||||
||>> chop_suffix (white_token o fst)
|
||||
||>> chop_prefix (white_comment_token o fst)
|
||||
||> chop_newline;
|
||||
in Span (cmd, (src_prefix, src_main, src_suffix1 @ src_suffix2, src_appendix), newline) end;
|
||||
|
||||
|
||||
(* present spans *)
|
||||
|
||||
local
|
||||
|
||||
fun err_bad_nesting pos =
|
||||
error ("Bad nesting of commands in presentation" ^ pos);
|
||||
|
||||
fun edge which f (x: string option, y) =
|
||||
if x = y then I
|
||||
else (case which (x, y) of NONE => I | SOME txt => cons (Latex.string (f txt)));
|
||||
|
||||
val begin_tag = edge #2 Latex.begin_tag;
|
||||
val end_tag = edge #1 Latex.end_tag;
|
||||
fun open_delim delim e = edge #2 Latex.begin_delim e #> delim #> edge #2 Latex.end_delim e;
|
||||
fun close_delim delim e = edge #1 Latex.begin_delim e #> delim #> edge #1 Latex.end_delim e;
|
||||
|
||||
fun read_tag s =
|
||||
(case space_explode "%" s of
|
||||
["", b] => (SOME b, NONE)
|
||||
| [a, b] => (NONE, SOME (a, b))
|
||||
| _ => error ("Bad document_tags specification: " ^ quote s));
|
||||
|
||||
in
|
||||
|
||||
fun make_command_tag options keywords =
|
||||
let
|
||||
val document_tags =
|
||||
map read_tag (space_explode "," (Options.string options \<^system_option>\<open>document_tags\<close>));
|
||||
val document_tags_default = map_filter #1 document_tags;
|
||||
val document_tags_command = map_filter #2 document_tags;
|
||||
in
|
||||
fn {cmd_name, cmd_tags, tag, active_tag} => fn state => fn state' =>
|
||||
let
|
||||
val tag' = try hd (fold (update (op =)) cmd_tags (the_list tag));
|
||||
|
||||
val keyword_tags =
|
||||
if cmd_name = "end" andalso Toplevel.is_end_theory state' then ["theory"]
|
||||
else Keyword.command_tags keywords cmd_name;
|
||||
val command_tags =
|
||||
the_list (AList.lookup (op =) document_tags_command cmd_name) @
|
||||
keyword_tags @ document_tags_default;
|
||||
|
||||
val active_tag' =
|
||||
if is_some tag' then tag'
|
||||
else
|
||||
(case command_tags of
|
||||
default_tag :: _ => SOME default_tag
|
||||
| [] =>
|
||||
if Keyword.is_vacuous keywords cmd_name andalso Toplevel.is_proof state
|
||||
then active_tag
|
||||
else NONE);
|
||||
in {tag' = tag', active_tag' = active_tag'} end
|
||||
end;
|
||||
|
||||
fun present_span thy command_tag span state state'
|
||||
(tag_stack, active_tag, newline, latex, present_cont) =
|
||||
let
|
||||
val ctxt' =
|
||||
Toplevel.presentation_context state'
|
||||
handle Toplevel.UNDEF => Proof_Context.get_global thy Context.PureN;
|
||||
val present = fold (fn (tok, (flag, 0)) =>
|
||||
fold cons (present_token ctxt' tok)
|
||||
#> cons (Latex.string flag)
|
||||
| _ => I);
|
||||
|
||||
val Span ((cmd_name, cmd_pos, cmd_tags), srcs, span_newline) = span;
|
||||
|
||||
val (tag, tags) = tag_stack;
|
||||
val {tag', active_tag'} =
|
||||
command_tag {cmd_name = cmd_name, cmd_tags = cmd_tags, tag = tag, active_tag = active_tag}
|
||||
state state';
|
||||
val edge = (active_tag, active_tag');
|
||||
|
||||
val nesting = Toplevel.level state' - Toplevel.level state;
|
||||
|
||||
val newline' =
|
||||
if is_none active_tag' then span_newline else newline;
|
||||
|
||||
val tag_stack' =
|
||||
if nesting = 0 andalso not (Toplevel.is_proof state) then tag_stack
|
||||
else if nesting >= 0 then (tag', replicate nesting tag @ tags)
|
||||
else
|
||||
(case drop (~ nesting - 1) tags of
|
||||
tg :: tgs => (tg, tgs)
|
||||
| [] => err_bad_nesting (Position.here cmd_pos));
|
||||
|
||||
val latex' =
|
||||
latex
|
||||
|> end_tag edge
|
||||
|> close_delim (fst present_cont) edge
|
||||
|> snd present_cont
|
||||
|> open_delim (present (#1 srcs)) edge
|
||||
|> begin_tag edge
|
||||
|> present (#2 srcs);
|
||||
val present_cont' =
|
||||
if newline then (present (#3 srcs), present (#4 srcs))
|
||||
else (I, present (#3 srcs) #> present (#4 srcs));
|
||||
in (tag_stack', active_tag', newline', latex', present_cont') end;
|
||||
|
||||
fun present_trailer ((_, tags), active_tag, _, latex, present_cont) =
|
||||
if not (null tags) then err_bad_nesting " at end of theory"
|
||||
else
|
||||
latex
|
||||
|> end_tag (active_tag, NONE)
|
||||
|> close_delim (fst present_cont) (active_tag, NONE)
|
||||
|> snd present_cont;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* present_thy *)
|
||||
|
||||
local
|
||||
|
||||
val markup_true = "\\isamarkuptrue%\n";
|
||||
val markup_false = "\\isamarkupfalse%\n";
|
||||
|
||||
val space_proper =
|
||||
Scan.one Token.is_blank -- Scan.many is_white_comment -- Scan.one is_black;
|
||||
|
||||
val is_improper = not o (is_black orf Token.is_begin_ignore orf Token.is_end_ignore);
|
||||
val improper = Scan.many is_improper;
|
||||
val improper_end = Scan.repeat (Scan.unless space_proper (Scan.one is_improper));
|
||||
val blank_end = Scan.repeat (Scan.unless space_proper (Scan.one Token.is_blank));
|
||||
|
||||
val opt_newline = Scan.option (Scan.one Token.is_newline);
|
||||
|
||||
val ignore =
|
||||
Scan.depend (fn d => opt_newline |-- Scan.one Token.is_begin_ignore
|
||||
>> pair (d + 1)) ||
|
||||
Scan.depend (fn d => Scan.one Token.is_end_ignore --|
|
||||
(if d = 0 then Scan.fail_with (K (fn () => "Bad nesting of meta-comments")) else opt_newline)
|
||||
>> pair (d - 1));
|
||||
|
||||
val tag = (improper -- Parse.$$$ "%" -- improper) |-- Parse.!!! (Parse.tag_name --| blank_end);
|
||||
|
||||
val locale =
|
||||
Scan.option ((Parse.$$$ "(" -- improper -- Parse.$$$ "in") |--
|
||||
Parse.!!! (improper |-- Parse.name --| (improper -- Parse.$$$ ")")));
|
||||
|
||||
val meta_args_parser_hook = Unsynchronized.ref((fn thy => fn s => ("",s)): theory -> string parser)
|
||||
|
||||
|
||||
in
|
||||
|
||||
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state};
|
||||
|
||||
fun present_thy options thy (segments: segment list) =
|
||||
let
|
||||
val keywords = Thy_Header.get_keywords thy;
|
||||
|
||||
|
||||
(* tokens *)
|
||||
|
||||
val ignored = Scan.state --| ignore
|
||||
>> (fn d => (NONE, (Ignore_Token, ("", d))));
|
||||
|
||||
fun markup pred mk flag = Scan.peek (fn d =>
|
||||
improper |--
|
||||
Parse.position (Scan.one (fn tok =>
|
||||
Token.is_command tok andalso pred keywords (Token.content_of tok))) --
|
||||
Scan.repeat tag --
|
||||
Parse.!!!! ((improper -- locale -- improper) |-- Parse.document_source --| improper_end)
|
||||
>> (fn (((tok, pos'), tags), source) =>
|
||||
let val name = Token.content_of tok
|
||||
in (SOME (name, pos', tags), (mk (name, source), (flag, d))) end));
|
||||
|
||||
val command = Scan.peek (fn d =>
|
||||
Scan.optional (Scan.one Token.is_command_modifier ::: improper) [] --
|
||||
Scan.one Token.is_command -- Scan.repeat tag
|
||||
>> (fn ((cmd_mod, cmd), tags) =>
|
||||
map (fn tok => (NONE, (Basic_Token tok, ("", d)))) cmd_mod @
|
||||
[(SOME (Token.content_of cmd, Token.pos_of cmd, tags),
|
||||
(Basic_Token cmd, (markup_false, d)))]));
|
||||
(*
|
||||
val cmt = Scan.peek (fn d =>
|
||||
Scan.one is_black_comment >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
||||
*)
|
||||
val cmt = Scan.peek (fn d =>
|
||||
(Parse.$$$ "--" || Parse.$$$ Symbol.comment) |--
|
||||
Parse.!!!! (improper |-- Parse.document_source) >>
|
||||
(fn source => (NONE, (Markup_Token ("cmt", "", source), ("", d)))));
|
||||
|
||||
|
||||
val other = Scan.peek (fn d =>
|
||||
Parse.not_eof >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
||||
|
||||
val tokens =
|
||||
(ignored ||
|
||||
markup Keyword.is_document_heading Markup_Token markup_true ||
|
||||
markup Keyword.is_document_body Markup_Env_Token markup_true ||
|
||||
markup Keyword.is_document_raw (Raw_Token o #2) "") >> single ||
|
||||
command ||
|
||||
(cmt || other) >> single;
|
||||
|
||||
|
||||
(* spans *)
|
||||
|
||||
val is_eof = fn (_, (Basic_Token x, _)) => Token.is_eof x | _ => false;
|
||||
val stopper = Scan.stopper (K (NONE, (Basic_Token Token.eof, ("", 0)))) is_eof;
|
||||
|
||||
val cmd = Scan.one (is_some o fst);
|
||||
val non_cmd = Scan.one (is_none o fst andf not o is_eof) >> #2;
|
||||
|
||||
val white_comments = Scan.many (white_comment_token o fst o snd);
|
||||
val blank = Scan.one (blank_token o fst o snd);
|
||||
val newline = Scan.one (newline_token o fst o snd);
|
||||
val before_cmd =
|
||||
Scan.option (newline -- white_comments) --
|
||||
Scan.option (newline -- white_comments) --
|
||||
Scan.option (blank -- white_comments) -- cmd;
|
||||
|
||||
val span =
|
||||
Scan.repeat non_cmd -- cmd --
|
||||
Scan.repeat (Scan.unless before_cmd non_cmd) --
|
||||
Scan.option (newline >> (single o snd))
|
||||
>> (fn (((toks1, (cmd, tok2)), toks3), tok4) =>
|
||||
make_span (the cmd) (toks1 @ (tok2 :: (toks3 @ the_default [] tok4))));
|
||||
|
||||
val spans = segments
|
||||
|> maps (Command_Span.content o #span)
|
||||
|> drop_suffix Token.is_space
|
||||
|> Source.of_list
|
||||
|> Source.source' 0 Token.stopper (Scan.error (Scan.bulk tokens >> flat))
|
||||
|> Source.source stopper (Scan.error (Scan.bulk span))
|
||||
|> Source.exhaust;
|
||||
|
||||
val command_results =
|
||||
segments |> map_filter (fn {command, state, ...} =>
|
||||
if Toplevel.is_ignored command then NONE else SOME (command, state));
|
||||
|
||||
|
||||
(* present commands *)
|
||||
|
||||
val command_tag = make_command_tag options keywords;
|
||||
|
||||
fun present_command tr span st st' =
|
||||
Toplevel.setmp_thread_position tr (present_span thy command_tag span st st');
|
||||
|
||||
fun present _ [] = I
|
||||
| present st ((span, (tr, st')) :: rest) = present_command tr span st st' #> present st' rest;
|
||||
in
|
||||
if length command_results = length spans then
|
||||
((NONE, []), NONE, true, [], (I, I))
|
||||
|> present Toplevel.toplevel (spans ~~ command_results)
|
||||
|> present_trailer
|
||||
|> rev
|
||||
else error "Messed-up outer syntax for presentation"
|
||||
end;
|
||||
|
||||
fun set_meta_args_parser f = (meta_args_parser_hook:= f)
|
||||
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** standard output operations **)
|
||||
|
||||
(* pretty printing *)
|
||||
|
||||
fun pretty_term ctxt t =
|
||||
Syntax.pretty_term (Variable.auto_fixes t ctxt) t;
|
||||
|
||||
fun pretty_thm ctxt = pretty_term ctxt o Thm.full_prop_of;
|
||||
|
||||
|
||||
(* default output *)
|
||||
|
||||
val lines = separate (Latex.string "\\isanewline%\n");
|
||||
val items = separate (Latex.string "\\isasep\\isanewline%\n");
|
||||
|
||||
fun isabelle ctxt body =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Latex.environment_block "isabelle" body
|
||||
else Latex.block (Latex.enclose_body "\\isa{" "}" body);
|
||||
|
||||
fun isabelle_typewriter ctxt body =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Latex.environment_block "isabellett" body
|
||||
else Latex.block (Latex.enclose_body "\\isatt{" "}" body);
|
||||
|
||||
fun typewriter ctxt s =
|
||||
isabelle_typewriter ctxt [Latex.string (Latex.output_ascii s)];
|
||||
|
||||
fun verbatim ctxt =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Document_Antiquotation.indent_lines ctxt #> typewriter ctxt
|
||||
else split_lines #> map (typewriter ctxt) #> lines #> Latex.block;
|
||||
|
||||
fun source ctxt =
|
||||
Token.args_of_src
|
||||
#> map (Token.unparse #> Document_Antiquotation.prepare_lines ctxt)
|
||||
#> space_implode " "
|
||||
#> output_source ctxt
|
||||
#> isabelle ctxt;
|
||||
|
||||
fun pretty ctxt =
|
||||
Document_Antiquotation.output ctxt #> Latex.string #> single #> isabelle ctxt;
|
||||
|
||||
fun pretty_source ctxt src prt =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_source
|
||||
then source ctxt src else pretty ctxt prt;
|
||||
|
||||
fun pretty_items ctxt =
|
||||
map (Document_Antiquotation.output ctxt #> Latex.string) #> items #> isabelle ctxt;
|
||||
|
||||
fun pretty_items_source ctxt src prts =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_source
|
||||
then source ctxt src else pretty_items ctxt prts;
|
||||
|
||||
|
||||
(* antiquotation variants *)
|
||||
|
||||
fun antiquotation_pretty name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, argument = x, ...} => pretty ctxt (f ctxt x));
|
||||
|
||||
fun antiquotation_pretty_source name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, source = src, argument = x} => pretty_source ctxt src (f ctxt x));
|
||||
|
||||
fun antiquotation_raw name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, argument = x, ...} => f ctxt x);
|
||||
|
||||
fun antiquotation_verbatim name scan f =
|
||||
antiquotation_raw name scan (fn ctxt => verbatim ctxt o f ctxt);
|
||||
|
||||
end;
|
|
@ -0,0 +1,543 @@
|
|||
(* Title: Pure/Thy/thy_output.ML
|
||||
Author: Makarius
|
||||
|
||||
Theory document output.
|
||||
*)
|
||||
|
||||
signature THY_OUTPUT =
|
||||
sig
|
||||
val output_document: Proof.context -> {markdown: bool} -> Input.source -> Latex.text list
|
||||
val check_comments: Proof.context -> Symbol_Pos.T list -> unit
|
||||
val output_token: Proof.context -> Token.T -> Latex.text list
|
||||
val output_source: Proof.context -> string -> Latex.text list
|
||||
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state}
|
||||
val present_thy: Options.T -> theory -> segment list -> Latex.text list
|
||||
val pretty_term: Proof.context -> term -> Pretty.T
|
||||
val pretty_thm: Proof.context -> thm -> Pretty.T
|
||||
val lines: Latex.text list -> Latex.text list
|
||||
val items: Latex.text list -> Latex.text list
|
||||
val isabelle: Proof.context -> Latex.text list -> Latex.text
|
||||
val isabelle_typewriter: Proof.context -> Latex.text list -> Latex.text
|
||||
val typewriter: Proof.context -> string -> Latex.text
|
||||
val verbatim: Proof.context -> string -> Latex.text
|
||||
val source: Proof.context -> Token.src -> Latex.text
|
||||
val pretty: Proof.context -> Pretty.T -> Latex.text
|
||||
val pretty_source: Proof.context -> Token.src -> Pretty.T -> Latex.text
|
||||
val pretty_items: Proof.context -> Pretty.T list -> Latex.text
|
||||
val pretty_items_source: Proof.context -> Token.src -> Pretty.T list -> Latex.text
|
||||
val antiquotation_pretty:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
||||
val antiquotation_pretty_source:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Pretty.T) -> theory -> theory
|
||||
val antiquotation_raw:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> Latex.text) -> theory -> theory
|
||||
val antiquotation_verbatim:
|
||||
binding -> 'a context_parser -> (Proof.context -> 'a -> string) -> theory -> theory
|
||||
end;
|
||||
|
||||
structure Thy_Output: THY_OUTPUT =
|
||||
struct
|
||||
|
||||
(* output document source *)
|
||||
|
||||
val output_symbols = single o Latex.symbols_output;
|
||||
|
||||
fun output_comment ctxt (kind, syms) =
|
||||
(case kind of
|
||||
Comment.Comment =>
|
||||
Input.cartouche_content syms
|
||||
|> output_document (ctxt |> Config.put Document_Antiquotation.thy_output_display false)
|
||||
{markdown = false}
|
||||
|> Latex.enclose_body "%\n\\isamarkupcmt{" "%\n}"
|
||||
| Comment.Cancel =>
|
||||
Symbol_Pos.cartouche_content syms
|
||||
|> output_symbols
|
||||
|> Latex.enclose_body "%\n\\isamarkupcancel{" "}"
|
||||
| Comment.Latex =>
|
||||
[Latex.symbols (Symbol_Pos.cartouche_content syms)])
|
||||
and output_comment_document ctxt (comment, syms) =
|
||||
(case comment of
|
||||
SOME kind => output_comment ctxt (kind, syms)
|
||||
| NONE => [Latex.symbols syms])
|
||||
and output_document_text ctxt syms =
|
||||
Comment.read_body syms |> maps (output_comment_document ctxt)
|
||||
and output_document ctxt {markdown} source =
|
||||
let
|
||||
val pos = Input.pos_of source;
|
||||
val syms = Input.source_explode source;
|
||||
|
||||
val output_antiquotes =
|
||||
maps (Document_Antiquotation.evaluate (output_document_text ctxt) ctxt);
|
||||
|
||||
fun output_line line =
|
||||
(if Markdown.line_is_item line then [Latex.string "\\item "] else []) @
|
||||
output_antiquotes (Markdown.line_content line);
|
||||
|
||||
fun output_block (Markdown.Par lines) =
|
||||
Latex.block (separate (Latex.string "\n") (map (Latex.block o output_line) lines))
|
||||
| output_block (Markdown.List {kind, body, ...}) =
|
||||
Latex.environment_block (Markdown.print_kind kind) (output_blocks body)
|
||||
and output_blocks blocks = separate (Latex.string "\n\n") (map output_block blocks);
|
||||
in
|
||||
if Toplevel.is_skipped_proof (Toplevel.presentation_state ctxt) then []
|
||||
else if markdown andalso exists (Markdown.is_control o Symbol_Pos.symbol) syms
|
||||
then
|
||||
let
|
||||
val ants = Antiquote.parse_comments pos syms;
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val blocks = Markdown.read_antiquotes ants;
|
||||
val _ = Context_Position.reports ctxt (reports @ Markdown.reports blocks);
|
||||
in output_blocks blocks end
|
||||
else
|
||||
let
|
||||
val ants = Antiquote.parse_comments pos (trim (Symbol.is_blank o Symbol_Pos.symbol) syms);
|
||||
val reports = Antiquote.antiq_reports ants;
|
||||
val _ = Context_Position.reports ctxt (reports @ Markdown.text_reports ants);
|
||||
in output_antiquotes ants end
|
||||
end;
|
||||
|
||||
|
||||
(* output tokens with formal comments *)
|
||||
|
||||
local
|
||||
|
||||
val output_symbols_antiq =
|
||||
(fn Antiquote.Text syms => output_symbols syms
|
||||
| Antiquote.Control {name = (name, _), body, ...} =>
|
||||
Latex.string (Latex.output_symbols [Symbol.encode (Symbol.Control name)]) ::
|
||||
output_symbols body
|
||||
| Antiquote.Antiq {body, ...} =>
|
||||
Latex.enclose_body "%\n\\isaantiq\n" "{}%\n\\endisaantiq\n" (output_symbols body));
|
||||
|
||||
fun output_comment_symbols ctxt {antiq} (comment, syms) =
|
||||
(case (comment, antiq) of
|
||||
(NONE, false) => output_symbols syms
|
||||
| (NONE, true) =>
|
||||
Antiquote.parse_comments (#1 (Symbol_Pos.range syms)) syms
|
||||
|> maps output_symbols_antiq
|
||||
| (SOME comment, _) => output_comment ctxt (comment, syms));
|
||||
|
||||
fun output_body ctxt antiq bg en syms =
|
||||
Comment.read_body syms
|
||||
|> maps (output_comment_symbols ctxt {antiq = antiq})
|
||||
|> Latex.enclose_body bg en;
|
||||
|
||||
in
|
||||
|
||||
fun output_token ctxt tok =
|
||||
let
|
||||
fun output antiq bg en =
|
||||
output_body ctxt antiq bg en (Input.source_explode (Token.input_of tok));
|
||||
in
|
||||
(case Token.kind_of tok of
|
||||
Token.Comment NONE => []
|
||||
| Token.Command => output false "\\isacommand{" "}"
|
||||
| Token.Keyword =>
|
||||
if Symbol.is_ascii_identifier (Token.content_of tok)
|
||||
then output false "\\isakeyword{" "}"
|
||||
else output false "" ""
|
||||
| Token.String => output false "{\\isachardoublequoteopen}" "{\\isachardoublequoteclose}"
|
||||
| Token.Alt_String => output false "{\\isacharbackquoteopen}" "{\\isacharbackquoteclose}"
|
||||
| Token.Verbatim => output true "{\\isacharverbatimopen}" "{\\isacharverbatimclose}"
|
||||
| Token.Cartouche => output false "{\\isacartoucheopen}" "{\\isacartoucheclose}"
|
||||
| _ => output false "" "")
|
||||
end handle ERROR msg => error (msg ^ Position.here (Token.pos_of tok));
|
||||
|
||||
fun output_source ctxt s =
|
||||
output_body ctxt false "" "" (Symbol_Pos.explode (s, Position.none));
|
||||
|
||||
fun check_comments ctxt =
|
||||
Comment.read_body #> List.app (fn (comment, syms) =>
|
||||
let
|
||||
val pos = #1 (Symbol_Pos.range syms);
|
||||
val _ =
|
||||
comment |> Option.app (fn kind =>
|
||||
Context_Position.reports ctxt (map (pair pos) (Markup.cartouche :: Comment.markups kind)));
|
||||
val _ = output_comment_symbols ctxt {antiq = false} (comment, syms);
|
||||
in if comment = SOME Comment.Comment then check_comments ctxt syms else () end);
|
||||
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** present theory source **)
|
||||
|
||||
(*NB: arranging white space around command spans is a black art*)
|
||||
|
||||
val is_white = Token.is_space orf Token.is_informal_comment;
|
||||
val is_black = not o is_white;
|
||||
|
||||
val is_white_comment = Token.is_informal_comment;
|
||||
val is_black_comment = Token.is_formal_comment;
|
||||
|
||||
|
||||
(* presentation tokens *)
|
||||
|
||||
datatype token =
|
||||
Ignore_Token
|
||||
| Basic_Token of Token.T
|
||||
| Markup_Token of string * Input.source
|
||||
| Markup_Env_Token of string * Input.source
|
||||
| Raw_Token of Input.source;
|
||||
|
||||
fun basic_token pred (Basic_Token tok) = pred tok
|
||||
| basic_token _ _ = false;
|
||||
|
||||
val white_token = basic_token is_white;
|
||||
val white_comment_token = basic_token is_white_comment;
|
||||
val blank_token = basic_token Token.is_blank;
|
||||
val newline_token = basic_token Token.is_newline;
|
||||
|
||||
fun present_token ctxt tok =
|
||||
(case tok of
|
||||
Ignore_Token => []
|
||||
| Basic_Token tok => output_token ctxt tok
|
||||
| Markup_Token (cmd, source) =>
|
||||
Latex.enclose_body ("%\n\\isamarkup" ^ cmd ^ "{") "%\n}\n"
|
||||
(output_document ctxt {markdown = false} source)
|
||||
| Markup_Env_Token (cmd, source) =>
|
||||
[Latex.environment_block ("isamarkup" ^ cmd) (output_document ctxt {markdown = true} source)]
|
||||
| Raw_Token source =>
|
||||
Latex.string "%\n" :: output_document ctxt {markdown = true} source @ [Latex.string "\n"]);
|
||||
|
||||
|
||||
(* command spans *)
|
||||
|
||||
type command = string * Position.T * string list; (*name, position, tags*)
|
||||
type source = (token * (string * int)) list; (*token, markup flag, meta-comment depth*)
|
||||
|
||||
datatype span = Span of command * (source * source * source * source) * bool;
|
||||
|
||||
fun make_span cmd src =
|
||||
let
|
||||
fun chop_newline (tok :: toks) =
|
||||
if newline_token (fst tok) then ([tok], toks, true)
|
||||
else ([], tok :: toks, false)
|
||||
| chop_newline [] = ([], [], false);
|
||||
val (((src_prefix, src_main), src_suffix1), (src_suffix2, src_appendix, newline)) =
|
||||
src
|
||||
|> chop_prefix (white_token o fst)
|
||||
||>> chop_suffix (white_token o fst)
|
||||
||>> chop_prefix (white_comment_token o fst)
|
||||
||> chop_newline;
|
||||
in Span (cmd, (src_prefix, src_main, src_suffix1 @ src_suffix2, src_appendix), newline) end;
|
||||
|
||||
|
||||
(* present spans *)
|
||||
|
||||
local
|
||||
|
||||
fun err_bad_nesting pos =
|
||||
error ("Bad nesting of commands in presentation" ^ pos);
|
||||
|
||||
fun edge which f (x: string option, y) =
|
||||
if x = y then I
|
||||
else (case which (x, y) of NONE => I | SOME txt => cons (Latex.string (f txt)));
|
||||
|
||||
val begin_tag = edge #2 Latex.begin_tag;
|
||||
val end_tag = edge #1 Latex.end_tag;
|
||||
fun open_delim delim e = edge #2 Latex.begin_delim e #> delim #> edge #2 Latex.end_delim e;
|
||||
fun close_delim delim e = edge #1 Latex.begin_delim e #> delim #> edge #1 Latex.end_delim e;
|
||||
|
||||
fun read_tag s =
|
||||
(case space_explode "%" s of
|
||||
["", b] => (SOME b, NONE)
|
||||
| [a, b] => (NONE, SOME (a, b))
|
||||
| _ => error ("Bad document_tags specification: " ^ quote s));
|
||||
|
||||
in
|
||||
|
||||
fun make_command_tag options keywords =
|
||||
let
|
||||
val document_tags =
|
||||
map read_tag (space_explode "," (Options.string options \<^system_option>\<open>document_tags\<close>));
|
||||
val document_tags_default = map_filter #1 document_tags;
|
||||
val document_tags_command = map_filter #2 document_tags;
|
||||
in
|
||||
fn {cmd_name, cmd_tags, tag, active_tag} => fn state => fn state' =>
|
||||
let
|
||||
val tag' = try hd (fold (update (op =)) cmd_tags (the_list tag));
|
||||
|
||||
val keyword_tags =
|
||||
if cmd_name = "end" andalso Toplevel.is_end_theory state' then ["theory"]
|
||||
else Keyword.command_tags keywords cmd_name;
|
||||
val command_tags =
|
||||
the_list (AList.lookup (op =) document_tags_command cmd_name) @
|
||||
keyword_tags @ document_tags_default;
|
||||
|
||||
val active_tag' =
|
||||
if is_some tag' then tag'
|
||||
else
|
||||
(case command_tags of
|
||||
default_tag :: _ => SOME default_tag
|
||||
| [] =>
|
||||
if Keyword.is_vacuous keywords cmd_name andalso Toplevel.is_proof state
|
||||
then active_tag
|
||||
else NONE);
|
||||
in {tag' = tag', active_tag' = active_tag'} end
|
||||
end;
|
||||
|
||||
fun present_span thy command_tag span state state'
|
||||
(tag_stack, active_tag, newline, latex, present_cont) =
|
||||
let
|
||||
val ctxt' =
|
||||
Toplevel.presentation_context state'
|
||||
handle Toplevel.UNDEF => Proof_Context.get_global thy Context.PureN;
|
||||
val present = fold (fn (tok, (flag, 0)) =>
|
||||
fold cons (present_token ctxt' tok)
|
||||
#> cons (Latex.string flag)
|
||||
| _ => I);
|
||||
|
||||
val Span ((cmd_name, cmd_pos, cmd_tags), srcs, span_newline) = span;
|
||||
|
||||
val (tag, tags) = tag_stack;
|
||||
val {tag', active_tag'} =
|
||||
command_tag {cmd_name = cmd_name, cmd_tags = cmd_tags, tag = tag, active_tag = active_tag}
|
||||
state state';
|
||||
val edge = (active_tag, active_tag');
|
||||
|
||||
val nesting = Toplevel.level state' - Toplevel.level state;
|
||||
|
||||
val newline' =
|
||||
if is_none active_tag' then span_newline else newline;
|
||||
|
||||
val tag_stack' =
|
||||
if nesting = 0 andalso not (Toplevel.is_proof state) then tag_stack
|
||||
else if nesting >= 0 then (tag', replicate nesting tag @ tags)
|
||||
else
|
||||
(case drop (~ nesting - 1) tags of
|
||||
tg :: tgs => (tg, tgs)
|
||||
| [] => err_bad_nesting (Position.here cmd_pos));
|
||||
|
||||
val latex' =
|
||||
latex
|
||||
|> end_tag edge
|
||||
|> close_delim (fst present_cont) edge
|
||||
|> snd present_cont
|
||||
|> open_delim (present (#1 srcs)) edge
|
||||
|> begin_tag edge
|
||||
|> present (#2 srcs);
|
||||
val present_cont' =
|
||||
if newline then (present (#3 srcs), present (#4 srcs))
|
||||
else (I, present (#3 srcs) #> present (#4 srcs));
|
||||
in (tag_stack', active_tag', newline', latex', present_cont') end;
|
||||
|
||||
fun present_trailer ((_, tags), active_tag, _, latex, present_cont) =
|
||||
if not (null tags) then err_bad_nesting " at end of theory"
|
||||
else
|
||||
latex
|
||||
|> end_tag (active_tag, NONE)
|
||||
|> close_delim (fst present_cont) (active_tag, NONE)
|
||||
|> snd present_cont;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
(* present_thy *)
|
||||
|
||||
local
|
||||
|
||||
val markup_true = "\\isamarkuptrue%\n";
|
||||
val markup_false = "\\isamarkupfalse%\n";
|
||||
|
||||
val space_proper =
|
||||
Scan.one Token.is_blank -- Scan.many is_white_comment -- Scan.one is_black;
|
||||
|
||||
val is_improper = not o (is_black orf Token.is_begin_ignore orf Token.is_end_ignore);
|
||||
val improper = Scan.many is_improper;
|
||||
val improper_end = Scan.repeat (Scan.unless space_proper (Scan.one is_improper));
|
||||
val blank_end = Scan.repeat (Scan.unless space_proper (Scan.one Token.is_blank));
|
||||
|
||||
val opt_newline = Scan.option (Scan.one Token.is_newline);
|
||||
|
||||
val ignore =
|
||||
Scan.depend (fn d => opt_newline |-- Scan.one Token.is_begin_ignore
|
||||
>> pair (d + 1)) ||
|
||||
Scan.depend (fn d => Scan.one Token.is_end_ignore --|
|
||||
(if d = 0 then Scan.fail_with (K (fn () => "Bad nesting of meta-comments")) else opt_newline)
|
||||
>> pair (d - 1));
|
||||
|
||||
val tag = (improper -- Parse.$$$ "%" -- improper) |-- Parse.!!! (Parse.tag_name --| blank_end);
|
||||
|
||||
val locale =
|
||||
Scan.option ((Parse.$$$ "(" -- improper -- Parse.$$$ "in") |--
|
||||
Parse.!!! (improper |-- Parse.name --| (improper -- Parse.$$$ ")")));
|
||||
|
||||
in
|
||||
|
||||
type segment = {span: Command_Span.span, command: Toplevel.transition, state: Toplevel.state};
|
||||
|
||||
fun present_thy options thy (segments: segment list) =
|
||||
let
|
||||
val keywords = Thy_Header.get_keywords thy;
|
||||
|
||||
|
||||
(* tokens *)
|
||||
|
||||
val ignored = Scan.state --| ignore
|
||||
>> (fn d => (NONE, (Ignore_Token, ("", d))));
|
||||
|
||||
fun markup pred mk flag = Scan.peek (fn d =>
|
||||
improper |--
|
||||
Parse.position (Scan.one (fn tok =>
|
||||
Token.is_command tok andalso pred keywords (Token.content_of tok))) --
|
||||
Scan.repeat tag --
|
||||
Parse.!!!! ((improper -- locale -- improper) |-- Parse.document_source --| improper_end)
|
||||
>> (fn (((tok, pos'), tags), source) =>
|
||||
let val name = Token.content_of tok
|
||||
in (SOME (name, pos', tags), (mk (name, source), (flag, d))) end));
|
||||
|
||||
val command = Scan.peek (fn d =>
|
||||
Scan.optional (Scan.one Token.is_command_modifier ::: improper) [] --
|
||||
Scan.one Token.is_command -- Scan.repeat tag
|
||||
>> (fn ((cmd_mod, cmd), tags) =>
|
||||
map (fn tok => (NONE, (Basic_Token tok, ("", d)))) cmd_mod @
|
||||
[(SOME (Token.content_of cmd, Token.pos_of cmd, tags),
|
||||
(Basic_Token cmd, (markup_false, d)))]));
|
||||
|
||||
val cmt = Scan.peek (fn d =>
|
||||
Scan.one is_black_comment >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
||||
|
||||
val other = Scan.peek (fn d =>
|
||||
Parse.not_eof >> (fn tok => (NONE, (Basic_Token tok, ("", d)))));
|
||||
|
||||
val tokens =
|
||||
(ignored ||
|
||||
markup Keyword.is_document_heading Markup_Token markup_true ||
|
||||
markup Keyword.is_document_body Markup_Env_Token markup_true ||
|
||||
markup Keyword.is_document_raw (Raw_Token o #2) "") >> single ||
|
||||
command ||
|
||||
(cmt || other) >> single;
|
||||
|
||||
|
||||
(* spans *)
|
||||
|
||||
val is_eof = fn (_, (Basic_Token x, _)) => Token.is_eof x | _ => false;
|
||||
val stopper = Scan.stopper (K (NONE, (Basic_Token Token.eof, ("", 0)))) is_eof;
|
||||
|
||||
val cmd = Scan.one (is_some o fst);
|
||||
val non_cmd = Scan.one (is_none o fst andf not o is_eof) >> #2;
|
||||
|
||||
val white_comments = Scan.many (white_comment_token o fst o snd);
|
||||
val blank = Scan.one (blank_token o fst o snd);
|
||||
val newline = Scan.one (newline_token o fst o snd);
|
||||
val before_cmd =
|
||||
Scan.option (newline -- white_comments) --
|
||||
Scan.option (newline -- white_comments) --
|
||||
Scan.option (blank -- white_comments) -- cmd;
|
||||
|
||||
val span =
|
||||
Scan.repeat non_cmd -- cmd --
|
||||
Scan.repeat (Scan.unless before_cmd non_cmd) --
|
||||
Scan.option (newline >> (single o snd))
|
||||
>> (fn (((toks1, (cmd, tok2)), toks3), tok4) =>
|
||||
make_span (the cmd) (toks1 @ (tok2 :: (toks3 @ the_default [] tok4))));
|
||||
|
||||
val spans = segments
|
||||
|> maps (Command_Span.content o #span)
|
||||
|> drop_suffix Token.is_space
|
||||
|> Source.of_list
|
||||
|> Source.source' 0 Token.stopper (Scan.error (Scan.bulk tokens >> flat))
|
||||
|> Source.source stopper (Scan.error (Scan.bulk span))
|
||||
|> Source.exhaust;
|
||||
|
||||
val command_results =
|
||||
segments |> map_filter (fn {command, state, ...} =>
|
||||
if Toplevel.is_ignored command then NONE else SOME (command, state));
|
||||
|
||||
|
||||
(* present commands *)
|
||||
|
||||
val command_tag = make_command_tag options keywords;
|
||||
|
||||
fun present_command tr span st st' =
|
||||
Toplevel.setmp_thread_position tr (present_span thy command_tag span st st');
|
||||
|
||||
fun present _ [] = I
|
||||
| present st ((span, (tr, st')) :: rest) = present_command tr span st st' #> present st' rest;
|
||||
in
|
||||
if length command_results = length spans then
|
||||
((NONE, []), NONE, true, [], (I, I))
|
||||
|> present Toplevel.toplevel (spans ~~ command_results)
|
||||
|> present_trailer
|
||||
|> rev
|
||||
else error "Messed-up outer syntax for presentation"
|
||||
end;
|
||||
|
||||
end;
|
||||
|
||||
|
||||
|
||||
(** standard output operations **)
|
||||
|
||||
(* pretty printing *)
|
||||
|
||||
fun pretty_term ctxt t =
|
||||
Syntax.pretty_term (Variable.auto_fixes t ctxt) t;
|
||||
|
||||
fun pretty_thm ctxt = pretty_term ctxt o Thm.full_prop_of;
|
||||
|
||||
|
||||
(* default output *)
|
||||
|
||||
val lines = separate (Latex.string "\\isanewline%\n");
|
||||
val items = separate (Latex.string "\\isasep\\isanewline%\n");
|
||||
|
||||
fun isabelle ctxt body =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Latex.environment_block "isabelle" body
|
||||
else Latex.block (Latex.enclose_body "\\isa{" "}" body);
|
||||
|
||||
fun isabelle_typewriter ctxt body =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Latex.environment_block "isabellett" body
|
||||
else Latex.block (Latex.enclose_body "\\isatt{" "}" body);
|
||||
|
||||
fun typewriter ctxt s =
|
||||
isabelle_typewriter ctxt [Latex.string (Latex.output_ascii s)];
|
||||
|
||||
fun verbatim ctxt =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_display
|
||||
then Document_Antiquotation.indent_lines ctxt #> typewriter ctxt
|
||||
else split_lines #> map (typewriter ctxt) #> lines #> Latex.block;
|
||||
|
||||
fun source ctxt =
|
||||
Token.args_of_src
|
||||
#> map (Token.unparse #> Document_Antiquotation.prepare_lines ctxt)
|
||||
#> space_implode " "
|
||||
#> output_source ctxt
|
||||
#> isabelle ctxt;
|
||||
|
||||
fun pretty ctxt =
|
||||
Document_Antiquotation.output ctxt #> Latex.string #> single #> isabelle ctxt;
|
||||
|
||||
fun pretty_source ctxt src prt =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_source
|
||||
then source ctxt src else pretty ctxt prt;
|
||||
|
||||
fun pretty_items ctxt =
|
||||
map (Document_Antiquotation.output ctxt #> Latex.string) #> items #> isabelle ctxt;
|
||||
|
||||
fun pretty_items_source ctxt src prts =
|
||||
if Config.get ctxt Document_Antiquotation.thy_output_source
|
||||
then source ctxt src else pretty_items ctxt prts;
|
||||
|
||||
|
||||
(* antiquotation variants *)
|
||||
|
||||
fun antiquotation_pretty name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, argument = x, ...} => pretty ctxt (f ctxt x));
|
||||
|
||||
fun antiquotation_pretty_source name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, source = src, argument = x} => pretty_source ctxt src (f ctxt x));
|
||||
|
||||
fun antiquotation_raw name scan f =
|
||||
Document_Antiquotation.setup name scan
|
||||
(fn {context = ctxt, argument = x, ...} => f ctxt x);
|
||||
|
||||
fun antiquotation_verbatim name scan f =
|
||||
antiquotation_raw name scan (fn ctxt => verbatim ctxt o f ctxt);
|
||||
|
||||
end;
|
Loading…
Reference in New Issue