summary |
shortlog |
changelog |
graph |
tags |
bookmarks |
branches |
files |
changeset |
raw | gz |
help

author | wenzelm |

Thu, 21 Jul 2016 10:52:27 +0200 | |

changeset 63533 | 42b6186fc0e4 |

parent 63525 | f01d1e393f3f (current diff) |

parent 63532 | b01154b74314 (diff) |

child 63536 | 8cecf0100996 |

child 63539 | 70d4d9e5707b |

merged

--- a/NEWS Wed Jul 20 14:52:09 2016 +0200 +++ b/NEWS Thu Jul 21 10:52:27 2016 +0200 @@ -52,6 +52,11 @@ * Proof method "blast" is more robust wrt. corner cases of Pure statements without object-logic judgment. +* Pure provides basic versions of proof methods "simp" and "simp_all" +that only know about meta-equality (==). Potential INCOMPATIBILITY in +theory imports that merge Pure with e.g. Main of Isabelle/HOL: the order +is relevant to avoid confusion of Pure.simp vs. HOL.simp. + *** Prover IDE -- Isabelle/Scala/jEdit *** @@ -103,6 +108,9 @@ * Command 'proof' provides information about proof outline with cases, e.g. for proof methods "cases", "induct", "goal_cases". +* Completion templates for commands involving "begin ... end" blocks, +e.g. 'context', 'notepad'. + *** Isar *** @@ -145,8 +153,8 @@ "method_facts". This is particularly useful for Eisbach method definitions. -* Eisbach provides method "use" to modify the main facts of a given -method expression, e.g. +* Proof method "use" allows to modify the main facts of a given method +expression, e.g. (use facts in simp) (use facts in \<open>simp add: ...\<close>)

--- a/src/Doc/Isar_Ref/Document_Preparation.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Document_Preparation.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Document_Preparation -imports Base Main + imports Main Base begin chapter \<open>Document preparation \label{ch:document-prep}\<close>

--- a/src/Doc/Isar_Ref/Framework.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Framework.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Framework -imports Base Main + imports Main Base begin chapter \<open>The Isabelle/Isar Framework \label{ch:isar-framework}\<close>

--- a/src/Doc/Isar_Ref/Generic.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Generic.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,23 +1,22 @@ (*:maxLineLen=78:*) theory Generic -imports Base Main +imports Main Base begin chapter \<open>Generic tools and packages \label{ch:gen-tools}\<close> section \<open>Configuration options \label{sec:config}\<close> -text \<open>Isabelle/Pure maintains a record of named configuration - options within the theory or proof context, with values of type - @{ML_type bool}, @{ML_type int}, @{ML_type real}, or @{ML_type - string}. Tools may declare options in ML, and then refer to these - values (relative to the context). Thus global reference variables - are easily avoided. The user may change the value of a - configuration option by means of an associated attribute of the same - name. This form of context declaration works particularly well with - commands such as @{command "declare"} or @{command "using"} like - this: +text \<open> + Isabelle/Pure maintains a record of named configuration options within the + theory or proof context, with values of type @{ML_type bool}, @{ML_type + int}, @{ML_type real}, or @{ML_type string}. Tools may declare options in + ML, and then refer to these values (relative to the context). Thus global + reference variables are easily avoided. The user may change the value of a + configuration option by means of an associated attribute of the same name. + This form of context declaration works particularly well with commands such + as @{command "declare"} or @{command "using"} like this: \<close> (*<*)experiment begin(*>*) @@ -40,14 +39,14 @@ @{syntax name} ('=' ('true' | 'false' | @{syntax int} | @{syntax float} | @{syntax name}))? \<close>} - \<^descr> @{command "print_options"} prints the available configuration - options, with names, types, and current values; the ``\<open>!\<close>'' option - indicates extra verbosity. + \<^descr> @{command "print_options"} prints the available configuration options, + with names, types, and current values; the ``\<open>!\<close>'' option indicates extra + verbosity. - \<^descr> \<open>name = value\<close> as an attribute expression modifies the - named option, with the syntax of the value depending on the option's - type. For @{ML_type bool} the default value is \<open>true\<close>. Any - attempt to change a global option in a local context is ignored. + \<^descr> \<open>name = value\<close> as an attribute expression modifies the named option, with + the syntax of the value depending on the option's type. For @{ML_type bool} + the default value is \<open>true\<close>. Any attempt to change a global option in a + local context is ignored. \<close> @@ -81,46 +80,41 @@ @@{method sleep} @{syntax real} \<close>} - \<^descr> @{method unfold}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> and @{method fold}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> expand (or fold back) the given definitions throughout - all goals; any chained facts provided are inserted into the goal and - subject to rewriting as well. - - \<^descr> @{method insert}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> inserts theorems as facts - into all goals of the proof state. Note that current facts - indicated for forward chaining are ignored. + \<^descr> @{method unfold}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> and @{method fold}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> expand (or + fold back) the given definitions throughout all goals; any chained facts + provided are inserted into the goal and subject to rewriting as well. - \<^descr> @{method erule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close>, @{method - drule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close>, and @{method frule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> are similar to the basic @{method rule} - method (see \secref{sec:pure-meth-att}), but apply rules by - elim-resolution, destruct-resolution, and forward-resolution, - respectively @{cite "isabelle-implementation"}. The optional natural - number argument (default 0) specifies additional assumption steps to - be performed here. + \<^descr> @{method insert}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> inserts theorems as facts into all goals of + the proof state. Note that current facts indicated for forward chaining are + ignored. + + \<^descr> @{method erule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close>, @{method drule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close>, and @{method + frule}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> are similar to the basic @{method rule} method (see + \secref{sec:pure-meth-att}), but apply rules by elim-resolution, + destruct-resolution, and forward-resolution, respectively @{cite + "isabelle-implementation"}. The optional natural number argument (default 0) + specifies additional assumption steps to be performed here. Note that these methods are improper ones, mainly serving for - experimentation and tactic script emulation. Different modes of - basic rule application are usually expressed in Isar at the proof - language level, rather than via implicit proof state manipulations. - For example, a proper single-step elimination would be done using - the plain @{method rule} method, with forward chaining of current - facts. + experimentation and tactic script emulation. Different modes of basic rule + application are usually expressed in Isar at the proof language level, + rather than via implicit proof state manipulations. For example, a proper + single-step elimination would be done using the plain @{method rule} method, + with forward chaining of current facts. - \<^descr> @{method intro} and @{method elim} repeatedly refine some goal - by intro- or elim-resolution, after having inserted any chained - facts. Exactly the rules given as arguments are taken into account; - this allows fine-tuned decomposition of a proof problem, in contrast - to common automated tools. + \<^descr> @{method intro} and @{method elim} repeatedly refine some goal by intro- + or elim-resolution, after having inserted any chained facts. Exactly the + rules given as arguments are taken into account; this allows fine-tuned + decomposition of a proof problem, in contrast to common automated tools. - \<^descr> @{method fail} yields an empty result sequence; it is the - identity of the ``\<open>|\<close>'' method combinator (cf.\ - \secref{sec:proof-meth}). + \<^descr> @{method fail} yields an empty result sequence; it is the identity of the + ``\<open>|\<close>'' method combinator (cf.\ \secref{sec:proof-meth}). - \<^descr> @{method succeed} yields a single (unchanged) result; it is - the identity of the ``\<open>,\<close>'' method combinator (cf.\ - \secref{sec:proof-meth}). + \<^descr> @{method succeed} yields a single (unchanged) result; it is the identity + of the ``\<open>,\<close>'' method combinator (cf.\ \secref{sec:proof-meth}). - \<^descr> @{method sleep}~\<open>s\<close> succeeds after a real-time delay of \<open>s\<close> seconds. This is occasionally useful for demonstration and testing - purposes. + \<^descr> @{method sleep}~\<open>s\<close> succeeds after a real-time delay of \<open>s\<close> seconds. This + is occasionally useful for demonstration and testing purposes. \begin{matharray}{rcl} @@ -147,38 +141,35 @@ @@{attribute rotated} @{syntax int}? \<close>} - \<^descr> @{attribute tagged}~\<open>name value\<close> and @{attribute - untagged}~\<open>name\<close> add and remove \<^emph>\<open>tags\<close> of some theorem. - Tags may be any list of string pairs that serve as formal comment. - The first string is considered the tag name, the second its value. - Note that @{attribute untagged} removes any tags of the same name. + \<^descr> @{attribute tagged}~\<open>name value\<close> and @{attribute untagged}~\<open>name\<close> add and + remove \<^emph>\<open>tags\<close> of some theorem. Tags may be any list of string pairs that + serve as formal comment. The first string is considered the tag name, the + second its value. Note that @{attribute untagged} removes any tags of the + same name. - \<^descr> @{attribute THEN}~\<open>a\<close> composes rules by resolution; it - resolves with the first premise of \<open>a\<close> (an alternative - position may be also specified). See also @{ML_op "RS"} in - @{cite "isabelle-implementation"}. + \<^descr> @{attribute THEN}~\<open>a\<close> composes rules by resolution; it resolves with the + first premise of \<open>a\<close> (an alternative position may be also specified). See + also @{ML_op "RS"} in @{cite "isabelle-implementation"}. - \<^descr> @{attribute unfolded}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> and @{attribute - folded}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> expand and fold back again the given - definitions throughout a rule. + \<^descr> @{attribute unfolded}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> and @{attribute folded}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> + expand and fold back again the given definitions throughout a rule. - \<^descr> @{attribute abs_def} turns an equation of the form @{prop "f x - y \<equiv> t"} into @{prop "f \<equiv> \<lambda>x y. t"}, which ensures that @{method - simp} or @{method unfold} steps always expand it. This also works - for object-logic equality. + \<^descr> @{attribute abs_def} turns an equation of the form @{prop "f x y \<equiv> t"} + into @{prop "f \<equiv> \<lambda>x y. t"}, which ensures that @{method simp} or @{method + unfold} steps always expand it. This also works for object-logic equality. - \<^descr> @{attribute rotated}~\<open>n\<close> rotate the premises of a - theorem by \<open>n\<close> (default 1). + \<^descr> @{attribute rotated}~\<open>n\<close> rotate the premises of a theorem by \<open>n\<close> (default + 1). - \<^descr> @{attribute (Pure) elim_format} turns a destruction rule into - elimination rule format, by resolving with the rule @{prop "PROP A \<Longrightarrow> - (PROP A \<Longrightarrow> PROP B) \<Longrightarrow> PROP B"}. + \<^descr> @{attribute (Pure) elim_format} turns a destruction rule into elimination + rule format, by resolving with the rule @{prop "PROP A \<Longrightarrow> (PROP A \<Longrightarrow> PROP B) \<Longrightarrow> + PROP B"}. - Note that the Classical Reasoner (\secref{sec:classical}) provides - its own version of this operation. + Note that the Classical Reasoner (\secref{sec:classical}) provides its own + version of this operation. - \<^descr> @{attribute no_vars} replaces schematic variables by free - ones; this is mainly for tuning output of pretty printed theorems. + \<^descr> @{attribute no_vars} replaces schematic variables by free ones; this is + mainly for tuning output of pretty printed theorems. \<close> @@ -197,51 +188,45 @@ @@{method split} @{syntax thms} \<close>} - These methods provide low-level facilities for equational reasoning - that are intended for specialized applications only. Normally, - single step calculations would be performed in a structured text - (see also \secref{sec:calculation}), while the Simplifier methods - provide the canonical way for automated normalization (see - \secref{sec:simplifier}). + These methods provide low-level facilities for equational reasoning that are + intended for specialized applications only. Normally, single step + calculations would be performed in a structured text (see also + \secref{sec:calculation}), while the Simplifier methods provide the + canonical way for automated normalization (see \secref{sec:simplifier}). + + \<^descr> @{method subst}~\<open>eq\<close> performs a single substitution step using rule \<open>eq\<close>, + which may be either a meta or object equality. - \<^descr> @{method subst}~\<open>eq\<close> performs a single substitution step - using rule \<open>eq\<close>, which may be either a meta or object - equality. - - \<^descr> @{method subst}~\<open>(asm) eq\<close> substitutes in an - assumption. + \<^descr> @{method subst}~\<open>(asm) eq\<close> substitutes in an assumption. - \<^descr> @{method subst}~\<open>(i \<dots> j) eq\<close> performs several - substitutions in the conclusion. The numbers \<open>i\<close> to \<open>j\<close> - indicate the positions to substitute at. Positions are ordered from - the top of the term tree moving down from left to right. For - example, in \<open>(a + b) + (c + d)\<close> there are three positions - where commutativity of \<open>+\<close> is applicable: 1 refers to \<open>a + b\<close>, 2 to the whole term, and 3 to \<open>c + d\<close>. + \<^descr> @{method subst}~\<open>(i \<dots> j) eq\<close> performs several substitutions in the + conclusion. The numbers \<open>i\<close> to \<open>j\<close> indicate the positions to substitute at. + Positions are ordered from the top of the term tree moving down from left to + right. For example, in \<open>(a + b) + (c + d)\<close> there are three positions where + commutativity of \<open>+\<close> is applicable: 1 refers to \<open>a + b\<close>, 2 to the whole + term, and 3 to \<open>c + d\<close>. - If the positions in the list \<open>(i \<dots> j)\<close> are non-overlapping - (e.g.\ \<open>(2 3)\<close> in \<open>(a + b) + (c + d)\<close>) you may - assume all substitutions are performed simultaneously. Otherwise - the behaviour of \<open>subst\<close> is not specified. + If the positions in the list \<open>(i \<dots> j)\<close> are non-overlapping (e.g.\ \<open>(2 3)\<close> in + \<open>(a + b) + (c + d)\<close>) you may assume all substitutions are performed + simultaneously. Otherwise the behaviour of \<open>subst\<close> is not specified. - \<^descr> @{method subst}~\<open>(asm) (i \<dots> j) eq\<close> performs the - substitutions in the assumptions. The positions refer to the - assumptions in order from left to right. For example, given in a - goal of the form \<open>P (a + b) \<Longrightarrow> P (c + d) \<Longrightarrow> \<dots>\<close>, position 1 of - commutativity of \<open>+\<close> is the subterm \<open>a + b\<close> and - position 2 is the subterm \<open>c + d\<close>. + \<^descr> @{method subst}~\<open>(asm) (i \<dots> j) eq\<close> performs the substitutions in the + assumptions. The positions refer to the assumptions in order from left to + right. For example, given in a goal of the form \<open>P (a + b) \<Longrightarrow> P (c + d) \<Longrightarrow> \<dots>\<close>, + position 1 of commutativity of \<open>+\<close> is the subterm \<open>a + b\<close> and position 2 is + the subterm \<open>c + d\<close>. - \<^descr> @{method hypsubst} performs substitution using some - assumption; this only works for equations of the form \<open>x = - t\<close> where \<open>x\<close> is a free or bound variable. + \<^descr> @{method hypsubst} performs substitution using some assumption; this only + works for equations of the form \<open>x = t\<close> where \<open>x\<close> is a free or bound + variable. - \<^descr> @{method split}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> performs single-step case - splitting using the given rules. Splitting is performed in the - conclusion or some assumption of the subgoal, depending of the - structure of the rule. + \<^descr> @{method split}~\<open>a\<^sub>1 \<dots> a\<^sub>n\<close> performs single-step case splitting using the + given rules. Splitting is performed in the conclusion or some assumption of + the subgoal, depending of the structure of the rule. - Note that the @{method simp} method already involves repeated - application of split rules as declared in the current context, using - @{attribute split}, for example. + Note that the @{method simp} method already involves repeated application of + split rules as declared in the current context, using @{attribute split}, + for example. \<close> @@ -273,6 +258,8 @@ \begin{tabular}{rcll} @{method_def simp} & : & \<open>method\<close> \\ @{method_def simp_all} & : & \<open>method\<close> \\ + \<open>Pure.\<close>@{method_def (Pure) simp} & : & \<open>method\<close> \\ + \<open>Pure.\<close>@{method_def (Pure) simp_all} & : & \<open>method\<close> \\ @{attribute_def simp_depth_limit} & : & \<open>attribute\<close> & default \<open>100\<close> \\ \end{tabular} \<^medskip> @@ -369,6 +356,12 @@ \end{tabular} \end{center} + + \<^medskip> + In Isabelle/Pure, proof methods @{method (Pure) simp} and @{method (Pure) + simp_all} only know about meta-equality \<open>\<equiv>\<close>. Any new object-logic needs to + re-define these methods via @{ML Simplifier.method_setup} in ML: + Isabelle/FOL or Isabelle/HOL may serve as blue-prints. \<close>

--- a/src/Doc/Isar_Ref/HOL_Specific.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/HOL_Specific.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,68 +1,65 @@ (*:maxLineLen=78:*) theory HOL_Specific -imports - Base - "~~/src/HOL/Library/Old_Datatype" - "~~/src/HOL/Library/Old_Recdef" - "~~/src/Tools/Adhoc_Overloading" - "~~/src/HOL/Library/Dlist" - "~~/src/HOL/Library/FSet" + imports + Main + "~~/src/HOL/Library/Old_Datatype" + "~~/src/HOL/Library/Old_Recdef" + "~~/src/Tools/Adhoc_Overloading" + "~~/src/HOL/Library/Dlist" + "~~/src/HOL/Library/FSet" + Base begin chapter \<open>Higher-Order Logic\<close> -text \<open>Isabelle/HOL is based on Higher-Order Logic, a polymorphic - version of Church's Simple Theory of Types. HOL can be best - understood as a simply-typed version of classical set theory. The - logic was first implemented in Gordon's HOL system - @{cite "mgordon-hol"}. It extends Church's original logic - @{cite "church40"} by explicit type variables (naive polymorphism) and - a sound axiomatization scheme for new types based on subsets of - existing types. - - Andrews's book @{cite andrews86} is a full description of the - original Church-style higher-order logic, with proofs of correctness - and completeness wrt.\ certain set-theoretic interpretations. The - particular extensions of Gordon-style HOL are explained semantically - in two chapters of the 1993 HOL book @{cite pitts93}. - - Experience with HOL over decades has demonstrated that higher-order - logic is widely applicable in many areas of mathematics and computer - science. In a sense, Higher-Order Logic is simpler than First-Order - Logic, because there are fewer restrictions and special cases. Note - that HOL is \<^emph>\<open>weaker\<close> than FOL with axioms for ZF set theory, - which is traditionally considered the standard foundation of regular - mathematics, but for most applications this does not matter. If you - prefer ML to Lisp, you will probably prefer HOL to ZF. - - \<^medskip> - The syntax of HOL follows \<open>\<lambda>\<close>-calculus and - functional programming. Function application is curried. To apply - the function \<open>f\<close> of type \<open>\<tau>\<^sub>1 \<Rightarrow> \<tau>\<^sub>2 \<Rightarrow> \<tau>\<^sub>3\<close> to the - arguments \<open>a\<close> and \<open>b\<close> in HOL, you simply write \<open>f - a b\<close> (as in ML or Haskell). There is no ``apply'' operator; the - existing application of the Pure \<open>\<lambda>\<close>-calculus is re-used. - Note that in HOL \<open>f (a, b)\<close> means ``\<open>f\<close> applied to - the pair \<open>(a, b)\<close> (which is notation for \<open>Pair a - b\<close>). The latter typically introduces extra formal efforts that can - be avoided by currying functions by default. Explicit tuples are as - infrequent in HOL formalizations as in good ML or Haskell programs. - - \<^medskip> - Isabelle/HOL has a distinct feel, compared to other - object-logics like Isabelle/ZF. It identifies object-level types - with meta-level types, taking advantage of the default - type-inference mechanism of Isabelle/Pure. HOL fully identifies - object-level functions with meta-level functions, with native - abstraction and application. - - These identifications allow Isabelle to support HOL particularly - nicely, but they also mean that HOL requires some sophistication - from the user. In particular, an understanding of Hindley-Milner - type-inference with type-classes, which are both used extensively in - the standard libraries and applications.\<close> +text \<open> + Isabelle/HOL is based on Higher-Order Logic, a polymorphic version of + Church's Simple Theory of Types. HOL can be best understood as a + simply-typed version of classical set theory. The logic was first + implemented in Gordon's HOL system @{cite "mgordon-hol"}. It extends + Church's original logic @{cite "church40"} by explicit type variables (naive + polymorphism) and a sound axiomatization scheme for new types based on + subsets of existing types. + + Andrews's book @{cite andrews86} is a full description of the original + Church-style higher-order logic, with proofs of correctness and completeness + wrt.\ certain set-theoretic interpretations. The particular extensions of + Gordon-style HOL are explained semantically in two chapters of the 1993 HOL + book @{cite pitts93}. + + Experience with HOL over decades has demonstrated that higher-order logic is + widely applicable in many areas of mathematics and computer science. In a + sense, Higher-Order Logic is simpler than First-Order Logic, because there + are fewer restrictions and special cases. Note that HOL is \<^emph>\<open>weaker\<close> than + FOL with axioms for ZF set theory, which is traditionally considered the + standard foundation of regular mathematics, but for most applications this + does not matter. If you prefer ML to Lisp, you will probably prefer HOL to + ZF. + + \<^medskip> The syntax of HOL follows \<open>\<lambda>\<close>-calculus and functional programming. + Function application is curried. To apply the function \<open>f\<close> of type \<open>\<tau>\<^sub>1 \<Rightarrow> + \<tau>\<^sub>2 \<Rightarrow> \<tau>\<^sub>3\<close> to the arguments \<open>a\<close> and \<open>b\<close> in HOL, you simply write \<open>f a b\<close> (as + in ML or Haskell). There is no ``apply'' operator; the existing application + of the Pure \<open>\<lambda>\<close>-calculus is re-used. Note that in HOL \<open>f (a, b)\<close> means ``\<open>f\<close> + applied to the pair \<open>(a, b)\<close> (which is notation for \<open>Pair a b\<close>). The latter + typically introduces extra formal efforts that can be avoided by currying + functions by default. Explicit tuples are as infrequent in HOL + formalizations as in good ML or Haskell programs. + + \<^medskip> Isabelle/HOL has a distinct feel, compared to other object-logics like + Isabelle/ZF. It identifies object-level types with meta-level types, taking + advantage of the default type-inference mechanism of Isabelle/Pure. HOL + fully identifies object-level functions with meta-level functions, with + native abstraction and application. + + These identifications allow Isabelle to support HOL particularly nicely, but + they also mean that HOL requires some sophistication from the user. In + particular, an understanding of Hindley-Milner type-inference with + type-classes, which are both used extensively in the standard libraries and + applications. +\<close> chapter \<open>Derived specification elements\<close> @@ -79,31 +76,27 @@ @{attribute_def (HOL) mono} & : & \<open>attribute\<close> \\ \end{matharray} - An \<^emph>\<open>inductive definition\<close> specifies the least predicate or set - \<open>R\<close> closed under given rules: applying a rule to elements of - \<open>R\<close> yields a result within \<open>R\<close>. For example, a - structural operational semantics is an inductive definition of an - evaluation relation. - - Dually, a \<^emph>\<open>coinductive definition\<close> specifies the greatest - predicate or set \<open>R\<close> that is consistent with given rules: - every element of \<open>R\<close> can be seen as arising by applying a rule - to elements of \<open>R\<close>. An important example is using - bisimulation relations to formalise equivalence of processes and - infinite data structures. - - Both inductive and coinductive definitions are based on the - Knaster-Tarski fixed-point theorem for complete lattices. The - collection of introduction rules given by the user determines a - functor on subsets of set-theoretic relations. The required - monotonicity of the recursion scheme is proven as a prerequisite to - the fixed-point definition and the resulting consequences. This - works by pushing inclusion through logical connectives and any other - operator that might be wrapped around recursive occurrences of the - defined relation: there must be a monotonicity theorem of the form - \<open>A \<le> B \<Longrightarrow> \<M> A \<le> \<M> B\<close>, for each premise \<open>\<M> R t\<close> in an - introduction rule. The default rule declarations of Isabelle/HOL - already take care of most common situations. + An \<^emph>\<open>inductive definition\<close> specifies the least predicate or set \<open>R\<close> closed + under given rules: applying a rule to elements of \<open>R\<close> yields a result within + \<open>R\<close>. For example, a structural operational semantics is an inductive + definition of an evaluation relation. + + Dually, a \<^emph>\<open>coinductive definition\<close> specifies the greatest predicate or set + \<open>R\<close> that is consistent with given rules: every element of \<open>R\<close> can be seen as + arising by applying a rule to elements of \<open>R\<close>. An important example is using + bisimulation relations to formalise equivalence of processes and infinite + data structures. + + Both inductive and coinductive definitions are based on the Knaster-Tarski + fixed-point theorem for complete lattices. The collection of introduction + rules given by the user determines a functor on subsets of set-theoretic + relations. The required monotonicity of the recursion scheme is proven as a + prerequisite to the fixed-point definition and the resulting consequences. + This works by pushing inclusion through logical connectives and any other + operator that might be wrapped around recursive occurrences of the defined + relation: there must be a monotonicity theorem of the form \<open>A \<le> B \<Longrightarrow> \<M> A \<le> \<M> + B\<close>, for each premise \<open>\<M> R t\<close> in an introduction rule. The default rule + declarations of Isabelle/HOL already take care of most common situations. @{rail \<open> (@@{command (HOL) inductive} | @@{command (HOL) inductive_set} | @@ -116,101 +109,98 @@ @@{attribute (HOL) mono} (() | 'add' | 'del') \<close>} - \<^descr> @{command (HOL) "inductive"} and @{command (HOL) - "coinductive"} define (co)inductive predicates from the introduction - rules. - - The propositions given as \<open>clauses\<close> in the @{keyword - "where"} part are either rules of the usual \<open>\<And>/\<Longrightarrow>\<close> format - (with arbitrary nesting), or equalities using \<open>\<equiv>\<close>. The - latter specifies extra-logical abbreviations in the sense of - @{command_ref abbreviation}. Introducing abstract syntax - simultaneously with the actual introduction rules is occasionally - useful for complex specifications. - - The optional @{keyword "for"} part contains a list of parameters of - the (co)inductive predicates that remain fixed throughout the - definition, in contrast to arguments of the relation that may vary - in each occurrence within the given \<open>clauses\<close>. + \<^descr> @{command (HOL) "inductive"} and @{command (HOL) "coinductive"} define + (co)inductive predicates from the introduction rules. + + The propositions given as \<open>clauses\<close> in the @{keyword "where"} part are + either rules of the usual \<open>\<And>/\<Longrightarrow>\<close> format (with arbitrary nesting), or + equalities using \<open>\<equiv>\<close>. The latter specifies extra-logical abbreviations in + the sense of @{command_ref abbreviation}. Introducing abstract syntax + simultaneously with the actual introduction rules is occasionally useful for + complex specifications. + + The optional @{keyword "for"} part contains a list of parameters of the + (co)inductive predicates that remain fixed throughout the definition, in + contrast to arguments of the relation that may vary in each occurrence + within the given \<open>clauses\<close>. The optional @{keyword "monos"} declaration contains additional - \<^emph>\<open>monotonicity theorems\<close>, which are required for each operator - applied to a recursive set in the introduction rules. - - \<^descr> @{command (HOL) "inductive_set"} and @{command (HOL) - "coinductive_set"} are wrappers for to the previous commands for - native HOL predicates. This allows to define (co)inductive sets, - where multiple arguments are simulated via tuples. + \<^emph>\<open>monotonicity theorems\<close>, which are required for each operator applied to a + recursive set in the introduction rules. + + \<^descr> @{command (HOL) "inductive_set"} and @{command (HOL) "coinductive_set"} + are wrappers for to the previous commands for native HOL predicates. This + allows to define (co)inductive sets, where multiple arguments are simulated + via tuples. \<^descr> @{command "print_inductives"} prints (co)inductive definitions and monotonicity rules; the ``\<open>!\<close>'' option indicates extra verbosity. - \<^descr> @{attribute (HOL) mono} declares monotonicity rules in the - context. These rule are involved in the automated monotonicity - proof of the above inductive and coinductive definitions. + \<^descr> @{attribute (HOL) mono} declares monotonicity rules in the context. These + rule are involved in the automated monotonicity proof of the above inductive + and coinductive definitions. \<close> subsection \<open>Derived rules\<close> -text \<open>A (co)inductive definition of \<open>R\<close> provides the following - main theorems: - - \<^descr> \<open>R.intros\<close> is the list of introduction rules as proven - theorems, for the recursive predicates (or sets). The rules are - also available individually, using the names given them in the - theory file; +text \<open> + A (co)inductive definition of \<open>R\<close> provides the following main theorems: + + \<^descr> \<open>R.intros\<close> is the list of introduction rules as proven theorems, for the + recursive predicates (or sets). The rules are also available individually, + using the names given them in the theory file; \<^descr> \<open>R.cases\<close> is the case analysis (or elimination) rule; - \<^descr> \<open>R.induct\<close> or \<open>R.coinduct\<close> is the (co)induction - rule; - - \<^descr> \<open>R.simps\<close> is the equation unrolling the fixpoint of the - predicate one step. - - - When several predicates \<open>R\<^sub>1, \<dots>, R\<^sub>n\<close> are defined simultaneously, - the list of introduction rules is called \<open>R\<^sub>1_\<dots>_R\<^sub>n.intros\<close>, the - case analysis rules are called \<open>R\<^sub>1.cases, \<dots>, R\<^sub>n.cases\<close>, and the - list of mutual induction rules is called \<open>R\<^sub>1_\<dots>_R\<^sub>n.inducts\<close>. + \<^descr> \<open>R.induct\<close> or \<open>R.coinduct\<close> is the (co)induction rule; + + \<^descr> \<open>R.simps\<close> is the equation unrolling the fixpoint of the predicate one + step. + + + When several predicates \<open>R\<^sub>1, \<dots>, R\<^sub>n\<close> are defined simultaneously, the list + of introduction rules is called \<open>R\<^sub>1_\<dots>_R\<^sub>n.intros\<close>, the case analysis rules + are called \<open>R\<^sub>1.cases, \<dots>, R\<^sub>n.cases\<close>, and the list of mutual induction rules + is called \<open>R\<^sub>1_\<dots>_R\<^sub>n.inducts\<close>. \<close> subsection \<open>Monotonicity theorems\<close> -text \<open>The context maintains a default set of theorems that are used - in monotonicity proofs. New rules can be declared via the - @{attribute (HOL) mono} attribute. See the main Isabelle/HOL - sources for some examples. The general format of such monotonicity - theorems is as follows: - - \<^item> Theorems of the form \<open>A \<le> B \<Longrightarrow> \<M> A \<le> \<M> B\<close>, for proving - monotonicity of inductive definitions whose introduction rules have - premises involving terms such as \<open>\<M> R t\<close>. - - \<^item> Monotonicity theorems for logical operators, which are of the - general form \<open>(\<dots> \<longrightarrow> \<dots>) \<Longrightarrow> \<dots> (\<dots> \<longrightarrow> \<dots>) \<Longrightarrow> \<dots> \<longrightarrow> \<dots>\<close>. For example, in - the case of the operator \<open>\<or>\<close>, the corresponding theorem is +text \<open> + The context maintains a default set of theorems that are used in + monotonicity proofs. New rules can be declared via the @{attribute (HOL) + mono} attribute. See the main Isabelle/HOL sources for some examples. The + general format of such monotonicity theorems is as follows: + + \<^item> Theorems of the form \<open>A \<le> B \<Longrightarrow> \<M> A \<le> \<M> B\<close>, for proving monotonicity of + inductive definitions whose introduction rules have premises involving terms + such as \<open>\<M> R t\<close>. + + \<^item> Monotonicity theorems for logical operators, which are of the general form + \<open>(\<dots> \<longrightarrow> \<dots>) \<Longrightarrow> \<dots> (\<dots> \<longrightarrow> \<dots>) \<Longrightarrow> \<dots> \<longrightarrow> \<dots>\<close>. For example, in the case of the operator \<open>\<or>\<close>, + the corresponding theorem is \[ \infer{\<open>P\<^sub>1 \<or> P\<^sub>2 \<longrightarrow> Q\<^sub>1 \<or> Q\<^sub>2\<close>}{\<open>P\<^sub>1 \<longrightarrow> Q\<^sub>1\<close> & \<open>P\<^sub>2 \<longrightarrow> Q\<^sub>2\<close>} \] - \<^item> De Morgan style equations for reasoning about the ``polarity'' - of expressions, e.g. + \<^item> De Morgan style equations for reasoning about the ``polarity'' of + expressions, e.g. \[ @{prop "\<not> \<not> P \<longleftrightarrow> P"} \qquad\qquad @{prop "\<not> (P \<and> Q) \<longleftrightarrow> \<not> P \<or> \<not> Q"} \] - \<^item> Equations for reducing complex operators to more primitive - ones whose monotonicity can easily be proved, e.g. + \<^item> Equations for reducing complex operators to more primitive ones whose + monotonicity can easily be proved, e.g. \[ @{prop "(P \<longrightarrow> Q) \<longleftrightarrow> \<not> P \<or> Q"} \qquad\qquad @{prop "Ball A P \<equiv> \<forall>x. x \<in> A \<longrightarrow> P x"} \] \<close> + subsubsection \<open>Examples\<close> text \<open>The finite powerset operator can be defined inductively like this:\<close> @@ -228,9 +218,10 @@ where acc: "(\<And>y. y \<prec> x \<Longrightarrow> acc r y) \<Longrightarrow> acc r x" (*<*)end(*>*) -text \<open>Common logical connectives can be easily characterized as - non-recursive inductive definitions with parameters, but without - arguments.\<close> +text \<open> + Common logical connectives can be easily characterized as non-recursive + inductive definitions with parameters, but without arguments. +\<close> (*<*)experiment begin(*>*) inductive AND for A B :: bool @@ -244,12 +235,13 @@ where "B a \<Longrightarrow> EXISTS B" (*<*)end(*>*) -text \<open>Here the \<open>cases\<close> or \<open>induct\<close> rules produced by the - @{command inductive} package coincide with the expected elimination rules - for Natural Deduction. Already in the original article by Gerhard Gentzen - @{cite "Gentzen:1935"} there is a hint that each connective can be - characterized by its introductions, and the elimination can be constructed - systematically.\<close> +text \<open> + Here the \<open>cases\<close> or \<open>induct\<close> rules produced by the @{command inductive} + package coincide with the expected elimination rules for Natural Deduction. + Already in the original article by Gerhard Gentzen @{cite "Gentzen:1935"} + there is a hint that each connective can be characterized by its + introductions, and the elimination can be constructed systematically. +\<close> section \<open>Recursive functions \label{sec:recursion}\<close> @@ -275,22 +267,21 @@ @@{command (HOL) fun_cases} (@{syntax thmdecl}? @{syntax prop} + @'and') \<close>} - \<^descr> @{command (HOL) "primrec"} defines primitive recursive functions - over datatypes (see also @{command_ref (HOL) datatype}). The given \<open>equations\<close> specify reduction rules that are produced by instantiating the - generic combinator for primitive recursion that is available for each - datatype. + \<^descr> @{command (HOL) "primrec"} defines primitive recursive functions over + datatypes (see also @{command_ref (HOL) datatype}). The given \<open>equations\<close> + specify reduction rules that are produced by instantiating the generic + combinator for primitive recursion that is available for each datatype. Each equation needs to be of the form: @{text [display] "f x\<^sub>1 \<dots> x\<^sub>m (C y\<^sub>1 \<dots> y\<^sub>k) z\<^sub>1 \<dots> z\<^sub>n = rhs"} - such that \<open>C\<close> is a datatype constructor, \<open>rhs\<close> contains only - the free variables on the left-hand side (or from the context), and all - recursive occurrences of \<open>f\<close> in \<open>rhs\<close> are of the form - \<open>f \<dots> y\<^sub>i \<dots>\<close> for some \<open>i\<close>. At most one reduction rule for - each constructor can be given. The order does not matter. For missing - constructors, the function is defined to return a default value, but this - equation is made difficult to access for users. + such that \<open>C\<close> is a datatype constructor, \<open>rhs\<close> contains only the free + variables on the left-hand side (or from the context), and all recursive + occurrences of \<open>f\<close> in \<open>rhs\<close> are of the form \<open>f \<dots> y\<^sub>i \<dots>\<close> for some \<open>i\<close>. At + most one reduction rule for each constructor can be given. The order does + not matter. For missing constructors, the function is defined to return a + default value, but this equation is made difficult to access for users. The reduction rules are declared as @{attribute simp} by default, which enables standard proof methods like @{method simp} and @{method auto} to @@ -304,62 +295,62 @@ command generates proof obligations for the completeness and the compatibility of patterns. - The defined function is considered partial, and the resulting - simplification rules (named \<open>f.psimps\<close>) and induction rule (named - \<open>f.pinduct\<close>) are guarded by a generated domain predicate \<open>f_dom\<close>. The @{command (HOL) "termination"} command can then be used to - establish that the function is total. + The defined function is considered partial, and the resulting simplification + rules (named \<open>f.psimps\<close>) and induction rule (named \<open>f.pinduct\<close>) are guarded + by a generated domain predicate \<open>f_dom\<close>. The @{command (HOL) "termination"} + command can then be used to establish that the function is total. \<^descr> @{command (HOL) "fun"} is a shorthand notation for ``@{command (HOL) - "function"}~\<open>(sequential)\<close>'', followed by automated proof attempts - regarding pattern matching and termination. See @{cite - "isabelle-function"} for further details. - - \<^descr> @{command (HOL) "termination"}~\<open>f\<close> commences a termination - proof for the previously defined function \<open>f\<close>. If this is omitted, - the command refers to the most recent function definition. After the proof - is closed, the recursive equations and the induction principle is - established. - - \<^descr> @{command (HOL) "fun_cases"} generates specialized elimination rules - for function equations. It expects one or more function equations and - produces rules that eliminate the given equalities, following the cases - given in the function definition. - - - Recursive definitions introduced by the @{command (HOL) "function"} - command accommodate reasoning by induction (cf.\ @{method induct}): rule - \<open>f.induct\<close> refers to a specific induction rule, with parameters - named according to the user-specified equations. Cases are numbered - starting from 1. For @{command (HOL) "primrec"}, the induction principle - coincides with structural recursion on the datatype where the recursion is - carried out. + "function"}~\<open>(sequential)\<close>'', followed by automated proof attempts regarding + pattern matching and termination. See @{cite "isabelle-function"} for + further details. + + \<^descr> @{command (HOL) "termination"}~\<open>f\<close> commences a termination proof for the + previously defined function \<open>f\<close>. If this is omitted, the command refers to + the most recent function definition. After the proof is closed, the + recursive equations and the induction principle is established. + + \<^descr> @{command (HOL) "fun_cases"} generates specialized elimination rules for + function equations. It expects one or more function equations and produces + rules that eliminate the given equalities, following the cases given in the + function definition. + + + Recursive definitions introduced by the @{command (HOL) "function"} command + accommodate reasoning by induction (cf.\ @{method induct}): rule \<open>f.induct\<close> + refers to a specific induction rule, with parameters named according to the + user-specified equations. Cases are numbered starting from 1. For @{command + (HOL) "primrec"}, the induction principle coincides with structural + recursion on the datatype where the recursion is carried out. The equations provided by these packages may be referred later as theorem - list \<open>f.simps\<close>, where \<open>f\<close> is the (collective) name of the - functions defined. Individual equations may be named explicitly as well. + list \<open>f.simps\<close>, where \<open>f\<close> is the (collective) name of the functions defined. + Individual equations may be named explicitly as well. The @{command (HOL) "function"} command accepts the following options. - \<^descr> \<open>sequential\<close> enables a preprocessor which disambiguates - overlapping patterns by making them mutually disjoint. Earlier equations - take precedence over later ones. This allows to give the specification in - a format very similar to functional programming. Note that the resulting - simplification and induction rules correspond to the transformed - specification, not the one given originally. This usually means that each - equation given by the user may result in several theorems. Also note that - this automatic transformation only works for ML-style datatype patterns. - - \<^descr> \<open>domintros\<close> enables the automated generation of introduction - rules for the domain predicate. While mostly not needed, they can be - helpful in some proofs about partial functions. + \<^descr> \<open>sequential\<close> enables a preprocessor which disambiguates overlapping + patterns by making them mutually disjoint. Earlier equations take precedence + over later ones. This allows to give the specification in a format very + similar to functional programming. Note that the resulting simplification + and induction rules correspond to the transformed specification, not the one + given originally. This usually means that each equation given by the user + may result in several theorems. Also note that this automatic transformation + only works for ML-style datatype patterns. + + \<^descr> \<open>domintros\<close> enables the automated generation of introduction rules for the + domain predicate. While mostly not needed, they can be helpful in some + proofs about partial functions. \<close> subsubsection \<open>Example: evaluation of expressions\<close> -text \<open>Subsequently, we define mutual datatypes for arithmetic and boolean - expressions, and use @{command primrec} for evaluation functions that - follow the same recursive structure.\<close> +text \<open> + Subsequently, we define mutual datatypes for arithmetic and boolean + expressions, and use @{command primrec} for evaluation functions that follow + the same recursive structure. +\<close> (*<*)experiment begin(*>*) datatype 'a aexp = @@ -387,16 +378,15 @@ | "evalb env (And b1 b2) = (evalb env b1 \<and> evalb env b2)" | "evalb env (Neg b) = (\<not> evalb env b)" -text \<open>Since the value of an expression depends on the value of its - variables, the functions @{const evala} and @{const evalb} take an - additional parameter, an \<^emph>\<open>environment\<close> that maps variables to their - values. +text \<open> + Since the value of an expression depends on the value of its variables, the + functions @{const evala} and @{const evalb} take an additional parameter, an + \<^emph>\<open>environment\<close> that maps variables to their values. \<^medskip> - Substitution on expressions can be defined similarly. The mapping - \<open>f\<close> of type @{typ "'a \<Rightarrow> 'a aexp"} given as a parameter is lifted - canonically on the types @{typ "'a aexp"} and @{typ "'a bexp"}, - respectively. + Substitution on expressions can be defined similarly. The mapping \<open>f\<close> of + type @{typ "'a \<Rightarrow> 'a aexp"} given as a parameter is lifted canonically on the + types @{typ "'a aexp"} and @{typ "'a bexp"}, respectively. \<close> primrec substa :: "('a \<Rightarrow> 'b aexp) \<Rightarrow> 'a aexp \<Rightarrow> 'b aexp" @@ -411,10 +401,11 @@ | "substb f (And b1 b2) = And (substb f b1) (substb f b2)" | "substb f (Neg b) = Neg (substb f b)" -text \<open>In textbooks about semantics one often finds substitution theorems, - which express the relationship between substitution and evaluation. For - @{typ "'a aexp"} and @{typ "'a bexp"}, we can prove such a theorem by - mutual induction, followed by simplification. +text \<open> + In textbooks about semantics one often finds substitution theorems, which + express the relationship between substitution and evaluation. For @{typ "'a + aexp"} and @{typ "'a bexp"}, we can prove such a theorem by mutual + induction, followed by simplification. \<close> lemma subst_one: @@ -437,9 +428,10 @@ (*<*)experiment begin(*>*) datatype ('a, 'b) "term" = Var 'a | App 'b "('a, 'b) term list" -text \<open>A substitution function on type @{typ "('a, 'b) term"} can be - defined as follows, by working simultaneously on @{typ "('a, 'b) - term list"}:\<close> +text \<open> + A substitution function on type @{typ "('a, 'b) term"} can be defined as + follows, by working simultaneously on @{typ "('a, 'b) term list"}: +\<close> primrec subst_term :: "('a \<Rightarrow> ('a, 'b) term) \<Rightarrow> ('a, 'b) term \<Rightarrow> ('a, 'b) term" and subst_term_list :: "('a \<Rightarrow> ('a, 'b) term) \<Rightarrow> ('a, 'b) term list \<Rightarrow> ('a, 'b) term list" @@ -449,9 +441,10 @@ | "subst_term_list f [] = []" | "subst_term_list f (t # ts) = subst_term f t # subst_term_list f ts" -text \<open>The recursion scheme follows the structure of the unfolded - definition of type @{typ "('a, 'b) term"}. To prove properties of this - substitution function, mutual induction is needed: +text \<open> + The recursion scheme follows the structure of the unfolded definition of + type @{typ "('a, 'b) term"}. To prove properties of this substitution + function, mutual induction is needed: \<close> lemma "subst_term (subst_term f1 \<circ> f2) t = @@ -475,9 +468,9 @@ "map_tree f (Atom a) = Atom (f a)" | "map_tree f (Branch ts) = Branch (\<lambda>x. map_tree f (ts x))" -text \<open>Note that all occurrences of functions such as \<open>ts\<close> above must - be applied to an argument. In particular, @{term "map_tree f \<circ> ts"} is not - allowed here. +text \<open> + Note that all occurrences of functions such as \<open>ts\<close> above must be applied to + an argument. In particular, @{term "map_tree f \<circ> ts"} is not allowed here. \<^medskip> Here is a simple composition lemma for @{term map_tree}: @@ -511,45 +504,42 @@ orders: ( 'max' | 'min' | 'ms' ) * \<close>} - \<^descr> @{method (HOL) pat_completeness} is a specialized method to - solve goals regarding the completeness of pattern matching, as - required by the @{command (HOL) "function"} package (cf.\ - @{cite "isabelle-function"}). - - \<^descr> @{method (HOL) relation}~\<open>R\<close> introduces a termination - proof using the relation \<open>R\<close>. The resulting proof state will - contain goals expressing that \<open>R\<close> is wellfounded, and that the - arguments of recursive calls decrease with respect to \<open>R\<close>. - Usually, this method is used as the initial proof step of manual - termination proofs. - - \<^descr> @{method (HOL) "lexicographic_order"} attempts a fully - automated termination proof by searching for a lexicographic - combination of size measures on the arguments of the function. The - method accepts the same arguments as the @{method auto} method, - which it uses internally to prove local descents. The @{syntax - clasimpmod} modifiers are accepted (as for @{method auto}). - - In case of failure, extensive information is printed, which can help - to analyse the situation (cf.\ @{cite "isabelle-function"}). - - \<^descr> @{method (HOL) "size_change"} also works on termination goals, - using a variation of the size-change principle, together with a - graph decomposition technique (see @{cite krauss_phd} for details). - Three kinds of orders are used internally: \<open>max\<close>, \<open>min\<close>, - and \<open>ms\<close> (multiset), which is only available when the theory - \<open>Multiset\<close> is loaded. When no order kinds are given, they are - tried in order. The search for a termination proof uses SAT solving + \<^descr> @{method (HOL) pat_completeness} is a specialized method to solve goals + regarding the completeness of pattern matching, as required by the @{command + (HOL) "function"} package (cf.\ @{cite "isabelle-function"}). + + \<^descr> @{method (HOL) relation}~\<open>R\<close> introduces a termination proof using the + relation \<open>R\<close>. The resulting proof state will contain goals expressing that + \<open>R\<close> is wellfounded, and that the arguments of recursive calls decrease with + respect to \<open>R\<close>. Usually, this method is used as the initial proof step of + manual termination proofs. + + \<^descr> @{method (HOL) "lexicographic_order"} attempts a fully automated + termination proof by searching for a lexicographic combination of size + measures on the arguments of the function. The method accepts the same + arguments as the @{method auto} method, which it uses internally to prove + local descents. The @{syntax clasimpmod} modifiers are accepted (as for + @{method auto}). + + In case of failure, extensive information is printed, which can help to + analyse the situation (cf.\ @{cite "isabelle-function"}). + + \<^descr> @{method (HOL) "size_change"} also works on termination goals, using a + variation of the size-change principle, together with a graph decomposition + technique (see @{cite krauss_phd} for details). Three kinds of orders are + used internally: \<open>max\<close>, \<open>min\<close>, and \<open>ms\<close> (multiset), which is only available + when the theory \<open>Multiset\<close> is loaded. When no order kinds are given, they + are tried in order. The search for a termination proof uses SAT solving internally. - For local descent proofs, the @{syntax clasimpmod} modifiers are - accepted (as for @{method auto}). - - \<^descr> @{method (HOL) induction_schema} derives user-specified - induction rules from well-founded induction and completeness of - patterns. This factors out some operations that are done internally - by the function package and makes them available separately. See - @{file "~~/src/HOL/ex/Induction_Schema.thy"} for examples. + For local descent proofs, the @{syntax clasimpmod} modifiers are accepted + (as for @{method auto}). + + \<^descr> @{method (HOL) induction_schema} derives user-specified induction rules + from well-founded induction and completeness of patterns. This factors out + some operations that are done internally by the function package and makes + them available separately. See @{file "~~/src/HOL/ex/Induction_Schema.thy"} + for examples. \<close> @@ -566,51 +556,45 @@ @{syntax specification} \<close>} - \<^descr> @{command (HOL) "partial_function"}~\<open>(mode)\<close> defines - recursive functions based on fixpoints in complete partial - orders. No termination proof is required from the user or - constructed internally. Instead, the possibility of non-termination - is modelled explicitly in the result type, which contains an - explicit bottom element. - - Pattern matching and mutual recursion are currently not supported. - Thus, the specification consists of a single function described by a - single recursive equation. - - There are no fixed syntactic restrictions on the body of the - function, but the induced functional must be provably monotonic - wrt.\ the underlying order. The monotonicity proof is performed - internally, and the definition is rejected when it fails. The proof - can be influenced by declaring hints using the - @{attribute (HOL) partial_function_mono} attribute. - - The mandatory \<open>mode\<close> argument specifies the mode of operation - of the command, which directly corresponds to a complete partial - order on the result type. By default, the following modes are - defined: - - \<^descr> \<open>option\<close> defines functions that map into the @{type - option} type. Here, the value @{term None} is used to model a - non-terminating computation. Monotonicity requires that if @{term - None} is returned by a recursive call, then the overall result must - also be @{term None}. This is best achieved through the use of the - monadic operator @{const "Option.bind"}. - - \<^descr> \<open>tailrec\<close> defines functions with an arbitrary result - type and uses the slightly degenerated partial order where @{term - "undefined"} is the bottom element. Now, monotonicity requires that - if @{term undefined} is returned by a recursive call, then the - overall result must also be @{term undefined}. In practice, this is - only satisfied when each recursive call is a tail call, whose result - is directly returned. Thus, this mode of operation allows the - definition of arbitrary tail-recursive functions. - - Experienced users may define new modes by instantiating the locale - @{const "partial_function_definitions"} appropriately. - - \<^descr> @{attribute (HOL) partial_function_mono} declares rules for - use in the internal monotonicity proofs of partial function - definitions. + \<^descr> @{command (HOL) "partial_function"}~\<open>(mode)\<close> defines recursive functions + based on fixpoints in complete partial orders. No termination proof is + required from the user or constructed internally. Instead, the possibility + of non-termination is modelled explicitly in the result type, which contains + an explicit bottom element. + + Pattern matching and mutual recursion are currently not supported. Thus, the + specification consists of a single function described by a single recursive + equation. + + There are no fixed syntactic restrictions on the body of the function, but + the induced functional must be provably monotonic wrt.\ the underlying + order. The monotonicity proof is performed internally, and the definition is + rejected when it fails. The proof can be influenced by declaring hints using + the @{attribute (HOL) partial_function_mono} attribute. + + The mandatory \<open>mode\<close> argument specifies the mode of operation of the + command, which directly corresponds to a complete partial order on the + result type. By default, the following modes are defined: + + \<^descr> \<open>option\<close> defines functions that map into the @{type option} type. Here, + the value @{term None} is used to model a non-terminating computation. + Monotonicity requires that if @{term None} is returned by a recursive + call, then the overall result must also be @{term None}. This is best + achieved through the use of the monadic operator @{const "Option.bind"}. + + \<^descr> \<open>tailrec\<close> defines functions with an arbitrary result type and uses the + slightly degenerated partial order where @{term "undefined"} is the bottom + element. Now, monotonicity requires that if @{term undefined} is returned + by a recursive call, then the overall result must also be @{term + undefined}. In practice, this is only satisfied when each recursive call + is a tail call, whose result is directly returned. Thus, this mode of + operation allows the definition of arbitrary tail-recursive functions. + + Experienced users may define new modes by instantiating the locale @{const + "partial_function_definitions"} appropriately. + + \<^descr> @{attribute (HOL) partial_function_mono} declares rules for use in the + internal monotonicity proofs of partial function definitions. \<close> @@ -635,21 +619,19 @@ (() | 'add' | 'del') ':' @{syntax thms}) | @{syntax clasimpmod} \<close>} - \<^descr> @{command (HOL) "recdef"} defines general well-founded - recursive functions (using the TFL package). The - ``\<open>(permissive)\<close>'' option tells TFL to recover from - failed proof attempts, returning unfinished results. The - \<open>recdef_simp\<close>, \<open>recdef_cong\<close>, and - \<open>recdef_wf\<close> hints refer to auxiliary rules to be used - in the internal automated proof process of TFL. Additional - @{syntax clasimpmod} declarations may be given to tune the context - of the Simplifier (cf.\ \secref{sec:simplifier}) and Classical - reasoner (cf.\ \secref{sec:classical}). + \<^descr> @{command (HOL) "recdef"} defines general well-founded recursive functions + (using the TFL package). The ``\<open>(permissive)\<close>'' option tells TFL to recover + from failed proof attempts, returning unfinished results. The \<open>recdef_simp\<close>, + \<open>recdef_cong\<close>, and \<open>recdef_wf\<close> hints refer to auxiliary rules to be used in + the internal automated proof process of TFL. Additional @{syntax clasimpmod} + declarations may be given to tune the context of the Simplifier (cf.\ + \secref{sec:simplifier}) and Classical reasoner (cf.\ + \secref{sec:classical}). \<^medskip> - Hints for @{command (HOL) "recdef"} may be also declared - globally, using the following attributes. + Hints for @{command (HOL) "recdef"} may be also declared globally, using the + following attributes. \begin{matharray}{rcl} @{attribute_def (HOL) recdef_simp} & : & \<open>attribute\<close> \\ @@ -674,11 +656,10 @@ \end{tabular} \<^medskip> - Adhoc overloading allows to overload a constant depending on - its type. Typically this involves the introduction of an - uninterpreted constant (used for input and output) and the addition - of some variants (used internally). For examples see - @{file "~~/src/HOL/ex/Adhoc_Overloading_Examples.thy"} and + Adhoc overloading allows to overload a constant depending on its type. + Typically this involves the introduction of an uninterpreted constant (used + for input and output) and the addition of some variants (used internally). + For examples see @{file "~~/src/HOL/ex/Adhoc_Overloading_Examples.thy"} and @{file "~~/src/HOL/Library/Monad_Syntax.thy"}. @{rail \<open> @@ -686,18 +667,17 @@ (@{syntax name} (@{syntax term} + ) + @'and') \<close>} - \<^descr> @{command "adhoc_overloading"}~\<open>c v\<^sub>1 ... v\<^sub>n\<close> - associates variants with an existing constant. - - \<^descr> @{command "no_adhoc_overloading"} is similar to - @{command "adhoc_overloading"}, but removes the specified variants - from the present context. - - \<^descr> @{attribute "show_variants"} controls printing of variants - of overloaded constants. If enabled, the internally used variants - are printed instead of their respective overloaded constants. This - is occasionally useful to check whether the system agrees with a - user's expectations about derived variants. + \<^descr> @{command "adhoc_overloading"}~\<open>c v\<^sub>1 ... v\<^sub>n\<close> associates variants with an + existing constant. + + \<^descr> @{command "no_adhoc_overloading"} is similar to @{command + "adhoc_overloading"}, but removes the specified variants from the present + context. + + \<^descr> @{attribute "show_variants"} controls printing of variants of overloaded + constants. If enabled, the internally used variants are printed instead of + their respective overloaded constants. This is occasionally useful to check + whether the system agrees with a user's expectations about derived variants. \<close> @@ -715,17 +695,16 @@ decl: (@{syntax name} ':')? @{syntax term} ('(' @'overloaded' ')')? \<close>} - \<^descr> @{command (HOL) "specification"}~\<open>decls \<phi>\<close> sets up a - goal stating the existence of terms with the properties specified to - hold for the constants given in \<open>decls\<close>. After finishing the - proof, the theory will be augmented with definitions for the given - constants, as well as with theorems stating the properties for these - constants. - - \<open>decl\<close> declares a constant to be defined by the - specification given. The definition for the constant \<open>c\<close> is - bound to the name \<open>c_def\<close> unless a theorem name is given in - the declaration. Overloaded constants should be declared as such. + \<^descr> @{command (HOL) "specification"}~\<open>decls \<phi>\<close> sets up a goal stating the + existence of terms with the properties specified to hold for the constants + given in \<open>decls\<close>. After finishing the proof, the theory will be augmented + with definitions for the given constants, as well as with theorems stating + the properties for these constants. + + \<open>decl\<close> declares a constant to be defined by the specification given. The + definition for the constant \<open>c\<close> is bound to the name \<open>c_def\<close> unless a + theorem name is given in the declaration. Overloaded constants should be + declared as such. \<close> @@ -755,22 +734,24 @@ old-style datatypes. - These commands are mostly obsolete; @{command (HOL) "datatype"} - should be used instead. - - See @{cite "isabelle-datatypes"} for more details on datatypes. Apart from proper - proof methods for case analysis and induction, there are also + These commands are mostly obsolete; @{command (HOL) "datatype"} should be + used instead. + + See @{cite "isabelle-datatypes"} for more details on datatypes. Apart from + proper proof methods for case analysis and induction, there are also emulations of ML tactics @{method (HOL) case_tac} and @{method (HOL) - induct_tac} available, see \secref{sec:hol-induct-tac}; these admit - to refer directly to the internal structure of subgoals (including - internally bound parameters). + induct_tac} available, see \secref{sec:hol-induct-tac}; these admit to refer + directly to the internal structure of subgoals (including internally bound + parameters). \<close> subsubsection \<open>Examples\<close> -text \<open>We define a type of finite sequences, with slightly different names - than the existing @{typ "'a list"} that is already in @{theory Main}:\<close> +text \<open> + We define a type of finite sequences, with slightly different names than the + existing @{typ "'a list"} that is already in @{theory Main}: +\<close> (*<*)experiment begin(*>*) datatype 'a seq = Empty | Seq 'a "'a seq" @@ -803,21 +784,21 @@ text \<open> In principle, records merely generalize the concept of tuples, where - components may be addressed by labels instead of just position. The - logical infrastructure of records in Isabelle/HOL is slightly more - advanced, though, supporting truly extensible record schemes. This - admits operations that are polymorphic with respect to record - extension, yielding ``object-oriented'' effects like (single) - inheritance. See also @{cite "NaraschewskiW-TPHOLs98"} for more - details on object-oriented verification and record subtyping in HOL. + components may be addressed by labels instead of just position. The logical + infrastructure of records in Isabelle/HOL is slightly more advanced, though, + supporting truly extensible record schemes. This admits operations that are + polymorphic with respect to record extension, yielding ``object-oriented'' + effects like (single) inheritance. See also @{cite "NaraschewskiW-TPHOLs98"} + for more details on object-oriented verification and record subtyping in + HOL. \<close> subsection \<open>Basic concepts\<close> text \<open> - Isabelle/HOL supports both \<^emph>\<open>fixed\<close> and \<^emph>\<open>schematic\<close> records - at the level of terms and types. The notation is as follows: + Isabelle/HOL supports both \<^emph>\<open>fixed\<close> and \<^emph>\<open>schematic\<close> records at the level of + terms and types. The notation is as follows: \begin{center} \begin{tabular}{l|l|l} @@ -830,47 +811,41 @@ The ASCII representation of \<open>\<lparr>x = a\<rparr>\<close> is \<open>(| x = a |)\<close>. - A fixed record \<open>\<lparr>x = a, y = b\<rparr>\<close> has field \<open>x\<close> of value - \<open>a\<close> and field \<open>y\<close> of value \<open>b\<close>. The corresponding - type is \<open>\<lparr>x :: A, y :: B\<rparr>\<close>, assuming that \<open>a :: A\<close> - and \<open>b :: B\<close>. - - A record scheme like \<open>\<lparr>x = a, y = b, \<dots> = m\<rparr>\<close> contains fields - \<open>x\<close> and \<open>y\<close> as before, but also possibly further fields - as indicated by the ``\<open>\<dots>\<close>'' notation (which is actually part - of the syntax). The improper field ``\<open>\<dots>\<close>'' of a record - scheme is called the \<^emph>\<open>more part\<close>. Logically it is just a free - variable, which is occasionally referred to as ``row variable'' in - the literature. The more part of a record scheme may be - instantiated by zero or more further components. For example, the - previous scheme may get instantiated to \<open>\<lparr>x = a, y = b, z = - c, \<dots> = m'\<rparr>\<close>, where \<open>m'\<close> refers to a different more part. - Fixed records are special instances of record schemes, where - ``\<open>\<dots>\<close>'' is properly terminated by the \<open>() :: unit\<close> - element. In fact, \<open>\<lparr>x = a, y = b\<rparr>\<close> is just an abbreviation - for \<open>\<lparr>x = a, y = b, \<dots> = ()\<rparr>\<close>. + A fixed record \<open>\<lparr>x = a, y = b\<rparr>\<close> has field \<open>x\<close> of value \<open>a\<close> and field \<open>y\<close> of + value \<open>b\<close>. The corresponding type is \<open>\<lparr>x :: A, y :: B\<rparr>\<close>, assuming that \<open>a :: + A\<close> and \<open>b :: B\<close>. + + A record scheme like \<open>\<lparr>x = a, y = b, \<dots> = m\<rparr>\<close> contains fields \<open>x\<close> and \<open>y\<close> as + before, but also possibly further fields as indicated by the ``\<open>\<dots>\<close>'' + notation (which is actually part of the syntax). The improper field ``\<open>\<dots>\<close>'' + of a record scheme is called the \<^emph>\<open>more part\<close>. Logically it is just a free + variable, which is occasionally referred to as ``row variable'' in the + literature. The more part of a record scheme may be instantiated by zero or + more further components. For example, the previous scheme may get + instantiated to \<open>\<lparr>x = a, y = b, z = c, \<dots> = m'\<rparr>\<close>, where \<open>m'\<close> refers to a + different more part. Fixed records are special instances of record schemes, + where ``\<open>\<dots>\<close>'' is properly terminated by the \<open>() :: unit\<close> element. In fact, + \<open>\<lparr>x = a, y = b\<rparr>\<close> is just an abbreviation for \<open>\<lparr>x = a, y = b, \<dots> = ()\<rparr>\<close>. \<^medskip> - Two key observations make extensible records in a simply - typed language like HOL work out: - - \<^enum> the more part is internalized, as a free term or type - variable, - - \<^enum> field names are externalized, they cannot be accessed within - the logic as first-class values. + Two key observations make extensible records in a simply typed language like + HOL work out: + + \<^enum> the more part is internalized, as a free term or type variable, + + \<^enum> field names are externalized, they cannot be accessed within the logic as + first-class values. \<^medskip> - In Isabelle/HOL record types have to be defined explicitly, - fixing their field names and types, and their (optional) parent - record. Afterwards, records may be formed using above syntax, while - obeying the canonical order of fields as given by their declaration. - The record package provides several standard operations like - selectors and updates. The common setup for various generic proof - tools enable succinct reasoning patterns. See also the Isabelle/HOL - tutorial @{cite "isabelle-hol-book"} for further instructions on using - records in practice. + In Isabelle/HOL record types have to be defined explicitly, fixing their + field names and types, and their (optional) parent record. Afterwards, + records may be formed using above syntax, while obeying the canonical order + of fields as given by their declaration. The record package provides several + standard operations like selectors and updates. The common setup for various + generic proof tools enable succinct reasoning patterns. See also the + Isabelle/HOL tutorial @{cite "isabelle-hol-book"} for further instructions + on using records in practice. \<close> @@ -893,45 +868,40 @@ modes: '(' (@{syntax name} +) ')' \<close>} - \<^descr> @{command (HOL) "record"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t = \<tau> + c\<^sub>1 :: \<sigma>\<^sub>1 - \<dots> c\<^sub>n :: \<sigma>\<^sub>n\<close> defines extensible record type \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close>, - derived from the optional parent record \<open>\<tau>\<close> by adding new - field components \<open>c\<^sub>i :: \<sigma>\<^sub>i\<close> etc. - - The type variables of \<open>\<tau>\<close> and \<open>\<sigma>\<^sub>i\<close> need to be - covered by the (distinct) parameters \<open>\<alpha>\<^sub>1, \<dots>, - \<alpha>\<^sub>m\<close>. Type constructor \<open>t\<close> has to be new, while \<open>\<tau>\<close> needs to specify an instance of an existing record type. At - least one new field \<open>c\<^sub>i\<close> has to be specified. - Basically, field names need to belong to a unique record. This is - not a real restriction in practice, since fields are qualified by - the record name internally. - - The parent record specification \<open>\<tau>\<close> is optional; if omitted - \<open>t\<close> becomes a root record. The hierarchy of all records - declared within a theory context forms a forest structure, i.e.\ a - set of trees starting with a root record each. There is no way to - merge multiple parent records! - - For convenience, \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> is made a - type abbreviation for the fixed record type \<open>\<lparr>c\<^sub>1 :: - \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n\<rparr>\<close>, likewise is \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m, \<zeta>) t_scheme\<close> made an abbreviation for - \<open>\<lparr>c\<^sub>1 :: \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n, \<dots> :: - \<zeta>\<rparr>\<close>. - - \<^descr> @{command (HOL) "print_record"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> prints the definition - of record \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close>. Optionally \<open>modes\<close> can be specified, which are - appended to the current print mode; see \secref{sec:print-modes}. + \<^descr> @{command (HOL) "record"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t = \<tau> + c\<^sub>1 :: \<sigma>\<^sub>1 \<dots> c\<^sub>n :: \<sigma>\<^sub>n\<close> + defines extensible record type \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close>, derived from the optional + parent record \<open>\<tau>\<close> by adding new field components \<open>c\<^sub>i :: \<sigma>\<^sub>i\<close> etc. + + The type variables of \<open>\<tau>\<close> and \<open>\<sigma>\<^sub>i\<close> need to be covered by the (distinct) + parameters \<open>\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m\<close>. Type constructor \<open>t\<close> has to be new, while \<open>\<tau>\<close> + needs to specify an instance of an existing record type. At least one new + field \<open>c\<^sub>i\<close> has to be specified. Basically, field names need to belong to a + unique record. This is not a real restriction in practice, since fields are + qualified by the record name internally. + + The parent record specification \<open>\<tau>\<close> is optional; if omitted \<open>t\<close> becomes a + root record. The hierarchy of all records declared within a theory context + forms a forest structure, i.e.\ a set of trees starting with a root record + each. There is no way to merge multiple parent records! + + For convenience, \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> is made a type abbreviation for the fixed + record type \<open>\<lparr>c\<^sub>1 :: \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n\<rparr>\<close>, likewise is \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m, \<zeta>) + t_scheme\<close> made an abbreviation for \<open>\<lparr>c\<^sub>1 :: \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n, \<dots> :: \<zeta>\<rparr>\<close>. + + \<^descr> @{command (HOL) "print_record"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> prints the definition of + record \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close>. Optionally \<open>modes\<close> can be specified, which are + appended to the current print mode; see \secref{sec:print-modes}. \<close> subsection \<open>Record operations\<close> -text \<open>Any record definition of the form presented above produces certain - standard operations. Selectors and updates are provided for any field, - including the improper one ``\<open>more\<close>''. There are also cumulative - record constructor functions. To simplify the presentation below, we - assume for now that \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> is a root record with fields - \<open>c\<^sub>1 :: \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n\<close>. +text \<open> + Any record definition of the form presented above produces certain standard + operations. Selectors and updates are provided for any field, including the + improper one ``\<open>more\<close>''. There are also cumulative record constructor + functions. To simplify the presentation below, we assume for now that \<open>(\<alpha>\<^sub>1, + \<dots>, \<alpha>\<^sub>m) t\<close> is a root record with fields \<open>c\<^sub>1 :: \<sigma>\<^sub>1, \<dots>, c\<^sub>n :: \<sigma>\<^sub>n\<close>. \<^medskip> \<^bold>\<open>Selectors\<close> and \<^bold>\<open>updates\<close> are available for any @@ -942,33 +912,32 @@ \<open>c\<^sub>i_update\<close> & \<open>::\<close> & \<open>\<sigma>\<^sub>i \<Rightarrow> \<lparr>\<^vec>c :: \<^vec>\<sigma>, \<dots> :: \<zeta>\<rparr> \<Rightarrow> \<lparr>\<^vec>c :: \<^vec>\<sigma>, \<dots> :: \<zeta>\<rparr>\<close> \\ \end{matharray} - There is special syntax for application of updates: \<open>r\<lparr>x := a\<rparr>\<close> - abbreviates term \<open>x_update a r\<close>. Further notation for repeated - updates is also available: \<open>r\<lparr>x := a\<rparr>\<lparr>y := b\<rparr>\<lparr>z := c\<rparr>\<close> may be - written \<open>r\<lparr>x := a, y := b, z := c\<rparr>\<close>. Note that because of postfix - notation the order of fields shown here is reverse than in the actual - term. Since repeated updates are just function applications, fields may be - freely permuted in \<open>\<lparr>x := a, y := b, z := c\<rparr>\<close>, as far as logical - equality is concerned. Thus commutativity of independent updates can be - proven within the logic for any two fields, but not as a general theorem. + There is special syntax for application of updates: \<open>r\<lparr>x := a\<rparr>\<close> abbreviates + term \<open>x_update a r\<close>. Further notation for repeated updates is also + available: \<open>r\<lparr>x := a\<rparr>\<lparr>y := b\<rparr>\<lparr>z := c\<rparr>\<close> may be written \<open>r\<lparr>x := a, y := b, z + := c\<rparr>\<close>. Note that because of postfix notation the order of fields shown here + is reverse than in the actual term. Since repeated updates are just function + applications, fields may be freely permuted in \<open>\<lparr>x := a, y := b, z := c\<rparr>\<close>, + as far as logical equality is concerned. Thus commutativity of independent + updates can be proven within the logic for any two fields, but not as a + general theorem. \<^medskip> - The \<^bold>\<open>make\<close> operation provides a cumulative record - constructor function: + The \<^bold>\<open>make\<close> operation provides a cumulative record constructor function: \begin{matharray}{lll} \<open>t.make\<close> & \<open>::\<close> & \<open>\<sigma>\<^sub>1 \<Rightarrow> \<dots> \<sigma>\<^sub>n \<Rightarrow> \<lparr>\<^vec>c :: \<^vec>\<sigma>\<rparr>\<close> \\ \end{matharray} \<^medskip> - We now reconsider the case of non-root records, which are derived - of some parent. In general, the latter may depend on another parent as - well, resulting in a list of \<^emph>\<open>ancestor records\<close>. Appending the lists - of fields of all ancestors results in a certain field prefix. The record - package automatically takes care of this by lifting operations over this - context of ancestor fields. Assuming that \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> has - ancestor fields \<open>b\<^sub>1 :: \<rho>\<^sub>1, \<dots>, b\<^sub>k :: \<rho>\<^sub>k\<close>, the above record - operations will get the following types: + We now reconsider the case of non-root records, which are derived of some + parent. In general, the latter may depend on another parent as well, + resulting in a list of \<^emph>\<open>ancestor records\<close>. Appending the lists of fields of + all ancestors results in a certain field prefix. The record package + automatically takes care of this by lifting operations over this context of + ancestor fields. Assuming that \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>m) t\<close> has ancestor fields \<open>b\<^sub>1 :: + \<rho>\<^sub>1, \<dots>, b\<^sub>k :: \<rho>\<^sub>k\<close>, the above record operations will get the following + types: \<^medskip> \begin{tabular}{lll} @@ -981,12 +950,11 @@ \end{tabular} \<^medskip> - Some further operations address the extension aspect of a - derived record scheme specifically: \<open>t.fields\<close> produces a record - fragment consisting of exactly the new fields introduced here (the result - may serve as a more part elsewhere); \<open>t.extend\<close> takes a fixed - record and adds a given more part; \<open>t.truncate\<close> restricts a record - scheme to a fixed record. + Some further operations address the extension aspect of a derived record + scheme specifically: \<open>t.fields\<close> produces a record fragment consisting of + exactly the new fields introduced here (the result may serve as a more part + elsewhere); \<open>t.extend\<close> takes a fixed record and adds a given more part; + \<open>t.truncate\<close> restricts a record scheme to a fixed record. \<^medskip> \begin{tabular}{lll} @@ -997,52 +965,49 @@ \end{tabular} \<^medskip> - Note that \<open>t.make\<close> and \<open>t.fields\<close> coincide for - root records. + Note that \<open>t.make\<close> and \<open>t.fields\<close> coincide for root records. \<close> subsection \<open>Derived rules and proof tools\<close> text \<open> - The record package proves several results internally, declaring - these facts to appropriate proof tools. This enables users to - reason about record structures quite conveniently. Assume that - \<open>t\<close> is a record type as specified above. + The record package proves several results internally, declaring these facts + to appropriate proof tools. This enables users to reason about record + structures quite conveniently. Assume that \<open>t\<close> is a record type as specified + above. \<^enum> Standard conversions for selectors or updates applied to record constructor terms are made part of the default Simplifier context; thus proofs by reduction of basic operations merely require the @{method simp} - method without further arguments. These rules are available as \<open>t.simps\<close>, too. + method without further arguments. These rules are available as \<open>t.simps\<close>, + too. \<^enum> Selectors applied to updated records are automatically reduced by an internal simplification procedure, which is also part of the standard Simplifier setup. - \<^enum> Inject equations of a form analogous to @{prop "(x, y) = (x', y') \<equiv> - x = x' \<and> y = y'"} are declared to the Simplifier and Classical Reasoner as + \<^enum> Inject equations of a form analogous to @{prop "(x, y) = (x', y') \<equiv> x = x' + \<and> y = y'"} are declared to the Simplifier and Classical Reasoner as @{attribute iff} rules. These rules are available as \<open>t.iffs\<close>. - \<^enum> The introduction rule for record equality analogous to \<open>x r = - x r' \<Longrightarrow> y r = y r' \<dots> \<Longrightarrow> r = r'\<close> is declared to the Simplifier, and as the - basic rule context as ``@{attribute intro}\<open>?\<close>''. The rule is - called \<open>t.equality\<close>. - - \<^enum> Representations of arbitrary record expressions as canonical - constructor terms are provided both in @{method cases} and @{method - induct} format (cf.\ the generic proof methods of the same name, - \secref{sec:cases-induct}). Several variations are available, for fixed - records, record schemes, more parts etc. + \<^enum> The introduction rule for record equality analogous to \<open>x r = x r' \<Longrightarrow> y r = + y r' \<dots> \<Longrightarrow> r = r'\<close> is declared to the Simplifier, and as the basic rule + context as ``@{attribute intro}\<open>?\<close>''. The rule is called \<open>t.equality\<close>. + + \<^enum> Representations of arbitrary record expressions as canonical constructor + terms are provided both in @{method cases} and @{method induct} format (cf.\ + the generic proof methods of the same name, \secref{sec:cases-induct}). + Several variations are available, for fixed records, record schemes, more + parts etc. The generic proof methods are sufficiently smart to pick the most sensible rule according to the type of the indicated record expression: users just - need to apply something like ``\<open>(cases r)\<close>'' to a certain proof - problem. - - \<^enum> The derived record operations \<open>t.make\<close>, \<open>t.fields\<close>, - \<open>t.extend\<close>, \<open>t.truncate\<close> are \<^emph>\<open>not\<close> treated - automatically, but usually need to be expanded by hand, using the - collective fact \<open>t.defs\<close>. + need to apply something like ``\<open>(cases r)\<close>'' to a certain proof problem. + + \<^enum> The derived record operations \<open>t.make\<close>, \<open>t.fields\<close>, \<open>t.extend\<close>, + \<open>t.truncate\<close> are \<^emph>\<open>not\<close> treated automatically, but usually need to be + expanded by hand, using the collective fact \<open>t.defs\<close>. \<close> @@ -1060,13 +1025,12 @@ A type definition identifies a new type with a non-empty subset of an existing type. More precisely, the new type is defined by exhibiting an - existing type \<open>\<tau>\<close>, a set \<open>A :: \<tau> set\<close>, and proving @{prop - "\<exists>x. x \<in> A"}. Thus \<open>A\<close> is a non-empty subset of \<open>\<tau>\<close>, and the - new type denotes this subset. New functions are postulated that establish - an isomorphism between the new type and the subset. In general, the type - \<open>\<tau>\<close> may involve type variables \<open>\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n\<close> which means - that the type definition produces a type constructor \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n) - t\<close> depending on those type arguments. + existing type \<open>\<tau>\<close>, a set \<open>A :: \<tau> set\<close>, and proving @{prop "\<exists>x. x \<in> A"}. Thus + \<open>A\<close> is a non-empty subset of \<open>\<tau>\<close>, and the new type denotes this subset. New + functions are postulated that establish an isomorphism between the new type + and the subset. In general, the type \<open>\<tau>\<close> may involve type variables \<open>\<alpha>\<^sub>1, \<dots>, + \<alpha>\<^sub>n\<close> which means that the type definition produces a type constructor \<open>(\<alpha>\<^sub>1, + \<dots>, \<alpha>\<^sub>n) t\<close> depending on those type arguments. @{rail \<open> @@{command (HOL) typedef} @{syntax "overloaded"}? abs_type '=' rep_set @@ -1078,98 +1042,94 @@ rep_set: @{syntax term} (@'morphisms' @{syntax name} @{syntax name})? \<close>} - To understand the concept of type definition better, we need to recount - its somewhat complex history. The HOL logic goes back to the ``Simple - Theory of Types'' (STT) of A. Church @{cite "church40"}, which is further - explained in the book by P. Andrews @{cite "andrews86"}. The overview - article by W. Farmer @{cite "Farmer:2008"} points out the ``seven - virtues'' of this relatively simple family of logics. STT has only ground - types, without polymorphism and without type definitions. + To understand the concept of type definition better, we need to recount its + somewhat complex history. The HOL logic goes back to the ``Simple Theory of + Types'' (STT) of A. Church @{cite "church40"}, which is further explained in + the book by P. Andrews @{cite "andrews86"}. The overview article by W. + Farmer @{cite "Farmer:2008"} points out the ``seven virtues'' of this + relatively simple family of logics. STT has only ground types, without + polymorphism and without type definitions. \<^medskip> - M. Gordon @{cite "Gordon:1985:HOL"} augmented Church's STT by - adding schematic polymorphism (type variables and type constructors) and a - facility to introduce new types as semantic subtypes from existing types. - This genuine extension of the logic was explained semantically by A. Pitts - in the book of the original Cambridge HOL88 system @{cite "pitts93"}. Type - definitions work in this setting, because the general model-theory of STT - is restricted to models that ensure that the universe of type - interpretations is closed by forming subsets (via predicates taken from - the logic). + M. Gordon @{cite "Gordon:1985:HOL"} augmented Church's STT by adding + schematic polymorphism (type variables and type constructors) and a facility + to introduce new types as semantic subtypes from existing types. This + genuine extension of the logic was explained semantically by A. Pitts in the + book of the original Cambridge HOL88 system @{cite "pitts93"}. Type + definitions work in this setting, because the general model-theory of STT is + restricted to models that ensure that the universe of type interpretations + is closed by forming subsets (via predicates taken from the logic). \<^medskip> - Isabelle/HOL goes beyond Gordon-style HOL by admitting overloaded - constant definitions @{cite "Wenzel:1997:TPHOL" and - "Haftmann-Wenzel:2006:classes"}, which are actually a concept of - Isabelle/Pure and do not depend on particular set-theoretic semantics of - HOL. Over many years, there was no formal checking of semantic type - definitions in Isabelle/HOL versus syntactic constant definitions in - Isabelle/Pure. So the @{command typedef} command was described as - ``axiomatic'' in the sense of \secref{sec:axiomatizations}, only with some - local checks of the given type and its representing set. + Isabelle/HOL goes beyond Gordon-style HOL by admitting overloaded constant + definitions @{cite "Wenzel:1997:TPHOL" and "Haftmann-Wenzel:2006:classes"}, + which are actually a concept of Isabelle/Pure and do not depend on + particular set-theoretic semantics of HOL. Over many years, there was no + formal checking of semantic type definitions in Isabelle/HOL versus + syntactic constant definitions in Isabelle/Pure. So the @{command typedef} + command was described as ``axiomatic'' in the sense of + \secref{sec:axiomatizations}, only with some local checks of the given type + and its representing set. Recent clarification of overloading in the HOL logic proper @{cite "Kuncar-Popescu:2015"} demonstrates how the dissimilar concepts of constant definitions versus type definitions may be understood uniformly. This requires an interpretation of Isabelle/HOL that substantially reforms the set-theoretic model of A. Pitts @{cite "pitts93"}, by taking a schematic - view on polymorphism and interpreting only ground types in the - set-theoretic sense of HOL88. Moreover, type-constructors may be - explicitly overloaded, e.g.\ by making the subset depend on type-class - parameters (cf.\ \secref{sec:class}). This is semantically like a - dependent type: the meaning relies on the operations provided by different - type-class instances. - - \<^descr> @{command (HOL) "typedef"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n) t = A\<close> defines a - new type \<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n) t\<close> from the set \<open>A\<close> over an existing - type. The set \<open>A\<close> may contain type variables \<open>\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n\<close> - as specified on the LHS, but no term variables. Non-emptiness of \<open>A\<close> - needs to be proven on the spot, in order to turn the internal conditional - characterization into usable theorems. - - The ``\<open>(overloaded)\<close>'' option allows the @{command - "typedef"} specification to depend on constants that are not (yet) - specified and thus left open as parameters, e.g.\ type-class parameters. + view on polymorphism and interpreting only ground types in the set-theoretic + sense of HOL88. Moreover, type-constructors may be explicitly overloaded, + e.g.\ by making the subset depend on type-class parameters (cf.\ + \secref{sec:class}). This is semantically like a dependent type: the meaning + relies on the operations provided by different type-class instances. + + \<^descr> @{command (HOL) "typedef"}~\<open>(\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n) t = A\<close> defines a new type \<open>(\<alpha>\<^sub>1, + \<dots>, \<alpha>\<^sub>n) t\<close> from the set \<open>A\<close> over an existing type. The set \<open>A\<close> may contain + type variables \<open>\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n\<close> as specified on the LHS, but no term variables. + Non-emptiness of \<open>A\<close> needs to be proven on the spot, in order to turn the + internal conditional characterization into usable theorems. + + The ``\<open>(overloaded)\<close>'' option allows the @{command "typedef"} specification + to depend on constants that are not (yet) specified and thus left open as + parameters, e.g.\ type-class parameters. Within a local theory specification, the newly introduced type constructor cannot depend on parameters or assumptions of the context: this is - syntactically impossible in HOL. The non-emptiness proof may formally - depend on local assumptions, but this has little practical relevance. - - For @{command (HOL) "typedef"}~\<open>t = A\<close> the newly introduced - type \<open>t\<close> is accompanied by a pair of morphisms to relate it to - the representing set over the old type. By default, the injection - from type to set is called \<open>Rep_t\<close> and its inverse \<open>Abs_t\<close>: An explicit @{keyword (HOL) "morphisms"} specification - allows to provide alternative names. + syntactically impossible in HOL. The non-emptiness proof may formally depend + on local assumptions, but this has little practical relevance. + + For @{command (HOL) "typedef"}~\<open>t = A\<close> the newly introduced type \<open>t\<close> is + accompanied by a pair of morphisms to relate it to the representing set over + the old type. By default, the injection from type to set is called \<open>Rep_t\<close> + and its inverse \<open>Abs_t\<close>: An explicit @{keyword (HOL) "morphisms"} + specification allows to provide alternative names. The logical characterization of @{command typedef} uses the predicate of locale @{const type_definition} that is defined in Isabelle/HOL. Various - basic consequences of that are instantiated accordingly, re-using the - locale facts with names derived from the new type constructor. Thus the - generic theorem @{thm type_definition.Rep} is turned into the specific - \<open>Rep_t\<close>, for example. - - Theorems @{thm type_definition.Rep}, @{thm - type_definition.Rep_inverse}, and @{thm type_definition.Abs_inverse} - provide the most basic characterization as a corresponding - injection/surjection pair (in both directions). The derived rules - @{thm type_definition.Rep_inject} and @{thm + basic consequences of that are instantiated accordingly, re-using the locale + facts with names derived from the new type constructor. Thus the generic + theorem @{thm type_definition.Rep} is turned into the specific \<open>Rep_t\<close>, for + example. + + Theorems @{thm type_definition.Rep}, @{thm type_definition.Rep_inverse}, and + @{thm type_definition.Abs_inverse} provide the most basic characterization + as a corresponding injection/surjection pair (in both directions). The + derived rules @{thm type_definition.Rep_inject} and @{thm type_definition.Abs_inject} provide a more convenient version of - injectivity, suitable for automated proof tools (e.g.\ in - declarations involving @{attribute simp} or @{attribute iff}). - Furthermore, the rules @{thm type_definition.Rep_cases}~/ @{thm - type_definition.Rep_induct}, and @{thm type_definition.Abs_cases}~/ - @{thm type_definition.Abs_induct} provide alternative views on - surjectivity. These rules are already declared as set or type rules - for the generic @{method cases} and @{method induct} methods, + injectivity, suitable for automated proof tools (e.g.\ in declarations + involving @{attribute simp} or @{attribute iff}). Furthermore, the rules + @{thm type_definition.Rep_cases}~/ @{thm type_definition.Rep_induct}, and + @{thm type_definition.Abs_cases}~/ @{thm type_definition.Abs_induct} provide + alternative views on surjectivity. These rules are already declared as set + or type rules for the generic @{method cases} and @{method induct} methods, respectively. \<close> subsubsection \<open>Examples\<close> -text \<open>The following trivial example pulls a three-element type into -existence within the formal logical environment of Isabelle/HOL.\<close> +text \<open> + The following trivial example pulls a three-element type into existence + within the formal logical environment of Isabelle/HOL.\<close> (*<*)experiment begin(*>*) typedef three = "{(True, True), (True, False), (False, True)}" @@ -1210,37 +1170,38 @@ @@{command (HOL) functor} (@{syntax name} ':')? @{syntax term} \<close>} - \<^descr> @{command (HOL) "functor"}~\<open>prefix: m\<close> allows to prove and - register properties about the functorial structure of type constructors. - These properties then can be used by other packages to deal with those - type constructors in certain type constructions. Characteristic theorems - are noted in the current local theory. By default, they are prefixed with - the base name of the type constructor, an explicit prefix can be given + \<^descr> @{command (HOL) "functor"}~\<open>prefix: m\<close> allows to prove and register + properties about the functorial structure of type constructors. These + properties then can be used by other packages to deal with those type + constructors in certain type constructions. Characteristic theorems are + noted in the current local theory. By default, they are prefixed with the + base name of the type constructor, an explicit prefix can be given alternatively. - The given term \<open>m\<close> is considered as \<^emph>\<open>mapper\<close> for the - corresponding type constructor and must conform to the following type - pattern: + The given term \<open>m\<close> is considered as \<^emph>\<open>mapper\<close> for the corresponding type + constructor and must conform to the following type pattern: \begin{matharray}{lll} \<open>m\<close> & \<open>::\<close> & \<open>\<sigma>\<^sub>1 \<Rightarrow> \<dots> \<sigma>\<^sub>k \<Rightarrow> (\<^vec>\<alpha>\<^sub>n) t \<Rightarrow> (\<^vec>\<beta>\<^sub>n) t\<close> \\ \end{matharray} - where \<open>t\<close> is the type constructor, \<open>\<^vec>\<alpha>\<^sub>n\<close> - and \<open>\<^vec>\<beta>\<^sub>n\<close> are distinct type variables free in the local - theory and \<open>\<sigma>\<^sub>1\<close>, \ldots, \<open>\<sigma>\<^sub>k\<close> is a subsequence of \<open>\<alpha>\<^sub>1 \<Rightarrow> \<beta>\<^sub>1\<close>, \<open>\<beta>\<^sub>1 \<Rightarrow> \<alpha>\<^sub>1\<close>, \ldots, \<open>\<alpha>\<^sub>n \<Rightarrow> \<beta>\<^sub>n\<close>, \<open>\<beta>\<^sub>n \<Rightarrow> \<alpha>\<^sub>n\<close>. + where \<open>t\<close> is the type constructor, \<open>\<^vec>\<alpha>\<^sub>n\<close> and \<open>\<^vec>\<beta>\<^sub>n\<close> are + distinct type variables free in the local theory and \<open>\<sigma>\<^sub>1\<close>, \ldots, \<open>\<sigma>\<^sub>k\<close> is + a subsequence of \<open>\<alpha>\<^sub>1 \<Rightarrow> \<beta>\<^sub>1\<close>, \<open>\<beta>\<^sub>1 \<Rightarrow> \<alpha>\<^sub>1\<close>, \ldots, \<open>\<alpha>\<^sub>n \<Rightarrow> \<beta>\<^sub>n\<close>, \<open>\<beta>\<^sub>n \<Rightarrow> \<alpha>\<^sub>n\<close>. \<close> section \<open>Quotient types with lifting and transfer\<close> -text \<open>The quotient package defines a new quotient type given a raw type and - a partial equivalence relation (\secref{sec:quotient-type}). The package - also historically includes automation for transporting definitions and - theorems (\secref{sec:old-quotient}), but most of this automation was - superseded by the Lifting (\secref{sec:lifting}) and Transfer - (\secref{sec:transfer}) packages.\<close> +text \<open> + The quotient package defines a new quotient type given a raw type and a + partial equivalence relation (\secref{sec:quotient-type}). The package also + historically includes automation for transporting definitions and theorems + (\secref{sec:old-quotient}), but most of this automation was superseded by + the Lifting (\secref{sec:lifting}) and Transfer (\secref{sec:transfer}) + packages. +\<close> subsection \<open>Quotient type definition \label{sec:quotient-type}\<close> @@ -1262,21 +1223,23 @@ quot_parametric: @'parametric' @{syntax thm} \<close>} - \<^descr> @{command (HOL) "quotient_type"} defines a new quotient type \<open>\<tau>\<close>. The injection from a quotient type to a raw type is called \<open>rep_\<tau>\<close>, its inverse \<open>abs_\<tau>\<close> unless explicit @{keyword (HOL) - "morphisms"} specification provides alternative names. @{command (HOL) - "quotient_type"} requires the user to prove that the relation is an - equivalence relation (predicate \<open>equivp\<close>), unless the user specifies - explicitly \<open>partial\<close> in which case the obligation is \<open>part_equivp\<close>. A quotient defined with \<open>partial\<close> is weaker in the - sense that less things can be proved automatically. + \<^descr> @{command (HOL) "quotient_type"} defines a new quotient type \<open>\<tau>\<close>. The + injection from a quotient type to a raw type is called \<open>rep_\<tau>\<close>, its inverse + \<open>abs_\<tau>\<close> unless explicit @{keyword (HOL) "morphisms"} specification provides + alternative names. @{command (HOL) "quotient_type"} requires the user to + prove that the relation is an equivalence relation (predicate \<open>equivp\<close>), + unless the user specifies explicitly \<open>partial\<close> in which case the obligation + is \<open>part_equivp\<close>. A quotient defined with \<open>partial\<close> is weaker in the sense + that less things can be proved automatically. The command internally proves a Quotient theorem and sets up the Lifting - package by the command @{command (HOL) setup_lifting}. Thus the Lifting - and Transfer packages can be used also with quotient types defined by - @{command (HOL) "quotient_type"} without any extra set-up. The - parametricity theorem for the equivalence relation R can be provided as an - extra argument of the command and is passed to the corresponding internal - call of @{command (HOL) setup_lifting}. This theorem allows the Lifting - package to generate a stronger transfer rule for equality. + package by the command @{command (HOL) setup_lifting}. Thus the Lifting and + Transfer packages can be used also with quotient types defined by @{command + (HOL) "quotient_type"} without any extra set-up. The parametricity theorem + for the equivalence relation R can be provided as an extra argument of the + command and is passed to the corresponding internal call of @{command (HOL) + setup_lifting}. This theorem allows the Lifting package to generate a + stronger transfer rule for equality. \<close> @@ -1286,9 +1249,9 @@ The Lifting package allows users to lift terms of the raw type to the abstract type, which is a necessary step in building a library for an abstract type. Lifting defines a new constant by combining coercion - functions (@{term Abs} and @{term Rep}) with the raw term. It also proves - an appropriate transfer rule for the Transfer (\secref{sec:transfer}) - package and, if possible, an equation for the code generator. + functions (@{term Abs} and @{term Rep}) with the raw term. It also proves an + appropriate transfer rule for the Transfer (\secref{sec:transfer}) package + and, if possible, an equation for the code generator. The Lifting package provides two main commands: @{command (HOL) "setup_lifting"} for initializing the package to work with a new type, and @@ -1330,12 +1293,12 @@ @{syntax thm} (@{syntax thm} @{syntax thm})? \<close>} - \<^descr> @{command (HOL) "setup_lifting"} Sets up the Lifting package to work - with a user-defined type. The command supports two modes. - - \<^enum> The first one is a low-level mode when the user must provide as a - first argument of @{command (HOL) "setup_lifting"} a quotient theorem - @{term "Quotient R Abs Rep T"}. The package configures a transfer rule for + \<^descr> @{command (HOL) "setup_lifting"} Sets up the Lifting package to work with + a user-defined type. The command supports two modes. + + \<^enum> The first one is a low-level mode when the user must provide as a first + argument of @{command (HOL) "setup_lifting"} a quotient theorem @{term + "Quotient R Abs Rep T"}. The package configures a transfer rule for equality, a domain transfer rules and sets up the @{command_def (HOL) "lift_definition"} command to work with the abstract type. An optional theorem @{term "reflp R"}, which certifies that the equivalence relation R @@ -1344,8 +1307,8 @@ for @{term R} can be provided as a third argument. This allows the package to generate a stronger transfer rule for equality. - Users generally will not prove the \<open>Quotient\<close> theorem manually for - new types, as special commands exist to automate the process. + Users generally will not prove the \<open>Quotient\<close> theorem manually for new + types, as special commands exist to automate the process. \<^enum> When a new subtype is defined by @{command (HOL) typedef}, @{command (HOL) "lift_definition"} can be used in its second mode, where only the @@ -1355,166 +1318,162 @@ (HOL) setup_lifting} using its first mode. For quotients, the command @{command (HOL) quotient_type} can be used. The - command defines a new quotient type and similarly to the previous case, - the corresponding Quotient theorem is proved and registered by @{command - (HOL) setup_lifting}. + command defines a new quotient type and similarly to the previous case, the + corresponding Quotient theorem is proved and registered by @{command (HOL) + setup_lifting}. \<^medskip> - The command @{command (HOL) "setup_lifting"} also sets up the - code generator for the new type. Later on, when a new constant is defined - by @{command (HOL) "lift_definition"}, the Lifting package proves and - registers a code equation (if there is one) for the new constant. - - \<^descr> @{command (HOL) "lift_definition"} \<open>f :: \<tau>\<close> @{keyword (HOL) - "is"} \<open>t\<close> Defines a new function \<open>f\<close> with an abstract type - \<open>\<tau>\<close> in terms of a corresponding operation \<open>t\<close> on a - representation type. More formally, if \<open>t :: \<sigma>\<close>, then the command - builds a term \<open>F\<close> as a corresponding combination of abstraction - and representation functions such that \<open>F :: \<sigma> \<Rightarrow> \<tau>\<close> and defines - \<open>f \<equiv> F t\<close>. The term \<open>t\<close> does not have to be necessarily a - constant but it can be any term. - - The command opens a proof and the user must discharge a respectfulness - proof obligation. For a type copy, i.e.\ a typedef with \<open>UNIV\<close>, the - obligation is discharged automatically. The proof goal is presented in a - user-friendly, readable form. A respectfulness theorem in the standard - format \<open>f.rsp\<close> and a transfer rule \<open>f.transfer\<close> for the - Transfer package are generated by the package. - - The user can specify a parametricity theorems for \<open>t\<close> after the - keyword @{keyword "parametric"}, which allows the command to generate - parametric transfer rules for \<open>f\<close>. + The command @{command (HOL) "setup_lifting"} also sets up the code generator + for the new type. Later on, when a new constant is defined by @{command + (HOL) "lift_definition"}, the Lifting package proves and registers a code + equation (if there is one) for the new constant. + + \<^descr> @{command (HOL) "lift_definition"} \<open>f :: \<tau>\<close> @{keyword (HOL) "is"} \<open>t\<close> + Defines a new function \<open>f\<close> with an abstract type \<open>\<tau>\<close> in terms of a + corresponding operation \<open>t\<close> on a representation type. More formally, if \<open>t + :: \<sigma>\<close>, then the command builds a term \<open>F\<close> as a corresponding combination of + abstraction and representation functions such that \<open>F :: \<sigma> \<Rightarrow> \<tau>\<close> and defines + \<open>f \<equiv> F t\<close>. The term \<open>t\<close> does not have to be necessarily a constant but it + can be any term. + + The command opens a proof and the user must discharge a respectfulness proof + obligation. For a type copy, i.e.\ a typedef with \<open>UNIV\<close>, the obligation is + discharged automatically. The proof goal is presented in a user-friendly, + readable form. A respectfulness theorem in the standard format \<open>f.rsp\<close> and a + transfer rule \<open>f.transfer\<close> for the Transfer package are generated by the + package. + + The user can specify a parametricity theorems for \<open>t\<close> after the keyword + @{keyword "parametric"}, which allows the command to generate parametric + transfer rules for \<open>f\<close>. For each constant defined through trivial quotients (type copies or - subtypes) \<open>f.rep_eq\<close> is generated. The equation is a code - certificate that defines \<open>f\<close> using the representation function. - - For each constant \<open>f.abs_eq\<close> is generated. The equation is - unconditional for total quotients. The equation defines \<open>f\<close> using - the abstraction function. + subtypes) \<open>f.rep_eq\<close> is generated. The equation is a code certificate that + defines \<open>f\<close> using the representation function. + + For each constant \<open>f.abs_eq\<close> is generated. The equation is unconditional for + total quotients. The equation defines \<open>f\<close> using the abstraction function. \<^medskip> - Integration with [@{attribute code} abstract]: For subtypes - (e.g.\ corresponding to a datatype invariant, such as @{typ "'a dlist"}), - @{command (HOL) "lift_definition"} uses a code certificate theorem \<open>f.rep_eq\<close> as a code equation. Because of the limitation of the code - generator, \<open>f.rep_eq\<close> cannot be used as a code equation if the - subtype occurs inside the result type rather than at the top level (e.g.\ - function returning @{typ "'a dlist option"} vs. @{typ "'a dlist"}). + Integration with [@{attribute code} abstract]: For subtypes (e.g.\ + corresponding to a datatype invariant, such as @{typ "'a dlist"}), @{command + (HOL) "lift_definition"} uses a code certificate theorem \<open>f.rep_eq\<close> as a + code equation. Because of the limitation of the code generator, \<open>f.rep_eq\<close> + cannot be used as a code equation if the subtype occurs inside the result + type rather than at the top level (e.g.\ function returning @{typ "'a dlist + option"} vs. @{typ "'a dlist"}). In this case, an extension of @{command (HOL) "lift_definition"} can be - invoked by specifying the flag \<open>code_dt\<close>. This extension enables - code execution through series of internal type and lifting definitions if - the return type \<open>\<tau>\<close> meets the following inductive conditions: + invoked by specifying the flag \<open>code_dt\<close>. This extension enables code + execution through series of internal type and lifting definitions if the + return type \<open>\<tau>\<close> meets the following inductive conditions: \<^descr> \<open>\<tau>\<close> is a type variable - \<^descr> \<open>\<tau> = \<tau>\<^sub>1 \<dots> \<tau>\<^sub>n \<kappa>\<close>, - where \<open>\<kappa>\<close> is an abstract type constructor and \<open>\<tau>\<^sub>1 \<dots> \<tau>\<^sub>n\<close> - do not contain abstract types (i.e.\ @{typ "int dlist"} is allowed whereas - @{typ "int dlist dlist"} not) - - \<^descr> \<open>\<tau> = \<tau>\<^sub>1 \<dots> \<tau>\<^sub>n \<kappa>\<close>, \<open>\<kappa>\<close> is a type constructor that - was defined as a (co)datatype whose constructor argument types do not - contain either non-free datatypes or the function type. + \<^descr> \<open>\<tau> = \<tau>\<^sub>1 \<dots> \<tau>\<^sub>n \<kappa>\<close>, where \<open>\<kappa>\<close> is an abstract type constructor and \<open>\<tau>\<^sub>1 \<dots> + \<tau>\<^sub>n\<close> do not contain abstract types (i.e.\ @{typ "int dlist"} is allowed + whereas @{typ "int dlist dlist"} not) + + \<^descr> \<open>\<tau> = \<tau>\<^sub>1 \<dots> \<tau>\<^sub>n \<kappa>\<close>, \<open>\<kappa>\<close> is a type constructor that was defined as a + (co)datatype whose constructor argument types do not contain either + non-free datatypes or the function type. Integration with [@{attribute code} equation]: For total quotients, - @{command (HOL) "lift_definition"} uses \<open>f.abs_eq\<close> as a code - equation. - - \<^descr> @{command (HOL) lifting_forget} and @{command (HOL) lifting_update} - These two commands serve for storing and deleting the set-up of the - Lifting package and corresponding transfer rules defined by this package. - This is useful for hiding of type construction details of an abstract type - when the construction is finished but it still allows additions to this - construction when this is later necessary. - - Whenever the Lifting package is set up with a new abstract type \<open>\<tau>\<close> by @{command_def (HOL) "lift_definition"}, the package defines a new - bundle that is called \<open>\<tau>.lifting\<close>. This bundle already includes - set-up for the Lifting package. The new transfer rules introduced by - @{command (HOL) "lift_definition"} can be stored in the bundle by the - command @{command (HOL) "lifting_update"} \<open>\<tau>.lifting\<close>. - - The command @{command (HOL) "lifting_forget"} \<open>\<tau>.lifting\<close> deletes - set-up of the Lifting package for \<open>\<tau>\<close> and deletes all the transfer - rules that were introduced by @{command (HOL) "lift_definition"} using - \<open>\<tau>\<close> as an abstract type. + @{command (HOL) "lift_definition"} uses \<open>f.abs_eq\<close> as a code equation. + + \<^descr> @{command (HOL) lifting_forget} and @{command (HOL) lifting_update} These + two commands serve for storing and deleting the set-up of the Lifting + package and corresponding transfer rules defined by this package. This is + useful for hiding of type construction details of an abstract type when the + construction is finished but it still allows additions to this construction + when this is later necessary. + + Whenever the Lifting package is set up with a new abstract type \<open>\<tau>\<close> by + @{command_def (HOL) "lift_definition"}, the package defines a new bundle + that is called \<open>\<tau>.lifting\<close>. This bundle already includes set-up for the + Lifting package. The new transfer rules introduced by @{command (HOL) + "lift_definition"} can be stored in the bundle by the command @{command + (HOL) "lifting_update"} \<open>\<tau>.lifting\<close>. + + The command @{command (HOL) "lifting_forget"} \<open>\<tau>.lifting\<close> deletes set-up of + the Lifting package for \<open>\<tau>\<close> and deletes all the transfer rules that were + introduced by @{command (HOL) "lift_definition"} using \<open>\<tau>\<close> as an abstract + type. The stored set-up in a bundle can be reintroduced by the Isar commands for including a bundle (@{command "include"}, @{keyword "includes"} and @{command "including"}). - \<^descr> @{command (HOL) "print_quot_maps"} prints stored quotient map - theorems. + \<^descr> @{command (HOL) "print_quot_maps"} prints stored quotient map theorems. \<^descr> @{command (HOL) "print_quotients"} prints stored quotient theorems. - \<^descr> @{attribute (HOL) quot_map} registers a quotient map theorem, a - theorem showing how to ``lift'' quotients over type constructors. E.g.\ - @{term "Quotient R Abs Rep T \<Longrightarrow> Quotient (rel_set R) (image Abs) (image - Rep) (rel_set T)"}. For examples see @{file "~~/src/HOL/Lifting_Set.thy"} - or @{file "~~/src/HOL/Lifting.thy"}. This property is proved automatically - if the involved type is BNF without dead variables. - - \<^descr> @{attribute (HOL) relator_eq_onp} registers a theorem that shows - that a relator applied to an equality restricted by a predicate @{term P} - (i.e.\ @{term "eq_onp P"}) is equal to a predicator applied to the @{term - P}. The combinator @{const eq_onp} is used for internal encoding of proper - subtypes. Such theorems allows the package to hide \<open>eq_onp\<close> from a - user in a user-readable form of a respectfulness theorem. For examples see - @{file "~~/src/HOL/Lifting_Set.thy"} or @{file "~~/src/HOL/Lifting.thy"}. - This property is proved automatically if the involved type is BNF without - dead variables. + \<^descr> @{attribute (HOL) quot_map} registers a quotient map theorem, a theorem + showing how to ``lift'' quotients over type constructors. E.g.\ @{term + "Quotient R Abs Rep T \<Longrightarrow> Quotient (rel_set R) (image Abs) (image Rep) + (rel_set T)"}. For examples see @{file "~~/src/HOL/Lifting_Set.thy"} or + @{file "~~/src/HOL/Lifting.thy"}. This property is proved automatically if + the involved type is BNF without dead variables. + + \<^descr> @{attribute (HOL) relator_eq_onp} registers a theorem that shows that a + relator applied to an equality restricted by a predicate @{term P} (i.e.\ + @{term "eq_onp P"}) is equal to a predicator applied to the @{term P}. The + combinator @{const eq_onp} is used for internal encoding of proper subtypes. + Such theorems allows the package to hide \<open>eq_onp\<close> from a user in a + user-readable form of a respectfulness theorem. For examples see @{file + "~~/src/HOL/Lifting_Set.thy"} or @{file "~~/src/HOL/Lifting.thy"}. This + property is proved automatically if the involved type is BNF without dead + variables. \<^descr> @{attribute (HOL) "relator_mono"} registers a property describing a monotonicity of a relator. E.g.\ @{prop "A \<le> B \<Longrightarrow> rel_set A \<le> rel_set B"}. This property is needed for proving a stronger transfer rule in - @{command_def (HOL) "lift_definition"} when a parametricity theorem for - the raw term is specified and also for the reflexivity prover. For - examples see @{file "~~/src/HOL/Lifting_Set.thy"} or @{file - "~~/src/HOL/Lifting.thy"}. This property is proved automatically if the - involved type is BNF without dead variables. + @{command_def (HOL) "lift_definition"} when a parametricity theorem for the + raw term is specified and also for the reflexivity prover. For examples see + @{file "~~/src/HOL/Lifting_Set.thy"} or @{file "~~/src/HOL/Lifting.thy"}. + This property is proved automatically if the involved type is BNF without + dead variables. \<^descr> @{attribute (HOL) "relator_distr"} registers a property describing a - distributivity of the relation composition and a relator. E.g.\ \<open>rel_set R \<circ>\<circ> rel_set S = rel_set (R \<circ>\<circ> S)\<close>. This property is needed for - proving a stronger transfer rule in @{command_def (HOL) "lift_definition"} - when a parametricity theorem for the raw term is specified. When this - equality does not hold unconditionally (e.g.\ for the function type), the - user can specified each direction separately and also register multiple - theorems with different set of assumptions. This attribute can be used - only after the monotonicity property was already registered by @{attribute - (HOL) "relator_mono"}. For examples see @{file - "~~/src/HOL/Lifting_Set.thy"} or @{file "~~/src/HOL/Lifting.thy"}. This - property is proved automatically if the involved type is BNF without dead - variables. - - \<^descr> @{attribute (HOL) quot_del} deletes a corresponding Quotient theorem - from the Lifting infrastructure and thus de-register the corresponding - quotient. This effectively causes that @{command (HOL) lift_definition} - will not do any lifting for the corresponding type. This attribute is - rather used for low-level manipulation with set-up of the Lifting package - because @{command (HOL) lifting_forget} is preferred for normal usage. - - \<^descr> @{attribute (HOL) lifting_restore} \<open>Quotient_thm pcr_def - pcr_cr_eq_thm\<close> registers the Quotient theorem \<open>Quotient_thm\<close> in the - Lifting infrastructure and thus sets up lifting for an abstract type - \<open>\<tau>\<close> (that is defined by \<open>Quotient_thm\<close>). Optional theorems - \<open>pcr_def\<close> and \<open>pcr_cr_eq_thm\<close> can be specified to register the - parametrized correspondence relation for \<open>\<tau>\<close>. E.g.\ for @{typ "'a - dlist"}, \<open>pcr_def\<close> is \<open>pcr_dlist A \<equiv> list_all2 A \<circ>\<circ> - cr_dlist\<close> and \<open>pcr_cr_eq_thm\<close> is \<open>pcr_dlist (op =) = (op - =)\<close>. This attribute is rather used for low-level manipulation with set-up - of the Lifting package because using of the bundle \<open>\<tau>.lifting\<close> - together with the commands @{command (HOL) lifting_forget} and @{command - (HOL) lifting_update} is preferred for normal usage. - - \<^descr> Integration with the BNF package @{cite "isabelle-datatypes"}: As - already mentioned, the theorems that are registered by the following - attributes are proved and registered automatically if the involved type is - BNF without dead variables: @{attribute (HOL) quot_map}, @{attribute (HOL) - relator_eq_onp}, @{attribute (HOL) "relator_mono"}, @{attribute (HOL) - "relator_distr"}. Also the definition of a relator and predicator is - provided automatically. Moreover, if the BNF represents a datatype, - simplification rules for a predicator are again proved automatically. + distributivity of the relation composition and a relator. E.g.\ \<open>rel_set R + \<circ>\<circ> rel_set S = rel_set (R \<circ>\<circ> S)\<close>. This property is needed for proving a + stronger transfer rule in @{command_def (HOL) "lift_definition"} when a + parametricity theorem for the raw term is specified. When this equality does + not hold unconditionally (e.g.\ for the function type), the user can + specified each direction separately and also register multiple theorems with + different set of assumptions. This attribute can be used only after the + monotonicity property was already registered by @{attribute (HOL) + "relator_mono"}. For examples see @{file "~~/src/HOL/Lifting_Set.thy"} or + @{file "~~/src/HOL/Lifting.thy"}. This property is proved automatically if + the involved type is BNF without dead variables. + + \<^descr> @{attribute (HOL) quot_del} deletes a corresponding Quotient theorem from + the Lifting infrastructure and thus de-register the corresponding quotient. + This effectively causes that @{command (HOL) lift_definition} will not do + any lifting for the corresponding type. This attribute is rather used for + low-level manipulation with set-up of the Lifting package because @{command + (HOL) lifting_forget} is preferred for normal usage. + + \<^descr> @{attribute (HOL) lifting_restore} \<open>Quotient_thm pcr_def pcr_cr_eq_thm\<close> + registers the Quotient theorem \<open>Quotient_thm\<close> in the Lifting infrastructure + and thus sets up lifting for an abstract type \<open>\<tau>\<close> (that is defined by + \<open>Quotient_thm\<close>). Optional theorems \<open>pcr_def\<close> and \<open>pcr_cr_eq_thm\<close> can be + specified to register the parametrized correspondence relation for \<open>\<tau>\<close>. + E.g.\ for @{typ "'a dlist"}, \<open>pcr_def\<close> is \<open>pcr_dlist A \<equiv> list_all2 A \<circ>\<circ> + cr_dlist\<close> and \<open>pcr_cr_eq_thm\<close> is \<open>pcr_dlist (op =) = (op =)\<close>. This attribute + is rather used for low-level manipulation with set-up of the Lifting package + because using of the bundle \<open>\<tau>.lifting\<close> together with the commands @{command + (HOL) lifting_forget} and @{command (HOL) lifting_update} is preferred for + normal usage. + + \<^descr> Integration with the BNF package @{cite "isabelle-datatypes"}: As already + mentioned, the theorems that are registered by the following attributes are + proved and registered automatically if the involved type is BNF without dead + variables: @{attribute (HOL) quot_map}, @{attribute (HOL) relator_eq_onp}, + @{attribute (HOL) "relator_mono"}, @{attribute (HOL) "relator_distr"}. Also + the definition of a relator and predicator is provided automatically. + Moreover, if the BNF represents a datatype, simplification rules for a + predicator are again proved automatically. \<close> @@ -1538,47 +1497,46 @@ @{attribute_def (HOL) "relator_domain"} & : & \<open>attribute\<close> \\ \end{matharray} - \<^descr> @{method (HOL) "transfer"} method replaces the current subgoal with - a logically equivalent one that uses different types and constants. The + \<^descr> @{method (HOL) "transfer"} method replaces the current subgoal with a + logically equivalent one that uses different types and constants. The replacement of types and constants is guided by the database of transfer rules. Goals are generalized over all free variables by default; this is necessary for variables whose types change, but can be overridden for specific variables with e.g. \<open>transfer fixing: x y z\<close>. - \<^descr> @{method (HOL) "transfer'"} is a variant of @{method (HOL) transfer} - that allows replacing a subgoal with one that is logically stronger - (rather than equivalent). For example, a subgoal involving equality on a - quotient type could be replaced with a subgoal involving equality (instead - of the corresponding equivalence relation) on the underlying raw type. - - \<^descr> @{method (HOL) "transfer_prover"} method assists with proving a - transfer rule for a new constant, provided the constant is defined in - terms of other constants that already have transfer rules. It should be - applied after unfolding the constant definitions. + \<^descr> @{method (HOL) "transfer'"} is a variant of @{method (HOL) transfer} that + allows replacing a subgoal with one that is logically stronger (rather than + equivalent). For example, a subgoal involving equality on a quotient type + could be replaced with a subgoal involving equality (instead of the + corresponding equivalence relation) on the underlying raw type. + + \<^descr> @{method (HOL) "transfer_prover"} method assists with proving a transfer + rule for a new constant, provided the constant is defined in terms of other + constants that already have transfer rules. It should be applied after + unfolding the constant definitions. \<^descr> @{method (HOL) "transfer_start"}, @{method (HOL) "transfer_step"}, - @{method (HOL) "transfer_end"}, @{method (HOL) "transfer_prover_start"} - and @{method (HOL) "transfer_prover_end"} methods are meant to be used - for debugging of @{method (HOL) "transfer"} and @{method (HOL) "transfer_prover"}, - which we can decompose as follows: - @{method (HOL) "transfer"} = (@{method (HOL) "transfer_start"}, - @{method (HOL) "transfer_step"}+, @{method (HOL) "transfer_end"}) and - @{method (HOL) "transfer_prover"} = (@{method (HOL) "transfer_prover_start"}, - @{method (HOL) "transfer_step"}+, @{method (HOL) "transfer_prover_end"}). - For usage examples see @{file "~~/src/HOL/ex/Transfer_Debug.thy"} - - \<^descr> @{attribute (HOL) "untransferred"} proves the same equivalent - theorem as @{method (HOL) "transfer"} internally does. - - \<^descr> @{attribute (HOL) Transfer.transferred} works in the opposite - direction than @{method (HOL) "transfer'"}. E.g.\ given the transfer - relation \<open>ZN x n \<equiv> (x = int n)\<close>, corresponding transfer rules and - the theorem \<open>\<forall>x::int \<in> {0..}. x < x + 1\<close>, the attribute would - prove \<open>\<forall>n::nat. n < n + 1\<close>. The attribute is still in experimental - phase of development. - - \<^descr> @{attribute (HOL) "transfer_rule"} attribute maintains a collection - of transfer rules, which relate constants at two different types. Typical + @{method (HOL) "transfer_end"}, @{method (HOL) "transfer_prover_start"} and + @{method (HOL) "transfer_prover_end"} methods are meant to be used for + debugging of @{method (HOL) "transfer"} and @{method (HOL) + "transfer_prover"}, which we can decompose as follows: @{method (HOL) + "transfer"} = (@{method (HOL) "transfer_start"}, @{method (HOL) + "transfer_step"}+, @{method (HOL) "transfer_end"}) and @{method (HOL) + "transfer_prover"} = (@{method (HOL) "transfer_prover_start"}, @{method + (HOL) "transfer_step"}+, @{method (HOL) "transfer_prover_end"}). For usage + examples see @{file "~~/src/HOL/ex/Transfer_Debug.thy"} + + \<^descr> @{attribute (HOL) "untransferred"} proves the same equivalent theorem as + @{method (HOL) "transfer"} internally does. + + \<^descr> @{attribute (HOL) Transfer.transferred} works in the opposite direction + than @{method (HOL) "transfer'"}. E.g.\ given the transfer relation \<open>ZN x n + \<equiv> (x = int n)\<close>, corresponding transfer rules and the theorem \<open>\<forall>x::int \<in> + {0..}. x < x + 1\<close>, the attribute would prove \<open>\<forall>n::nat. n < n + 1\<close>. The + attribute is still in experimental phase of development. + + \<^descr> @{attribute (HOL) "transfer_rule"} attribute maintains a collection of + transfer rules, which relate constants at two different types. Typical transfer rules may relate different type instances of the same polymorphic constant, or they may relate an operation on a raw type to a corresponding operation on an abstract type (quotient or subtype). For example: @@ -1592,17 +1550,17 @@ \<open>bi_unique A \<Longrightarrow> (list_all2 A ===> op =) distinct distinct\<close> \\ \<open>\<lbrakk>bi_unique A; bi_unique B\<rbrakk> \<Longrightarrow> bi_unique (rel_prod A B)\<close> - Preservation of predicates on relations (\<open>bi_unique, bi_total, - right_unique, right_total, left_unique, left_total\<close>) with the respect to - a relator is proved automatically if the involved type is BNF @{cite + Preservation of predicates on relations (\<open>bi_unique, bi_total, right_unique, + right_total, left_unique, left_total\<close>) with the respect to a relator is + proved automatically if the involved type is BNF @{cite "isabelle-datatypes"} without dead variables. - \<^descr> @{attribute (HOL) "transfer_domain_rule"} attribute maintains a - collection of rules, which specify a domain of a transfer relation by a - predicate. E.g.\ given the transfer relation \<open>ZN x n \<equiv> (x = int - n)\<close>, one can register the following transfer domain rule: \<open>Domainp - ZN = (\<lambda>x. x \<ge> 0)\<close>. The rules allow the package to produce more readable - transferred goals, e.g.\ when quantifiers are transferred. + \<^descr> @{attribute (HOL) "transfer_domain_rule"} attribute maintains a collection + of rules, which specify a domain of a transfer relation by a predicate. + E.g.\ given the transfer relation \<open>ZN x n \<equiv> (x = int n)\<close>, one can register + the following transfer domain rule: \<open>Domainp ZN = (\<lambda>x. x \<ge> 0)\<close>. The rules + allow the package to produce more readable transferred goals, e.g.\ when + quantifiers are transferred. \<^descr> @{attribute (HOL) relator_eq} attribute collects identity laws for relators of various type constructors, e.g. @{term "rel_set (op =) = (op @@ -1660,8 +1618,8 @@ @@{method (HOL) lifting_setup} @{syntax thms}? \<close>} - \<^descr> @{command (HOL) "quotient_definition"} defines a constant on the - quotient type. + \<^descr> @{command (HOL) "quotient_definition"} defines a constant on the quotient + type. \<^descr> @{command (HOL) "print_quotmapsQ3"} prints quotient map functions. @@ -1669,42 +1627,42 @@ \<^descr> @{command (HOL) "print_quotconsts"} prints quotient constants. - \<^descr> @{method (HOL) "lifting"} and @{method (HOL) "lifting_setup"} - methods match the current goal with the given raw theorem to be lifted - producing three new subgoals: regularization, injection and cleaning - subgoals. @{method (HOL) "lifting"} tries to apply the heuristics for - automatically solving these three subgoals and leaves only the subgoals - unsolved by the heuristics to the user as opposed to @{method (HOL) - "lifting_setup"} which leaves the three subgoals unsolved. - - \<^descr> @{method (HOL) "descending"} and @{method (HOL) "descending_setup"} - try to guess a raw statement that would lift to the current subgoal. Such - statement is assumed as a new subgoal and @{method (HOL) "descending"} - continues in the same way as @{method (HOL) "lifting"} does. @{method - (HOL) "descending"} tries to solve the arising regularization, injection - and cleaning subgoals with the analogous method @{method (HOL) - "descending_setup"} which leaves the four unsolved subgoals. - - \<^descr> @{method (HOL) "partiality_descending"} finds the regularized - theorem that would lift to the current subgoal, lifts it and leaves as a - subgoal. This method can be used with partial equivalence quotients where - the non regularized statements would not be true. @{method (HOL) + \<^descr> @{method (HOL) "lifting"} and @{method (HOL) "lifting_setup"} methods + match the current goal with the given raw theorem to be lifted producing + three new subgoals: regularization, injection and cleaning subgoals. + @{method (HOL) "lifting"} tries to apply the heuristics for automatically + solving these three subgoals and leaves only the subgoals unsolved by the + heuristics to the user as opposed to @{method (HOL) "lifting_setup"} which + leaves the three subgoals unsolved. + + \<^descr> @{method (HOL) "descending"} and @{method (HOL) "descending_setup"} try to + guess a raw statement that would lift to the current subgoal. Such statement + is assumed as a new subgoal and @{method (HOL) "descending"} continues in + the same way as @{method (HOL) "lifting"} does. @{method (HOL) "descending"} + tries to solve the arising regularization, injection and cleaning subgoals + with the analogous method @{method (HOL) "descending_setup"} which leaves + the four unsolved subgoals. + + \<^descr> @{method (HOL) "partiality_descending"} finds the regularized theorem that + would lift to the current subgoal, lifts it and leaves as a subgoal. This + method can be used with partial equivalence quotients where the non + regularized statements would not be true. @{method (HOL) "partiality_descending_setup"} leaves the injection and cleaning subgoals unchanged. - \<^descr> @{method (HOL) "regularize"} applies the regularization heuristics - to the current subgoal. + \<^descr> @{method (HOL) "regularize"} applies the regularization heuristics to the + current subgoal. \<^descr> @{method (HOL) "injection"} applies the injection heuristics to the current goal using the stored quotient respectfulness theorems. - \<^descr> @{method (HOL) "cleaning"} applies the injection cleaning heuristics - to the current subgoal using the stored quotient preservation theorems. - - \<^descr> @{attribute (HOL) quot_lifted} attribute tries to automatically - transport the theorem to the quotient type. The attribute uses all the - defined quotients types and quotient constants often producing undesired - results or theorems that cannot be lifted. + \<^descr> @{method (HOL) "cleaning"} applies the injection cleaning heuristics to + the current subgoal using the stored quotient preservation theorems. + + \<^descr> @{attribute (HOL) quot_lifted} attribute tries to automatically transport + the theorem to the quotient type. The attribute uses all the defined + quotients types and quotient constants often producing undesired results or + theorems that cannot be lifted. \<^descr> @{attribute (HOL) quot_respect} and @{attribute (HOL) quot_preserve} attributes declare a theorem as a respectfulness and preservation theorem @@ -1712,14 +1670,14 @@ @{method (HOL) "injection"} and @{method (HOL) "cleaning"} methods respectively. - \<^descr> @{attribute (HOL) quot_thm} declares that a certain theorem is a - quotient extension theorem. Quotient extension theorems allow for - quotienting inside container types. Given a polymorphic type that serves - as a container, a map function defined for this container using @{command - (HOL) "functor"} and a relation map defined for for the container type, - the quotient extension theorem should be @{term "Quotient3 R Abs Rep \<Longrightarrow> - Quotient3 (rel_map R) (map Abs) (map Rep)"}. Quotient extension theorems - are stored in a database and are used all the steps of lifting theorems. + \<^descr> @{attribute (HOL) quot_thm} declares that a certain theorem is a quotient + extension theorem. Quotient extension theorems allow for quotienting inside + container types. Given a polymorphic type that serves as a container, a map + function defined for this container using @{command (HOL) "functor"} and a + relation map defined for for the container type, the quotient extension + theorem should be @{term "Quotient3 R Abs Rep \<Longrightarrow> Quotient3 (rel_map R) (map + Abs) (map Rep)"}. Quotient extension theorems are stored in a database and + are used all the steps of lifting theorems. \<close> @@ -1728,9 +1686,9 @@ section \<open>Proving propositions\<close> text \<open> - In addition to the standard proof methods, a number of diagnosis - tools search for proofs and provide an Isar proof snippet on success. - These tools are available via the following commands. + In addition to the standard proof methods, a number of diagnosis tools + search for proofs and provide an Isar proof snippet on success. These tools + are available via the following commands. \begin{matharray}{rcl} @{command_def (HOL) "solve_direct"}\<open>\<^sup>*\<close> & : & \<open>proof \<rightarrow>\<close> \\ @@ -1758,25 +1716,23 @@ facts: '(' ( ( ( ( 'add' | 'del' ) ':' ) ? @{syntax thms} ) + ) ? ')' \<close>} % FIXME check args "value" - \<^descr> @{command (HOL) "solve_direct"} checks whether the current - subgoals can be solved directly by an existing theorem. Duplicate - lemmas can be detected in this way. - - \<^descr> @{command (HOL) "try0"} attempts to prove a subgoal - using a combination of standard proof methods (@{method auto}, - @{method simp}, @{method blast}, etc.). Additional facts supplied - via \<open>simp:\<close>, \<open>intro:\<close>, \<open>elim:\<close>, and \<open>dest:\<close> are passed to the appropriate proof methods. - - \<^descr> @{command (HOL) "try"} attempts to prove or disprove a subgoal - using a combination of provers and disprovers (@{command (HOL) - "solve_direct"}, @{command (HOL) "quickcheck"}, @{command (HOL) - "try0"}, @{command (HOL) "sledgehammer"}, @{command (HOL) - "nitpick"}). - - \<^descr> @{command (HOL) "sledgehammer"} attempts to prove a subgoal - using external automatic provers (resolution provers and SMT - solvers). See the Sledgehammer manual @{cite "isabelle-sledgehammer"} - for details. + \<^descr> @{command (HOL) "solve_direct"} checks whether the current subgoals can be + solved directly by an existing theorem. Duplicate lemmas can be detected in + this way. + + \<^descr> @{command (HOL) "try0"} attempts to prove a subgoal using a combination of + standard proof methods (@{method auto}, @{method simp}, @{method blast}, + etc.). Additional facts supplied via \<open>simp:\<close>, \<open>intro:\<close>, \<open>elim:\<close>, and \<open>dest:\<close> + are passed to the appropriate proof methods. + + \<^descr> @{command (HOL) "try"} attempts to prove or disprove a subgoal using a + combination of provers and disprovers (@{command (HOL) "solve_direct"}, + @{command (HOL) "quickcheck"}, @{command (HOL) "try0"}, @{command (HOL) + "sledgehammer"}, @{command (HOL) "nitpick"}). + + \<^descr> @{command (HOL) "sledgehammer"} attempts to prove a subgoal using external + automatic provers (resolution provers and SMT solvers). See the Sledgehammer + manual @{cite "isabelle-sledgehammer"} for details. \<^descr> @{command (HOL) "sledgehammer_params"} changes @{command (HOL) "sledgehammer"} configuration options persistently. @@ -1786,9 +1742,9 @@ section \<open>Checking and refuting propositions\<close> text \<open> - Identifying incorrect propositions usually involves evaluation of - particular assignments and systematic counterexample search. This - is supported by the following commands. + Identifying incorrect propositions usually involves evaluation of particular + assignments and systematic counterexample search. This is supported by the + following commands. \begin{matharray}{rcl} @{command_def (HOL) "value"}\<open>\<^sup>*\<close> & : & \<open>context \<rightarrow>\<close> \\ @@ -1827,130 +1783,121 @@ args: ( @{syntax name} '=' value + ',' ) \<close>} % FIXME check "value" - \<^descr> @{command (HOL) "value"}~\<open>t\<close> evaluates and prints a - term; optionally \<open>modes\<close> can be specified, which are appended - to the current print mode; see \secref{sec:print-modes}. - Evaluation is tried first using ML, falling - back to normalization by evaluation if this fails. - Alternatively a specific evaluator can be selected using square - brackets; typical evaluators use the current set of code equations - to normalize and include \<open>simp\<close> for fully symbolic evaluation - using the simplifier, \<open>nbe\<close> for \<^emph>\<open>normalization by + \<^descr> @{command (HOL) "value"}~\<open>t\<close> evaluates and prints a term; optionally + \<open>modes\<close> can be specified, which are appended to the current print mode; see + \secref{sec:print-modes}. Evaluation is tried first using ML, falling back + to normalization by evaluation if this fails. Alternatively a specific + evaluator can be selected using square brackets; typical evaluators use the + current set of code equations to normalize and include \<open>simp\<close> for fully + symbolic evaluation using the simplifier, \<open>nbe\<close> for \<^emph>\<open>normalization by evaluation\<close> and \<^emph>\<open>code\<close> for code generation in SML. - \<^descr> @{command (HOL) "values"}~\<open>t\<close> enumerates a set - comprehension by evaluation and prints its values up to the given - number of solutions; optionally \<open>modes\<close> can be specified, - which are appended to the current print mode; see + \<^descr> @{command (HOL) "values"}~\<open>t\<close> enumerates a set comprehension by evaluation + and prints its values up to the given number of solutions; optionally + \<open>modes\<close> can be specified, which are appended to the current print mode; see \secref{sec:print-modes}. - \<^descr> @{command (HOL) "quickcheck"} tests the current goal for - counterexamples using a series of assignments for its free - variables; by default the first subgoal is tested, an other can be - selected explicitly using an optional goal index. Assignments can - be chosen exhausting the search space up to a given size, or using a - fixed number of random assignments in the search space, or exploring - the search space symbolically using narrowing. By default, - quickcheck uses exhaustive testing. A number of configuration + \<^descr> @{command (HOL) "quickcheck"} tests the current goal for counterexamples + using a series of assignments for its free variables; by default the first + subgoal is tested, an other can be selected explicitly using an optional + goal index. Assignments can be chosen exhausting the search space up to a + given size, or using a fixed number of random assignments in the search + space, or exploring the search space symbolically using narrowing. By + default, quickcheck uses exhaustive testing. A number of configuration options are supported for @{command (HOL) "quickcheck"}, notably: - \<^descr>[\<open>tester\<close>] specifies which testing approach to apply. - There are three testers, \<open>exhaustive\<close>, \<open>random\<close>, and - \<open>narrowing\<close>. An unknown configuration option is treated as - an argument to tester, making \<open>tester =\<close> optional. When - multiple testers are given, these are applied in parallel. If no - tester is specified, quickcheck uses the testers that are set - active, i.e.\ configurations @{attribute - quickcheck_exhaustive_active}, @{attribute - quickcheck_random_active}, @{attribute + \<^descr>[\<open>tester\<close>] specifies which testing approach to apply. There are three + testers, \<open>exhaustive\<close>, \<open>random\<close>, and \<open>narrowing\<close>. An unknown configuration + option is treated as an argument to tester, making \<open>tester =\<close> optional. + When multiple testers are given, these are applied in parallel. If no + tester is specified, quickcheck uses the testers that are set active, + i.e.\ configurations @{attribute quickcheck_exhaustive_active}, + @{attribute quickcheck_random_active}, @{attribute quickcheck_narrowing_active} are set to true. - \<^descr>[\<open>size\<close>] specifies the maximum size of the search space - for assignment values. - - \<^descr>[\<open>genuine_only\<close>] sets quickcheck only to return genuine - counterexample, but not potentially spurious counterexamples due - to underspecified functions. - - \<^descr>[\<open>abort_potential\<close>] sets quickcheck to abort once it - found a potentially spurious counterexample and to not continue - to search for a further genuine counterexample. - For this option to be effective, the \<open>genuine_only\<close> option - must be set to false. - - \<^descr>[\<open>eval\<close>] takes a term or a list of terms and evaluates - these terms under the variable assignment found by quickcheck. - This option is currently only supported by the default - (exhaustive) tester. - - \<^descr>[\<open>iterations\<close>] sets how many sets of assignments are - generated for each particular size. - - \<^descr>[\<open>no_assms\<close>] specifies whether assumptions in - structured proofs should be ignored. - - \<^descr>[\<open>locale\<close>] specifies how to process conjectures in - a locale context, i.e.\ they can be interpreted or expanded. - The option is a whitespace-separated list of the two words - \<open>interpret\<close> and \<open>expand\<close>. The list determines the - order they are employed. The default setting is to first use - interpretations and then test the expanded conjecture. - The option is only provided as attribute declaration, but not - as parameter to the command. + \<^descr>[\<open>size\<close>] specifies the maximum size of the search space for assignment + values. + + \<^descr>[\<open>genuine_only\<close>] sets quickcheck only to return genuine counterexample, + but not potentially spurious counterexamples due to underspecified + functions. + + \<^descr>[\<open>abort_potential\<close>] sets quickcheck to abort once it found a potentially + spurious counterexample and to not continue to search for a further + genuine counterexample. For this option to be effective, the + \<open>genuine_only\<close> option must be set to false. + + \<^descr>[\<open>eval\<close>] takes a term or a list of terms and evaluates these terms under + the variable assignment found by quickcheck. This option is currently only + supported by the default (exhaustive) tester. + + \<^descr>[\<open>iterations\<close>] sets how many sets of assignments are generated for each + particular size. + + \<^descr>[\<open>no_assms\<close>] specifies whether assumptions in structured proofs should be + ignored. + + \<^descr>[\<open>locale\<close>] specifies how to process conjectures in a locale context, + i.e.\ they can be interpreted or expanded. The option is a + whitespace-separated list of the two words \<open>interpret\<close> and \<open>expand\<close>. The + list determines the order they are employed. The default setting is to + first use interpretations and then test the expanded conjecture. The + option is only provided as attribute declaration, but not as parameter to + the command. \<^descr>[\<open>timeout\<close>] sets the time limit in seconds. - \<^descr>[\<open>default_type\<close>] sets the type(s) generally used to - instantiate type variables. - - \<^descr>[\<open>report\<close>] if set quickcheck reports how many tests - fulfilled the preconditions. - - \<^descr>[\<open>use_subtype\<close>] if set quickcheck automatically lifts - conjectures to registered subtypes if possible, and tests the - lifted conjecture. - - \<^descr>[\<open>quiet\<close>] if set quickcheck does not output anything - while testing. - - \<^descr>[\<open>verbose\<close>] if set quickcheck informs about the current - size and cardinality while testing. - - \<^descr>[\<open>expect\<close>] can be used to check if the user's - expectation was met (\<open>no_expectation\<close>, \<open>no_counterexample\<close>, or \<open>counterexample\<close>). + \<^descr>[\<open>default_type\<close>] sets the type(s) generally used to instantiate type + variables. + + \<^descr>[\<open>report\<close>] if set quickcheck reports how many tests fulfilled the + preconditions. + + \<^descr>[\<open>use_subtype\<close>] if set quickcheck automatically lifts conjectures to + registered subtypes if possible, and tests the lifted conjecture. + + \<^descr>[\<open>quiet\<close>] if set quickcheck does not output anything while testing. + + \<^descr>[\<open>verbose\<close>] if set quickcheck informs about the current size and + cardinality while testing. + + \<^descr>[\<open>expect\<close>] can be used to check if the user's expectation was met + (\<open>no_expectation\<close>, \<open>no_counterexample\<close>, or \<open>counterexample\<close>). These option can be given within square brackets. Using the following type classes, the testers generate values and convert them back into Isabelle terms for displaying counterexamples. - \<^descr>[\<open>exhaustive\<close>] The parameters of the type classes @{class exhaustive} - and @{class full_exhaustive} implement the testing. They take a - testing function as a parameter, which takes a value of type @{typ "'a"} - and optionally produces a counterexample, and a size parameter for the test values. - In @{class full_exhaustive}, the testing function parameter additionally - expects a lazy term reconstruction in the type @{typ Code_Evaluation.term} - of the tested value. + \<^descr>[\<open>exhaustive\<close>] The parameters of the type classes @{class exhaustive} and + @{class full_exhaustive} implement the testing. They take a testing + function as a parameter, which takes a value of type @{typ "'a"} and + optionally produces a counterexample, and a size parameter for the test + values. In @{class full_exhaustive}, the testing function parameter + additionally expects a lazy term reconstruction in the type @{typ + Code_Evaluation.term} of the tested value. The canonical implementation for \<open>exhaustive\<close> testers calls the given - testing function on all values up to the given size and stops as soon - as a counterexample is found. - - \<^descr>[\<open>random\<close>] The operation @{const Quickcheck_Random.random} - of the type class @{class random} generates a pseudo-random - value of the given size and a lazy term reconstruction of the value - in the type @{typ Code_Evaluation.term}. A pseudo-randomness generator - is defined in theory @{theory Random}. - - \<^descr>[\<open>narrowing\<close>] implements Haskell's Lazy Smallcheck @{cite "runciman-naylor-lindblad"} - using the type classes @{class narrowing} and @{class partial_term_of}. - Variables in the current goal are initially represented as symbolic variables. - If the execution of the goal tries to evaluate one of them, the test engine - replaces it with refinements provided by @{const narrowing}. - Narrowing views every value as a sum-of-products which is expressed using the operations - @{const Quickcheck_Narrowing.cons} (embedding a value), - @{const Quickcheck_Narrowing.apply} (product) and @{const Quickcheck_Narrowing.sum} (sum). - The refinement should enable further evaluation of the goal. + testing function on all values up to the given size and stops as soon as a + counterexample is found. + + \<^descr>[\<open>random\<close>] The operation @{const Quickcheck_Random.random} of the type + class @{class random} generates a pseudo-random value of the given size + and a lazy term reconstruction of the value in the type @{typ + Code_Evaluation.term}. A pseudo-randomness generator is defined in theory + @{theory Random}. + + \<^descr>[\<open>narrowing\<close>] implements Haskell's Lazy Smallcheck @{cite + "runciman-naylor-lindblad"} using the type classes @{class narrowing} and + @{class partial_term_of}. Variables in the current goal are initially + represented as symbolic variables. If the execution of the goal tries to + evaluate one of them, the test engine replaces it with refinements + provided by @{const narrowing}. Narrowing views every value as a + sum-of-products which is expressed using the operations @{const + Quickcheck_Narrowing.cons} (embedding a value), @{const + Quickcheck_Narrowing.apply} (product) and @{const + Quickcheck_Narrowing.sum} (sum). The refinement should enable further + evaluation of the goal. For example, @{const narrowing} for the list type @{typ "'a :: narrowing list"} can be recursively defined as @@ -1960,45 +1907,44 @@ (Quickcheck_Narrowing.cons (op #)) narrowing) narrowing)"}. - If a symbolic variable of type @{typ "_ list"} is evaluated, it is replaced by (i)~the empty - list @{term "[]"} and (ii)~by a non-empty list whose head and tail can then be recursively - refined if needed. - - To reconstruct counterexamples, the operation @{const partial_term_of} transforms - \<open>narrowing\<close>'s deep representation of terms to the type @{typ Code_Evaluation.term}. - The deep representation models symbolic variables as - @{const Quickcheck_Narrowing.Narrowing_variable}, which are normally converted to - @{const Code_Evaluation.Free}, and refined values as - @{term "Quickcheck_Narrowing.Narrowing_constructor i args"}, where @{term "i :: integer"} - denotes the index in the sum of refinements. In the above example for lists, - @{term "0"} corresponds to @{term "[]"} and @{term "1"} + If a symbolic variable of type @{typ "_ list"} is evaluated, it is + replaced by (i)~the empty list @{term "[]"} and (ii)~by a non-empty list + whose head and tail can then be recursively refined if needed. + + To reconstruct counterexamples, the operation @{const partial_term_of} + transforms \<open>narrowing\<close>'s deep representation of terms to the type @{typ + Code_Evaluation.term}. The deep representation models symbolic variables + as @{const Quickcheck_Narrowing.Narrowing_variable}, which are normally + converted to @{const Code_Evaluation.Free}, and refined values as @{term + "Quickcheck_Narrowing.Narrowing_constructor i args"}, where @{term "i :: + integer"} denotes the index in the sum of refinements. In the above + example for lists, @{term "0"} corresponds to @{term "[]"} and @{term "1"} to @{term "op #"}. - The command @{command (HOL) "code_datatype"} sets up @{const partial_term_of} - such that the @{term "i"}-th refinement is interpreted as the @{term "i"}-th constructor, - but it does not ensures consistency with @{const narrowing}. - - \<^descr> @{command (HOL) "quickcheck_params"} changes @{command (HOL) - "quickcheck"} configuration options persistently. - - \<^descr> @{command (HOL) "quickcheck_generator"} creates random and - exhaustive value generators for a given type and operations. It - generates values by using the operations as if they were - constructors of that type. - - \<^descr> @{command (HOL) "nitpick"} tests the current goal for - counterexamples using a reduction to first-order relational - logic. See the Nitpick manual @{cite "isabelle-nitpick"} for details. - - \<^descr> @{command (HOL) "nitpick_params"} changes @{command (HOL) - "nitpick"} configuration options persistently. + The command @{command (HOL) "code_datatype"} sets up @{const + partial_term_of} such that the @{term "i"}-th refinement is interpreted as + the @{term "i"}-th constructor, but it does not ensures consistency with + @{const narrowing}. + + \<^descr> @{command (HOL) "quickcheck_params"} changes @{command (HOL) "quickcheck"} + configuration options persistently. + + \<^descr> @{command (HOL) "quickcheck_generator"} creates random and exhaustive + value generators for a given type and operations. It generates values by + using the operations as if they were constructors of that type. + + \<^descr> @{command (HOL) "nitpick"} tests the current goal for counterexamples + using a reduction to first-order relational logic. See the Nitpick manual + @{cite "isabelle-nitpick"} for details. + + \<^descr> @{command (HOL) "nitpick_params"} changes @{command (HOL) "nitpick"} + configuration options persistently. \<^descr> @{command (HOL) "find_unused_assms"} finds potentially superfluous - assumptions in theorems using quickcheck. - It takes the theory name to be checked for superfluous assumptions as - optional argument. If not provided, it checks the current theory. - Options to the internal quickcheck invocations can be changed with - common configuration declarations. + assumptions in theorems using quickcheck. It takes the theory name to be + checked for superfluous assumptions as optional argument. If not provided, + it checks the current theory. Options to the internal quickcheck invocations + can be changed with common configuration declarations. \<close> @@ -2013,10 +1959,9 @@ @{attribute_def (HOL) coercion_args} & : & \<open>attribute\<close> \\ \end{matharray} - Coercive subtyping allows the user to omit explicit type - conversions, also called \<^emph>\<open>coercions\<close>. Type inference will add - them as necessary when parsing a term. See - @{cite "traytel-berghofer-nipkow-2011"} for details. + Coercive subtyping allows the user to omit explicit type conversions, also + called \<^emph>\<open>coercions\<close>. Type inference will add them as necessary when parsing + a term. See @{cite "traytel-berghofer-nipkow-2011"} for details. @{rail \<open> @@{attribute (HOL) coercion} (@{syntax term}) @@ -2028,48 +1973,44 @@ @@{attribute (HOL) coercion_args} (@{syntax const}) (('+' | '0' | '-')+) \<close>} - \<^descr> @{attribute (HOL) "coercion"}~\<open>f\<close> registers a new - coercion function \<open>f :: \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2\<close> where \<open>\<sigma>\<^sub>1\<close> and - \<open>\<sigma>\<^sub>2\<close> are type constructors without arguments. Coercions are - composed by the inference algorithm if needed. Note that the type - inference algorithm is complete only if the registered coercions - form a lattice. - - \<^descr> @{attribute (HOL) "coercion_delete"}~\<open>f\<close> deletes a - preceding declaration (using @{attribute (HOL) "coercion"}) of the - function \<open>f :: \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2\<close> as a coercion. - - \<^descr> @{attribute (HOL) "coercion_map"}~\<open>map\<close> registers a - new map function to lift coercions through type constructors. The - function \<open>map\<close> must conform to the following type pattern + \<^descr> @{attribute (HOL) "coercion"}~\<open>f\<close> registers a new coercion function \<open>f :: + \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2\<close> where \<open>\<sigma>\<^sub>1\<close> and \<open>\<sigma>\<^sub>2\<close> are type constructors without arguments. + Coercions are composed by the inference algorithm if needed. Note that the + type inference algorithm is complete only if the registered coercions form a + lattice. + + \<^descr> @{attribute (HOL) "coercion_delete"}~\<open>f\<close> deletes a preceding declaration + (using @{attribute (HOL) "coercion"}) of the function \<open>f :: \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2\<close> as a + coercion. + + \<^descr> @{attribute (HOL) "coercion_map"}~\<open>map\<close> registers a new map function to + lift coercions through type constructors. The function \<open>map\<close> must conform to + the following type pattern \begin{matharray}{lll} \<open>map\<close> & \<open>::\<close> & \<open>f\<^sub>1 \<Rightarrow> \<dots> \<Rightarrow> f\<^sub>n \<Rightarrow> (\<alpha>\<^sub>1, \<dots>, \<alpha>\<^sub>n) t \<Rightarrow> (\<beta>\<^sub>1, \<dots>, \<beta>\<^sub>n) t\<close> \\ \end{matharray} - where \<open>t\<close> is a type constructor and \<open>f\<^sub>i\<close> is of type - \<open>\<alpha>\<^sub>i \<Rightarrow> \<beta>\<^sub>i\<close> or \<open>\<beta>\<^sub>i \<Rightarrow> \<alpha>\<^sub>i\<close>. Registering a map function - overwrites any existing map function for this particular type - constructor. - - \<^descr> @{attribute (HOL) "coercion_args"} can be used to disallow - coercions to be inserted in certain positions in a term. For example, - given the constant \<open>c :: \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2 \<Rightarrow> \<sigma>\<^sub>3 \<Rightarrow> \<sigma>\<^sub>4\<close> and the list - of policies \<open>- + 0\<close> as arguments, coercions will not be - inserted in the first argument of \<open>c\<close> (policy \<open>-\<close>); - they may be inserted in the second argument (policy \<open>+\<close>) - even if the constant \<open>c\<close> itself is in a position where - coercions are disallowed; the third argument inherits the allowance - of coercsion insertion from the position of the constant \<open>c\<close> - (policy \<open>0\<close>). The standard usage of policies is the definition - of syntatic constructs (usually extralogical, i.e., processed and - stripped during type inference), that should not be destroyed by the - insertion of coercions (see, for example, the setup for the case syntax - in @{theory Ctr_Sugar}). - - \<^descr> @{attribute (HOL) "coercion_enabled"} enables the coercion - inference algorithm. + where \<open>t\<close> is a type constructor and \<open>f\<^sub>i\<close> is of type \<open>\<alpha>\<^sub>i \<Rightarrow> \<beta>\<^sub>i\<close> or \<open>\<beta>\<^sub>i \<Rightarrow> + \<alpha>\<^sub>i\<close>. Registering a map function overwrites any existing map function for + this particular type constructor. + + \<^descr> @{attribute (HOL) "coercion_args"} can be used to disallow coercions to be + inserted in certain positions in a term. For example, given the constant \<open>c + :: \<sigma>\<^sub>1 \<Rightarrow> \<sigma>\<^sub>2 \<Rightarrow> \<sigma>\<^sub>3 \<Rightarrow> \<sigma>\<^sub>4\<close> and the list of policies \<open>- + 0\<close> as arguments, + coercions will not be inserted in the first argument of \<open>c\<close> (policy \<open>-\<close>); + they may be inserted in the second argument (policy \<open>+\<close>) even if the + constant \<open>c\<close> itself is in a position where coercions are disallowed; the + third argument inherits the allowance of coercsion insertion from the + position of the constant \<open>c\<close> (policy \<open>0\<close>). The standard usage of policies is + the definition of syntatic constructs (usually extralogical, i.e., processed + and stripped during type inference), that should not be destroyed by the + insertion of coercions (see, for example, the setup for the case syntax in + @{theory Ctr_Sugar}). + + \<^descr> @{attribute (HOL) "coercion_enabled"} enables the coercion inference + algorithm. \<close> @@ -2082,19 +2023,19 @@ @{attribute_def (HOL) arith_split} & : & \<open>attribute\<close> \\ \end{matharray} - \<^descr> @{method (HOL) arith} decides linear arithmetic problems (on - types \<open>nat\<close>, \<open>int\<close>, \<open>real\<close>). Any current facts - are inserted into the goal before running the procedure. - - \<^descr> @{attribute (HOL) arith} declares facts that are supplied to - the arithmetic provers implicitly. - - \<^descr> @{attribute (HOL) arith_split} attribute declares case split - rules to be expanded before @{method (HOL) arith} is invoked. - - - Note that a simpler (but faster) arithmetic prover is already - invoked by the Simplifier. + \<^descr> @{method (HOL) arith} decides linear arithmetic problems (on types \<open>nat\<close>, + \<open>int\<close>, \<open>real\<close>). Any current facts are inserted into the goal before running + the procedure. + + \<^descr> @{attribute (HOL) arith} declares facts that are supplied to the + arithmetic provers implicitly. + + \<^descr> @{attribute (HOL) arith_split} attribute declares case split rules to be + expanded before @{method (HOL) arith} is invoked. + + + Note that a simpler (but faster) arithmetic prover is already invoked by the + Simplifier. \<close> @@ -2109,19 +2050,18 @@ @@{method (HOL) iprover} (@{syntax rulemod} *) \<close>} - \<^descr> @{method (HOL) iprover} performs intuitionistic proof search, - depending on specifically declared rules from the context, or given - as explicit arguments. Chained facts are inserted into the goal - before commencing proof search. - - Rules need to be classified as @{attribute (Pure) intro}, - @{attribute (Pure) elim}, or @{attribute (Pure) dest}; here the - ``\<open>!\<close>'' indicator refers to ``safe'' rules, which may be - applied aggressively (without considering back-tracking later). - Rules declared with ``\<open>?\<close>'' are ignored in proof search (the - single-step @{method (Pure) rule} method still observes these). An - explicit weight annotation may be given as well; otherwise the - number of rule premises will be taken into account here. + \<^descr> @{method (HOL) iprover} performs intuitionistic proof search, depending on + specifically declared rules from the context, or given as explicit + arguments. Chained facts are inserted into the goal before commencing proof + search. + + Rules need to be classified as @{attribute (Pure) intro}, @{attribute (Pure) + elim}, or @{attribute (Pure) dest}; here the ``\<open>!\<close>'' indicator refers to + ``safe'' rules, which may be applied aggressively (without considering + back-tracking later). Rules declared with ``\<open>?\<close>'' are ignored in proof + search (the single-step @{method (Pure) rule} method still observes these). + An explicit weight annotation may be given as well; otherwise the number of + rule premises will be taken into account here. \<close> @@ -2141,16 +2081,16 @@ @{syntax thms}? \<close>} - \<^descr> @{method (HOL) meson} implements Loveland's model elimination - procedure @{cite "loveland-78"}. See @{file - "~~/src/HOL/ex/Meson_Test.thy"} for examples. + \<^descr> @{method (HOL) meson} implements Loveland's model elimination procedure + @{cite "loveland-78"}. See @{file "~~/src/HOL/ex/Meson_Test.thy"} for + examples. \<^descr> @{method (HOL) metis} combines ordered resolution and ordered - paramodulation to find first-order (or mildly higher-order) proofs. - The first optional argument specifies a type encoding; see the - Sledgehammer manual @{cite "isabelle-sledgehammer"} for details. The - directory @{file "~~/src/HOL/Metis_Examples"} contains several small - theories developed to a large extent using @{method (HOL) metis}. + paramodulation to find first-order (or mildly higher-order) proofs. The + first optional argument specifies a type encoding; see the Sledgehammer + manual @{cite "isabelle-sledgehammer"} for details. The directory @{file + "~~/src/HOL/Metis_Examples"} contains several small theories developed to a + large extent using @{method (HOL) metis}. \<close> @@ -2170,23 +2110,21 @@ @@{attribute (HOL) algebra} (() | 'add' | 'del') \<close>} - \<^descr> @{method (HOL) algebra} performs algebraic reasoning via - Gr\"obner bases, see also @{cite "Chaieb-Wenzel:2007"} and - @{cite \<open>\S3.2\<close> "Chaieb-thesis"}. The method handles deals with two main - classes of problems: + \<^descr> @{method (HOL) algebra} performs algebraic reasoning via Gr\"obner bases, + see also @{cite "Chaieb-Wenzel:2007"} and @{cite \<open>\S3.2\<close> "Chaieb-thesis"}. + The method handles deals with two main classes of problems: \<^enum> Universal problems over multivariate polynomials in a (semi)-ring/field/idom; the capabilities of the method are augmented - according to properties of these structures. For this problem class - the method is only complete for algebraically closed fields, since - the underlying method is based on Hilbert's Nullstellensatz, where - the equivalence only holds for algebraically closed fields. - - The problems can contain equations \<open>p = 0\<close> or inequations - \<open>q \<noteq> 0\<close> anywhere within a universal problem statement. - - \<^enum> All-exists problems of the following restricted (but useful) - form: + according to properties of these structures. For this problem class the + method is only complete for algebraically closed fields, since the + underlying method is based on Hilbert's Nullstellensatz, where the + equivalence only holds for algebraically closed fields. + + The problems can contain equations \<open>p = 0\<close> or inequations \<open>q \<noteq> 0\<close> anywhere + within a universal problem statement. + + \<^enum> All-exists problems of the following restricted (but useful) form: @{text [display] "\<forall>x\<^sub>1 \<dots> x\<^sub>n. e\<^sub>1(x\<^sub>1, \<dots>, x\<^sub>n) = 0 \<and> \<dots> \<and> e\<^sub>m(x\<^sub>1, \<dots>, x\<^sub>n) = 0 \<longrightarrow> @@ -2195,23 +2133,25 @@ \<dots> \<and> p\<^sub>t\<^sub>1(x\<^sub>1, \<dots>, x\<^sub>n) * y\<^sub>1 + \<dots> + p\<^sub>t\<^sub>k(x\<^sub>1, \<dots>, x\<^sub>n) * y\<^sub>k = 0)"} - Here \<open>e\<^sub>1, \<dots>, e\<^sub>n\<close> and the \<open>p\<^sub>i\<^sub>j\<close> are multivariate - polynomials only in the variables mentioned as arguments. - - The proof method is preceded by a simplification step, which may be - modified by using the form \<open>(algebra add: ths\<^sub>1 del: ths\<^sub>2)\<close>. - This acts like declarations for the Simplifier - (\secref{sec:simplifier}) on a private simpset for this tool. - - \<^descr> @{attribute algebra} (as attribute) manages the default - collection of pre-simplification rules of the above proof method. + Here \<open>e\<^sub>1, \<dots>, e\<^sub>n\<close> and the \<open>p\<^sub>i\<^sub>j\<close> are multivariate polynomials only in + the variables mentioned as arguments. + + The proof method is preceded by a simplification step, which may be modified + by using the form \<open>(algebra add: ths\<^sub>1 del: ths\<^sub>2)\<close>. This acts like + declarations for the Simplifier (\secref{sec:simplifier}) on a private + simpset for this tool. + + \<^descr> @{attribute algebra} (as attribute) manages the default collection of + pre-simplification rules of the above proof method. \<close> subsubsection \<open>Example\<close> -text \<open>The subsequent example is from geometry: collinearity is - invariant by rotation.\<close> +text \<open> + The subsequent example is from geometry: collinearity is invariant by + rotation. +\<close> (*<*)experiment begin(*>*) type_synonym point = "int \<times> int" @@ -2243,19 +2183,19 @@ @@{method (HOL) coherent} @{syntax thms}? \<close>} - \<^descr> @{method (HOL) coherent} solves problems of \<^emph>\<open>Coherent - Logic\<close> @{cite "Bezem-Coquand:2005"}, which covers applications in - confluence theory, lattice theory and projective geometry. See - @{file "~~/src/HOL/ex/Coherent.thy"} for some examples. + \<^descr> @{method (HOL) coherent} solves problems of \<^emph>\<open>Coherent Logic\<close> @{cite + "Bezem-Coquand:2005"}, which covers applications in confluence theory, + lattice theory and projective geometry. See @{file + "~~/src/HOL/ex/Coherent.thy"} for some examples. \<close> section \<open>Unstructured case analysis and induction \label{sec:hol-induct-tac}\<close> text \<open> - The following tools of Isabelle/HOL support cases analysis and - induction in unstructured tactic scripts; see also - \secref{sec:cases-induct} for proper Isar versions of similar ideas. + The following tools of Isabelle/HOL support cases analysis and induction in + unstructured tactic scripts; see also \secref{sec:cases-induct} for proper + Isar versions of similar ideas. \begin{matharray}{rcl} @{method_def (HOL) case_tac}\<open>\<^sup>*\<close> & : & \<open>method\<close> \\ @@ -2276,31 +2216,29 @@ rule: 'rule' ':' @{syntax thm} \<close>} - \<^descr> @{method (HOL) case_tac} and @{method (HOL) induct_tac} admit - to reason about inductive types. Rules are selected according to - the declarations by the @{attribute cases} and @{attribute induct} - attributes, cf.\ \secref{sec:cases-induct}. The @{command (HOL) - datatype} package already takes care of this. + \<^descr> @{method (HOL) case_tac} and @{method (HOL) induct_tac} admit to reason + about inductive types. Rules are selected according to the declarations by + the @{attribute cases} and @{attribute induct} attributes, cf.\ + \secref{sec:cases-induct}. The @{command (HOL) datatype} package already + takes care of this. These unstructured tactics feature both goal addressing and dynamic - instantiation. Note that named rule cases are \<^emph>\<open>not\<close> provided - as would be by the proper @{method cases} and @{method induct} proof - methods (see \secref{sec:cases-induct}). Unlike the @{method - induct} method, @{method induct_tac} does not handle structured rule - statements, only the compact object-logic conclusion of the subgoal - being addressed. - - \<^descr> @{method (HOL) ind_cases} and @{command (HOL) - "inductive_cases"} provide an interface to the internal @{ML_text - mk_cases} operation. Rules are simplified in an unrestricted - forward manner. - - While @{method (HOL) ind_cases} is a proof method to apply the - result immediately as elimination rules, @{command (HOL) - "inductive_cases"} provides case split theorems at the theory level - for later use. The @{keyword "for"} argument of the @{method (HOL) - ind_cases} method allows to specify a list of variables that should - be generalized before applying the resulting rule. + instantiation. Note that named rule cases are \<^emph>\<open>not\<close> provided as would be by + the proper @{method cases} and @{method induct} proof methods (see + \secref{sec:cases-induct}). Unlike the @{method induct} method, @{method + induct_tac} does not handle structured rule statements, only the compact + object-logic conclusion of the subgoal being addressed. + + \<^descr> @{method (HOL) ind_cases} and @{command (HOL) "inductive_cases"} provide + an interface to the internal @{ML_text mk_cases} operation. Rules are + simplified in an unrestricted forward manner. + + While @{method (HOL) ind_cases} is a proof method to apply the result + immediately as elimination rules, @{command (HOL) "inductive_cases"} + provides case split theorems at the theory level for later use. The + @{keyword "for"} argument of the @{method (HOL) ind_cases} method allows to + specify a list of variables that should be generalized before applying the + resulting rule. \<close> @@ -2315,9 +2253,9 @@ @@{attribute (HOL) split_format} ('(' 'complete' ')')? \<close>} - \<^descr> @{attribute (HOL) split_format}\ \<open>(complete)\<close> causes - arguments in function applications to be represented canonically - according to their tuple type structure. + \<^descr> @{attribute (HOL) split_format}\ \<open>(complete)\<close> causes arguments in function + applications to be represented canonically according to their tuple type + structure. Note that this operation tends to invent funny names for new local parameters introduced. @@ -2326,24 +2264,25 @@ chapter \<open>Executable code\<close> -text \<open>For validation purposes, it is often useful to \<^emph>\<open>execute\<close> - specifications. In principle, execution could be simulated by Isabelle's - inference kernel, i.e. by a combination of resolution and simplification. - Unfortunately, this approach is rather inefficient. A more efficient way - of executing specifications is to translate them into a functional - programming language such as ML. +text \<open> + For validation purposes, it is often useful to \<^emph>\<open>execute\<close> specifications. In + principle, execution could be simulated by Isabelle's inference kernel, i.e. + by a combination of resolution and simplification. Unfortunately, this + approach is rather inefficient. A more efficient way of executing + specifications is to translate them into a functional programming language + such as ML. Isabelle provides a generic framework to support code generation from executable specifications. Isabelle/HOL instantiates these mechanisms in a way that is amenable to end-user applications. Code can be generated for - functional programs (including overloading using type classes) targeting - SML @{cite SML}, OCaml @{cite OCaml}, Haskell @{cite - "haskell-revised-report"} and Scala @{cite "scala-overview-tech-report"}. - Conceptually, code generation is split up in three steps: \<^emph>\<open>selection\<close> - of code theorems, \<^emph>\<open>translation\<close> into an abstract executable view and - \<^emph>\<open>serialization\<close> to a specific \<^emph>\<open>target language\<close>. Inductive - specifications can be executed using the predicate compiler which operates - within HOL. See @{cite "isabelle-codegen"} for an introduction. + functional programs (including overloading using type classes) targeting SML + @{cite SML}, OCaml @{cite OCaml}, Haskell @{cite "haskell-revised-report"} + and Scala @{cite "scala-overview-tech-report"}. Conceptually, code + generation is split up in three steps: \<^emph>\<open>selection\<close> of code theorems, + \<^emph>\<open>translation\<close> into an abstract executable view and \<^emph>\<open>serialization\<close> to a + specific \<^emph>\<open>target language\<close>. Inductive specifications can be executed using + the predicate compiler which operates within HOL. See @{cite + "isabelle-codegen"} for an introduction. \begin{matharray}{rcl} @{command_def (HOL) "export_code"}\<open>\<^sup>*\<close> & : & \<open>context \<rightarrow>\<close> \\ @@ -2456,22 +2395,20 @@ instruction is given, only abstract code is generated internally. Constants may be specified by giving them literally, referring to all - executable constants within a certain theory by giving \<open>name._\<close>, - or referring to \<^emph>\<open>all\<close> executable constants currently available by - giving \<open>_\<close>. + executable constants within a certain theory by giving \<open>name._\<close>, or + referring to \<^emph>\<open>all\<close> executable constants currently available by giving \<open>_\<close>. By default, exported identifiers are minimized per module. This can be suppressed by prepending @{keyword "open"} before the list of constants. - By default, for each involved theory one corresponding name space module - is generated. Alternatively, a module name may be specified after the - @{keyword "module_name"} keyword; then \<^emph>\<open>all\<close> code is placed in this - module. - - For \<^emph>\<open>SML\<close>, \<^emph>\<open>OCaml\<close> and \<^emph>\<open>Scala\<close> the file specification - refers to a single file; for \<^emph>\<open>Haskell\<close>, it refers to a whole - directory, where code is generated in multiple files reflecting the module - hierarchy. Omitting the file specification denotes standard output. + By default, for each involved theory one corresponding name space module is + generated. Alternatively, a module name may be specified after the @{keyword + "module_name"} keyword; then \<^emph>\<open>all\<close> code is placed in this module. + + For \<^emph>\<open>SML\<close>, \<^emph>\<open>OCaml\<close> and \<^emph>\<open>Scala\<close> the file specification refers to a single + file; for \<^emph>\<open>Haskell\<close>, it refers to a whole directory, where code is + generated in multiple files reflecting the module hierarchy. Omitting the + file specification denotes standard output. Serializers take an optional list of arguments in parentheses. For \<^emph>\<open>Haskell\<close> a module name prefix may be given using the ``\<open>root:\<close>'' argument; @@ -2479,82 +2416,80 @@ appropriate datatype declaration. \<^descr> @{attribute (HOL) code} declare code equations for code generation. - Variant \<open>code equation\<close> declares a conventional equation as code - equation. Variants \<open>code abstype\<close> and \<open>code abstract\<close> - declare abstract datatype certificates or code equations on abstract - datatype representations respectively. Vanilla \<open>code\<close> falls back - to \<open>code equation\<close> or \<open>code abstype\<close> depending on the - syntactic shape of the underlying equation. Variant \<open>code del\<close> - deselects a code equation for code generation. - - Variants \<open>code drop:\<close> and \<open>code abort:\<close> take a list of - constant as arguments and drop all code equations declared for them. In - the case of {text abort}, these constants then are are not required to - have a definition by means of code equations; if needed these are - implemented by program abort (exception) instead. + Variant \<open>code equation\<close> declares a conventional equation as code equation. + Variants \<open>code abstype\<close> and \<open>code abstract\<close> declare abstract datatype + certificates or code equations on abstract datatype representations + respectively. Vanilla \<open>code\<close> falls back to \<open>code equation\<close> or \<open>code abstype\<close> + depending on the syntactic shape of the underlying equation. Variant \<open>code + del\<close> deselects a code equation for code generation. + + Variants \<open>code drop:\<close> and \<open>code abort:\<close> take a list of constant as arguments + and drop all code equations declared for them. In the case of {text abort}, + these constants then are are not required to have a definition by means of + code equations; if needed these are implemented by program abort (exception) + instead. Usually packages introducing code equations provide a reasonable default setup for selection. - \<^descr> @{command (HOL) "code_datatype"} specifies a constructor set for a - logical type. - - \<^descr> @{command (HOL) "print_codesetup"} gives an overview on selected - code equations and code generator datatypes. - - \<^descr> @{attribute (HOL) code_unfold} declares (or with option ``\<open>del\<close>'' removes) theorems which during preprocessing are applied as - rewrite rules to any code equation or evaluation input. - - \<^descr> @{attribute (HOL) code_post} declares (or with option ``\<open>del\<close>'' removes) theorems which are applied as rewrite rules to any - result of an evaluation. - - \<^descr> @{attribute (HOL) code_abbrev} declares (or with option ``\<open>del\<close>'' removes) equations which are applied as rewrite rules to any - result of an evaluation and symmetrically during preprocessing to any code + \<^descr> @{command (HOL) "code_datatype"} specifies a constructor set for a logical + type. + + \<^descr> @{command (HOL) "print_codesetup"} gives an overview on selected code + equations and code generator datatypes. + + \<^descr> @{attribute (HOL) code_unfold} declares (or with option ``\<open>del\<close>'' removes) + theorems which during preprocessing are applied as rewrite rules to any code equation or evaluation input. - \<^descr> @{command (HOL) "print_codeproc"} prints the setup of the code - generator preprocessor. - - \<^descr> @{command (HOL) "code_thms"} prints a list of theorems representing - the corresponding program containing all given constants after - preprocessing. + \<^descr> @{attribute (HOL) code_post} declares (or with option ``\<open>del\<close>'' removes) + theorems which are applied as rewrite rules to any result of an evaluation. + + \<^descr> @{attribute (HOL) code_abbrev} declares (or with option ``\<open>del\<close>'' removes) + equations which are applied as rewrite rules to any result of an evaluation + and symmetrically during preprocessing to any code equation or evaluation + input. + + \<^descr> @{command (HOL) "print_codeproc"} prints the setup of the code generator + preprocessor. + + \<^descr> @{command (HOL) "code_thms"} prints a list of theorems representing the + corresponding program containing all given constants after preprocessing. \<^descr> @{command (HOL) "code_deps"} visualizes dependencies of theorems - representing the corresponding program containing all given constants - after preprocessing. - - \<^descr> @{command (HOL) "code_reserved"} declares a list of names as - reserved for a given target, preventing it to be shadowed by any generated - code. + representing the corresponding program containing all given constants after + preprocessing. + + \<^descr> @{command (HOL) "code_reserved"} declares a list of names as reserved for + a given target, preventing it to be shadowed by any generated code. \<^descr> @{command (HOL) "code_printing"} associates a series of symbols (constants, type constructors, classes, class relations, instances, module - names) with target-specific serializations; omitting a serialization - deletes an existing serialization. - - \<^descr> @{command (HOL) "code_monad"} provides an auxiliary mechanism to - generate monadic code for Haskell. + names) with target-specific serializations; omitting a serialization deletes + an existing serialization. + + \<^descr> @{command (HOL) "code_monad"} provides an auxiliary mechanism to generate + monadic code for Haskell. \<^descr> @{command (HOL) "code_identifier"} associates a a series of symbols (constants, type constructors, classes, class relations, instances, module names) with target-specific hints how these symbols shall be named. These hints gain precedence over names for symbols with no hints at all. - Conflicting hints are subject to name disambiguation. \<^emph>\<open>Warning:\<close> It - is at the discretion of the user to ensure that name prefixes of - identifiers in compound statements like type classes or datatypes are - still the same. - - \<^descr> @{command (HOL) "code_reflect"} without a ``\<open>file\<close>'' - argument compiles code into the system runtime environment and modifies - the code generator setup that future invocations of system runtime code - generation referring to one of the ``\<open>datatypes\<close>'' or ``\<open>functions\<close>'' entities use these precompiled entities. With a ``\<open>file\<close>'' argument, the corresponding code is generated into that - specified file without modifying the code generator setup. - - \<^descr> @{command (HOL) "code_pred"} creates code equations for a predicate - given a set of introduction rules. Optional mode annotations determine - which arguments are supposed to be input or output. If alternative - introduction rules are declared, one must prove a corresponding - elimination rule. + Conflicting hints are subject to name disambiguation. \<^emph>\<open>Warning:\<close> It is at + the discretion of the user to ensure that name prefixes of identifiers in + compound statements like type classes or datatypes are still the same. + + \<^descr> @{command (HOL) "code_reflect"} without a ``\<open>file\<close>'' argument compiles + code into the system runtime environment and modifies the code generator + setup that future invocations of system runtime code generation referring to + one of the ``\<open>datatypes\<close>'' or ``\<open>functions\<close>'' entities use these precompiled + entities. With a ``\<open>file\<close>'' argument, the corresponding code is generated + into that specified file without modifying the code generator setup. + + \<^descr> @{command (HOL) "code_pred"} creates code equations for a predicate given + a set of introduction rules. Optional mode annotations determine which + arguments are supposed to be input or output. If alternative introduction + rules are declared, one must prove a corresponding elimination rule. \<close> end

--- a/src/Doc/Isar_Ref/Inner_Syntax.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Inner_Syntax.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Inner_Syntax -imports Base Main + imports Main Base begin chapter \<open>Inner syntax --- the term language \label{ch:inner-syntax}\<close>

--- a/src/Doc/Isar_Ref/Outer_Syntax.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Outer_Syntax.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Outer_Syntax -imports Base Main + imports Main Base begin chapter \<open>Outer syntax --- the theory language \label{ch:outer-syntax}\<close>

--- a/src/Doc/Isar_Ref/Preface.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Preface.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Preface -imports Base Main + imports Main Base begin text \<open>

--- a/src/Doc/Isar_Ref/Proof.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Proof.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Proof -imports Base Main + imports Main Base begin chapter \<open>Proofs \label{ch:proofs}\<close> @@ -189,8 +189,8 @@ "and"}; the resulting list of current facts consists of all of these concatenated. - A structured assumption like \<^theory_text>\<open>assume "B x" and "A x" for x\<close> is equivalent - to \<^theory_text>\<open>assume "\<And>x. A x \<Longrightarrow> B x"\<close>, but vacuous quantification is avoided: a + A structured assumption like \<^theory_text>\<open>assume "B x" if "A x" for x\<close> is equivalent to + \<^theory_text>\<open>assume "\<And>x. A x \<Longrightarrow> B x"\<close>, but vacuous quantification is avoided: a for-context only effects propositions according to actual use of variables. \<^descr> \<^theory_text>\<open>define x where "x = t"\<close> introduces a local (non-polymorphic) definition. @@ -283,6 +283,8 @@ @{command_def "with"} & : & \<open>proof(state) \<rightarrow> proof(chain)\<close> \\ @{command_def "using"} & : & \<open>proof(prove) \<rightarrow> proof(prove)\<close> \\ @{command_def "unfolding"} & : & \<open>proof(prove) \<rightarrow> proof(prove)\<close> \\ + @{method_def "use"} & : & \<open>method\<close> \\ + @{fact_def "method_facts"} & : & \<open>fact\<close> \\ \end{matharray} New facts are established either by assumption or proof of local statements. @@ -300,6 +302,8 @@ ; (@@{command from} | @@{command with} | @@{command using} | @@{command unfolding}) (@{syntax thms} + @'and') + ; + @{method use} @{syntax thms} @'in' @{syntax method} \<close>} \<^descr> @{command "note"}~\<open>a = b\<^sub>1 \<dots> b\<^sub>n\<close> recalls existing facts \<open>b\<^sub>1, \<dots>, b\<^sub>n\<close>, @@ -323,14 +327,22 @@ \<AND> this\<close>''; thus the forward chaining is from earlier facts together with the current ones. - \<^descr> @{command "using"}~\<open>b\<^sub>1 \<dots> b\<^sub>n\<close> augments the facts being currently - indicated for use by a subsequent refinement step (such as @{command_ref - "apply"} or @{command_ref "proof"}). + \<^descr> @{command "using"}~\<open>b\<^sub>1 \<dots> b\<^sub>n\<close> augments the facts to be used by a + subsequent refinement step (such as @{command_ref "apply"} or @{command_ref + "proof"}). \<^descr> @{command "unfolding"}~\<open>b\<^sub>1 \<dots> b\<^sub>n\<close> is structurally similar to @{command - "using"}, but unfolds definitional equations \<open>b\<^sub>1, \<dots> b\<^sub>n\<close> throughout the + "using"}, but unfolds definitional equations \<open>b\<^sub>1 \<dots> b\<^sub>n\<close> throughout the goal state and facts. + \<^descr> \<^theory_text>\<open>(use b\<^sub>1 \<dots> b\<^sub>n in method)\<close> uses the facts in the given method + expression. The facts provided by the proof state (via @{command "using"} + etc.) are ignored, but it is possible to refer to @{fact method_facts} + explicitly. + + \<^descr> @{fact method_facts} is a dynamic fact that refers to the currently used + facts of the goal state. + Forward chaining with an empty list of theorems is the same as not chaining at all. Thus ``@{command "from"}~\<open>nothing\<close>'' has no effect apart from

--- a/src/Doc/Isar_Ref/Proof_Script.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Proof_Script.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Proof_Script -imports Base Main + imports Main Base begin chapter \<open>Proof scripts\<close>

--- a/src/Doc/Isar_Ref/Quick_Reference.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Quick_Reference.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Quick_Reference -imports Base Main + imports Main Base begin chapter \<open>Isabelle/Isar quick reference \label{ap:refcard}\<close>

--- a/src/Doc/Isar_Ref/Spec.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Spec.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Spec -imports Base Main + imports Main Base begin chapter \<open>Specifications\<close>

--- a/src/Doc/Isar_Ref/Symbols.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Symbols.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Symbols -imports Base Main + imports Main Base begin chapter \<open>Predefined Isabelle symbols \label{app:symbols}\<close>

--- a/src/Doc/Isar_Ref/Synopsis.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Doc/Isar_Ref/Synopsis.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1,7 +1,7 @@ (*:maxLineLen=78:*) theory Synopsis -imports Base Main + imports Main Base begin chapter \<open>Synopsis\<close>

--- a/src/HOL/Binomial.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Binomial.thy Thu Jul 21 10:52:27 2016 +0200 @@ -1424,18 +1424,6 @@ lemma choose_one: "n choose 1 = n" for n :: nat by simp -(*FIXME: messy and apparently unused*) -lemma binomial_induct [rule_format]: "(\<forall>n::nat. P n n) \<longrightarrow> - (\<forall>n. P (Suc n) 0) \<longrightarrow> (\<forall>n. (\<forall>k < n. P n k \<longrightarrow> P n (Suc k) \<longrightarrow> - P (Suc n) (Suc k))) \<longrightarrow> (\<forall>k \<le> n. P n k)" - apply (induct n) - apply auto - apply (case_tac "k = 0") - apply auto - apply (case_tac "k = Suc n") - apply (auto simp add: le_Suc_eq elim: lessE) - done - lemma card_UNION: assumes "finite A" and "\<forall>k \<in> A. finite k"

--- a/src/HOL/Eisbach/Eisbach.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Eisbach/Eisbach.thy Thu Jul 21 10:52:27 2016 +0200 @@ -22,9 +22,4 @@ method solves methods m = (m; fail) -method_setup use = \<open> - Attrib.thms -- (Scan.lift @{keyword "in"} |-- Method_Closure.method_text) >> - (fn (thms, text) => fn ctxt => fn _ => Method_Closure.method_evaluate text ctxt thms) -\<close> \<open>indicate method facts and context for method expression\<close> - end

--- a/src/HOL/Eisbach/Eisbach_Tools.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Eisbach/Eisbach_Tools.thy Thu Jul 21 10:52:27 2016 +0200 @@ -55,11 +55,11 @@ \<close> method_setup catch = \<open> - Method_Closure.method_text -- Method_Closure.method_text >> + Method.text_closure -- Method.text_closure >> (fn (text, text') => fn ctxt => fn using => fn st => let - val method = Method_Closure.method_evaluate text ctxt using; - val backup_results = Method_Closure.method_evaluate text' ctxt using st; + val method = Method.evaluate_runtime text ctxt using; + val backup_results = Method.evaluate_runtime text' ctxt using st; in (case try method st of SOME seq => try_map backup_results seq

--- a/src/HOL/Eisbach/match_method.ML Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Eisbach/match_method.ML Thu Jul 21 10:52:27 2016 +0200 @@ -100,7 +100,7 @@ (case Token.get_value body of SOME (Token.Source src) => let - val text = Method_Closure.read ctxt src; + val text = Method.read ctxt src; val ts' = map (fn (b, (Parse_Tools.Real_Val v, match_args)) => @@ -201,7 +201,7 @@ |> (fn ctxt => fold2 upd_ctxt binds pats ([], ctxt) |> apfst rev) ||> Proof_Context.restore_mode ctxt; - val (text, src) = Method_Closure.read_closure_input ctxt6 (Token.input_of body); + val (text, src) = Method.read_closure_input ctxt6 (Token.input_of body); val morphism = Variable.export_morphism ctxt6 @@ -645,12 +645,12 @@ Match_Fact net => make_fact_matches goal_ctxt (Item_Net.retrieve net) |> Seq.map (fn (text, ctxt') => - Method_Closure.method_evaluate text ctxt' using (ctxt', st) + Method.evaluate_runtime text ctxt' using (ctxt', st) |> Seq.filter_results |> Seq.map (fn (_, thm) => (outer_ctxt, thm))) | Match_Term net => make_term_matches goal_ctxt (Item_Net.retrieve net) |> Seq.map (fn (text, ctxt') => - Method_Closure.method_evaluate text ctxt' using (ctxt', st) + Method.evaluate_runtime text ctxt' using (ctxt', st) |> Seq.filter_results |> Seq.map (fn (_, thm) => (outer_ctxt, thm))) | match_kind => if Thm.no_prems st then Seq.empty @@ -721,7 +721,7 @@ end; fun apply_text (text, ctxt') = - Method_Closure.method_evaluate text ctxt' using (ctxt', focused_goal) + Method.evaluate_runtime text ctxt' using (ctxt', focused_goal) |> Seq.filter_results |> Seq.maps (Seq.DETERM do_retrofit)

--- a/src/HOL/Eisbach/method_closure.ML Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Eisbach/method_closure.ML Thu Jul 21 10:52:27 2016 +0200 @@ -10,11 +10,6 @@ signature METHOD_CLOSURE = sig - val read: Proof.context -> Token.src -> Method.text - val read_closure: Proof.context -> Token.src -> Method.text * Token.src - val read_closure_input: Proof.context -> Input.source -> Method.text * Token.src - val method_text: Method.text context_parser - val method_evaluate: Method.text -> Proof.context -> Method.method val apply_method: Proof.context -> string -> term list -> thm list list -> (Proof.context -> Method.method) list -> Proof.context -> thm list -> context_tactic val method: binding -> (binding * typ option * mixfix) list -> binding list -> @@ -82,51 +77,7 @@ end; -(* read method text *) - -fun read ctxt src = - (case Scan.read Token.stopper (Parse.!!! (Method.parser 0 --| Scan.ahead Parse.eof)) src of - SOME (text, range) => - if Method.checked_text text then text - else (Method.report (text, range); Method.check_text ctxt text) - | NONE => error ("Failed to parse method" ^ Position.here (#1 (Token.range_of src)))); - -fun read_closure ctxt src0 = - let - val src1 = map Token.init_assignable src0; - val text = read ctxt src1 |> Method.map_source (Method.method_closure ctxt); - val src2 = map Token.closure src1; - in (text, src2) end; - -fun read_closure_input ctxt = - Input.source_explode #> - Token.read_no_commands (Thy_Header.get_keywords' ctxt) (Scan.one Token.not_eof) #> - read_closure ctxt; - -val method_text = - Args.context -- Scan.lift (Parse.token Parse.text) >> (fn (ctxt, tok) => - (case Token.get_value tok of - SOME (Token.Source src) => read ctxt src - | _ => - let - val (text, src) = read_closure_input ctxt (Token.input_of tok); - val _ = Token.assign (SOME (Token.Source src)) tok; - in text end)); - - -(* evaluate method text *) - -fun method_evaluate text ctxt = - let - val text' = - text |> (Method.map_source o map o Token.map_facts) - (fn SOME name => - (case Proof_Context.lookup_fact ctxt name of - SOME {dynamic = true, thms} => K thms - | _ => I) - | NONE => I); - val ctxt' = Config.put Method.closure false ctxt; - in fn facts => Method.RUNTIME (fn st => Method.evaluate text' ctxt' facts st) end; +(* instantiate and evaluate method text *) fun method_instantiate vars body ts ctxt = let @@ -134,7 +85,7 @@ val subst = fold (Pattern.match thy) (vars ~~ ts) (Vartab.empty, Vartab.empty); val morphism = Morphism.term_morphism "method_instantiate" (Envir.subst_term subst); val body' = (Method.map_source o map) (Token.transform morphism) body; - in method_evaluate body' ctxt end; + in Method.evaluate_runtime body' ctxt end; @@ -215,12 +166,12 @@ let fun bind_method (name, text) ctxt = let - val method = method_evaluate text; + val method = Method.evaluate_runtime text; val inner_update = method o update_dynamic_method (name, K (method ctxt)); in update_dynamic_method (name, inner_update) ctxt end; fun rep [] x = Scan.succeed [] x - | rep (m :: ms) x = ((method_text >> pair m) ::: rep ms) x; + | rep (m :: ms) x = ((Method.text_closure >> pair m) ::: rep ms) x; in rep method_args >> fold bind_method end; fun gen_method add_fixes name vars uses declares methods source lthy = @@ -256,7 +207,7 @@ |> put_recursive_method (full_name, fn _ => fn _ => Method.fail); val (text, src) = - read_closure (Config.put Proof_Context.dynamic_facts_dummy true lthy3) source; + Method.read_closure (Config.put Proof_Context.dynamic_facts_dummy true lthy3) source; val morphism = Variable.export_morphism lthy3 @@ -273,7 +224,7 @@ |> Proof_Context.restore_naming lthy |> put_closure name {vars = term_args', named_thms = named_thms, methods = method_args, body = text'} - |> Method.local_setup name + |> Method.local_setup name (Args.context :|-- (fn ctxt => let val {body, vars, ...} = get_closure ctxt full_name in parser vars (recursive_method full_name vars body) end)) ""

--- a/src/HOL/Isar_Examples/Higher_Order_Logic.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/Isar_Examples/Higher_Order_Logic.thy Thu Jul 21 10:52:27 2016 +0200 @@ -82,7 +82,7 @@ assumes "\<bottom>" shows A proof - - from \<open>\<bottom>\<close> have "\<forall>A. A" unfolding false_def . + from \<open>\<bottom>\<close> have "\<forall>A. A" by (simp only: false_def) then show A .. qed @@ -108,7 +108,7 @@ assumes "\<not> A" and A shows B proof - - from \<open>\<not> A\<close> have "A \<longrightarrow> \<bottom>" unfolding not_def . + from \<open>\<not> A\<close> have "A \<longrightarrow> \<bottom>" by (simp only: not_def) from this and \<open>A\<close> have "\<bottom>" .. then show B .. qed

--- a/src/HOL/ex/Adhoc_Overloading_Examples.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/HOL/ex/Adhoc_Overloading_Examples.thy Thu Jul 21 10:52:27 2016 +0200 @@ -7,8 +7,8 @@ theory Adhoc_Overloading_Examples imports Main + "~~/src/HOL/Library/Infinite_Set" "~~/src/Tools/Adhoc_Overloading" - "~~/src/HOL/Library/Infinite_Set" begin text \<open>Adhoc overloading allows to overload a constant depending on

--- a/src/Pure/General/symbol.scala Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Pure/General/symbol.scala Thu Jul 21 10:52:27 2016 +0200 @@ -458,8 +458,9 @@ val symbolic = recode_set((for { (sym, _) <- symbols; if raw_symbolic(sym) } yield sym): _*) - /* comment */ + /* misc symbols */ + val newline_decoded = decode(newline) val comment_decoded = decode(comment) @@ -526,7 +527,15 @@ def is_blank(sym: Symbol): Boolean = symbols.blanks.contains(sym) - /* comment */ + /* misc symbols */ + + val newline: Symbol = "\\<newline>" + def newline_decoded: Symbol = symbols.newline_decoded + + def print_newlines(str: String): String = + if (str.contains('\n')) + (for (s <- iterator(str)) yield { if (s == "\n") newline_decoded else s }).mkString + else str val comment: Symbol = "\\<comment>" def comment_decoded: Symbol = symbols.comment_decoded

--- a/src/Pure/Isar/method.ML Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Pure/Isar/method.ML Thu Jul 21 10:52:27 2016 +0200 @@ -92,6 +92,7 @@ val RUNTIME: context_tactic -> context_tactic val sleep: Time.time -> context_tactic val evaluate: text -> Proof.context -> method + val evaluate_runtime: text -> Proof.context -> method type modifier = {init: Proof.context -> Proof.context, attribute: attribute, pos: Position.T} val modifier: attribute -> Position.T -> modifier val old_section_parser: bool Config.T @@ -103,6 +104,10 @@ val report: text_range -> unit val parser: int -> text_range parser val parse: text_range parser + val read: Proof.context -> Token.src -> text + val read_closure: Proof.context -> Token.src -> text * Token.src + val read_closure_input: Proof.context -> Input.source -> text * Token.src + val text_closure: text context_parser end; structure Method: METHOD = @@ -576,6 +581,18 @@ end; +fun evaluate_runtime text ctxt = + let + val text' = + text |> (map_source o map o Token.map_facts) + (fn SOME name => + (case Proof_Context.lookup_fact ctxt name of + SOME {dynamic = true, thms} => K thms + | _ => I) + | NONE => I); + val ctxt' = Config.put closure false ctxt; + in fn facts => RUNTIME (fn st => evaluate text' ctxt' facts st) end; + (** concrete syntax **) @@ -743,6 +760,38 @@ end; +(* read method text *) + +fun read ctxt src = + (case Scan.read Token.stopper (Parse.!!! (parser 0 --| Scan.ahead Parse.eof)) src of + SOME (text, range) => + if checked_text text then text + else (report (text, range); check_text ctxt text) + | NONE => error ("Failed to parse method" ^ Position.here (#1 (Token.range_of src)))); + +fun read_closure ctxt src0 = + let + val src1 = map Token.init_assignable src0; + val text = read ctxt src1 |> map_source (method_closure ctxt); + val src2 = map Token.closure src1; + in (text, src2) end; + +fun read_closure_input ctxt = + Input.source_explode #> + Token.read_no_commands (Thy_Header.get_keywords' ctxt) (Scan.one Token.not_eof) #> + read_closure ctxt; + +val text_closure = + Args.context -- Scan.lift (Parse.token Parse.text) >> (fn (ctxt, tok) => + (case Token.get_value tok of + SOME (Token.Source src) => read ctxt src + | _ => + let + val (text, src) = read_closure_input ctxt (Token.input_of tok); + val _ = Token.assign (SOME (Token.Source src)) tok; + in text end)); + + (* theory setup *) val _ = Theory.setup @@ -792,7 +841,11 @@ setup @{binding tactic} (parse_tactic >> (K o METHOD)) "ML tactic as proof method" #> setup @{binding raw_tactic} (parse_tactic >> (fn tac => fn _ => CONTEXT_TACTIC o tac)) - "ML tactic as raw proof method"); + "ML tactic as raw proof method" #> + setup @{binding use} + (Attrib.thms -- (Scan.lift (Parse.$$$ "in") |-- text_closure) >> + (fn (thms, text) => fn ctxt => fn _ => evaluate_runtime text ctxt thms)) + "indicate method facts and context for method expression"); (*final declarations of this structure!*)

--- a/src/Pure/Isar/outer_syntax.scala Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Pure/Isar/outer_syntax.scala Thu Jul 21 10:52:27 2016 +0200 @@ -91,6 +91,8 @@ val keywords1 = keywords + (name, kind, tags) val completion1 = if (replace == Some("")) completion + else if (replace.isEmpty && Keyword.theory_block.contains(kind)) + completion + (name, name + "\nbegin\n\u0007\nend") + (name, name) else completion + (name, replace getOrElse name) new Outer_Syntax(keywords1, completion1, language_context, true) }

--- a/src/Pure/simplifier.ML Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Pure/simplifier.ML Thu Jul 21 10:52:27 2016 +0200 @@ -69,7 +69,10 @@ val simp_modifiers': Method.modifier parser list val simp_modifiers: Method.modifier parser list val method_setup: Method.modifier parser list -> theory -> theory - val easy_setup: thm -> thm list -> theory -> theory + val unsafe_solver_tac: Proof.context -> int -> tactic + val unsafe_solver: solver + val safe_solver_tac: Proof.context -> int -> tactic + val safe_solver: solver end; structure Simplifier: SIMPLIFIER = @@ -376,31 +379,22 @@ (CHANGED_PROP o PARALLEL_ALLGOALS o tac) ctxt)) "simplification (all goals)"; -fun easy_setup reflect trivs = method_setup [] #> Context.theory_map (map_ss (fn ctxt0 => - let - val trivialities = Drule.reflexive_thm :: trivs; - - fun unsafe_solver_tac ctxt = - FIRST' [resolve_tac ctxt (trivialities @ Raw_Simplifier.prems_of ctxt), assume_tac ctxt]; - val unsafe_solver = mk_solver "easy unsafe" unsafe_solver_tac; - - (*no premature instantiation of variables during simplification*) - fun safe_solver_tac ctxt = - FIRST' [match_tac ctxt (trivialities @ Raw_Simplifier.prems_of ctxt), eq_assume_tac]; - val safe_solver = mk_solver "easy safe" safe_solver_tac; +fun unsafe_solver_tac ctxt = + FIRST' [resolve_tac ctxt (Drule.reflexive_thm :: Raw_Simplifier.prems_of ctxt), assume_tac ctxt]; +val unsafe_solver = mk_solver "Pure unsafe" unsafe_solver_tac; - fun mk_eq thm = - if can Logic.dest_equals (Thm.concl_of thm) then [thm] - else [thm RS reflect] handle THM _ => []; +(*no premature instantiation of variables during simplification*) +fun safe_solver_tac ctxt = + FIRST' [match_tac ctxt (Drule.reflexive_thm :: Raw_Simplifier.prems_of ctxt), eq_assume_tac]; +val safe_solver = mk_solver "Pure safe" safe_solver_tac; - fun mksimps thm = mk_eq (Thm.forall_elim_vars (Thm.maxidx_of thm + 1) thm); - in - empty_simpset ctxt0 - setSSolver safe_solver - setSolver unsafe_solver - |> set_subgoaler asm_simp_tac - |> set_mksimps (K mksimps) - end)); +val _ = + Theory.setup + (method_setup [] #> Context.theory_map (map_ss (fn ctxt => + empty_simpset ctxt + setSSolver safe_solver + setSolver unsafe_solver + |> set_subgoaler asm_simp_tac))); end;

--- a/src/Sequents/LK.thy Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Sequents/LK.thy Thu Jul 21 10:52:27 2016 +0200 @@ -205,7 +205,7 @@ setup \<open>Simplifier.method_setup []\<close> -text \<open>To create substition rules\<close> +text \<open>To create substitution rules\<close> lemma eq_imp_subst: "\<turnstile> a = b \<Longrightarrow> $H, A(a), $G \<turnstile> $E, A(b), $F" by simp

--- a/src/Tools/jEdit/src/completion_popup.scala Wed Jul 20 14:52:09 2016 +0200 +++ b/src/Tools/jEdit/src/completion_popup.scala Thu Jul 21 10:52:27 2016 +0200 @@ -35,8 +35,8 @@ private val html = item.description match { case a :: bs => - "<html><b>" + HTML.output(a) + "</b>" + - HTML.output(bs.map(" " + _).mkString) + "</html>" + "<html><b>" + HTML.output(Symbol.print_newlines(a)) + "</b>" + + HTML.output(bs.map(b => " " + Symbol.print_newlines(b)).mkString) + "</html>" case Nil => "" } override def toString: String = html