/* Title: Pure/Isar/outer_syntax.scala
Author: Makarius
Isabelle/Isar outer syntax.
*/
package isabelle
import scala.util.parsing.input.{Reader, CharSequenceReader}
import scala.collection.mutable
object Outer_Syntax
{
def quote_string(str: String): String =
{
val result = new StringBuilder(str.length + 10)
result += '"'
for (s <- Symbol.iterator(str)) {
if (s.length == 1) {
val c = s(0)
if (c < 32 && c != YXML.X && c != YXML.Y || c == '\\' || c == '"') {
result += '\\'
if (c < 10) result += '0'
if (c < 100) result += '0'
result ++= (c.asInstanceOf[Int].toString)
}
else result += c
}
else result ++= s
}
result += '"'
result.toString
}
def init(): Outer_Syntax = new Outer_Syntax()
}
class Outer_Syntax private(
keywords: Map[String, String] = Map((";" -> Keyword.DIAG)),
lexicon: Scan.Lexicon = Scan.Lexicon.empty,
val completion: Completion = Completion.init())
{
def keyword_kind(name: String): Option[String] = keywords.get(name)
def + (name: String, kind: String, replace: String): Outer_Syntax =
new Outer_Syntax(
keywords + (name -> kind),
lexicon + name,
if (Keyword.control(kind)) completion else completion + (name, replace))
def + (name: String, kind: String): Outer_Syntax = this + (name, kind, name)
def + (name: String): Outer_Syntax = this + (name, Keyword.MINOR)
def is_command(name: String): Boolean =
keyword_kind(name) match {
case Some(kind) => kind != Keyword.MINOR
case None => false
}
def heading_level(name: String): Option[Int] =
name match {
// FIXME avoid hard-wired info!?
case "header" => Some(1)
case "chapter" => Some(2)
case "section" | "sect" => Some(3)
case "subsection" | "subsect" => Some(4)
case "subsubsection" | "subsubsect" => Some(5)
case _ =>
keyword_kind(name) match {
case Some(kind) if Keyword.theory(kind) => Some(6)
case _ => None
}
}
def heading_level(command: Command): Option[Int] =
heading_level(command.name)
/* tokenize */
def scan(input: Reader[Char]): List[Token] =
{
import lexicon._
parseAll(rep(token(is_command)), input) match {
case Success(tokens, _) => tokens
case _ => error("Unexpected failure of tokenizing input:\n" + input.source.toString)
}
}
def scan(input: CharSequence): List[Token] =
scan(new CharSequenceReader(input))
def scan_context(input: CharSequence, context: Scan.Context): (List[Token], Scan.Context) =
{
import lexicon._
var in: Reader[Char] = new CharSequenceReader(input)
val toks = new mutable.ListBuffer[Token]
var ctxt = context
while (!in.atEnd) {
parse(token_context(is_command, ctxt), in) match {
case Success((x, c), rest) => { toks += x; ctxt = c; in = rest }
case NoSuccess(_, rest) =>
error("Unexpected failure of tokenizing input:\n" + rest.source.toString)
}
}
(toks.toList, ctxt)
}
}