Open main menu

SCI/Specifications/SCI in action/Parser

< SCI‎ | Specifications‎ | SCI in action
Revision as of 19:17, 18 January 2009 by Timofonic (talk | contribs) (Merging of the SCI documentation. Work in progress.)

Document conversion incomplete. Work in progress.

The Parser

Vocabulary file formats

Originally written by Lars Skovlund

The main vocabulary (VOCAB.000)

The file begins with a list of 26 offsets. Each index corresponds to a letter in the (English) alphabet, and points to the first word starting with that letter. The offset is set to 0 if no words start with that letter. If an input word starts with an alphabetical letter, this table is used to speed up the vocabulary searching - though not strictly necessary, this speeds up the lookup process somewhat.

After the offset table are the actual words. A word definition consists of two parts: The actual text of the word, compressed in a special way, and a 24-bit (yes, three bytes) ID. The ID divided in 2 12-bit quantities, a word class (grammatically speaking) mask, and a group number. The class mask is used, among other things, for throwing away unnecessary words. "Take book", for instance, is a valid sentence in parser'ese, while it isn't in English.

The possible values are arranged as a bit field to allow for class masks, see later. Only one bit is actually tested by the interpreter. If a word class equals to 0xff ("anyword"), the word is excluded (allowing for parser'ese). The values go like this:

0x001
Number (not found in the vocabulary, set internally)
0x002
Special
0x004
Special
0x008
Special[1]
0x010
Preposition
0x020
Article
0x040
Qualifying adjective
0x080
Relative pronoun
0x100
Noun
0x200
Indicative verb )such as "is","went" as opposed to _do_ this or that, which is imperative)
0x400
Adverb
0x800
Imperative verb

The group number is used to implement synonyms (words with the same meaning), as well as by the Said instruction to identify words. There is also a way of using synonyms in code, see the appropriate document.

The compression works in this way: Each string starts with a byte-sized copy count. This many characters are retained from the previous string. The actual text comes after, in normal low-ascii. The last character in the text has its high bit set (no null termination!).

Here is an example of the compression scheme:

apple 0,appl\0xE5


The byte count is 0 because we assume that "apple" is the first word beginning with an a (not likely, though!). 0xE5 is 0x65 (the ascii value for 'e') | 0x80. Watch now the next word:

athlete 1,thlet\0xE5


Here, the initial letter is identical to that of its predecessor, so the copy count is 1. Another example:

atrocious 2,rociou\0xF3


The suffix vocabulary (VOCAB.901)

For the following section, a reference to a grammar book may be advisable.

The suffix vocabulary is structurally much simpler. It consists of variably-sized records with this layout:

NULL-TERMINATED Suffix string
WORD The class mask for the suffix
NULL-TERMINATED Reduced string
WORD The output word class


The suffix vocabulary is used by the interpreter in order to parse compound words, and other words which consist of more than one part. For instance, a simple plural noun like "enemies" is reduced to its singular form "enemy", "stunning" is converted to "stun" etc. The point is that the interpreter gets a second chance at figuring out the meaning if the word can not be identified as entered. A word which changes its class does might end up as a different word class, the correct class is always retained. Thus, "carefully", an adverb, is reduced to its adjectival form "careful", and found in the vocabulary as such, but it is still marked as an adverb.

The suffix vocabulary consists of variably-sized records with this layout:

NULL-TERMINATED Suffix string
WORD The output word class
NULL-TERMINATED Reduced string
WORD The allowed class mask for the reduced

An asterisk (*) represents the word stem. Taking the above example with "enemies", the interpreter finds this record:

*ies
0x100
*y
0x100

Word class 0x100 being a noun.

The interpreter then tries to replace "enemies" with "enemy" and finds that word in the vocabulary. "Enemy" is a noun (class 1), which it is also supposed to be, according to the suffix vocabulary. Since we succeeded, the word class is set to the output value (which is, incidentally, also 1).

Numbers

If the word turns out to be a number (written with numbers, that is), and that number is not listed explicitly in the vocabulary, it gets an ID of 0xFFD, and a word class of 0x100.

The tree vocabulary (VOCAB.900)

This vocabulary is used solely for building parse trees. It consists of a series of word values which end up in the data nodes on the tree. It doesn't make much sense without the original parsing code.

The black box: The magic behind Sierra's text parser

Original document by Lars Skovlund. Document incomplete by the author.

This document describes the process of parsing user input and relating it to game actions. This document does not describe the process of the user typing his command; only the "behind-the-scenes" work is described, hence the title.

The process of parsing is two-fold, mainly for speed reasons. The Parse kernel function takes the actual input string and generates a special "said" event (type 0x80) from it. This function is only called once per line. Parse can either accept or reject the input.

A rejection can only occur if Parse fails to identify a word in the sentence.

Even if Parse accepts the sentence, it does not need to make sense. Still, syntax checks are made - see later.

Assuming that the parsing succeeded, the User object (which encapsulates the parser) then goes on to call the relevant event handlers. These event hand- lerrs in turn call the Said kernel function. This function is potentially called hundreds or even thousands of times, so it must execute as quickly as possible. Said simply determines from the pre-parsed input line whether or not a specific command is desired.

The Parse function must always work on an internal copy of the actual string, because the user must be able to recall his exact last input using the F3 key. The parser's first step is to convert the input line to pure lower case. This is because the vocabulary words are entered in lower case. The parser then searches the main vocabulary (VOCAB.000), hoping to find the word.

This doesn't necessarily happen yet. Consider, for example, the meaning of the word "carefully", which does not appear in the vocabulary, but is found anyway. This is due to the so-called suffix vocabulary, which is discussed in another document.

If the word still can't be found, the interpreter copies the failing word into a buffer temporarily allocated on the heap (remember, the interpreter operates on its own local buffers). It then calls the Game::wordFail method which prints an appropriate message. The interpreter then deallocates the buffer and exits (it does, however, still return an event. The claimed property of that event is set to TRUE to indicate that the event has already been responded to (error message printed)).

If the interpreter succeeds in identifying all the words, it then goes on to check the syntax of the sentence - it builds a parse tree. See the appropri- ate document.

If the syntax of the sentence is invalid, the interpreter calls Game::syntaxFail, passing the entire input line. As for the error situation, the event is claimed.

As mentioned in the beginning of this text, this function generates an event. This event, apart from its type id, does not contain any data. Rather, all pertinent data is kept in the interpreter.

The Said kernel function is called for each command which the game might respond to at any given time. Its only parameter is a pointer to a said information block which resides in script space. This block is described below (see the Said specs section).

The Said function first does some sanity checking on the event pointer which Parse stored earlier. It must be a said event (type property), and it must not have been handled by an earlier call to Said (claimed property).

It then word-extends the passed said block into a temporary buffer (command codes are byte-sized, remember?). This is supposedly just for convenience/speed, and not really needed.

The Parse tree

This and the two following sections borrow some ideas and structures from abstract language theory. Readers might want to consider related literature.

Most of the information explained here was gathered by Lars Skovlund, and, before that, Dark Minister.

After tokenizing, looking up, and finally aliasing the data found in the parsed input string, the interpreter proceeds to build a parse tree Tπ from the input tokens


I := ω0, ω12 … ωn-1

where

  • ωj ∈ W
  • γj ∈ Γ
  • μj ∈ 2c
  • ωj = (γj, μj

With W being the set of all words, Γ being the set of all word groups, C being the set of all class masks {1,2,4,8,10,20,40,80,100}, γj being the word group ωj belongs to, and ωj being its class mask, as described above.

For the following sections, we define

  • group W → Γ.group : (γ,μ) (γ,μ) ↦ γ
  • classes: W → C.classes: (γ,μ) ↦ μ
  • Cx = {ω|ω ∈ class(ω)}

To do that, it uses the class masks M as input for a pushdown automaton (PDA) A built from a parser grammar; if M was accepted by A, the parse tree Tπ will be built from the matching syntax tree to represent the semantics.

The PDA is defined by a grammar G = (V,Σ,P,s) most of which, along with its semantics, is stored in vocab.900. This resource contains a parser rule at every 20 bytes, starting with a non-terminal symbol υ (one word) and a null-terminated list of up to five tuples σi,mi , both of which are words. In these tuples, mi is a terminal or non-terminal symbol (determined by σi ), and σi is the meaning of mi :

σi Type Meaning
0x141 Non-terminal Predicate part: his identifies the first part of a sentence
0x142 Non-terminal Subject part: This identifies the second part of a sentence
0x143 Non-terminal Suffix part: This identifies the third and last part of a sentence
0x144 Non-terminal Reference part: This identifies words that reference another word in the same sentence part
0x146 Terminal Match on class mask: Matches if mi ∈ classes(ωj)
0x14d Terminal Match on word group: Matches if mi = group(ωj)
0x154 Terminal "Force storage": Apparently, this was only used for debugging

With the notable exception of the first rule, these rules constitute P. V := {χ|∃R ∈ P.χ &exist R}; typically, V = {0χ12ƒ…0χ13ƒ⋅ s = m0 of the first rule encountered; in all games observed, it was set to 0x13c. Σ contains all word groups and class masks. For the sake of simplicity, we will consider rules matching composite class masks to be several rules. Here is a simplified example of what such a grammar might look like (the hexadecimal prefix '0x' is omitted for brevity):

In addition to this grammar, each right-hand non-terminal mi carries its semantic value ρi , which is not relevant for constructing a syntax tree, but must be considered for the semantic tree Tπ. These values were omitted in the example above. As in the example above, the grammar is a context-free (type 2) grammar, almost in Chomsky Normal Form (CNF) in SCI; constructing a grammar with CNF rules from it would be trivial.[2]

Obviously, G is an ambiguous grammar. In SCI, rule precedence is implied by rule order, so the resulting left derivation tree is well-defined (in the example, it would be defined by D0.[3]

Parser grammar example

G = ⟨{⟩12ƒ…13e,(C1,C2,C4,…,C100},P,13c)

13c 13b 134
13c 13b 13d 133
13c 13b 13d
13c 13c
13c 13b 13d 13b 13d
13b 131 134
13b 131 13d 13d
P = { 13b 131
13d 134
131 C80
133 C20
134 C10}

Semantics

This is important, since the parser does much more than just accept or discard input. Using the semantic tags applied to each non-terminal on the right-hand side of a rule, it constructs what I will call the semantic parse tree Tπ , which attempts to describe what the input means. For each non-terminal rule

r = υ0 ↦ υ1υ2…υn

there are semantic tags σr,1r,2…σr,n ∈ S , as explained above. Tπ is now constructed from the resulting derivation and the semantic tags assiociated with each non-terminal of the rule used. The construction algorithm is explained below with Tπ being constructed from nodes, which have the following structure:

NODE = {◇} ∪ S × V × (NODE ∪ Γ);*

Where S is the set of possible semantic values, and V is the set of non-terminals as defined in the grammar. We will also use the sequence γ0,&gamma1,&gamma2 … &gammak-1 which will represent the word groups the input tokens belonged to (in the exact order they were accepted), and the sequence r0,r1,r2 … rl-1 , which will be the list of rules used to create the left derivation tree as described in the previous section.


<syntax type="?">

             Helper function sci_said_recursive: S \times V \times (V \cup \Sigma)* \to \Node
             Parameters: s \in S, Rule r \in V \times (V \cup \Sigma): v0 \to v1 v2 ... vi
             cnmr = cnr
             \Node n := s, v0
             FOR j := 1 TO i
                       IF (vj \in \Sigma) THEN
                               n := n, \gammacn\gamma
                               cn\gamma := cn\gamma + 1
                       ELSE
                               cnoldr := cnr
                               cnr := cnr + 1
                               n := n, sci_said_recursive(\sigmarmr,j, rcnoldr)
                       FI
             ROF
             RETURN (n)


             Helper function get_children: \Node \to \Node*
                             get_children((s, v, n0, n1 ... nm)) := n0, n1 ... nm


             Algorithm SCI-SAID-TREE
             cn\gamma := 0;
             cnr := 1;
             ntemp := ntemp, SCI-SAID-RECURSIVE(0, r0)
             root(T\Pi) := (141, 13f, get_children(ntemp))

</syntax>

Here is an example, based on the previous one:

Example: Semantic tree example

  • k = 2
  • γ0 = 842
  • γ1 = 917
  • l = 4
  • r0 = 13c ↦ 13b 134
  • σr0,1 = 141
  • σr0,2 = 142
  • r1 = 13b ↦ 131
  • σ11,1 = 141
  • r2 = 131 ↦ C80
  • r3 = 134 ↦ C10

The resulting tree would look like this:

(141 13f
	(141 13b
		(141 131 842)
	)
	(142 134 917)
)


Notes

  1. The three special classes are apparently used for words with very specific semantics, such as "if", "not", "and" etc. It is unknown as of yet whether they receive special treatment by the parser.
  2. FreeSCI constructs a GNF (Greibach Normal Form) representation from these rules for parsing.
  3. In FreeSCI, you can use the ”parse” console command to retreive all possible left derivation trees.