An interpretation is an assignment of meaning to the symbols of a formal language. Many formal languages used in mathematics, logic, and theoretical computer science are defined in solely syntactic terms, and as such do not have any meaning until they are given some interpretation. The general study of interpretations of formal languages is called formal semantics.
The most commonly studied formal logics are propositional logic, predicate logic and their modal analogs, and for these there are standard ways of presenting an interpretation. In these contexts an interpretation is a function that provides the extension of symbols and strings of symbols of an object language. For example, an interpretation function could take the predicate T (for "tall") and assign it the extension {a} (for "Abraham Lincoln"). Note that all our interpretation does is assign the extension {a} to the non-logical constant T, and does not make a claim about whether T is to stand for tall and 'a' for Abraham Lincoln. Nor does logical interpretation have anything to say about logical connectives like 'and', 'or' and 'not'. Though we may take these symbols to stand for certain things or concepts, this is not determined by the interpretation function.
An interpretation often (but not always) provides a way to determine the truth values of sentences in a language. If a given interpretation assigns the value True to a sentence or theory, the interpretation is called a model of that sentence or theory.
A formal language consists of a fixed collection of sentences (also called words or formulas, depending on the context) composed from a fixed set of letters or symbols. The inventory from which these letters are taken is called the alphabet over which the language is defined. To distinguish the strings of symbols that are in a formal language from arbitrary strings of symbols, the former are sometimes called well-formed formulæ (wff). The essential feature of a formal language is that its syntax can be defined without reference to interpretation. For example, we can determine that (P or Q) is a well-formed formula even without knowing whether it is true or false.
A formal language can be defined with the alphabet , and with a word being in if it begins with and is composed solely of the symbols and .
A possible interpretation of could assign the decimal digit '1' to and '0' to . Then would denote 101 under this interpretation of .
In the specific cases of propositional logic and predicate logic, the formal languages considered have alphabets that are divided into two sets: the logical symbols (logical constants) and the non-logical symbols. The idea behind this terminology is that logical symbols have the same meaning regardless of the subject matter being studied, while non-logical symbols change in meaning depending on the area of investigation.
Logical constants are always given the same meaning by every interpretation of the standard kind, so that only the meanings of the non-logical symbols are changed. Logical constants include quantifier symbols ? ("all") and ? ("some"), symbols for logical connectives ? ("and"), ? ("or"), ¬ ("not"), parentheses and other grouping symbols, and (in many treatments) the equality symbol =.
Many of the commonly studied interpretations associate each sentence in a formal language with a single truth value, either True or False. These interpretations are called truth functional;^{[dubious - discuss]} they include the usual interpretations of propositional and first-order logic. The sentences that are made true by a particular assignment are said to be satisfied by that assignment.
No sentence can be made both true and false by the same interpretation,^{[dubious - discuss]} but it is possible that the truth value of the same sentence can be different under different interpretations. A sentence is consistent if it is true under at least one interpretation; otherwise it is inconsistent. A sentence ? is said to be logically valid if it is satisfied by every interpretation (if ? is satisfied by every interpretation that satisfies ? then ? is said to be a logical consequence of ?).
Some of the logical symbols of a language (other than quantifiers) are truth-functional connectives that represent truth functions -- functions that take truth values as arguments and return truth values as outputs (in other words, these are operations on truth values of sentences).
The truth-functional connectives enable compound sentences to be built up from simpler sentences. In this way, the truth value of the compound sentence is defined as a certain truth function of the truth values of the simpler sentences. The connectives are usually taken to be logical constants, meaning that the meaning of the connectives is always the same, independent of what interpretations are given to the other symbols in a formula.
This is how we define logical connectives in propositional logic:
So under a given interpretation of all the sentence letters ? and ? (i.e., after assigning a truth-value to each sentence letter), we can determine the truth-values of all formulas that have them as constituents, as a function of the logical connectives. The following table shows how this kind of thing looks. The first two columns show the truth-values of the sentence letters as determined by the four possible interpretations. The other columns show the truth-values of formulas built from these sentence letters, with truth-values determined recursively.
Interpretation | ? | ? | ¬? | (? ? ?) | (? ? ?) | (? -> ?) | (? ?) |
---|---|---|---|---|---|---|---|
#1 | T | T | F | T | T | T | T |
#2 | T | F | F | F | T | F | F |
#3 | F | T | T | F | T | T | F |
#4 | F | F | T | F | F | T | T |
Now it is easier to see what makes a formula logically valid. Take the formula F: (? ? ¬?). If our interpretation function makes ? True, then ¬? is made False by the negation connective. Since the disjunct ? of F is True under that interpretation, F is True. Now the only other possible interpretation of ? makes it False, and if so, ¬? is made True by the negation function. That would make F True again, since one of Fs disjuncts, ¬?, would be true under this interpretation. Since these two interpretations for F are the only possible logical interpretations, and since F comes out True for both, we say that it is logically valid or tautologous.
An interpretation of a theory is the relationship between a theory and some subject matter when there is a many-to-one correspondence between certain elementary statements of the theory, and certain statements related to the subject matter. If every elementary statement in the theory has a correspondent it is called a full interpretation, otherwise it is called a partial interpretation.^{[1]}
The formal language for propositional logic consists of formulas built up from propositional symbols (also called sentential symbols, sentential variables, and propositional variables) and logical connectives. The only non-logical symbols in a formal language for propositional logic are the propositional symbols, which are often denoted by capital letters. To make the formal language precise, a specific set of propositional symbols must be fixed.
The standard kind of interpretation in this setting is a function that maps each propositional symbol to one of the truth values true and false. This function is known as a truth assignment or valuation function. In many presentations, it is literally a truth value that is assigned, but some presentations assign truthbearers instead.
For a language with n distinct propositional variables there are 2^{n} distinct possible interpretations. For any particular variable a, for example, there are 2^{1}=2 possible interpretations: 1) a is assigned T, or 2) a is assigned F. For the pair a, b there are 2^{2}=4 possible interpretations: 1) both are assigned T, 2) both are assigned F, 3) a is assigned T and b is assigned F, or 4) a is assigned F and b is assigned T.
Given any truth assignment for a set of propositional symbols, there is a unique extension to an interpretation for all the propositional formulas built up from those variables. This extended interpretation is defined inductively, using the truth-table definitions of the logical connectives discussed above.
Unlike propositional logic, where every language is the same apart from a choice of a different set of propositional variables, there are many different first-order languages. Each first-order language is defined by a signature. The signature consists of a set of non-logical symbols and an identification of each of these symbols as a constant symbol, a function symbol, or a predicate symbol. In the case of function and predicate symbols, a natural number arity is also assigned. The alphabet for the formal language consists of logical constants, the equality relation symbol =, all the symbols from the signature, and an additional infinite set of symbols known as variables.
For example, in the language of rings, there are constant symbols 0 and 1, two binary function symbols + and ·, and no binary relation symbols. (Here the equality relation is taken as a logical constant.)
Again, we might define a first-order language L, as consisting of individual symbols a, b, and c; predicate symbols F, G, H, I and J; variables x, y, z; no function letters; no sentential symbols.
Given a signature ?, the corresponding formal language is known as the set of ?-formulas. Each ?-formula is built up out of atomic formulas by means of logical connectives; atomic formulas are built from terms using predicate symbols. The formal definition of the set of ?-formulas proceeds in the other direction: first, terms are assembled from the constant and function symbols together with the variables. Then, terms can be combined into an atomic formula using a predicate symbol (relation symbol) from the signature or the special predicate symbol "=" for equality (see the section "Interpreting equality" below). Finally, the formulas of the language are assembled from atomic formulas using the logical connectives and quantifiers.
To ascribe meaning to all sentences of a first-order language, the following information is needed.
An object carrying this information is known as a structure (of signature ?, or ?-structure, or L-structure), or as a "model".
The information specified in the interpretation provides enough information to give a truth value to any atomic formula, after each of its free variables, if any, has been replaced by an element of the domain. The truth value of an arbitrary sentence is then defined inductively using the T-schema, which is a definition of first-order semantics developed by Alfred Tarski. The T-schema interprets the logical connectives using truth tables, as discussed above. Thus, for example, is satisfied if and only if both ? and ? are satisfied.
This leaves the issue of how to interpret formulas of the form and . The domain of discourse forms the range for these quantifiers. The idea is that the sentence is true under an interpretation exactly when every substitution instance of ?(x), where x is replaced by some element of the domain, is satisfied. The formula is satisfied if there is at least one element d of the domain such that ?(d) is satisfied.
Strictly speaking, a substitution instance such as the formula ?(d) mentioned above is not a formula in the original formal language of ?, because d is an element of the domain. There are two ways of handling this technical issue. The first is to pass to a larger language in which each element of the domain is named by a constant symbol. The second is to add to the interpretation a function that assigns each variable to an element of the domain. Then the T-schema can quantify over variations of the original interpretation in which this variable assignment function is changed, instead of quantifying over substitution instances.
Some authors also admit propositional variables in first-order logic, which must then also be interpreted. A propositional variable can stand on its own as an atomic formula. The interpretation of a propositional variable is one of the two truth values true and false.^{[3]}
Because the first-order interpretations described here are defined in set theory, they do not associate each predicate symbol with a property^{[4]}(or relation), but rather with the extension of that property (or relation). In other words, these first-order interpretations are extensional^{[5]} not intensional.
An example of interpretation of the language L described above is as follows.
In the interpretation of L:
As stated above, a first-order interpretation is usually required to specify a nonempty set as the domain of discourse. The reason for this requirement is to guarantee that equivalences such as
where x is not a free variable of ?, are logically valid. This equivalence holds in every interpretation with a nonempty domain, but does not always hold when empty domains are permitted. For example, the equivalence
fails in any structure with an empty domain. Thus the proof theory of first-order logic becomes more complicated when empty structures are permitted. However, the gain in allowing them is negligible, as both the intended interpretations and the interesting interpretations of the theories people study have non-empty domains.^{[6]}^{[7]}
Empty relations do not cause any problem for first-order interpretations, because there is no similar notion of passing a relation symbol across a logical connective, enlarging its scope in the process. Thus it is acceptable for relation symbols to be interpreted as being identically false. However, the interpretation of a function symbol must always assign a well-defined and total function to the symbol.
The equality relation is often treated specially in first order logic and other predicate logics. There are two general approaches.
The first approach is to treat equality as no different than any other binary relation. In this case, if an equality symbol is included in the signature, it is usually necessary to add various axioms about equality to axiom systems (for example, the substitution axiom saying that if a = b and R(a) holds then R(b) holds as well). This approach to equality is most useful when studying signatures that do not include the equality relation, such as the signature for set theory or the signature for second-order arithmetic in which there is only an equality relation for numbers, but not an equality relation for set of numbers.
The second approach is to treat the equality relation symbol as a logical constant that must be interpreted by the real equality relation in any interpretation. An interpretation that interprets equality this way is known as a normal model, so this second approach is the same as only studying interpretations that happen to be normal models. The advantage of this approach is that the axioms related to equality are automatically satisfied by every normal model, and so they do not need to be explicitly included in first-order theories when equality is treated this way. This second approach is sometimes called first order logic with equality, but many authors adopt it for the general study of first-order logic without comment.
There are a few other reasons to restrict study of first-order logic to normal models. First, it is known that any first-order interpretation in which equality is interpreted by an equivalence relation and satisfies the substitution axioms for equality can be cut down to an elementarily equivalent interpretation on a subset of the original domain. Thus there is little additional generality in studying non-normal models. Second, if non-normal models are considered, then every consistent theory has an infinite model; this affects the statements of results such as the Löwenheim-Skolem theorem, which are usually stated under the assumption that only normal models are considered.
A generalization of first order logic considers languages with more than one sort of variables. The idea is different sorts of variables represent different types of objects. Every sort of variable can be quantified; thus an interpretation for a many-sorted language has a separate domain for each of the sorts of variables to range over (there is an infinite collection of variables of each of the different sorts). Function and relation symbols, in addition to having arities, are specified so that each of their arguments must come from a certain sort.
One example of many-sorted logic is for planar Euclidean geometry. There are two sorts; points and lines. There is an equality relation symbol for points, an equality relation symbol for lines, and a binary incidence relation E which takes one point variable and one line variable. The intended interpretation of this language has the point variables range over all points on the Euclidean plane, the line variable range over all lines on the plane, and the incidence relation E(p,l) holds if and only if point p is on line l.
A formal language for higher-order predicate logic looks much the same as a formal language for first-order logic. The difference is that there are now many different types of variables. Some variables correspond to elements of the domain, as in first-order logic. Other variables correspond to objects of higher type: subsets of the domain, functions from the domain, functions that take a subset of the domain and return a function from the domain to subsets of the domain, etc. All of these types of variables can be quantified.
There are two kinds of interpretations commonly employed for higher-order logic. Full semantics require that, once the domain of discourse is satisfied, the higher-order variables range over all possible elements of the correct type (all subsets of the domain, all functions from the domain to itself, etc.). Thus the specification of a full interpretation is the same as the specification of a first-order interpretation. Henkin semantics, which are essentially multi-sorted first-order semantics, require the interpretation to specify a separate domain for each type of higher-order variable to range over. Thus an interpretation in Henkin semantics includes a domain D, a collection of subsets of D, a collection of functions from D to D, etc. The relationship between these two semantics is an important topic in higher order logic.
The interpretations of propositional logic and predicate logic described above are not the only possible interpretations. In particular, there are other types of interpretations that are used in the study of non-classical logic (such as intuitionistic logic), and in the study of modal logic.
Interpretations used to study non-classical logic include topological models, Boolean-valued models, and Kripke models. Modal logic is also studied using Kripke models.
Many formal languages are associated with a particular interpretation that is used to motivate them. For example, the first-order signature for set theory includes only one binary relation, ?, which is intended to represent set membership, and the domain of discourse in a first-order theory of the natural numbers is intended to be the set of natural numbers.
The intended interpretation is called the standard model (a term introduced by Abraham Robinson in 1960).^{[8]} In the context of Peano arithmetic, it consists of the natural numbers with their ordinary arithmetical operations. All models that are isomorphic to the one just given are also called standard; these models all satisfy the Peano axioms. There are also non-standard models of the (first-order version of the) Peano axioms, which contain elements not correlated with any natural number.
While the intended interpretation can have no explicit indication in the strictly formal syntactical rules, it naturally affects the choice of the formation and transformation rules of the syntactical system. For example, primitive signs must permit expression of the concepts to be modeled; sentential formulas are chosen so that their counterparts in the intended interpretation are meaningful declarative sentences; primitive sentences need to come out as true sentences in the interpretation; rules of inference must be such that, if the sentence is directly derivable from a sentence , then turns out to be a true sentence, with meaning implication, as usual. These requirements ensure that all provable sentences also come out to be true.^{[9]}
Most formal systems have many more models than they were intended to have (the existence of non-standard models is an example). When we speak about 'models' in empirical sciences, we mean, if we want reality to be a model of our science, to speak about an intended model. A model in the empirical sciences is an intended factually-true descriptive interpretation (or in other contexts: a non-intended arbitrary interpretation used to clarify such an intended factually-true descriptive interpretation.) All models are interpretations that have the same domain of discourse as the intended one, but other assignments for non-logical constants.^{[10]}^{[page needed]}
Given a simple formal system (we shall call this one ) whose alphabet ? consists only of three symbols and whose formation rule for formulas is:
The single axiom schema of is:
A formal proof can be constructed as follows:
In this example the theorem produced " " can be interpreted as meaning "One plus three equals four." A different interpretation would be to read it backwards as "Four minus three equals one."^{[11]}^{[page needed]}
There are other uses of the term "interpretation" that are commonly used, which do not refer to the assignment of meanings to formal languages.
In model theory, a structure A is said to interpret a structure B if there is a definable subset D of A, and definable relations and functions on D, such that B is isomorphic to the structure with domain D and these functions and relations. In some settings, it is not the domain D that is used, but rather D modulo an equivalence relation definable in A. For additional information, see Interpretation (model theory).
A theory T is said to interpret another theory S if there is a finite extension by definitions T? of T such that S is contained in T?.