|Edit this box|
In the framework of transformational-generative grammar (of which Government and Binding Theory and Minimalism are recent developments), the structure of a sentence is represented by phrase structure trees, otherwise known as phrase markers or tree diagrams. Such trees provide information about the sentences they represent by showing how, starting from an initial category S (or, for ID/LP grammar, Z), the various syntactic categories (e.g. noun phrase, verb phrase, etc.) are formed.
There are various theories as to how best to make grammars such that by systematic application of the rules, one can arrive at every phrase marker in a language (and hence every sentence in the language). The most common are Phrase structure grammars and ID/LP grammars, the latter having a slight explanatory advantage over the former.
In formal logic, computer science and linguistics , the term syntax is used to denote the literal text of something (that thing is typically called an expression) written in a formal language, programming language or natural language. This is opposed to its semantics or meaning. The syntax of expressions can be specified in different ways (for example, as parse trees--see below) whereas semantics is some assignment of meanings to expressions.
The analysis of programming language syntax usually entails the transformation of a linear sequence of tokens (a token is akin to an individual word or punctuation mark in a natural language) into a hierarchical syntax tree (abstract syntax trees are one convenient form of syntax tree). This process, called parsing, is in some respects analogous to syntactic analysis in linguistics; in fact, certain concepts, such as the Chomsky hierarchy and context-free grammars, are common to the study of syntax in both linguistics and computer science. However, the applications of these concepts vary widely between the two fields, and the practical resemblances are small.