Regular grammar

In theoretical computer science and formal language theory, a regular grammar is a formal grammar that is right-regular or left-regular. Every regular grammar describes a regular language.

Strictly regular grammars

A right regular grammar (also called right linear grammar) is a formal grammar (N, Σ, P, S) such that all the production rules in P are of one of the following forms:

  1. Aa, where A is a non-terminal in N and a is a terminal in Σ
  2. AaB, where A and B are non-terminals in N and a is in Σ
  3. A → ε, where A is in N and ε denotes the empty string, i.e. the string of length 0.

In a left regular grammar (also called left linear grammar), all rules obey the forms

  1. Aa, where A is a non-terminal in N and a is a terminal in Σ
  2. ABa, where A and B are in N and a is in Σ
  3. A → ε, where A is in N and ε is the empty string.

A regular grammar is a left or right regular grammar.

Some textbooks and articles disallow empty production rules, and assume that the empty string is not present in languages.

Extended regular grammars

An extended right regular grammar is one in which all rules obey one of

  1. Aw, where A is a non-terminal in N and w is in a (possibly empty) string of terminals Σ*
  2. AwB, where A and B are in N and w is in Σ*.

Some authors call this type of grammar a right regular grammar (or right linear grammar)[1] and the type above a strictly right regular grammar (or strictly right linear grammar).[2]

An extended left regular grammar is one in which all rules obey one of

  1. Aw, where A is a non-terminal in N and w is in Σ*
  2. ABw, where A and B are in N and w is in Σ*.

Examples

An example of a right regular grammar G with N = {S, A}, Σ = {a, b, c}, P consists of the following rules

S aS
S bA
A ε
A cA

and S is the start symbol. This grammar describes the same language as the regular expression a*bc*, viz. the set of all strings consisting of arbitrarily many "a"s, followed by a single "b", followed by arbitrarily many "c"s.

A somewhat longer but more explicit extended right regular grammar G for the same regular expression is given by N = {S, A, B, C}, Σ = {a, b, c}, where P consists of the following rules:

S A
A aA
A B
B bC
C ε
C cC

...where each uppercase letter corresponds to phrases starting at the next position in the regular expression.

As an example from the area of programming languages, the set of all strings denoting a floating point number can be described by an extended right regular grammar G with N = {S, A,B,C,D,E,F}, Σ = {0,1,2,3,4,5,6,7,8,9,+,-,.,e}, where S is the start symbol, and P consists of the following rules:

S +A      A 0A      B 0C      C 0C      D +E      E 0F      F 0F
S -AA 1AB 1CC 1CD -EE 1FF 1F
S AA 2AB 2CC 2CD EE 2FF 2F
A 3AB 3CC 3CE 3FF 3F
A 4AB 4CC 4CE 4FF 4F
A 5AB 5CC 5CE 5FF 5F
A 6AB 6CC 6CE 6FF 6F
A 7AB 7CC 7CE 7FF 7F
A 8AB 8CC 8CE 8FF 8F
A 9AB 9CC 9CE 9FF 9F
A .BC eDF ε
A BC ε

Expressive power

There is a direct one-to-one correspondence between the rules of a (strictly) right regular grammar and those of a nondeterministic finite automaton, such that the grammar generates exactly the language the automaton accepts.[3] Hence, the right regular grammars generate exactly all regular languages. The left regular grammars describe the reverses of all such languages, that is, exactly the regular languages as well.

Every strict right regular grammar is extended right regular, while every extended right regular grammar can be made strict by inserting new nonterminals, such that the result generates the same language; hence, extended right regular grammars generate the regular languages as well. Analogously, so do the extended left regular grammars.

If empty productions are disallowed, only all regular languages that do not include the empty string can be generated.[4]

While regular grammars can only describe regular languages, the converse is not true: regular languages can also be described by non-regular grammars.

Mixing left and right regular rules

If mixing of left-regular and right-regular rules is allowed, we still have a linear grammar, but not necessarily a regular one. What is more, such a grammar need not generate a regular language: all linear grammars can be easily brought into this form, and hence, such grammars can generate exactly all linear languages, including nonregular ones.

For instance, the grammar G with N = {S, A}, Σ = {a, b}, P with start symbol S and rules

S aA
A Sb
S ε

generates , the paradigmatic non-regular linear language.

See also

  • Regular expression, a compact notation for regular grammars
  • Regular tree grammar, a generalization from strings to trees
  • Prefix grammar
  • Chomsky hierarchy
  • Perrin, Dominique (1990), "Finite Automata", in Leeuwen, Jan van (ed.), Formal Models and Semantics, Handbook of Theoretical Computer Science, B, Elsevier, pp. 1–58
  • Pin, Jean-Éric (Oct 2012). Mathematical Foundations of Automata Theory (PDF)., chapter III

References

  1. John E. Hopcroft and Jeffrey D. Ullman (1979). Introduction to Automata Theory, Languages, and Computation. Reading/MA: Addison-Wesley. ISBN 0-201-02988-X. Here: p.217 (left, right regular grammars as subclasses of context-free grammars), p.79 (context-free grammars)
  2. Hopcroft and Ullman 1979 (p.229, exercise 9.2) call it a normal form for right-linear grammars.
  3. Hopcroft and Ullman 1979, p.218-219, Theorem 9.1 and 9.2
  4. Hopcroft and Ullman 1979, p.229, Exercise 9.2
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.