mirror of
https://github.com/GrammaticalFramework/gf-core.git
synced 2026-05-06 09:42:50 -06:00
GF/src is now for 2.9, and the new sources are in src-3.0 - keep it this way until the release of GF 3
This commit is contained in:
45
src-3.0/GF/CFGM/AbsCFG.hs
Normal file
45
src-3.0/GF/CFGM/AbsCFG.hs
Normal file
@@ -0,0 +1,45 @@
|
||||
module GF.CFGM.AbsCFG where
|
||||
|
||||
-- Haskell module generated by the BNF converter
|
||||
|
||||
newtype Ident = Ident String deriving (Eq,Ord,Show)
|
||||
newtype SingleQuoteString = SingleQuoteString String deriving (Eq,Ord,Show)
|
||||
data Grammars =
|
||||
Grammars [Grammar]
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Grammar =
|
||||
Grammar Ident [Flag] [Rule]
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Flag =
|
||||
StartCat Category
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Rule =
|
||||
Rule Fun Profiles Category [Symbol]
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Fun =
|
||||
Cons Ident
|
||||
| Coerce
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Profiles =
|
||||
Profiles [Profile]
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Profile =
|
||||
UnifyProfile [Integer]
|
||||
| ConstProfile Ident
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Symbol =
|
||||
CatS Category
|
||||
| TermS String
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
data Category =
|
||||
Category SingleQuoteString
|
||||
deriving (Eq,Ord,Show)
|
||||
|
||||
36
src-3.0/GF/CFGM/CFG.cf
Normal file
36
src-3.0/GF/CFGM/CFG.cf
Normal file
@@ -0,0 +1,36 @@
|
||||
entrypoints Grammars;
|
||||
|
||||
Grammars. Grammars ::= [Grammar];
|
||||
|
||||
Grammar. Grammar ::= "grammar" Ident [Flag] [Rule] "end" "grammar";
|
||||
separator Grammar "";
|
||||
|
||||
StartCat. Flag ::= "startcat" Category;
|
||||
terminator Flag ";";
|
||||
|
||||
Rule. Rule ::= Fun ":" Profiles "." Category "->" [Symbol];
|
||||
terminator Rule ";";
|
||||
|
||||
Cons. Fun ::= Ident ;
|
||||
Coerce. Fun ::= "_" ;
|
||||
|
||||
Profiles. Profiles ::= "[" [Profile] "]";
|
||||
|
||||
separator Profile ",";
|
||||
|
||||
UnifyProfile. Profile ::= "[" [Integer] "]";
|
||||
ConstProfile. Profile ::= Ident ;
|
||||
|
||||
separator Integer ",";
|
||||
|
||||
CatS. Symbol ::= Category;
|
||||
TermS. Symbol ::= String;
|
||||
|
||||
-- separator Symbol "";
|
||||
[]. [Symbol] ::= "." ;
|
||||
(:[]). [Symbol] ::= Symbol ;
|
||||
(:). [Symbol] ::= Symbol [Symbol] ;
|
||||
|
||||
Category. Category ::= SingleQuoteString ;
|
||||
|
||||
token SingleQuoteString '\'' ((char - ["'\\"]) | ('\\' ["'\\"]))* '\'' ;
|
||||
312
src-3.0/GF/CFGM/LexCFG.hs
Normal file
312
src-3.0/GF/CFGM/LexCFG.hs
Normal file
File diff suppressed because one or more lines are too long
135
src-3.0/GF/CFGM/LexCFG.x
Normal file
135
src-3.0/GF/CFGM/LexCFG.x
Normal file
@@ -0,0 +1,135 @@
|
||||
-- -*- haskell -*-
|
||||
-- This Alex file was machine-generated by the BNF converter
|
||||
{
|
||||
module LexCFG where
|
||||
|
||||
import ErrM
|
||||
|
||||
}
|
||||
|
||||
|
||||
$l = [a-zA-Z\192 - \255] # [\215 \247] -- isolatin1 letter FIXME
|
||||
$c = [A-Z\192-\221] # [\215] -- capital isolatin1 letter FIXME
|
||||
$s = [a-z\222-\255] # [\247] -- small isolatin1 letter FIXME
|
||||
$d = [0-9] -- digit
|
||||
$i = [$l $d _ '] -- identifier character
|
||||
$u = [\0-\255] -- universal: any character
|
||||
|
||||
@rsyms = -- reserved words consisting of special symbols
|
||||
\; | \: | \. | \- \> | \_ | \[ | \] | \,
|
||||
|
||||
:-
|
||||
|
||||
$white+ ;
|
||||
@rsyms { tok (\p s -> PT p (TS $ share s)) }
|
||||
\' ($u # [\' \\]| \\ [\' \\]) * \' { tok (\p s -> PT p (eitherResIdent (T_SingleQuoteString . share) s)) }
|
||||
|
||||
$l $i* { tok (\p s -> PT p (eitherResIdent (TV . share) s)) }
|
||||
\" ([$u # [\" \\ \n]] | (\\ (\" | \\ | \' | n | t)))* \"{ tok (\p s -> PT p (TL $ share $ unescapeInitTail s)) }
|
||||
|
||||
$d+ { tok (\p s -> PT p (TI $ share s)) }
|
||||
|
||||
|
||||
{
|
||||
|
||||
tok f p s = f p s
|
||||
|
||||
share :: String -> String
|
||||
share = id
|
||||
|
||||
data Tok =
|
||||
TS !String -- reserved words
|
||||
| TL !String -- string literals
|
||||
| TI !String -- integer literals
|
||||
| TV !String -- identifiers
|
||||
| TD !String -- double precision float literals
|
||||
| TC !String -- character literals
|
||||
| T_SingleQuoteString !String
|
||||
|
||||
deriving (Eq,Show,Ord)
|
||||
|
||||
data Token =
|
||||
PT Posn Tok
|
||||
| Err Posn
|
||||
deriving (Eq,Show,Ord)
|
||||
|
||||
tokenPos (PT (Pn _ l _) _ :_) = "line " ++ show l
|
||||
tokenPos (Err (Pn _ l _) :_) = "line " ++ show l
|
||||
tokenPos _ = "end of file"
|
||||
|
||||
posLineCol (Pn _ l c) = (l,c)
|
||||
mkPosToken t@(PT p _) = (posLineCol p, prToken t)
|
||||
|
||||
prToken t = case t of
|
||||
PT _ (TS s) -> s
|
||||
PT _ (TI s) -> s
|
||||
PT _ (TV s) -> s
|
||||
PT _ (TD s) -> s
|
||||
PT _ (TC s) -> s
|
||||
PT _ (T_SingleQuoteString s) -> s
|
||||
|
||||
_ -> show t
|
||||
|
||||
data BTree = N | B String Tok BTree BTree deriving (Show)
|
||||
|
||||
eitherResIdent :: (String -> Tok) -> String -> Tok
|
||||
eitherResIdent tv s = treeFind resWords
|
||||
where
|
||||
treeFind N = tv s
|
||||
treeFind (B a t left right) | s < a = treeFind left
|
||||
| s > a = treeFind right
|
||||
| s == a = t
|
||||
|
||||
resWords = b "grammar" (b "end" N N) (b "startcat" N N)
|
||||
where b s = B s (TS s)
|
||||
|
||||
unescapeInitTail :: String -> String
|
||||
unescapeInitTail = unesc . tail where
|
||||
unesc s = case s of
|
||||
'\\':c:cs | elem c ['\"', '\\', '\''] -> c : unesc cs
|
||||
'\\':'n':cs -> '\n' : unesc cs
|
||||
'\\':'t':cs -> '\t' : unesc cs
|
||||
'"':[] -> []
|
||||
c:cs -> c : unesc cs
|
||||
_ -> []
|
||||
|
||||
-------------------------------------------------------------------
|
||||
-- Alex wrapper code.
|
||||
-- A modified "posn" wrapper.
|
||||
-------------------------------------------------------------------
|
||||
|
||||
data Posn = Pn !Int !Int !Int
|
||||
deriving (Eq, Show,Ord)
|
||||
|
||||
alexStartPos :: Posn
|
||||
alexStartPos = Pn 0 1 1
|
||||
|
||||
alexMove :: Posn -> Char -> Posn
|
||||
alexMove (Pn a l c) '\t' = Pn (a+1) l (((c+7) `div` 8)*8+1)
|
||||
alexMove (Pn a l c) '\n' = Pn (a+1) (l+1) 1
|
||||
alexMove (Pn a l c) _ = Pn (a+1) l (c+1)
|
||||
|
||||
type AlexInput = (Posn, -- current position,
|
||||
Char, -- previous char
|
||||
String) -- current input string
|
||||
|
||||
tokens :: String -> [Token]
|
||||
tokens str = go (alexStartPos, '\n', str)
|
||||
where
|
||||
go :: (Posn, Char, String) -> [Token]
|
||||
go inp@(pos, _, str) =
|
||||
case alexScan inp 0 of
|
||||
AlexEOF -> []
|
||||
AlexError (pos, _, _) -> fail $ show pos ++ ": lexical error"
|
||||
AlexSkip inp' len -> go inp'
|
||||
AlexToken inp' len act -> act pos (take len str) : (go inp')
|
||||
|
||||
alexGetChar :: AlexInput -> Maybe (Char,AlexInput)
|
||||
alexGetChar (p, c, []) = Nothing
|
||||
alexGetChar (p, _, (c:s)) =
|
||||
let p' = alexMove p c
|
||||
in p' `seq` Just (c, (p', c, s))
|
||||
|
||||
alexInputPrevChar :: AlexInput -> Char
|
||||
alexInputPrevChar (p, c, s) = c
|
||||
}
|
||||
779
src-3.0/GF/CFGM/ParCFG.hs
Normal file
779
src-3.0/GF/CFGM/ParCFG.hs
Normal file
@@ -0,0 +1,779 @@
|
||||
{-# OPTIONS -fglasgow-exts -cpp #-}
|
||||
module GF.CFGM.ParCFG where
|
||||
import GF.CFGM.AbsCFG
|
||||
import GF.CFGM.LexCFG
|
||||
import GF.Data.ErrM
|
||||
import Array
|
||||
#if __GLASGOW_HASKELL__ >= 503
|
||||
import GHC.Exts
|
||||
#else
|
||||
import GlaExts
|
||||
#endif
|
||||
|
||||
-- parser produced by Happy Version 1.15
|
||||
|
||||
newtype HappyAbsSyn = HappyAbsSyn (() -> ())
|
||||
happyIn4 :: (Ident) -> (HappyAbsSyn )
|
||||
happyIn4 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn4 #-}
|
||||
happyOut4 :: (HappyAbsSyn ) -> (Ident)
|
||||
happyOut4 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut4 #-}
|
||||
happyIn5 :: (Integer) -> (HappyAbsSyn )
|
||||
happyIn5 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn5 #-}
|
||||
happyOut5 :: (HappyAbsSyn ) -> (Integer)
|
||||
happyOut5 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut5 #-}
|
||||
happyIn6 :: (String) -> (HappyAbsSyn )
|
||||
happyIn6 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn6 #-}
|
||||
happyOut6 :: (HappyAbsSyn ) -> (String)
|
||||
happyOut6 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut6 #-}
|
||||
happyIn7 :: (SingleQuoteString) -> (HappyAbsSyn )
|
||||
happyIn7 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn7 #-}
|
||||
happyOut7 :: (HappyAbsSyn ) -> (SingleQuoteString)
|
||||
happyOut7 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut7 #-}
|
||||
happyIn8 :: (Grammars) -> (HappyAbsSyn )
|
||||
happyIn8 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn8 #-}
|
||||
happyOut8 :: (HappyAbsSyn ) -> (Grammars)
|
||||
happyOut8 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut8 #-}
|
||||
happyIn9 :: (Grammar) -> (HappyAbsSyn )
|
||||
happyIn9 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn9 #-}
|
||||
happyOut9 :: (HappyAbsSyn ) -> (Grammar)
|
||||
happyOut9 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut9 #-}
|
||||
happyIn10 :: ([Grammar]) -> (HappyAbsSyn )
|
||||
happyIn10 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn10 #-}
|
||||
happyOut10 :: (HappyAbsSyn ) -> ([Grammar])
|
||||
happyOut10 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut10 #-}
|
||||
happyIn11 :: (Flag) -> (HappyAbsSyn )
|
||||
happyIn11 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn11 #-}
|
||||
happyOut11 :: (HappyAbsSyn ) -> (Flag)
|
||||
happyOut11 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut11 #-}
|
||||
happyIn12 :: ([Flag]) -> (HappyAbsSyn )
|
||||
happyIn12 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn12 #-}
|
||||
happyOut12 :: (HappyAbsSyn ) -> ([Flag])
|
||||
happyOut12 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut12 #-}
|
||||
happyIn13 :: (Rule) -> (HappyAbsSyn )
|
||||
happyIn13 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn13 #-}
|
||||
happyOut13 :: (HappyAbsSyn ) -> (Rule)
|
||||
happyOut13 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut13 #-}
|
||||
happyIn14 :: ([Rule]) -> (HappyAbsSyn )
|
||||
happyIn14 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn14 #-}
|
||||
happyOut14 :: (HappyAbsSyn ) -> ([Rule])
|
||||
happyOut14 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut14 #-}
|
||||
happyIn15 :: (Fun) -> (HappyAbsSyn )
|
||||
happyIn15 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn15 #-}
|
||||
happyOut15 :: (HappyAbsSyn ) -> (Fun)
|
||||
happyOut15 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut15 #-}
|
||||
happyIn16 :: (Profiles) -> (HappyAbsSyn )
|
||||
happyIn16 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn16 #-}
|
||||
happyOut16 :: (HappyAbsSyn ) -> (Profiles)
|
||||
happyOut16 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut16 #-}
|
||||
happyIn17 :: ([Profile]) -> (HappyAbsSyn )
|
||||
happyIn17 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn17 #-}
|
||||
happyOut17 :: (HappyAbsSyn ) -> ([Profile])
|
||||
happyOut17 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut17 #-}
|
||||
happyIn18 :: (Profile) -> (HappyAbsSyn )
|
||||
happyIn18 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn18 #-}
|
||||
happyOut18 :: (HappyAbsSyn ) -> (Profile)
|
||||
happyOut18 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut18 #-}
|
||||
happyIn19 :: ([Integer]) -> (HappyAbsSyn )
|
||||
happyIn19 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn19 #-}
|
||||
happyOut19 :: (HappyAbsSyn ) -> ([Integer])
|
||||
happyOut19 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut19 #-}
|
||||
happyIn20 :: (Symbol) -> (HappyAbsSyn )
|
||||
happyIn20 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn20 #-}
|
||||
happyOut20 :: (HappyAbsSyn ) -> (Symbol)
|
||||
happyOut20 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut20 #-}
|
||||
happyIn21 :: ([Symbol]) -> (HappyAbsSyn )
|
||||
happyIn21 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn21 #-}
|
||||
happyOut21 :: (HappyAbsSyn ) -> ([Symbol])
|
||||
happyOut21 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut21 #-}
|
||||
happyIn22 :: (Category) -> (HappyAbsSyn )
|
||||
happyIn22 x = unsafeCoerce# x
|
||||
{-# INLINE happyIn22 #-}
|
||||
happyOut22 :: (HappyAbsSyn ) -> (Category)
|
||||
happyOut22 x = unsafeCoerce# x
|
||||
{-# INLINE happyOut22 #-}
|
||||
happyInTok :: Token -> (HappyAbsSyn )
|
||||
happyInTok x = unsafeCoerce# x
|
||||
{-# INLINE happyInTok #-}
|
||||
happyOutTok :: (HappyAbsSyn ) -> Token
|
||||
happyOutTok x = unsafeCoerce# x
|
||||
{-# INLINE happyOutTok #-}
|
||||
|
||||
happyActOffsets :: HappyAddr
|
||||
happyActOffsets = HappyA# "\x00\x00\x36\x00\x00\x00\x29\x00\x35\x00\x00\x00\x32\x00\x00\x00\x30\x00\x38\x00\x19\x00\x2e\x00\x00\x00\x00\x00\x00\x00\x00\x00\x37\x00\x34\x00\x00\x00\x2d\x00\x00\x00\x00\x00\x2f\x00\x00\x00\x31\x00\xfd\xff\x00\x00\x2c\x00\x2a\x00\x23\x00\x22\x00\x2b\x00\x25\x00\x20\x00\x00\x00\xfd\xff\x00\x00\x00\x00\x00\x00\x17\x00\x1c\x00\x00\x00\x1c\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00"#
|
||||
|
||||
happyGotoOffsets :: HappyAddr
|
||||
happyGotoOffsets = HappyA# "\x28\x00\x00\x00\x00\x00\x00\x00\x1e\x00\x00\x00\x21\x00\x05\x00\x01\x00\x00\x00\x1d\x00\x04\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x14\x00\x00\x00\x00\x00\x0c\x00\x00\x00\x00\x00\x00\x00\x1a\x00\x03\x00\x00\x00\x00\x00\x00\x00\x00\x00\x0a\x00\x00\x00\x00\x00\x00\x00\x0d\x00\x02\x00\x00\x00\xff\xff\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00"#
|
||||
|
||||
happyDefActions :: HappyAddr
|
||||
happyDefActions = HappyA# "\xf8\xff\x00\x00\xfe\xff\x00\x00\xfa\xff\xf7\xff\x00\x00\xf5\xff\xf2\xff\x00\x00\x00\x00\x00\x00\xe0\xff\xf6\xff\xfb\xff\xf0\xff\x00\x00\x00\x00\xef\xff\x00\x00\xf4\xff\xf9\xff\x00\x00\xf1\xff\x00\x00\xed\xff\xe9\xff\x00\x00\xec\xff\xe8\xff\x00\x00\x00\x00\xe7\xff\x00\x00\xfd\xff\xed\xff\xee\xff\xeb\xff\xea\xff\xe8\xff\x00\x00\xe4\xff\xe2\xff\xf3\xff\xe5\xff\xe3\xff\xfc\xff\xe6\xff\xe1\xff"#
|
||||
|
||||
happyCheck :: HappyAddr
|
||||
happyCheck = HappyA# "\xff\xff\x02\x00\x03\x00\x06\x00\x02\x00\x03\x00\x03\x00\x03\x00\x07\x00\x0c\x00\x00\x00\x0a\x00\x00\x00\x08\x00\x01\x00\x10\x00\x11\x00\x12\x00\x10\x00\x11\x00\x12\x00\x12\x00\x12\x00\x0d\x00\x0e\x00\x0d\x00\x0e\x00\x01\x00\x0f\x00\x00\x00\x05\x00\x03\x00\x0c\x00\x00\x00\x09\x00\x05\x00\x0d\x00\x0c\x00\x09\x00\x07\x00\x0b\x00\x0f\x00\x0e\x00\x0f\x00\x04\x00\x08\x00\x06\x00\x04\x00\x0d\x00\x0f\x00\x08\x00\x07\x00\x03\x00\x06\x00\x02\x00\x0a\x00\x01\x00\x01\x00\x11\x00\x0b\x00\xff\xff\x0f\x00\x0c\x00\x0a\x00\xff\xff\xff\xff\x0c\x00\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff\xff"#
|
||||
|
||||
happyTable :: HappyAddr
|
||||
happyTable = HappyA# "\x00\x00\x29\x00\x0c\x00\x1e\x00\x29\x00\x0c\x00\x0c\x00\x0c\x00\x09\x00\x03\x00\x1a\x00\x0a\x00\x1a\x00\x08\x00\x20\x00\x2a\x00\x30\x00\x2c\x00\x2a\x00\x2b\x00\x2c\x00\x1f\x00\x0d\x00\x25\x00\x1c\x00\x1b\x00\x1c\x00\x20\x00\x2f\x00\x0f\x00\x13\x00\x2e\x00\x18\x00\x07\x00\x14\x00\x05\x00\x23\x00\x03\x00\x10\x00\x27\x00\x11\x00\x21\x00\x2f\x00\x0f\x00\x03\x00\x28\x00\x04\x00\x29\x00\x23\x00\x0f\x00\x24\x00\x25\x00\x1f\x00\x1a\x00\x17\x00\x16\x00\x18\x00\x15\x00\xff\xff\x0c\x00\x00\x00\x0f\x00\x03\x00\x07\x00\x00\x00\x00\x00\x03\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00"#
|
||||
|
||||
happyReduceArr = array (1, 31) [
|
||||
(1 , happyReduce_1),
|
||||
(2 , happyReduce_2),
|
||||
(3 , happyReduce_3),
|
||||
(4 , happyReduce_4),
|
||||
(5 , happyReduce_5),
|
||||
(6 , happyReduce_6),
|
||||
(7 , happyReduce_7),
|
||||
(8 , happyReduce_8),
|
||||
(9 , happyReduce_9),
|
||||
(10 , happyReduce_10),
|
||||
(11 , happyReduce_11),
|
||||
(12 , happyReduce_12),
|
||||
(13 , happyReduce_13),
|
||||
(14 , happyReduce_14),
|
||||
(15 , happyReduce_15),
|
||||
(16 , happyReduce_16),
|
||||
(17 , happyReduce_17),
|
||||
(18 , happyReduce_18),
|
||||
(19 , happyReduce_19),
|
||||
(20 , happyReduce_20),
|
||||
(21 , happyReduce_21),
|
||||
(22 , happyReduce_22),
|
||||
(23 , happyReduce_23),
|
||||
(24 , happyReduce_24),
|
||||
(25 , happyReduce_25),
|
||||
(26 , happyReduce_26),
|
||||
(27 , happyReduce_27),
|
||||
(28 , happyReduce_28),
|
||||
(29 , happyReduce_29),
|
||||
(30 , happyReduce_30),
|
||||
(31 , happyReduce_31)
|
||||
]
|
||||
|
||||
happy_n_terms = 18 :: Int
|
||||
happy_n_nonterms = 19 :: Int
|
||||
|
||||
happyReduce_1 = happySpecReduce_1 0# happyReduction_1
|
||||
happyReduction_1 happy_x_1
|
||||
= case happyOutTok happy_x_1 of { (PT _ (TV happy_var_1)) ->
|
||||
happyIn4
|
||||
(Ident happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_2 = happySpecReduce_1 1# happyReduction_2
|
||||
happyReduction_2 happy_x_1
|
||||
= case happyOutTok happy_x_1 of { (PT _ (TI happy_var_1)) ->
|
||||
happyIn5
|
||||
((read happy_var_1) :: Integer
|
||||
)}
|
||||
|
||||
happyReduce_3 = happySpecReduce_1 2# happyReduction_3
|
||||
happyReduction_3 happy_x_1
|
||||
= case happyOutTok happy_x_1 of { (PT _ (TL happy_var_1)) ->
|
||||
happyIn6
|
||||
(happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_4 = happySpecReduce_1 3# happyReduction_4
|
||||
happyReduction_4 happy_x_1
|
||||
= case happyOutTok happy_x_1 of { (PT _ (T_SingleQuoteString happy_var_1)) ->
|
||||
happyIn7
|
||||
(SingleQuoteString (happy_var_1)
|
||||
)}
|
||||
|
||||
happyReduce_5 = happySpecReduce_1 4# happyReduction_5
|
||||
happyReduction_5 happy_x_1
|
||||
= case happyOut10 happy_x_1 of { happy_var_1 ->
|
||||
happyIn8
|
||||
(Grammars (reverse happy_var_1)
|
||||
)}
|
||||
|
||||
happyReduce_6 = happyReduce 6# 5# happyReduction_6
|
||||
happyReduction_6 (happy_x_6 `HappyStk`
|
||||
happy_x_5 `HappyStk`
|
||||
happy_x_4 `HappyStk`
|
||||
happy_x_3 `HappyStk`
|
||||
happy_x_2 `HappyStk`
|
||||
happy_x_1 `HappyStk`
|
||||
happyRest)
|
||||
= case happyOut4 happy_x_2 of { happy_var_2 ->
|
||||
case happyOut12 happy_x_3 of { happy_var_3 ->
|
||||
case happyOut14 happy_x_4 of { happy_var_4 ->
|
||||
happyIn9
|
||||
(Grammar happy_var_2 (reverse happy_var_3) (reverse happy_var_4)
|
||||
) `HappyStk` happyRest}}}
|
||||
|
||||
happyReduce_7 = happySpecReduce_0 6# happyReduction_7
|
||||
happyReduction_7 = happyIn10
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_8 = happySpecReduce_2 6# happyReduction_8
|
||||
happyReduction_8 happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut10 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut9 happy_x_2 of { happy_var_2 ->
|
||||
happyIn10
|
||||
(flip (:) happy_var_1 happy_var_2
|
||||
)}}
|
||||
|
||||
happyReduce_9 = happySpecReduce_2 7# happyReduction_9
|
||||
happyReduction_9 happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut22 happy_x_2 of { happy_var_2 ->
|
||||
happyIn11
|
||||
(StartCat happy_var_2
|
||||
)}
|
||||
|
||||
happyReduce_10 = happySpecReduce_0 8# happyReduction_10
|
||||
happyReduction_10 = happyIn12
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_11 = happySpecReduce_3 8# happyReduction_11
|
||||
happyReduction_11 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut12 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut11 happy_x_2 of { happy_var_2 ->
|
||||
happyIn12
|
||||
(flip (:) happy_var_1 happy_var_2
|
||||
)}}
|
||||
|
||||
happyReduce_12 = happyReduce 7# 9# happyReduction_12
|
||||
happyReduction_12 (happy_x_7 `HappyStk`
|
||||
happy_x_6 `HappyStk`
|
||||
happy_x_5 `HappyStk`
|
||||
happy_x_4 `HappyStk`
|
||||
happy_x_3 `HappyStk`
|
||||
happy_x_2 `HappyStk`
|
||||
happy_x_1 `HappyStk`
|
||||
happyRest)
|
||||
= case happyOut15 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut16 happy_x_3 of { happy_var_3 ->
|
||||
case happyOut22 happy_x_5 of { happy_var_5 ->
|
||||
case happyOut21 happy_x_7 of { happy_var_7 ->
|
||||
happyIn13
|
||||
(Rule happy_var_1 happy_var_3 happy_var_5 happy_var_7
|
||||
) `HappyStk` happyRest}}}}
|
||||
|
||||
happyReduce_13 = happySpecReduce_0 10# happyReduction_13
|
||||
happyReduction_13 = happyIn14
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_14 = happySpecReduce_3 10# happyReduction_14
|
||||
happyReduction_14 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut14 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut13 happy_x_2 of { happy_var_2 ->
|
||||
happyIn14
|
||||
(flip (:) happy_var_1 happy_var_2
|
||||
)}}
|
||||
|
||||
happyReduce_15 = happySpecReduce_1 11# happyReduction_15
|
||||
happyReduction_15 happy_x_1
|
||||
= case happyOut4 happy_x_1 of { happy_var_1 ->
|
||||
happyIn15
|
||||
(Cons happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_16 = happySpecReduce_1 11# happyReduction_16
|
||||
happyReduction_16 happy_x_1
|
||||
= happyIn15
|
||||
(Coerce
|
||||
)
|
||||
|
||||
happyReduce_17 = happySpecReduce_3 12# happyReduction_17
|
||||
happyReduction_17 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut17 happy_x_2 of { happy_var_2 ->
|
||||
happyIn16
|
||||
(Profiles happy_var_2
|
||||
)}
|
||||
|
||||
happyReduce_18 = happySpecReduce_0 13# happyReduction_18
|
||||
happyReduction_18 = happyIn17
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_19 = happySpecReduce_1 13# happyReduction_19
|
||||
happyReduction_19 happy_x_1
|
||||
= case happyOut18 happy_x_1 of { happy_var_1 ->
|
||||
happyIn17
|
||||
((:[]) happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_20 = happySpecReduce_3 13# happyReduction_20
|
||||
happyReduction_20 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut18 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut17 happy_x_3 of { happy_var_3 ->
|
||||
happyIn17
|
||||
((:) happy_var_1 happy_var_3
|
||||
)}}
|
||||
|
||||
happyReduce_21 = happySpecReduce_3 14# happyReduction_21
|
||||
happyReduction_21 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut19 happy_x_2 of { happy_var_2 ->
|
||||
happyIn18
|
||||
(UnifyProfile happy_var_2
|
||||
)}
|
||||
|
||||
happyReduce_22 = happySpecReduce_1 14# happyReduction_22
|
||||
happyReduction_22 happy_x_1
|
||||
= case happyOut4 happy_x_1 of { happy_var_1 ->
|
||||
happyIn18
|
||||
(ConstProfile happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_23 = happySpecReduce_0 15# happyReduction_23
|
||||
happyReduction_23 = happyIn19
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_24 = happySpecReduce_1 15# happyReduction_24
|
||||
happyReduction_24 happy_x_1
|
||||
= case happyOut5 happy_x_1 of { happy_var_1 ->
|
||||
happyIn19
|
||||
((:[]) happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_25 = happySpecReduce_3 15# happyReduction_25
|
||||
happyReduction_25 happy_x_3
|
||||
happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut5 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut19 happy_x_3 of { happy_var_3 ->
|
||||
happyIn19
|
||||
((:) happy_var_1 happy_var_3
|
||||
)}}
|
||||
|
||||
happyReduce_26 = happySpecReduce_1 16# happyReduction_26
|
||||
happyReduction_26 happy_x_1
|
||||
= case happyOut22 happy_x_1 of { happy_var_1 ->
|
||||
happyIn20
|
||||
(CatS happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_27 = happySpecReduce_1 16# happyReduction_27
|
||||
happyReduction_27 happy_x_1
|
||||
= case happyOut6 happy_x_1 of { happy_var_1 ->
|
||||
happyIn20
|
||||
(TermS happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_28 = happySpecReduce_1 17# happyReduction_28
|
||||
happyReduction_28 happy_x_1
|
||||
= happyIn21
|
||||
([]
|
||||
)
|
||||
|
||||
happyReduce_29 = happySpecReduce_1 17# happyReduction_29
|
||||
happyReduction_29 happy_x_1
|
||||
= case happyOut20 happy_x_1 of { happy_var_1 ->
|
||||
happyIn21
|
||||
((:[]) happy_var_1
|
||||
)}
|
||||
|
||||
happyReduce_30 = happySpecReduce_2 17# happyReduction_30
|
||||
happyReduction_30 happy_x_2
|
||||
happy_x_1
|
||||
= case happyOut20 happy_x_1 of { happy_var_1 ->
|
||||
case happyOut21 happy_x_2 of { happy_var_2 ->
|
||||
happyIn21
|
||||
((:) happy_var_1 happy_var_2
|
||||
)}}
|
||||
|
||||
happyReduce_31 = happySpecReduce_1 18# happyReduction_31
|
||||
happyReduction_31 happy_x_1
|
||||
= case happyOut7 happy_x_1 of { happy_var_1 ->
|
||||
happyIn22
|
||||
(Category happy_var_1
|
||||
)}
|
||||
|
||||
happyNewToken action sts stk [] =
|
||||
happyDoAction 17# (error "reading EOF!") action sts stk []
|
||||
|
||||
happyNewToken action sts stk (tk:tks) =
|
||||
let cont i = happyDoAction i tk action sts stk tks in
|
||||
case tk of {
|
||||
PT _ (TS ";") -> cont 1#;
|
||||
PT _ (TS ":") -> cont 2#;
|
||||
PT _ (TS ".") -> cont 3#;
|
||||
PT _ (TS "->") -> cont 4#;
|
||||
PT _ (TS "_") -> cont 5#;
|
||||
PT _ (TS "[") -> cont 6#;
|
||||
PT _ (TS "]") -> cont 7#;
|
||||
PT _ (TS ",") -> cont 8#;
|
||||
PT _ (TS "end") -> cont 9#;
|
||||
PT _ (TS "grammar") -> cont 10#;
|
||||
PT _ (TS "startcat") -> cont 11#;
|
||||
PT _ (TV happy_dollar_dollar) -> cont 12#;
|
||||
PT _ (TI happy_dollar_dollar) -> cont 13#;
|
||||
PT _ (TL happy_dollar_dollar) -> cont 14#;
|
||||
PT _ (T_SingleQuoteString happy_dollar_dollar) -> cont 15#;
|
||||
_ -> cont 16#;
|
||||
_ -> happyError' (tk:tks)
|
||||
}
|
||||
|
||||
happyError_ tk tks = happyError' (tk:tks)
|
||||
|
||||
happyThen :: () => Err a -> (a -> Err b) -> Err b
|
||||
happyThen = (thenM)
|
||||
happyReturn :: () => a -> Err a
|
||||
happyReturn = (returnM)
|
||||
happyThen1 m k tks = (thenM) m (\a -> k a tks)
|
||||
happyReturn1 :: () => a -> b -> Err a
|
||||
happyReturn1 = \a tks -> (returnM) a
|
||||
happyError' :: () => [Token] -> Err a
|
||||
happyError' = happyError
|
||||
|
||||
pGrammars tks = happySomeParser where
|
||||
happySomeParser = happyThen (happyParse 0# tks) (\x -> happyReturn (happyOut8 x))
|
||||
|
||||
happySeq = happyDontSeq
|
||||
|
||||
returnM :: a -> Err a
|
||||
returnM = return
|
||||
|
||||
thenM :: Err a -> (a -> Err b) -> Err b
|
||||
thenM = (>>=)
|
||||
|
||||
happyError :: [Token] -> Err a
|
||||
happyError ts =
|
||||
Bad $ "syntax error at " ++ tokenPos ts ++ if null ts then [] else (" before " ++ unwords (map prToken (take 4 ts)))
|
||||
|
||||
myLexer = tokens
|
||||
{-# LINE 1 "GenericTemplate.hs" #-}
|
||||
-- $Id: ParCFG.hs,v 1.8 2005/05/17 14:04:37 bringert Exp $
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
{-# LINE 27 "GenericTemplate.hs" #-}
|
||||
|
||||
|
||||
|
||||
data Happy_IntList = HappyCons Int# Happy_IntList
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
infixr 9 `HappyStk`
|
||||
data HappyStk a = HappyStk a (HappyStk a)
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- starting the parse
|
||||
|
||||
happyParse start_state = happyNewToken start_state notHappyAtAll notHappyAtAll
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Accepting the parse
|
||||
|
||||
-- If the current token is 0#, it means we've just accepted a partial
|
||||
-- parse (a %partial parser). We must ignore the saved token on the top of
|
||||
-- the stack in this case.
|
||||
happyAccept 0# tk st sts (_ `HappyStk` ans `HappyStk` _) =
|
||||
happyReturn1 ans
|
||||
happyAccept j tk st sts (HappyStk ans _) =
|
||||
(happyTcHack j (happyTcHack st)) (happyReturn1 ans)
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Arrays only: do the next action
|
||||
|
||||
|
||||
|
||||
happyDoAction i tk st
|
||||
= {- nothing -}
|
||||
|
||||
|
||||
case action of
|
||||
0# -> {- nothing -}
|
||||
happyFail i tk st
|
||||
-1# -> {- nothing -}
|
||||
happyAccept i tk st
|
||||
n | (n <# (0# :: Int#)) -> {- nothing -}
|
||||
|
||||
(happyReduceArr ! rule) i tk st
|
||||
where rule = (I# ((negateInt# ((n +# (1# :: Int#))))))
|
||||
n -> {- nothing -}
|
||||
|
||||
|
||||
happyShift new_state i tk st
|
||||
where new_state = (n -# (1# :: Int#))
|
||||
where off = indexShortOffAddr happyActOffsets st
|
||||
off_i = (off +# i)
|
||||
check = if (off_i >=# (0# :: Int#))
|
||||
then (indexShortOffAddr happyCheck off_i ==# i)
|
||||
else False
|
||||
action | check = indexShortOffAddr happyTable off_i
|
||||
| otherwise = indexShortOffAddr happyDefActions st
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
indexShortOffAddr (HappyA# arr) off =
|
||||
#if __GLASGOW_HASKELL__ > 500
|
||||
narrow16Int# i
|
||||
#elif __GLASGOW_HASKELL__ == 500
|
||||
intToInt16# i
|
||||
#else
|
||||
(i `iShiftL#` 16#) `iShiftRA#` 16#
|
||||
#endif
|
||||
where
|
||||
#if __GLASGOW_HASKELL__ >= 503
|
||||
i = word2Int# ((high `uncheckedShiftL#` 8#) `or#` low)
|
||||
#else
|
||||
i = word2Int# ((high `shiftL#` 8#) `or#` low)
|
||||
#endif
|
||||
high = int2Word# (ord# (indexCharOffAddr# arr (off' +# 1#)))
|
||||
low = int2Word# (ord# (indexCharOffAddr# arr off'))
|
||||
off' = off *# 2#
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
data HappyAddr = HappyA# Addr#
|
||||
|
||||
|
||||
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- HappyState data type (not arrays)
|
||||
|
||||
{-# LINE 169 "GenericTemplate.hs" #-}
|
||||
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Shifting a token
|
||||
|
||||
happyShift new_state 0# tk st sts stk@(x `HappyStk` _) =
|
||||
let i = (case unsafeCoerce# x of { (I# (i)) -> i }) in
|
||||
-- trace "shifting the error token" $
|
||||
happyDoAction i tk new_state (HappyCons (st) (sts)) (stk)
|
||||
|
||||
happyShift new_state i tk st sts stk =
|
||||
happyNewToken new_state (HappyCons (st) (sts)) ((happyInTok (tk))`HappyStk`stk)
|
||||
|
||||
-- happyReduce is specialised for the common cases.
|
||||
|
||||
happySpecReduce_0 i fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happySpecReduce_0 nt fn j tk st@((action)) sts stk
|
||||
= happyGoto nt j tk st (HappyCons (st) (sts)) (fn `HappyStk` stk)
|
||||
|
||||
happySpecReduce_1 i fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happySpecReduce_1 nt fn j tk _ sts@((HappyCons (st@(action)) (_))) (v1`HappyStk`stk')
|
||||
= let r = fn v1 in
|
||||
happySeq r (happyGoto nt j tk st sts (r `HappyStk` stk'))
|
||||
|
||||
happySpecReduce_2 i fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happySpecReduce_2 nt fn j tk _ (HappyCons (_) (sts@((HappyCons (st@(action)) (_))))) (v1`HappyStk`v2`HappyStk`stk')
|
||||
= let r = fn v1 v2 in
|
||||
happySeq r (happyGoto nt j tk st sts (r `HappyStk` stk'))
|
||||
|
||||
happySpecReduce_3 i fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happySpecReduce_3 nt fn j tk _ (HappyCons (_) ((HappyCons (_) (sts@((HappyCons (st@(action)) (_))))))) (v1`HappyStk`v2`HappyStk`v3`HappyStk`stk')
|
||||
= let r = fn v1 v2 v3 in
|
||||
happySeq r (happyGoto nt j tk st sts (r `HappyStk` stk'))
|
||||
|
||||
happyReduce k i fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happyReduce k nt fn j tk st sts stk
|
||||
= case happyDrop (k -# (1# :: Int#)) sts of
|
||||
sts1@((HappyCons (st1@(action)) (_))) ->
|
||||
let r = fn stk in -- it doesn't hurt to always seq here...
|
||||
happyDoSeq r (happyGoto nt j tk st1 sts1 r)
|
||||
|
||||
happyMonadReduce k nt fn 0# tk st sts stk
|
||||
= happyFail 0# tk st sts stk
|
||||
happyMonadReduce k nt fn j tk st sts stk =
|
||||
happyThen1 (fn stk) (\r -> happyGoto nt j tk st1 sts1 (r `HappyStk` drop_stk))
|
||||
where sts1@((HappyCons (st1@(action)) (_))) = happyDrop k (HappyCons (st) (sts))
|
||||
drop_stk = happyDropStk k stk
|
||||
|
||||
happyDrop 0# l = l
|
||||
happyDrop n (HappyCons (_) (t)) = happyDrop (n -# (1# :: Int#)) t
|
||||
|
||||
happyDropStk 0# l = l
|
||||
happyDropStk n (x `HappyStk` xs) = happyDropStk (n -# (1#::Int#)) xs
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Moving to a new state after a reduction
|
||||
|
||||
|
||||
happyGoto nt j tk st =
|
||||
{- nothing -}
|
||||
happyDoAction j tk new_state
|
||||
where off = indexShortOffAddr happyGotoOffsets st
|
||||
off_i = (off +# nt)
|
||||
new_state = indexShortOffAddr happyTable off_i
|
||||
|
||||
|
||||
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Error recovery (0# is the error token)
|
||||
|
||||
-- parse error if we are in recovery and we fail again
|
||||
happyFail 0# tk old_st _ stk =
|
||||
-- trace "failing" $
|
||||
happyError_ tk
|
||||
|
||||
{- We don't need state discarding for our restricted implementation of
|
||||
"error". In fact, it can cause some bogus parses, so I've disabled it
|
||||
for now --SDM
|
||||
|
||||
-- discard a state
|
||||
happyFail 0# tk old_st (HappyCons ((action)) (sts))
|
||||
(saved_tok `HappyStk` _ `HappyStk` stk) =
|
||||
-- trace ("discarding state, depth " ++ show (length stk)) $
|
||||
happyDoAction 0# tk action sts ((saved_tok`HappyStk`stk))
|
||||
-}
|
||||
|
||||
-- Enter error recovery: generate an error token,
|
||||
-- save the old token and carry on.
|
||||
happyFail i tk (action) sts stk =
|
||||
-- trace "entering error recovery" $
|
||||
happyDoAction 0# tk action sts ( (unsafeCoerce# (I# (i))) `HappyStk` stk)
|
||||
|
||||
-- Internal happy errors:
|
||||
|
||||
notHappyAtAll = error "Internal Happy error\n"
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Hack to get the typechecker to accept our action functions
|
||||
|
||||
|
||||
happyTcHack :: Int# -> a -> a
|
||||
happyTcHack x y = y
|
||||
{-# INLINE happyTcHack #-}
|
||||
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Seq-ing. If the --strict flag is given, then Happy emits
|
||||
-- happySeq = happyDoSeq
|
||||
-- otherwise it emits
|
||||
-- happySeq = happyDontSeq
|
||||
|
||||
happyDoSeq, happyDontSeq :: a -> b -> b
|
||||
happyDoSeq a b = a `seq` b
|
||||
happyDontSeq a b = b
|
||||
|
||||
-----------------------------------------------------------------------------
|
||||
-- Don't inline any functions from the template. GHC has a nasty habit
|
||||
-- of deciding to inline happyGoto everywhere, which increases the size of
|
||||
-- the generated parser quite a bit.
|
||||
|
||||
|
||||
{-# NOINLINE happyDoAction #-}
|
||||
{-# NOINLINE happyTable #-}
|
||||
{-# NOINLINE happyCheck #-}
|
||||
{-# NOINLINE happyActOffsets #-}
|
||||
{-# NOINLINE happyGotoOffsets #-}
|
||||
{-# NOINLINE happyDefActions #-}
|
||||
|
||||
{-# NOINLINE happyShift #-}
|
||||
{-# NOINLINE happySpecReduce_0 #-}
|
||||
{-# NOINLINE happySpecReduce_1 #-}
|
||||
{-# NOINLINE happySpecReduce_2 #-}
|
||||
{-# NOINLINE happySpecReduce_3 #-}
|
||||
{-# NOINLINE happyReduce #-}
|
||||
{-# NOINLINE happyMonadReduce #-}
|
||||
{-# NOINLINE happyGoto #-}
|
||||
{-# NOINLINE happyFail #-}
|
||||
|
||||
-- end of Happy Template.
|
||||
129
src-3.0/GF/CFGM/ParCFG.y
Normal file
129
src-3.0/GF/CFGM/ParCFG.y
Normal file
@@ -0,0 +1,129 @@
|
||||
-- This Happy file was machine-generated by the BNF converter
|
||||
{
|
||||
module ParCFG where
|
||||
import AbsCFG
|
||||
import LexCFG
|
||||
import ErrM
|
||||
}
|
||||
|
||||
%name pGrammars Grammars
|
||||
|
||||
-- no lexer declaration
|
||||
%monad { Err } { thenM } { returnM }
|
||||
%tokentype { Token }
|
||||
|
||||
%token
|
||||
';' { PT _ (TS ";") }
|
||||
':' { PT _ (TS ":") }
|
||||
'.' { PT _ (TS ".") }
|
||||
'->' { PT _ (TS "->") }
|
||||
'_' { PT _ (TS "_") }
|
||||
'[' { PT _ (TS "[") }
|
||||
']' { PT _ (TS "]") }
|
||||
',' { PT _ (TS ",") }
|
||||
'end' { PT _ (TS "end") }
|
||||
'grammar' { PT _ (TS "grammar") }
|
||||
'startcat' { PT _ (TS "startcat") }
|
||||
|
||||
L_ident { PT _ (TV $$) }
|
||||
L_integ { PT _ (TI $$) }
|
||||
L_quoted { PT _ (TL $$) }
|
||||
L_SingleQuoteString { PT _ (T_SingleQuoteString $$) }
|
||||
L_err { _ }
|
||||
|
||||
|
||||
%%
|
||||
|
||||
Ident :: { Ident } : L_ident { Ident $1 }
|
||||
Integer :: { Integer } : L_integ { (read $1) :: Integer }
|
||||
String :: { String } : L_quoted { $1 }
|
||||
SingleQuoteString :: { SingleQuoteString} : L_SingleQuoteString { SingleQuoteString ($1)}
|
||||
|
||||
Grammars :: { Grammars }
|
||||
Grammars : ListGrammar { Grammars (reverse $1) }
|
||||
|
||||
|
||||
Grammar :: { Grammar }
|
||||
Grammar : 'grammar' Ident ListFlag ListRule 'end' 'grammar' { Grammar $2 (reverse $3) (reverse $4) }
|
||||
|
||||
|
||||
ListGrammar :: { [Grammar] }
|
||||
ListGrammar : {- empty -} { [] }
|
||||
| ListGrammar Grammar { flip (:) $1 $2 }
|
||||
|
||||
|
||||
Flag :: { Flag }
|
||||
Flag : 'startcat' Category { StartCat $2 }
|
||||
|
||||
|
||||
ListFlag :: { [Flag] }
|
||||
ListFlag : {- empty -} { [] }
|
||||
| ListFlag Flag ';' { flip (:) $1 $2 }
|
||||
|
||||
|
||||
Rule :: { Rule }
|
||||
Rule : Fun ':' Profiles '.' Category '->' ListSymbol { Rule $1 $3 $5 $7 }
|
||||
|
||||
|
||||
ListRule :: { [Rule] }
|
||||
ListRule : {- empty -} { [] }
|
||||
| ListRule Rule ';' { flip (:) $1 $2 }
|
||||
|
||||
|
||||
Fun :: { Fun }
|
||||
Fun : Ident { Cons $1 }
|
||||
| '_' { Coerce }
|
||||
|
||||
|
||||
Profiles :: { Profiles }
|
||||
Profiles : '[' ListProfile ']' { Profiles $2 }
|
||||
|
||||
|
||||
ListProfile :: { [Profile] }
|
||||
ListProfile : {- empty -} { [] }
|
||||
| Profile { (:[]) $1 }
|
||||
| Profile ',' ListProfile { (:) $1 $3 }
|
||||
|
||||
|
||||
Profile :: { Profile }
|
||||
Profile : '[' ListInteger ']' { UnifyProfile $2 }
|
||||
| Ident { ConstProfile $1 }
|
||||
|
||||
|
||||
ListInteger :: { [Integer] }
|
||||
ListInteger : {- empty -} { [] }
|
||||
| Integer { (:[]) $1 }
|
||||
| Integer ',' ListInteger { (:) $1 $3 }
|
||||
|
||||
|
||||
Symbol :: { Symbol }
|
||||
Symbol : Category { CatS $1 }
|
||||
| String { TermS $1 }
|
||||
|
||||
|
||||
ListSymbol :: { [Symbol] }
|
||||
ListSymbol : '.' { [] }
|
||||
| Symbol { (:[]) $1 }
|
||||
| Symbol ListSymbol { (:) $1 $2 }
|
||||
|
||||
|
||||
Category :: { Category }
|
||||
Category : SingleQuoteString { Category $1 }
|
||||
|
||||
|
||||
|
||||
{
|
||||
|
||||
returnM :: a -> Err a
|
||||
returnM = return
|
||||
|
||||
thenM :: Err a -> (a -> Err b) -> Err b
|
||||
thenM = (>>=)
|
||||
|
||||
happyError :: [Token] -> Err a
|
||||
happyError ts =
|
||||
Bad $ "syntax error at " ++ tokenPos ts ++ if null ts then [] else (" before " ++ unwords (map prToken (take 4 ts)))
|
||||
|
||||
myLexer = tokens
|
||||
}
|
||||
|
||||
157
src-3.0/GF/CFGM/PrintCFG.hs
Normal file
157
src-3.0/GF/CFGM/PrintCFG.hs
Normal file
@@ -0,0 +1,157 @@
|
||||
module GF.CFGM.PrintCFG where
|
||||
|
||||
-- pretty-printer generated by the BNF converter
|
||||
|
||||
import GF.CFGM.AbsCFG
|
||||
import Char
|
||||
|
||||
-- the top-level printing method
|
||||
printTree :: Print a => a -> String
|
||||
printTree = render . prt 0
|
||||
|
||||
type Doc = [ShowS] -> [ShowS]
|
||||
|
||||
doc :: ShowS -> Doc
|
||||
doc = (:)
|
||||
|
||||
render :: Doc -> String
|
||||
render d = rend 0 (map ($ "") $ d []) "" where
|
||||
rend i ss = case ss of
|
||||
"[" :ts -> showChar '[' . rend i ts
|
||||
"(" :ts -> showChar '(' . rend i ts
|
||||
"{" :ts -> showChar '{' . new (i+1) . rend (i+1) ts
|
||||
"}" : ";":ts -> new (i-1) . space "}" . showChar ';' . new (i-1) . rend (i-1) ts
|
||||
"}" :ts -> new (i-1) . showChar '}' . new (i-1) . rend (i-1) ts
|
||||
";" :ts -> showChar ';' . new i . rend i ts
|
||||
t : "," :ts -> showString t . space "," . rend i ts
|
||||
t : ")" :ts -> showString t . showChar ')' . rend i ts
|
||||
t : "]" :ts -> showString t . showChar ']' . rend i ts
|
||||
t :ts -> space t . rend i ts
|
||||
_ -> id
|
||||
new i = showChar '\n' . replicateS (2*i) (showChar ' ') . dropWhile isSpace
|
||||
space t = showString t . (\s -> if null s then "" else (' ':s))
|
||||
|
||||
parenth :: Doc -> Doc
|
||||
parenth ss = doc (showChar '(') . ss . doc (showChar ')')
|
||||
|
||||
concatS :: [ShowS] -> ShowS
|
||||
concatS = foldr (.) id
|
||||
|
||||
concatD :: [Doc] -> Doc
|
||||
concatD = foldr (.) id
|
||||
|
||||
replicateS :: Int -> ShowS -> ShowS
|
||||
replicateS n f = concatS (replicate n f)
|
||||
|
||||
-- the printer class does the job
|
||||
class Print a where
|
||||
prt :: Int -> a -> Doc
|
||||
prtList :: [a] -> Doc
|
||||
prtList = concatD . map (prt 0)
|
||||
|
||||
instance Print a => Print [a] where
|
||||
prt _ = prtList
|
||||
|
||||
instance Print Char where
|
||||
prt _ s = doc (showChar '\'' . mkEsc '\'' s . showChar '\'')
|
||||
prtList s = doc (showChar '"' . concatS (map (mkEsc '"') s) . showChar '"')
|
||||
|
||||
mkEsc :: Char -> Char -> ShowS
|
||||
mkEsc q s = case s of
|
||||
_ | s == q -> showChar '\\' . showChar s
|
||||
'\\'-> showString "\\\\"
|
||||
'\n' -> showString "\\n"
|
||||
'\t' -> showString "\\t"
|
||||
_ -> showChar s
|
||||
|
||||
prPrec :: Int -> Int -> Doc -> Doc
|
||||
prPrec i j = if j<i then parenth else id
|
||||
|
||||
|
||||
instance Print Integer where
|
||||
prt _ x = doc (shows x)
|
||||
prtList es = case es of
|
||||
[] -> (concatD [])
|
||||
[x] -> (concatD [prt 0 x])
|
||||
x:xs -> (concatD [prt 0 x , doc (showString ",") , prt 0 xs])
|
||||
|
||||
|
||||
instance Print Double where
|
||||
prt _ x = doc (shows x)
|
||||
|
||||
|
||||
instance Print Ident where
|
||||
prt _ (Ident i) = doc (showString i)
|
||||
|
||||
|
||||
instance Print SingleQuoteString where
|
||||
prt _ (SingleQuoteString i) = doc (showString i)
|
||||
|
||||
|
||||
|
||||
instance Print Grammars where
|
||||
prt i e = case e of
|
||||
Grammars grammars -> prPrec i 0 (concatD [prt 0 grammars])
|
||||
|
||||
|
||||
instance Print Grammar where
|
||||
prt i e = case e of
|
||||
Grammar id flags rules -> prPrec i 0 (concatD [doc (showString "grammar") , prt 0 id , prt 0 flags , prt 0 rules , doc (showString "end") , doc (showString "grammar")])
|
||||
|
||||
prtList es = case es of
|
||||
[] -> (concatD [])
|
||||
x:xs -> (concatD [prt 0 x , prt 0 xs])
|
||||
|
||||
instance Print Flag where
|
||||
prt i e = case e of
|
||||
StartCat category -> prPrec i 0 (concatD [doc (showString "startcat") , prt 0 category])
|
||||
|
||||
prtList es = case es of
|
||||
[] -> (concatD [])
|
||||
x:xs -> (concatD [prt 0 x , doc (showString ";") , prt 0 xs])
|
||||
|
||||
instance Print Rule where
|
||||
prt i e = case e of
|
||||
Rule fun profiles category symbols -> prPrec i 0 (concatD [prt 0 fun , doc (showString ":") , prt 0 profiles , doc (showString ".") , prt 0 category , doc (showString "->") , prt 0 symbols])
|
||||
|
||||
prtList es = case es of
|
||||
[] -> (concatD [])
|
||||
x:xs -> (concatD [prt 0 x , doc (showString ";") , prt 0 xs])
|
||||
|
||||
instance Print Fun where
|
||||
prt i e = case e of
|
||||
Cons id -> prPrec i 0 (concatD [prt 0 id])
|
||||
Coerce -> prPrec i 0 (concatD [doc (showString "_")])
|
||||
|
||||
|
||||
instance Print Profiles where
|
||||
prt i e = case e of
|
||||
Profiles profiles -> prPrec i 0 (concatD [doc (showString "[") , prt 0 profiles , doc (showString "]")])
|
||||
|
||||
|
||||
instance Print Profile where
|
||||
prt i e = case e of
|
||||
UnifyProfile ns -> prPrec i 0 (concatD [doc (showString "[") , prt 0 ns , doc (showString "]")])
|
||||
ConstProfile id -> prPrec i 0 (concatD [prt 0 id])
|
||||
|
||||
prtList es = case es of
|
||||
[] -> (concatD [])
|
||||
[x] -> (concatD [prt 0 x])
|
||||
x:xs -> (concatD [prt 0 x , doc (showString ",") , prt 0 xs])
|
||||
|
||||
instance Print Symbol where
|
||||
prt i e = case e of
|
||||
CatS category -> prPrec i 0 (concatD [prt 0 category])
|
||||
TermS str -> prPrec i 0 (concatD [prt 0 str])
|
||||
|
||||
prtList es = case es of
|
||||
[] -> (concatD [doc (showString ".")])
|
||||
[x] -> (concatD [prt 0 x])
|
||||
x:xs -> (concatD [prt 0 x , prt 0 xs])
|
||||
|
||||
instance Print Category where
|
||||
prt i e = case e of
|
||||
Category singlequotestring -> prPrec i 0 (concatD [prt 0 singlequotestring])
|
||||
|
||||
|
||||
|
||||
113
src-3.0/GF/CFGM/PrintCFGrammar.hs
Normal file
113
src-3.0/GF/CFGM/PrintCFGrammar.hs
Normal file
@@ -0,0 +1,113 @@
|
||||
----------------------------------------------------------------------
|
||||
-- |
|
||||
-- Module : PrintCFGrammar
|
||||
-- Maintainer : BB
|
||||
-- Stability : (stable)
|
||||
-- Portability : (portable)
|
||||
--
|
||||
-- > CVS $Date: 2005/05/17 14:04:38 $
|
||||
-- > CVS $Author: bringert $
|
||||
-- > CVS $Revision: 1.20 $
|
||||
--
|
||||
-- Handles printing a CFGrammar in CFGM format.
|
||||
-----------------------------------------------------------------------------
|
||||
|
||||
module GF.CFGM.PrintCFGrammar (prCanonAsCFGM) where
|
||||
|
||||
import GF.Canon.AbsGFC
|
||||
import qualified GF.CFGM.PrintCFG as PrintCFG
|
||||
import GF.Infra.Ident
|
||||
import GF.Canon.GFC
|
||||
import GF.Infra.Modules
|
||||
|
||||
import qualified GF.Conversion.GFC as Cnv
|
||||
import GF.Infra.Print (prt)
|
||||
import GF.Formalism.CFG (CFRule(..))
|
||||
import qualified GF.Formalism.Utilities as GU
|
||||
import qualified GF.Conversion.Types as GT
|
||||
import qualified GF.CFGM.AbsCFG as AbsCFG
|
||||
import GF.Formalism.Utilities (Symbol(..))
|
||||
|
||||
import GF.Data.ErrM
|
||||
import GF.Data.Utilities (compareBy)
|
||||
import qualified GF.Infra.Option as Option
|
||||
|
||||
import Data.List (intersperse, sortBy)
|
||||
import Data.Maybe (listToMaybe, maybeToList, maybe)
|
||||
|
||||
import GF.Infra.Print
|
||||
import GF.System.Tracing
|
||||
|
||||
-- | FIXME: should add an Options argument,
|
||||
-- to be able to decide which CFG conversion one wants to use
|
||||
prCanonAsCFGM :: Option.Options -> CanonGrammar -> String
|
||||
prCanonAsCFGM opts gr = unlines $ map (prLangAsCFGM gr) xs
|
||||
where
|
||||
cncs = maybe [] (allConcretes gr) (greatestAbstract gr)
|
||||
cncms = map (\i -> (i,fromOk (lookupModule gr i))) cncs
|
||||
fromOk (Ok x) = x
|
||||
fromOk (Bad y) = error y
|
||||
xs = tracePrt "CFGM languages" (prtBefore "\n")
|
||||
[ (i, getFlag fs "startcat", getFlag fs "conversion") |
|
||||
(i, ModMod (Module{flags=fs})) <- cncms ]
|
||||
|
||||
-- | FIXME: need to look in abstract module too
|
||||
getFlag :: [Flag] -> String -> Maybe String
|
||||
getFlag fs x = listToMaybe [v | Flg (IC k) (IC v) <- fs, k == x]
|
||||
|
||||
-- FIXME: (1) Should use 'ShellState.stateCFG'
|
||||
-- instead of 'Cnv.gfc2cfg' (which recalculates the grammar every time)
|
||||
--
|
||||
-- FIXME: (2) Should use the state options, when calculating the CFG
|
||||
-- (this is solved automatically if one solves (1) above)
|
||||
prLangAsCFGM :: CanonGrammar -> (Ident, Maybe String, Maybe String) -> String
|
||||
prLangAsCFGM gr (i, start, cnv) = prCFGrammarAsCFGM (Cnv.gfc2cfg opts (gr, i)) i start
|
||||
-- prLangAsCFGM gr i start = prCFGrammarAsCFGM (Cnv.cfg (Cnv.pInfo opts gr i)) i start
|
||||
where opts = Option.Opts $ maybeToList $ fmap Option.gfcConversion cnv
|
||||
|
||||
prCFGrammarAsCFGM :: GT.CGrammar -> Ident -> Maybe String -> String
|
||||
prCFGrammarAsCFGM gr i start = PrintCFG.printTree $ cfGrammarToCFGM gr i start
|
||||
|
||||
cfGrammarToCFGM :: GT.CGrammar -> Ident -> Maybe String -> AbsCFG.Grammar
|
||||
cfGrammarToCFGM gr i start =
|
||||
AbsCFG.Grammar (identToCFGMIdent i) flags $ sortCFGMRules $ map ruleToCFGMRule gr
|
||||
where flags = maybe [] (\c -> [AbsCFG.StartCat $ strToCFGMCat (c++"{}.s")]) start
|
||||
sortCFGMRules = sortBy (compareBy ruleKey)
|
||||
ruleKey (AbsCFG.Rule f ps cat rhs) = (cat,f)
|
||||
|
||||
ruleToCFGMRule :: GT.CRule -> AbsCFG.Rule
|
||||
ruleToCFGMRule (CFRule c rhs (GU.Name fun profile))
|
||||
= AbsCFG.Rule fun' p' c' rhs'
|
||||
where
|
||||
fun' = identToFun fun
|
||||
p' = profileToCFGMProfile profile
|
||||
c' = catToCFGMCat c
|
||||
rhs' = map symbolToGFCMSymbol rhs
|
||||
|
||||
profileToCFGMProfile :: [GU.Profile (GU.SyntaxForest GT.Fun)] -> AbsCFG.Profiles
|
||||
profileToCFGMProfile = AbsCFG.Profiles . map cnvProfile
|
||||
where cnvProfile (GU.Unify ns) = AbsCFG.UnifyProfile $ map fromIntegral ns
|
||||
-- FIXME: is it always FNode?
|
||||
cnvProfile (GU.Constant (GU.FNode c _)) = AbsCFG.ConstProfile $ identToCFGMIdent c
|
||||
|
||||
|
||||
identToCFGMIdent :: Ident -> AbsCFG.Ident
|
||||
identToCFGMIdent = AbsCFG.Ident . prt
|
||||
|
||||
identToFun :: Ident -> AbsCFG.Fun
|
||||
identToFun IW = AbsCFG.Coerce
|
||||
identToFun i = AbsCFG.Cons (identToCFGMIdent i)
|
||||
|
||||
strToCFGMCat :: String -> AbsCFG.Category
|
||||
strToCFGMCat = AbsCFG.Category . AbsCFG.SingleQuoteString . quoteSingle
|
||||
|
||||
catToCFGMCat :: GT.CCat -> AbsCFG.Category
|
||||
catToCFGMCat = strToCFGMCat . prt
|
||||
|
||||
symbolToGFCMSymbol :: Symbol GT.CCat GT.Token -> AbsCFG.Symbol
|
||||
symbolToGFCMSymbol (Cat c) = AbsCFG.CatS (catToCFGMCat c)
|
||||
symbolToGFCMSymbol (Tok t) = AbsCFG.TermS (prt t)
|
||||
|
||||
quoteSingle :: String -> String
|
||||
quoteSingle s = "'" ++ escapeSingle s ++ "'"
|
||||
where escapeSingle = concatMap (\c -> if c == '\'' then "\\'" else [c])
|
||||
Reference in New Issue
Block a user