# -*- coding: utf-8 -*-
#
# parselglossy -- Generic input parsing library, speaking in tongues
# Copyright (C) 2020 Roberto Di Remigio, Radovan Bast, and contributors.
#
# This file is part of parselglossy.
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
#
# For information on the complete list of contributors to the
# parselglossy library, see: <http://parselglossy.readthedocs.io/>
#
"""Top-level functions for parselglossy."""
import json
from pathlib import Path
from typing import TYPE_CHECKING, Any, List, Optional, Union
if TYPE_CHECKING:
try:
import pyparsing as pp
except ImportError:
# Import local copy
from . import pyparsing as pp # type: ignore
from ..exceptions import ParselglossyError
from ..utils import ComplexEncoder, JSONDict, dict_to_list, flatten_list, path_resolver
from . import getkw
# ->->-> SNIP <-<-<-
[docs]def parse_string_to_dict(lexer: "pp.ParserElement", s: str) -> JSONDict:
"""Helper function around parseString(s).asDict()
that checks whether some keywords or sections were accidentally
repeated and shadowing earlier keywords/sections.
Parameters
----------
lexer : pp.ParserElement
Nested dictionary
s : str
String to parse
Returns
-------
tokes_dict: JSONDict
Dictionary of tokens
Raises
------
:exc:`ParselglossyError`
"""
tokens_dict = lexer.parseString(s).asDict() # type: JSONDict
tokens_list = lexer.parseString(s).asList() # type: List[Any]
if flatten_list(tokens_list) != flatten_list(dict_to_list(tokens_dict)):
raise ParselglossyError("A keyword is repeated. Please check your input.")
return tokens_dict
# -<-<-< SNAP >->->-
[docs]def lex_from_str(
*,
in_str: str,
grammar: str = "standard",
ir_file: Optional[Union[str, Path]] = None,
) -> JSONDict:
"""Run grammar of choice on input string.
Parameters
----------
in_str : str
The string to be parsed.
grammar : str
Grammar to be used. Defaults to "standard".
ir_file : Optional[Union[str, Path]]
File to write intermediate representation to (JSON format).
None by default, which means file is not written out.
Returns
-------
The contents of the input string as a dictionary.
Raises
------
:exc:`ParselglossyError`
"""
try:
lexer = dispatch_grammar(grammar)
except KeyError:
raise ParselglossyError(f"Grammar {grammar} not available.")
ir = parse_string_to_dict(lexer, in_str)
if ir_file is not None:
ir_file = path_resolver(ir_file)
with ir_file.open("w") as out:
json.dump(ir, out, cls=ComplexEncoder, indent=4)
return ir
[docs]def dispatch_grammar(grammar: str) -> "pp.ParserElement":
available_grammars = {
"getkw": getkw.grammar(),
"standard": getkw.grammar(has_complex=True),
}
return available_grammars[grammar]