520 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			C++
		
	
	
	
		
		
			
		
	
	
			520 lines
		
	
	
		
			19 KiB
		
	
	
	
		
			C++
		
	
	
	
|  | //     __ _____ _____ _____
 | ||
|  | //  __|  |   __|     |   | |  JSON for Modern C++
 | ||
|  | // |  |  |__   |  |  | | | |  version 3.11.3
 | ||
|  | // |_____|_____|_____|_|___|  https://github.com/nlohmann/json
 | ||
|  | //
 | ||
|  | // SPDX-FileCopyrightText: 2013-2023 Niels Lohmann <https://nlohmann.me>
 | ||
|  | // SPDX-License-Identifier: MIT
 | ||
|  | 
 | ||
|  | #pragma once
 | ||
|  | 
 | ||
|  | #include <cmath> // isfinite
 | ||
|  | #include <cstdint> // uint8_t
 | ||
|  | #include <functional> // function
 | ||
|  | #include <string> // string
 | ||
|  | #include <utility> // move
 | ||
|  | #include <vector> // vector
 | ||
|  | 
 | ||
|  | #include <nlohmann/detail/exceptions.hpp>
 | ||
|  | #include <nlohmann/detail/input/input_adapters.hpp>
 | ||
|  | #include <nlohmann/detail/input/json_sax.hpp>
 | ||
|  | #include <nlohmann/detail/input/lexer.hpp>
 | ||
|  | #include <nlohmann/detail/macro_scope.hpp>
 | ||
|  | #include <nlohmann/detail/meta/is_sax.hpp>
 | ||
|  | #include <nlohmann/detail/string_concat.hpp>
 | ||
|  | #include <nlohmann/detail/value_t.hpp>
 | ||
|  | 
 | ||
|  | NLOHMANN_JSON_NAMESPACE_BEGIN | ||
|  | namespace detail | ||
|  | { | ||
|  | ////////////
 | ||
|  | // parser //
 | ||
|  | ////////////
 | ||
|  | 
 | ||
|  | enum class parse_event_t : std::uint8_t | ||
|  | { | ||
|  |     /// the parser read `{` and started to process a JSON object
 | ||
|  |     object_start, | ||
|  |     /// the parser read `}` and finished processing a JSON object
 | ||
|  |     object_end, | ||
|  |     /// the parser read `[` and started to process a JSON array
 | ||
|  |     array_start, | ||
|  |     /// the parser read `]` and finished processing a JSON array
 | ||
|  |     array_end, | ||
|  |     /// the parser read a key of a value in an object
 | ||
|  |     key, | ||
|  |     /// the parser finished reading a JSON value
 | ||
|  |     value | ||
|  | }; | ||
|  | 
 | ||
|  | template<typename BasicJsonType> | ||
|  | using parser_callback_t = | ||
|  |     std::function<bool(int /*depth*/, parse_event_t /*event*/, BasicJsonType& /*parsed*/)>; | ||
|  | 
 | ||
|  | /*!
 | ||
|  | @brief syntax analysis | ||
|  | 
 | ||
|  | This class implements a recursive descent parser. | ||
|  | */ | ||
|  | template<typename BasicJsonType, typename InputAdapterType> | ||
|  | class parser | ||
|  | { | ||
|  |     using number_integer_t = typename BasicJsonType::number_integer_t; | ||
|  |     using number_unsigned_t = typename BasicJsonType::number_unsigned_t; | ||
|  |     using number_float_t = typename BasicJsonType::number_float_t; | ||
|  |     using string_t = typename BasicJsonType::string_t; | ||
|  |     using lexer_t = lexer<BasicJsonType, InputAdapterType>; | ||
|  |     using token_type = typename lexer_t::token_type; | ||
|  | 
 | ||
|  |   public: | ||
|  |     /// a parser reading from an input adapter
 | ||
|  |     explicit parser(InputAdapterType&& adapter, | ||
|  |                     const parser_callback_t<BasicJsonType> cb = nullptr, | ||
|  |                     const bool allow_exceptions_ = true, | ||
|  |                     const bool skip_comments = false) | ||
|  |         : callback(cb) | ||
|  |         , m_lexer(std::move(adapter), skip_comments) | ||
|  |         , allow_exceptions(allow_exceptions_) | ||
|  |     { | ||
|  |         // read first token
 | ||
|  |         get_token(); | ||
|  |     } | ||
|  | 
 | ||
|  |     /*!
 | ||
|  |     @brief public parser interface | ||
|  | 
 | ||
|  |     @param[in] strict      whether to expect the last token to be EOF | ||
|  |     @param[in,out] result  parsed JSON value | ||
|  | 
 | ||
|  |     @throw parse_error.101 in case of an unexpected token | ||
|  |     @throw parse_error.102 if to_unicode fails or surrogate error | ||
|  |     @throw parse_error.103 if to_unicode fails | ||
|  |     */ | ||
|  |     void parse(const bool strict, BasicJsonType& result) | ||
|  |     { | ||
|  |         if (callback) | ||
|  |         { | ||
|  |             json_sax_dom_callback_parser<BasicJsonType> sdp(result, callback, allow_exceptions); | ||
|  |             sax_parse_internal(&sdp); | ||
|  | 
 | ||
|  |             // in strict mode, input must be completely read
 | ||
|  |             if (strict && (get_token() != token_type::end_of_input)) | ||
|  |             { | ||
|  |                 sdp.parse_error(m_lexer.get_position(), | ||
|  |                                 m_lexer.get_token_string(), | ||
|  |                                 parse_error::create(101, m_lexer.get_position(), | ||
|  |                                                     exception_message(token_type::end_of_input, "value"), nullptr)); | ||
|  |             } | ||
|  | 
 | ||
|  |             // in case of an error, return discarded value
 | ||
|  |             if (sdp.is_errored()) | ||
|  |             { | ||
|  |                 result = value_t::discarded; | ||
|  |                 return; | ||
|  |             } | ||
|  | 
 | ||
|  |             // set top-level value to null if it was discarded by the callback
 | ||
|  |             // function
 | ||
|  |             if (result.is_discarded()) | ||
|  |             { | ||
|  |                 result = nullptr; | ||
|  |             } | ||
|  |         } | ||
|  |         else | ||
|  |         { | ||
|  |             json_sax_dom_parser<BasicJsonType> sdp(result, allow_exceptions); | ||
|  |             sax_parse_internal(&sdp); | ||
|  | 
 | ||
|  |             // in strict mode, input must be completely read
 | ||
|  |             if (strict && (get_token() != token_type::end_of_input)) | ||
|  |             { | ||
|  |                 sdp.parse_error(m_lexer.get_position(), | ||
|  |                                 m_lexer.get_token_string(), | ||
|  |                                 parse_error::create(101, m_lexer.get_position(), exception_message(token_type::end_of_input, "value"), nullptr)); | ||
|  |             } | ||
|  | 
 | ||
|  |             // in case of an error, return discarded value
 | ||
|  |             if (sdp.is_errored()) | ||
|  |             { | ||
|  |                 result = value_t::discarded; | ||
|  |                 return; | ||
|  |             } | ||
|  |         } | ||
|  | 
 | ||
|  |         result.assert_invariant(); | ||
|  |     } | ||
|  | 
 | ||
|  |     /*!
 | ||
|  |     @brief public accept interface | ||
|  | 
 | ||
|  |     @param[in] strict  whether to expect the last token to be EOF | ||
|  |     @return whether the input is a proper JSON text | ||
|  |     */ | ||
|  |     bool accept(const bool strict = true) | ||
|  |     { | ||
|  |         json_sax_acceptor<BasicJsonType> sax_acceptor; | ||
|  |         return sax_parse(&sax_acceptor, strict); | ||
|  |     } | ||
|  | 
 | ||
|  |     template<typename SAX> | ||
|  |     JSON_HEDLEY_NON_NULL(2) | ||
|  |     bool sax_parse(SAX* sax, const bool strict = true) | ||
|  |     { | ||
|  |         (void)detail::is_sax_static_asserts<SAX, BasicJsonType> {}; | ||
|  |         const bool result = sax_parse_internal(sax); | ||
|  | 
 | ||
|  |         // strict mode: next byte must be EOF
 | ||
|  |         if (result && strict && (get_token() != token_type::end_of_input)) | ||
|  |         { | ||
|  |             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                     m_lexer.get_token_string(), | ||
|  |                                     parse_error::create(101, m_lexer.get_position(), exception_message(token_type::end_of_input, "value"), nullptr)); | ||
|  |         } | ||
|  | 
 | ||
|  |         return result; | ||
|  |     } | ||
|  | 
 | ||
|  |   private: | ||
|  |     template<typename SAX> | ||
|  |     JSON_HEDLEY_NON_NULL(2) | ||
|  |     bool sax_parse_internal(SAX* sax) | ||
|  |     { | ||
|  |         // stack to remember the hierarchy of structured values we are parsing
 | ||
|  |         // true = array; false = object
 | ||
|  |         std::vector<bool> states; | ||
|  |         // value to avoid a goto (see comment where set to true)
 | ||
|  |         bool skip_to_state_evaluation = false; | ||
|  | 
 | ||
|  |         while (true) | ||
|  |         { | ||
|  |             if (!skip_to_state_evaluation) | ||
|  |             { | ||
|  |                 // invariant: get_token() was called before each iteration
 | ||
|  |                 switch (last_token) | ||
|  |                 { | ||
|  |                     case token_type::begin_object: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->start_object(static_cast<std::size_t>(-1)))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // closing } -> we are done
 | ||
|  |                         if (get_token() == token_type::end_object) | ||
|  |                         { | ||
|  |                             if (JSON_HEDLEY_UNLIKELY(!sax->end_object())) | ||
|  |                             { | ||
|  |                                 return false; | ||
|  |                             } | ||
|  |                             break; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // parse key
 | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(last_token != token_type::value_string)) | ||
|  |                         { | ||
|  |                             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                     m_lexer.get_token_string(), | ||
|  |                                                     parse_error::create(101, m_lexer.get_position(), exception_message(token_type::value_string, "object key"), nullptr)); | ||
|  |                         } | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->key(m_lexer.get_string()))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // parse separator (:)
 | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(get_token() != token_type::name_separator)) | ||
|  |                         { | ||
|  |                             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                     m_lexer.get_token_string(), | ||
|  |                                                     parse_error::create(101, m_lexer.get_position(), exception_message(token_type::name_separator, "object separator"), nullptr)); | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // remember we are now inside an object
 | ||
|  |                         states.push_back(false); | ||
|  | 
 | ||
|  |                         // parse values
 | ||
|  |                         get_token(); | ||
|  |                         continue; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::begin_array: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->start_array(static_cast<std::size_t>(-1)))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // closing ] -> we are done
 | ||
|  |                         if (get_token() == token_type::end_array) | ||
|  |                         { | ||
|  |                             if (JSON_HEDLEY_UNLIKELY(!sax->end_array())) | ||
|  |                             { | ||
|  |                                 return false; | ||
|  |                             } | ||
|  |                             break; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         // remember we are now inside an array
 | ||
|  |                         states.push_back(true); | ||
|  | 
 | ||
|  |                         // parse values (no need to call get_token)
 | ||
|  |                         continue; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::value_float: | ||
|  |                     { | ||
|  |                         const auto res = m_lexer.get_number_float(); | ||
|  | 
 | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!std::isfinite(res))) | ||
|  |                         { | ||
|  |                             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                     m_lexer.get_token_string(), | ||
|  |                                                     out_of_range::create(406, concat("number overflow parsing '", m_lexer.get_token_string(), '\''), nullptr)); | ||
|  |                         } | ||
|  | 
 | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->number_float(res, m_lexer.get_string()))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  | 
 | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::literal_false: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->boolean(false))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::literal_null: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->null())) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::literal_true: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->boolean(true))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::value_integer: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->number_integer(m_lexer.get_number_integer()))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::value_string: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->string(m_lexer.get_string()))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::value_unsigned: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(!sax->number_unsigned(m_lexer.get_number_unsigned()))) | ||
|  |                         { | ||
|  |                             return false; | ||
|  |                         } | ||
|  |                         break; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     case token_type::parse_error: | ||
|  |                     { | ||
|  |                         // using "uninitialized" to avoid "expected" message
 | ||
|  |                         return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                 m_lexer.get_token_string(), | ||
|  |                                                 parse_error::create(101, m_lexer.get_position(), exception_message(token_type::uninitialized, "value"), nullptr)); | ||
|  |                     } | ||
|  |                     case token_type::end_of_input: | ||
|  |                     { | ||
|  |                         if (JSON_HEDLEY_UNLIKELY(m_lexer.get_position().chars_read_total == 1)) | ||
|  |                         { | ||
|  |                             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                     m_lexer.get_token_string(), | ||
|  |                                                     parse_error::create(101, m_lexer.get_position(), | ||
|  |                                                             "attempting to parse an empty input; check that your input string or stream contains the expected JSON", nullptr)); | ||
|  |                         } | ||
|  | 
 | ||
|  |                         return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                 m_lexer.get_token_string(), | ||
|  |                                                 parse_error::create(101, m_lexer.get_position(), exception_message(token_type::literal_or_value, "value"), nullptr)); | ||
|  |                     } | ||
|  |                     case token_type::uninitialized: | ||
|  |                     case token_type::end_array: | ||
|  |                     case token_type::end_object: | ||
|  |                     case token_type::name_separator: | ||
|  |                     case token_type::value_separator: | ||
|  |                     case token_type::literal_or_value: | ||
|  |                     default: // the last token was unexpected
 | ||
|  |                     { | ||
|  |                         return sax->parse_error(m_lexer.get_position(), | ||
|  |                                                 m_lexer.get_token_string(), | ||
|  |                                                 parse_error::create(101, m_lexer.get_position(), exception_message(token_type::literal_or_value, "value"), nullptr)); | ||
|  |                     } | ||
|  |                 } | ||
|  |             } | ||
|  |             else | ||
|  |             { | ||
|  |                 skip_to_state_evaluation = false; | ||
|  |             } | ||
|  | 
 | ||
|  |             // we reached this line after we successfully parsed a value
 | ||
|  |             if (states.empty()) | ||
|  |             { | ||
|  |                 // empty stack: we reached the end of the hierarchy: done
 | ||
|  |                 return true; | ||
|  |             } | ||
|  | 
 | ||
|  |             if (states.back())  // array
 | ||
|  |             { | ||
|  |                 // comma -> next value
 | ||
|  |                 if (get_token() == token_type::value_separator) | ||
|  |                 { | ||
|  |                     // parse a new value
 | ||
|  |                     get_token(); | ||
|  |                     continue; | ||
|  |                 } | ||
|  | 
 | ||
|  |                 // closing ]
 | ||
|  |                 if (JSON_HEDLEY_LIKELY(last_token == token_type::end_array)) | ||
|  |                 { | ||
|  |                     if (JSON_HEDLEY_UNLIKELY(!sax->end_array())) | ||
|  |                     { | ||
|  |                         return false; | ||
|  |                     } | ||
|  | 
 | ||
|  |                     // We are done with this array. Before we can parse a
 | ||
|  |                     // new value, we need to evaluate the new state first.
 | ||
|  |                     // By setting skip_to_state_evaluation to false, we
 | ||
|  |                     // are effectively jumping to the beginning of this if.
 | ||
|  |                     JSON_ASSERT(!states.empty()); | ||
|  |                     states.pop_back(); | ||
|  |                     skip_to_state_evaluation = true; | ||
|  |                     continue; | ||
|  |                 } | ||
|  | 
 | ||
|  |                 return sax->parse_error(m_lexer.get_position(), | ||
|  |                                         m_lexer.get_token_string(), | ||
|  |                                         parse_error::create(101, m_lexer.get_position(), exception_message(token_type::end_array, "array"), nullptr)); | ||
|  |             } | ||
|  | 
 | ||
|  |             // states.back() is false -> object
 | ||
|  | 
 | ||
|  |             // comma -> next value
 | ||
|  |             if (get_token() == token_type::value_separator) | ||
|  |             { | ||
|  |                 // parse key
 | ||
|  |                 if (JSON_HEDLEY_UNLIKELY(get_token() != token_type::value_string)) | ||
|  |                 { | ||
|  |                     return sax->parse_error(m_lexer.get_position(), | ||
|  |                                             m_lexer.get_token_string(), | ||
|  |                                             parse_error::create(101, m_lexer.get_position(), exception_message(token_type::value_string, "object key"), nullptr)); | ||
|  |                 } | ||
|  | 
 | ||
|  |                 if (JSON_HEDLEY_UNLIKELY(!sax->key(m_lexer.get_string()))) | ||
|  |                 { | ||
|  |                     return false; | ||
|  |                 } | ||
|  | 
 | ||
|  |                 // parse separator (:)
 | ||
|  |                 if (JSON_HEDLEY_UNLIKELY(get_token() != token_type::name_separator)) | ||
|  |                 { | ||
|  |                     return sax->parse_error(m_lexer.get_position(), | ||
|  |                                             m_lexer.get_token_string(), | ||
|  |                                             parse_error::create(101, m_lexer.get_position(), exception_message(token_type::name_separator, "object separator"), nullptr)); | ||
|  |                 } | ||
|  | 
 | ||
|  |                 // parse values
 | ||
|  |                 get_token(); | ||
|  |                 continue; | ||
|  |             } | ||
|  | 
 | ||
|  |             // closing }
 | ||
|  |             if (JSON_HEDLEY_LIKELY(last_token == token_type::end_object)) | ||
|  |             { | ||
|  |                 if (JSON_HEDLEY_UNLIKELY(!sax->end_object())) | ||
|  |                 { | ||
|  |                     return false; | ||
|  |                 } | ||
|  | 
 | ||
|  |                 // We are done with this object. Before we can parse a
 | ||
|  |                 // new value, we need to evaluate the new state first.
 | ||
|  |                 // By setting skip_to_state_evaluation to false, we
 | ||
|  |                 // are effectively jumping to the beginning of this if.
 | ||
|  |                 JSON_ASSERT(!states.empty()); | ||
|  |                 states.pop_back(); | ||
|  |                 skip_to_state_evaluation = true; | ||
|  |                 continue; | ||
|  |             } | ||
|  | 
 | ||
|  |             return sax->parse_error(m_lexer.get_position(), | ||
|  |                                     m_lexer.get_token_string(), | ||
|  |                                     parse_error::create(101, m_lexer.get_position(), exception_message(token_type::end_object, "object"), nullptr)); | ||
|  |         } | ||
|  |     } | ||
|  | 
 | ||
|  |     /// get next token from lexer
 | ||
|  |     token_type get_token() | ||
|  |     { | ||
|  |         return last_token = m_lexer.scan(); | ||
|  |     } | ||
|  | 
 | ||
|  |     std::string exception_message(const token_type expected, const std::string& context) | ||
|  |     { | ||
|  |         std::string error_msg = "syntax error "; | ||
|  | 
 | ||
|  |         if (!context.empty()) | ||
|  |         { | ||
|  |             error_msg += concat("while parsing ", context, ' '); | ||
|  |         } | ||
|  | 
 | ||
|  |         error_msg += "- "; | ||
|  | 
 | ||
|  |         if (last_token == token_type::parse_error) | ||
|  |         { | ||
|  |             error_msg += concat(m_lexer.get_error_message(), "; last read: '", | ||
|  |                                 m_lexer.get_token_string(), '\''); | ||
|  |         } | ||
|  |         else | ||
|  |         { | ||
|  |             error_msg += concat("unexpected ", lexer_t::token_type_name(last_token)); | ||
|  |         } | ||
|  | 
 | ||
|  |         if (expected != token_type::uninitialized) | ||
|  |         { | ||
|  |             error_msg += concat("; expected ", lexer_t::token_type_name(expected)); | ||
|  |         } | ||
|  | 
 | ||
|  |         return error_msg; | ||
|  |     } | ||
|  | 
 | ||
|  |   private: | ||
|  |     /// callback function
 | ||
|  |     const parser_callback_t<BasicJsonType> callback = nullptr; | ||
|  |     /// the type of the last read token
 | ||
|  |     token_type last_token = token_type::uninitialized; | ||
|  |     /// the lexer
 | ||
|  |     lexer_t m_lexer; | ||
|  |     /// whether to throw exceptions in case of errors
 | ||
|  |     const bool allow_exceptions = true; | ||
|  | }; | ||
|  | 
 | ||
|  | }  // namespace detail
 | ||
|  | NLOHMANN_JSON_NAMESPACE_END |