Commit bbc85f30 authored by Benjamin Peterson's avatar Benjamin Peterson

port simplejson upgrade from the trunk #4136

json also now works only with unicode strings

Patch by Antoine Pitrou; updated by me
parent 9d5e41f9
...@@ -112,7 +112,7 @@ Using json.tool from the shell to validate and pretty-print:: ...@@ -112,7 +112,7 @@ Using json.tool from the shell to validate and pretty-print::
Basic Usage Basic Usage
----------- -----------
.. function:: dump(obj, fp[, skipkeys[, ensure_ascii[, check_circular[, allow_nan[, cls[, indent[, separators[, encoding[, default[, **kw]]]]]]]]]]) .. function:: dump(obj, fp[, skipkeys[, ensure_ascii[, check_circular[, allow_nan[, cls[, indent[, separators[, default[, **kw]]]]]]]]]])
Serialize *obj* as a JSON formatted stream to *fp* (a ``.write()``-supporting Serialize *obj* as a JSON formatted stream to *fp* (a ``.write()``-supporting
file-like object). file-like object).
...@@ -122,11 +122,10 @@ Basic Usage ...@@ -122,11 +122,10 @@ Basic Usage
:class:`float`, :class:`bool`, ``None``) will be skipped instead of raising a :class:`float`, :class:`bool`, ``None``) will be skipped instead of raising a
:exc:`TypeError`. :exc:`TypeError`.
If *ensure_ascii* is ``False`` (default: ``True``), then some chunks written The :mod:`json` module always produces :class:`str` objects, not
to *fp* may be :class:`unicode` instances, subject to normal Python :class:`bytes` objects. Therefore, ``fp.write()`` must support :class:`str`
:class:`str` to :class:`unicode` coercion rules. Unless ``fp.write()`` input.
explicitly understands :class:`unicode` (as in :func:`codecs.getwriter`) this
is likely to cause an error.
If *check_circular* is ``False`` (default: ``True``), then the circular If *check_circular* is ``False`` (default: ``True``), then the circular
reference check for container types will be skipped and a circular reference reference check for container types will be skipped and a circular reference
...@@ -146,8 +145,6 @@ Basic Usage ...@@ -146,8 +145,6 @@ Basic Usage
will be used instead of the default ``(', ', ': ')`` separators. ``(',', will be used instead of the default ``(', ', ': ')`` separators. ``(',',
':')`` is the most compact JSON representation. ':')`` is the most compact JSON representation.
*encoding* is the character encoding for str instances, default is UTF-8.
*default(obj)* is a function that should return a serializable version of *default(obj)* is a function that should return a serializable version of
*obj* or raise :exc:`TypeError`. The default simply raises :exc:`TypeError`. *obj* or raise :exc:`TypeError`. The default simply raises :exc:`TypeError`.
...@@ -156,26 +153,17 @@ Basic Usage ...@@ -156,26 +153,17 @@ Basic Usage
*cls* kwarg. *cls* kwarg.
.. function:: dumps(obj[, skipkeys[, ensure_ascii[, check_circular[, allow_nan[, cls[, indent[, separators[, encoding[, default[, **kw]]]]]]]]]]) .. function:: dumps(obj[, skipkeys[, ensure_ascii[, check_circular[, allow_nan[, cls[, indent[, separators[, default[, **kw]]]]]]]]]])
Serialize *obj* to a JSON formatted :class:`str`. Serialize *obj* to a JSON formatted :class:`str`. The arguments have the
same meaning as in :func:`dump`.
If *ensure_ascii* is ``False``, then the return value will be a
:class:`unicode` instance. The other arguments have the same meaning as in
:func:`dump`.
.. function:: load(fp[, cls[, object_hook[, parse_float[, parse_int[, parse_constant[, object_pairs_hook[, **kw]]]]]]]])
.. function:: load(fp[, encoding[, cls[, object_hook[, parse_float[, parse_int[, parse_constant[, object_pairs_hook[, **kw]]]]]]]])
Deserialize *fp* (a ``.read()``-supporting file-like object containing a JSON Deserialize *fp* (a ``.read()``-supporting file-like object containing a JSON
document) to a Python object. document) to a Python object.
If the contents of *fp* are encoded with an ASCII based encoding other than
UTF-8 (e.g. latin-1), then an appropriate *encoding* name must be specified.
Encodings that are not ASCII based (such as UCS-2) are not allowed, and
should be wrapped with ``codecs.getreader(encoding)(fp)``, or simply decoded
to a :class:`unicode` object and passed to :func:`loads`.
*object_hook* is an optional function that will be called with the result of *object_hook* is an optional function that will be called with the result of
any object literal decode (a :class:`dict`). The return value of any object literal decode (a :class:`dict`). The return value of
*object_hook* will be used instead of the :class:`dict`. This feature can be used *object_hook* will be used instead of the :class:`dict`. This feature can be used
...@@ -241,7 +229,7 @@ Encoders and decoders ...@@ -241,7 +229,7 @@ Encoders and decoders
+---------------+-------------------+ +---------------+-------------------+
| array | list | | array | list |
+---------------+-------------------+ +---------------+-------------------+
| string | unicode | | string | str |
+---------------+-------------------+ +---------------+-------------------+
| number (int) | int | | number (int) | int |
+---------------+-------------------+ +---------------+-------------------+
...@@ -257,13 +245,6 @@ Encoders and decoders ...@@ -257,13 +245,6 @@ Encoders and decoders
It also understands ``NaN``, ``Infinity``, and ``-Infinity`` as their It also understands ``NaN``, ``Infinity``, and ``-Infinity`` as their
corresponding ``float`` values, which is outside the JSON spec. corresponding ``float`` values, which is outside the JSON spec.
*encoding* determines the encoding used to interpret any :class:`str` objects
decoded by this instance (UTF-8 by default). It has no effect when decoding
:class:`unicode` objects.
Note that currently only encodings that are a superset of ASCII work, strings
of other encodings should be passed in as :class:`unicode`.
*object_hook*, if specified, will be called with the result of every JSON *object_hook*, if specified, will be called with the result of every JSON
object decoded and its return value will be used in place of the given object decoded and its return value will be used in place of the given
:class:`dict`. This can be used to provide custom deserializations (e.g. to :class:`dict`. This can be used to provide custom deserializations (e.g. to
...@@ -298,20 +279,20 @@ Encoders and decoders ...@@ -298,20 +279,20 @@ Encoders and decoders
.. method:: decode(s) .. method:: decode(s)
Return the Python representation of *s* (a :class:`str` or Return the Python representation of *s* (a :class:`str` instance
:class:`unicode` instance containing a JSON document) containing a JSON document)
.. method:: raw_decode(s) .. method:: raw_decode(s)
Decode a JSON document from *s* (a :class:`str` or :class:`unicode` Decode a JSON document from *s* (a :class:`str` beginning with a
beginning with a JSON document) and return a 2-tuple of the Python JSON document) and return a 2-tuple of the Python representation
representation and the index in *s* where the document ended. and the index in *s* where the document ended.
This can be used to decode a JSON document from a string that may have This can be used to decode a JSON document from a string that may have
extraneous data at the end. extraneous data at the end.
.. class:: JSONEncoder([skipkeys[, ensure_ascii[, check_circular[, allow_nan[, sort_keys[, indent[, separators[, encoding[, default]]]]]]]]]) .. class:: JSONEncoder([skipkeys[, ensure_ascii[, check_circular[, allow_nan[, sort_keys[, indent[, separators[, default]]]]]]]])
Extensible JSON encoder for Python data structures. Extensible JSON encoder for Python data structures.
...@@ -324,7 +305,7 @@ Encoders and decoders ...@@ -324,7 +305,7 @@ Encoders and decoders
+-------------------+---------------+ +-------------------+---------------+
| list, tuple | array | | list, tuple | array |
+-------------------+---------------+ +-------------------+---------------+
| str, unicode | string | | str | string |
+-------------------+---------------+ +-------------------+---------------+
| int, float | number | | int, float | number |
+-------------------+---------------+ +-------------------+---------------+
...@@ -344,9 +325,9 @@ Encoders and decoders ...@@ -344,9 +325,9 @@ Encoders and decoders
attempt encoding of keys that are not str, int, float or None. If attempt encoding of keys that are not str, int, float or None. If
*skipkeys* is ``True``, such items are simply skipped. *skipkeys* is ``True``, such items are simply skipped.
If *ensure_ascii* is ``True`` (the default), the output is guaranteed to be If *ensure_ascii* is ``True`` (the default), the output is guaranteed to
:class:`str` objects with all incoming unicode characters escaped. If have all incoming non-ASCII characters escaped. If *ensure_ascii* is
*ensure_ascii* is ``False``, the output will be a unicode object. ``False``, these characters will be output as-is.
If *check_circular* is ``True`` (the default), then lists, dicts, and custom If *check_circular* is ``True`` (the default), then lists, dicts, and custom
encoded objects will be checked for circular references during encoding to encoded objects will be checked for circular references during encoding to
...@@ -376,10 +357,6 @@ Encoders and decoders ...@@ -376,10 +357,6 @@ Encoders and decoders
otherwise be serialized. It should return a JSON encodable version of the otherwise be serialized. It should return a JSON encodable version of the
object or raise a :exc:`TypeError`. object or raise a :exc:`TypeError`.
If *encoding* is not ``None``, then all input strings will be transformed
into unicode using that encoding prior to JSON-encoding. The default is
UTF-8.
.. method:: default(o) .. method:: default(o)
......
r"""A simple, fast, extensible JSON encoder and decoder r"""JSON (JavaScript Object Notation) <http://json.org> is a subset of
JSON (JavaScript Object Notation) <http://json.org> is a subset of
JavaScript syntax (ECMA-262 3rd edition) used as a lightweight data JavaScript syntax (ECMA-262 3rd edition) used as a lightweight data
interchange format. interchange format.
json exposes an API familiar to uses of the standard library :mod:`json` exposes an API familiar to users of the standard library
marshal and pickle modules. :mod:`marshal` and :mod:`pickle` modules. It is the externally maintained
version of the :mod:`json` library contained in Python 2.6, but maintains
compatibility with Python 2.4 and Python 2.5 and (currently) has
significant performance advantages, even without using the optional C
extension for speedups.
Encoding basic Python object hierarchies:: Encoding basic Python object hierarchies::
...@@ -32,23 +34,28 @@ Compact encoding:: ...@@ -32,23 +34,28 @@ Compact encoding::
>>> json.dumps([1,2,3,{'4': 5, '6': 7}], separators=(',',':')) >>> json.dumps([1,2,3,{'4': 5, '6': 7}], separators=(',',':'))
'[1,2,3,{"4":5,"6":7}]' '[1,2,3,{"4":5,"6":7}]'
Pretty printing (using repr() because of extraneous whitespace in the output):: Pretty printing::
>>> import json >>> import json
>>> print(repr(json.dumps({'4': 5, '6': 7}, sort_keys=True, indent=4))) >>> s = json.dumps({'4': 5, '6': 7}, sort_keys=True, indent=4)
'{\n "4": 5, \n "6": 7\n}' >>> print('\n'.join([l.rstrip() for l in s.splitlines()]))
{
"4": 5,
"6": 7
}
Decoding JSON:: Decoding JSON::
>>> import json >>> import json
>>> json.loads('["foo", {"bar":["baz", null, 1.0, 2]}]') >>> obj = ['foo', {'bar': ['baz', None, 1.0, 2]}]
['foo', {'bar': ['baz', None, 1.0, 2]}] >>> json.loads('["foo", {"bar":["baz", null, 1.0, 2]}]') == obj
>>> json.loads('"\\"foo\\bar"') True
'"foo\x08ar' >>> json.loads('"\\"foo\\bar"') == '"foo\x08ar'
True
>>> from io import StringIO >>> from io import StringIO
>>> io = StringIO('["streaming API"]') >>> io = StringIO('["streaming API"]')
>>> json.load(io) >>> json.load(io)[0] == 'streaming API'
['streaming API'] True
Specializing JSON object decoding:: Specializing JSON object decoding::
...@@ -61,43 +68,36 @@ Specializing JSON object decoding:: ...@@ -61,43 +68,36 @@ Specializing JSON object decoding::
>>> json.loads('{"__complex__": true, "real": 1, "imag": 2}', >>> json.loads('{"__complex__": true, "real": 1, "imag": 2}',
... object_hook=as_complex) ... object_hook=as_complex)
(1+2j) (1+2j)
>>> import decimal >>> from decimal import Decimal
>>> json.loads('1.1', parse_float=decimal.Decimal) >>> json.loads('1.1', parse_float=Decimal) == Decimal('1.1')
Decimal('1.1') True
Extending JSONEncoder:: Specializing JSON object encoding::
>>> import json >>> import json
>>> class ComplexEncoder(json.JSONEncoder): >>> def encode_complex(obj):
... def default(self, obj): ... if isinstance(obj, complex):
... if isinstance(obj, complex): ... return [obj.real, obj.imag]
... return [obj.real, obj.imag] ... raise TypeError(repr(o) + " is not JSON serializable")
... return json.JSONEncoder.default(self, obj)
... ...
>>> dumps(2 + 1j, cls=ComplexEncoder) >>> json.dumps(2 + 1j, default=encode_complex)
'[2.0, 1.0]'
>>> json.JSONEncoder(default=encode_complex).encode(2 + 1j)
'[2.0, 1.0]' '[2.0, 1.0]'
>>> ComplexEncoder().encode(2 + 1j) >>> ''.join(json.JSONEncoder(default=encode_complex).iterencode(2 + 1j))
'[2.0, 1.0]' '[2.0, 1.0]'
>>> list(ComplexEncoder().iterencode(2 + 1j))
['[', '2.0', ', ', '1.0', ']']
Using json.tool from the shell to validate and Using json.tool from the shell to validate and pretty-print::
pretty-print::
$ echo '{"json":"obj"}' | python -mjson.tool $ echo '{"json":"obj"}' | python -m json.tool
{ {
"json": "obj" "json": "obj"
} }
$ echo '{ 1.2:3.4}' | python -mjson.tool $ echo '{ 1.2:3.4}' | python -m json.tool
Expecting property name: line 1 column 2 (char 2) Expecting property name: line 1 column 2 (char 2)
Note that the JSON produced by this module's default settings
is a subset of YAML, so it may be used as a serializer for that as well.
""" """
__version__ = '2.0.9'
__version__ = '1.9'
__all__ = [ __all__ = [
'dump', 'dumps', 'load', 'loads', 'dump', 'dumps', 'load', 'loads',
'JSONDecoder', 'JSONEncoder', 'JSONDecoder', 'JSONEncoder',
...@@ -115,45 +115,43 @@ _default_encoder = JSONEncoder( ...@@ -115,45 +115,43 @@ _default_encoder = JSONEncoder(
allow_nan=True, allow_nan=True,
indent=None, indent=None,
separators=None, separators=None,
encoding='utf-8',
default=None, default=None,
) )
def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True, def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True,
allow_nan=True, cls=None, indent=None, separators=None, allow_nan=True, cls=None, indent=None, separators=None,
encoding='utf-8', default=None, **kw): default=None, **kw):
"""Serialize ``obj`` as a JSON formatted stream to ``fp`` (a """Serialize ``obj`` as a JSON formatted stream to ``fp`` (a
``.write()``-supporting file-like object). ``.write()``-supporting file-like object).
If ``skipkeys`` is ``True`` then ``dict`` keys that are not basic types If ``skipkeys`` is true then ``dict`` keys that are not basic types
(``str``, ``unicode``, ``int``, ``long``, ``float``, ``bool``, ``None``) (``str``, ``unicode``, ``int``, ``float``, ``bool``, ``None``) will be
will be skipped instead of raising a ``TypeError``. skipped instead of raising a ``TypeError``.
If ``ensure_ascii`` is ``False``, then the some chunks written to ``fp`` If ``ensure_ascii`` is false, then the some chunks written to ``fp``
may be ``unicode`` instances, subject to normal Python ``str`` to may be ``unicode`` instances, subject to normal Python ``str`` to
``unicode`` coercion rules. Unless ``fp.write()`` explicitly ``unicode`` coercion rules. Unless ``fp.write()`` explicitly
understands ``unicode`` (as in ``codecs.getwriter()``) this is likely understands ``unicode`` (as in ``codecs.getwriter()``) this is likely
to cause an error. to cause an error.
If ``check_circular`` is ``False``, then the circular reference check If ``check_circular`` is false, then the circular reference check
for container types will be skipped and a circular reference will for container types will be skipped and a circular reference will
result in an ``OverflowError`` (or worse). result in an ``OverflowError`` (or worse).
If ``allow_nan`` is ``False``, then it will be a ``ValueError`` to If ``allow_nan`` is false, then it will be a ``ValueError`` to
serialize out of range ``float`` values (``nan``, ``inf``, ``-inf``) serialize out of range ``float`` values (``nan``, ``inf``, ``-inf``)
in strict compliance of the JSON specification, instead of using the in strict compliance of the JSON specification, instead of using the
JavaScript equivalents (``NaN``, ``Infinity``, ``-Infinity``). JavaScript equivalents (``NaN``, ``Infinity``, ``-Infinity``).
If ``indent`` is a non-negative integer, then JSON array elements and object If ``indent`` is a non-negative integer, then JSON array elements and
members will be pretty-printed with that indent level. An indent level object members will be pretty-printed with that indent level. An indent
of 0 will only insert newlines. ``None`` is the most compact representation. level of 0 will only insert newlines. ``None`` is the most compact
representation.
If ``separators`` is an ``(item_separator, dict_separator)`` tuple If ``separators`` is an ``(item_separator, dict_separator)`` tuple
then it will be used instead of the default ``(', ', ': ')`` separators. then it will be used instead of the default ``(', ', ': ')`` separators.
``(',', ':')`` is the most compact JSON representation. ``(',', ':')`` is the most compact JSON representation.
``encoding`` is the character encoding for str instances, default is UTF-8.
``default(obj)`` is a function that should return a serializable version ``default(obj)`` is a function that should return a serializable version
of obj or raise TypeError. The default simply raises TypeError. of obj or raise TypeError. The default simply raises TypeError.
...@@ -163,17 +161,17 @@ def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True, ...@@ -163,17 +161,17 @@ def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True,
""" """
# cached encoder # cached encoder
if (skipkeys is False and ensure_ascii is True and if (not skipkeys and ensure_ascii and
check_circular is True and allow_nan is True and check_circular and allow_nan and
cls is None and indent is None and separators is None and cls is None and indent is None and separators is None and
encoding == 'utf-8' and default is None and not kw): default is None and not kw):
iterable = _default_encoder.iterencode(obj) iterable = _default_encoder.iterencode(obj)
else: else:
if cls is None: if cls is None:
cls = JSONEncoder cls = JSONEncoder
iterable = cls(skipkeys=skipkeys, ensure_ascii=ensure_ascii, iterable = cls(skipkeys=skipkeys, ensure_ascii=ensure_ascii,
check_circular=check_circular, allow_nan=allow_nan, indent=indent, check_circular=check_circular, allow_nan=allow_nan, indent=indent,
separators=separators, encoding=encoding, separators=separators,
default=default, **kw).iterencode(obj) default=default, **kw).iterencode(obj)
# could accelerate with writelines in some versions of Python, at # could accelerate with writelines in some versions of Python, at
# a debuggability cost # a debuggability cost
...@@ -183,22 +181,22 @@ def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True, ...@@ -183,22 +181,22 @@ def dump(obj, fp, skipkeys=False, ensure_ascii=True, check_circular=True,
def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True, def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True,
allow_nan=True, cls=None, indent=None, separators=None, allow_nan=True, cls=None, indent=None, separators=None,
encoding='utf-8', default=None, **kw): default=None, **kw):
"""Serialize ``obj`` to a JSON formatted ``str``. """Serialize ``obj`` to a JSON formatted ``str``.
If ``skipkeys`` is ``True`` then ``dict`` keys that are not basic types If ``skipkeys`` is false then ``dict`` keys that are not basic types
(``str``, ``unicode``, ``int``, ``long``, ``float``, ``bool``, ``None``) (``str``, ``unicode``, ``int``, ``float``, ``bool``, ``None``) will be
will be skipped instead of raising a ``TypeError``. skipped instead of raising a ``TypeError``.
If ``ensure_ascii`` is ``False``, then the return value will be a If ``ensure_ascii`` is false, then the return value will be a
``unicode`` instance subject to normal Python ``str`` to ``unicode`` ``unicode`` instance subject to normal Python ``str`` to ``unicode``
coercion rules instead of being escaped to an ASCII ``str``. coercion rules instead of being escaped to an ASCII ``str``.
If ``check_circular`` is ``False``, then the circular reference check If ``check_circular`` is false, then the circular reference check
for container types will be skipped and a circular reference will for container types will be skipped and a circular reference will
result in an ``OverflowError`` (or worse). result in an ``OverflowError`` (or worse).
If ``allow_nan`` is ``False``, then it will be a ``ValueError`` to If ``allow_nan`` is false, then it will be a ``ValueError`` to
serialize out of range ``float`` values (``nan``, ``inf``, ``-inf``) in serialize out of range ``float`` values (``nan``, ``inf``, ``-inf``) in
strict compliance of the JSON specification, instead of using the strict compliance of the JSON specification, instead of using the
JavaScript equivalents (``NaN``, ``Infinity``, ``-Infinity``). JavaScript equivalents (``NaN``, ``Infinity``, ``-Infinity``).
...@@ -212,8 +210,6 @@ def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True, ...@@ -212,8 +210,6 @@ def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True,
then it will be used instead of the default ``(', ', ': ')`` separators. then it will be used instead of the default ``(', ', ': ')`` separators.
``(',', ':')`` is the most compact JSON representation. ``(',', ':')`` is the most compact JSON representation.
``encoding`` is the character encoding for str instances, default is UTF-8.
``default(obj)`` is a function that should return a serializable version ``default(obj)`` is a function that should return a serializable version
of obj or raise TypeError. The default simply raises TypeError. of obj or raise TypeError. The default simply raises TypeError.
...@@ -223,35 +219,27 @@ def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True, ...@@ -223,35 +219,27 @@ def dumps(obj, skipkeys=False, ensure_ascii=True, check_circular=True,
""" """
# cached encoder # cached encoder
if (skipkeys is False and ensure_ascii is True and if (not skipkeys and ensure_ascii and
check_circular is True and allow_nan is True and check_circular and allow_nan and
cls is None and indent is None and separators is None and cls is None and indent is None and separators is None and
encoding == 'utf-8' and default is None and not kw): default is None and not kw):
return _default_encoder.encode(obj) return _default_encoder.encode(obj)
if cls is None: if cls is None:
cls = JSONEncoder cls = JSONEncoder
return cls( return cls(
skipkeys=skipkeys, ensure_ascii=ensure_ascii, skipkeys=skipkeys, ensure_ascii=ensure_ascii,
check_circular=check_circular, allow_nan=allow_nan, indent=indent, check_circular=check_circular, allow_nan=allow_nan, indent=indent,
separators=separators, encoding=encoding, default=default, separators=separators, default=default,
**kw).encode(obj) **kw).encode(obj)
_default_decoder = JSONDecoder(encoding=None, object_hook=None, _default_decoder = JSONDecoder(object_hook=None, object_pairs_hook=None)
object_pairs_hook=None)
def load(fp, encoding=None, cls=None, object_hook=None, parse_float=None, def load(fp, cls=None, object_hook=None, parse_float=None,
parse_int=None, parse_constant=None, object_pairs_hook=None, **kw): parse_int=None, parse_constant=None, object_pairs_hook=None, **kw):
"""Deserialize ``fp`` (a ``.read()``-supporting file-like object """Deserialize ``fp`` (a ``.read()``-supporting file-like object containing
containing a JSON document) to a Python object. a JSON document) to a Python object.
If the contents of ``fp`` is encoded with an ASCII based encoding other
than utf-8 (e.g. latin-1), then an appropriate ``encoding`` name must
be specified. Encodings that are not ASCII based (such as UCS-2) are
not allowed, and should be wrapped with
``codecs.getreader(fp)(encoding)``, or simply decoded to a ``unicode``
object and passed to ``loads()``
``object_hook`` is an optional function that will be called with the ``object_hook`` is an optional function that will be called with the
result of any object literal decode (a ``dict``). The return value of result of any object literal decode (a ``dict``). The return value of
...@@ -263,21 +251,16 @@ def load(fp, encoding=None, cls=None, object_hook=None, parse_float=None, ...@@ -263,21 +251,16 @@ def load(fp, encoding=None, cls=None, object_hook=None, parse_float=None,
""" """
return loads(fp.read(), return loads(fp.read(),
encoding=encoding, cls=cls, object_hook=object_hook, cls=cls, object_hook=object_hook,
parse_float=parse_float, parse_int=parse_int, parse_float=parse_float, parse_int=parse_int,
parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw) parse_constant=parse_constant, object_pairs_hook=object_pairs_hook, **kw)
def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None, def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None,
parse_int=None, parse_constant=None, object_pairs_hook=None, **kw): parse_int=None, parse_constant=None, object_pairs_hook=None, **kw):
"""Deserialize ``s`` (a ``str`` or ``unicode`` instance containing a JSON """Deserialize ``s`` (a ``str`` instance containing a JSON
document) to a Python object. document) to a Python object.
If ``s`` is a ``str`` instance and is encoded with an ASCII based encoding
other than utf-8 (e.g. latin-1) then an appropriate ``encoding`` name
must be specified. Encodings that are not ASCII based (such as UCS-2)
are not allowed and should be decoded to ``unicode`` first.
``object_hook`` is an optional function that will be called with the ``object_hook`` is an optional function that will be called with the
result of any object literal decode (a ``dict``). The return value of result of any object literal decode (a ``dict``). The return value of
``object_hook`` will be used instead of the ``dict``. This feature ``object_hook`` will be used instead of the ``dict``. This feature
...@@ -302,7 +285,7 @@ def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None, ...@@ -302,7 +285,7 @@ def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None,
kwarg. kwarg.
""" """
if (cls is None and encoding is None and object_hook is None and if (cls is None and object_hook is None and
parse_int is None and parse_float is None and parse_int is None and parse_float is None and
parse_constant is None and object_pairs_hook is None and not kw): parse_constant is None and object_pairs_hook is None and not kw):
return _default_decoder.decode(s) return _default_decoder.decode(s)
...@@ -318,4 +301,4 @@ def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None, ...@@ -318,4 +301,4 @@ def loads(s, encoding=None, cls=None, object_hook=None, parse_float=None,
kw['parse_int'] = parse_int kw['parse_int'] = parse_int
if parse_constant is not None: if parse_constant is not None:
kw['parse_constant'] = parse_constant kw['parse_constant'] = parse_constant
return cls(encoding=encoding, **kw).decode(s) return cls(**kw).decode(s)
"""Implementation of JSONDecoder """Implementation of JSONDecoder
""" """
import binascii
import re import re
import sys import sys
import struct
from json.scanner import Scanner, pattern from json.scanner import make_scanner
try: try:
from _json import scanstring as c_scanstring from _json import scanstring as c_scanstring
except ImportError: except ImportError:
...@@ -14,7 +15,14 @@ __all__ = ['JSONDecoder'] ...@@ -14,7 +15,14 @@ __all__ = ['JSONDecoder']
FLAGS = re.VERBOSE | re.MULTILINE | re.DOTALL FLAGS = re.VERBOSE | re.MULTILINE | re.DOTALL
NaN, PosInf, NegInf = float('nan'), float('inf'), float('-inf') def _floatconstants():
_BYTES = binascii.unhexlify(b'7FF80000000000007FF0000000000000')
if sys.byteorder != 'big':
_BYTES = _BYTES[:8][::-1] + _BYTES[8:][::-1]
nan, inf = struct.unpack('dd', _BYTES)
return nan, inf, -inf
NaN, PosInf, NegInf = _floatconstants()
def linecol(doc, pos): def linecol(doc, pos):
...@@ -31,61 +39,43 @@ def linecol(doc, pos): ...@@ -31,61 +39,43 @@ def linecol(doc, pos):
def errmsg(msg, doc, pos, end=None): def errmsg(msg, doc, pos, end=None):
# Note that this function is called from _json
lineno, colno = linecol(doc, pos) lineno, colno = linecol(doc, pos)
if end is None: if end is None:
fmt = '{0}: line {1} column {2} (char {3})' fmt = '{0}: line {1} column {2} (char {3})'
return fmt.format(msg, lineno, colno, pos) return fmt.format(msg, lineno, colno, pos)
#fmt = '%s: line %d column %d (char %d)'
#return fmt % (msg, lineno, colno, pos)
endlineno, endcolno = linecol(doc, end) endlineno, endcolno = linecol(doc, end)
fmt = '{0}: line {1} column {2} - line {3} column {4} (char {5} - {6})' fmt = '{0}: line {1} column {2} - line {3} column {4} (char {5} - {6})'
return fmt.format(msg, lineno, colno, endlineno, endcolno, pos, end) return fmt.format(msg, lineno, colno, endlineno, endcolno, pos, end)
#fmt = '%s: line %d column %d - line %d column %d (char %d - %d)'
#return fmt % (msg, lineno, colno, endlineno, endcolno, pos, end)
_CONSTANTS = { _CONSTANTS = {
'-Infinity': NegInf, '-Infinity': NegInf,
'Infinity': PosInf, 'Infinity': PosInf,
'NaN': NaN, 'NaN': NaN,
'true': True,
'false': False,
'null': None,
} }
def JSONConstant(match, context, c=_CONSTANTS):
s = match.group(0)
fn = getattr(context, 'parse_constant', None)
if fn is None:
rval = c[s]
else:
rval = fn(s)
return rval, None
pattern('(-?Infinity|NaN|true|false|null)')(JSONConstant)
def JSONNumber(match, context):
match = JSONNumber.regex.match(match.string, *match.span())
integer, frac, exp = match.groups()
if frac or exp:
fn = getattr(context, 'parse_float', None) or float
res = fn(integer + (frac or '') + (exp or ''))
else:
fn = getattr(context, 'parse_int', None) or int
res = fn(integer)
return res, None
pattern(r'(-?(?:0|[1-9][0-9]*))(\.[0-9]+)?([eE][-+]?[0-9]+)?')(JSONNumber)
STRINGCHUNK = re.compile(r'(.*?)(["\\\x00-\x1f])', FLAGS) STRINGCHUNK = re.compile(r'(.*?)(["\\\x00-\x1f])', FLAGS)
BACKSLASH = { BACKSLASH = {
'"': '"', '\\': '\\', '/': '/', '"': '"', '\\': '\\', '/': '/',
'b': '\b', 'f': '\f', 'n': '\n', 'r': '\r', 't': '\t', 'b': '\b', 'f': '\f', 'n': '\n', 'r': '\r', 't': '\t',
} }
DEFAULT_ENCODING = "utf-8" def py_scanstring(s, end, strict=True,
_b=BACKSLASH, _m=STRINGCHUNK.match):
"""Scan the string s for a JSON string. End is the index of the
character in s after the quote that started the JSON string.
Unescapes all valid JSON string escape sequences and raises ValueError
on attempt to decode an invalid string. If strict is False then literal
control characters are allowed in the string.
def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHUNK.match): Returns a tuple of the decoded string and the index of the character in s
if encoding is None: after the end quote."""
encoding = DEFAULT_ENCODING
chunks = [] chunks = []
_append = chunks.append _append = chunks.append
begin = end - 1 begin = end - 1
...@@ -96,14 +86,16 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU ...@@ -96,14 +86,16 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU
errmsg("Unterminated string starting at", s, begin)) errmsg("Unterminated string starting at", s, begin))
end = chunk.end() end = chunk.end()
content, terminator = chunk.groups() content, terminator = chunk.groups()
# Content is contains zero or more unescaped string characters
if content: if content:
if not isinstance(content, str):
content = str(content, encoding)
_append(content) _append(content)
# Terminator is the end of string, a literal control character,
# or a backslash denoting that an escape sequence follows
if terminator == '"': if terminator == '"':
break break
elif terminator != '\\': elif terminator != '\\':
if strict: if strict:
#msg = "Invalid control character %r at" % (terminator,)
msg = "Invalid control character {0!r} at".format(terminator) msg = "Invalid control character {0!r} at".format(terminator)
raise ValueError(errmsg(msg, s, end)) raise ValueError(errmsg(msg, s, end))
else: else:
...@@ -114,9 +106,10 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU ...@@ -114,9 +106,10 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU
except IndexError: except IndexError:
raise ValueError( raise ValueError(
errmsg("Unterminated string starting at", s, begin)) errmsg("Unterminated string starting at", s, begin))
# If not a unicode escape sequence, must be in the lookup table
if esc != 'u': if esc != 'u':
try: try:
m = _b[esc] char = _b[esc]
except KeyError: except KeyError:
msg = "Invalid \\escape: {0!r}".format(esc) msg = "Invalid \\escape: {0!r}".format(esc)
raise ValueError(errmsg(msg, s, end)) raise ValueError(errmsg(msg, s, end))
...@@ -124,131 +117,138 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU ...@@ -124,131 +117,138 @@ def py_scanstring(s, end, encoding=None, strict=True, _b=BACKSLASH, _m=STRINGCHU
else: else:
esc = s[end + 1:end + 5] esc = s[end + 1:end + 5]
next_end = end + 5 next_end = end + 5
msg = "Invalid \\uXXXX escape" if len(esc) != 4:
try: msg = "Invalid \\uXXXX escape"
if len(esc) != 4:
raise ValueError
uni = int(esc, 16)
if 0xd800 <= uni <= 0xdbff and sys.maxunicode > 65535:
msg = "Invalid \\uXXXX\\uXXXX surrogate pair"
if not s[end + 5:end + 7] == '\\u':
raise ValueError
esc2 = s[end + 7:end + 11]
if len(esc2) != 4:
raise ValueError
uni2 = int(esc2, 16)
uni = 0x10000 + (((uni - 0xd800) << 10) | (uni2 - 0xdc00))
next_end += 6
m = chr(uni)
except ValueError:
raise ValueError(errmsg(msg, s, end)) raise ValueError(errmsg(msg, s, end))
uni = int(esc, 16)
# Check for surrogate pair on UCS-4 systems
if 0xd800 <= uni <= 0xdbff and sys.maxunicode > 65535:
msg = "Invalid \\uXXXX\\uXXXX surrogate pair"
if not s[end + 5:end + 7] == '\\u':
raise ValueError(errmsg(msg, s, end))
esc2 = s[end + 7:end + 11]
if len(esc2) != 4:
raise ValueError(errmsg(msg, s, end))
uni2 = int(esc2, 16)
uni = 0x10000 + (((uni - 0xd800) << 10) | (uni2 - 0xdc00))
next_end += 6
char = chr(uni)
end = next_end end = next_end
_append(m) _append(char)
return ''.join(chunks), end return ''.join(chunks), end
# Use speedup # Use speedup if available
if c_scanstring is not None: scanstring = c_scanstring or py_scanstring
scanstring = c_scanstring
else:
scanstring = py_scanstring
def JSONString(match, context):
encoding = getattr(context, 'encoding', None)
strict = getattr(context, 'strict', True)
return scanstring(match.string, match.end(), encoding, strict)
pattern(r'"')(JSONString)
WHITESPACE = re.compile(r'[ \t\n\r]*', FLAGS)
WHITESPACE_STR = ' \t\n\r'
WHITESPACE = re.compile(r'\s*', FLAGS)
def JSONObject(s_and_end, strict, scan_once, object_hook, object_pairs_hook,
def JSONObject(match, context, _w=WHITESPACE.match): _w=WHITESPACE.match, _ws=WHITESPACE_STR):
s, end = s_and_end
pairs = [] pairs = []
pairs_append = pairs.append pairs_append = pairs.append
s = match.string # Use a slice to prevent IndexError from being raised, the following
end = _w(s, match.end()).end() # check will raise a more specific ValueError if the string is empty
nextchar = s[end:end + 1] nextchar = s[end:end + 1]
# Trivial empty object # Normally we expect nextchar == '"'
if nextchar == '}':
return pairs, end + 1
if nextchar != '"': if nextchar != '"':
raise ValueError(errmsg("Expecting property name", s, end)) if nextchar in _ws:
end = _w(s, end).end()
nextchar = s[end:end + 1]
# Trivial empty object
if nextchar == '}':
return pairs, end + 1
elif nextchar != '"':
raise ValueError(errmsg("Expecting property name", s, end))
end += 1 end += 1
encoding = getattr(context, 'encoding', None)
strict = getattr(context, 'strict', True)
iterscan = JSONScanner.iterscan
while True: while True:
key, end = scanstring(s, end, encoding, strict) key, end = scanstring(s, end, strict)
end = _w(s, end).end() # To skip some function call overhead we optimize the fast paths where
# the JSON key separator is ": " or just ":".
if s[end:end + 1] != ':': if s[end:end + 1] != ':':
raise ValueError(errmsg("Expecting : delimiter", s, end)) end = _w(s, end).end()
end = _w(s, end + 1).end() if s[end:end + 1] != ':':
raise ValueError(errmsg("Expecting : delimiter", s, end))
end += 1
try: try:
value, end = next(iterscan(s, idx=end, context=context)) if s[end] in _ws:
end += 1
if s[end] in _ws:
end = _w(s, end + 1).end()
except IndexError:
pass
try:
value, end = scan_once(s, end)
except StopIteration: except StopIteration:
raise ValueError(errmsg("Expecting object", s, end)) raise ValueError(errmsg("Expecting object", s, end))
pairs_append((key, value)) pairs_append((key, value))
end = _w(s, end).end() try:
nextchar = s[end:end + 1] nextchar = s[end]
if nextchar in _ws:
end = _w(s, end + 1).end()
nextchar = s[end]
except IndexError:
nextchar = ''
end += 1 end += 1
if nextchar == '}': if nextchar == '}':
break break
if nextchar != ',': elif nextchar != ',':
raise ValueError(errmsg("Expecting , delimiter", s, end - 1)) raise ValueError(errmsg("Expecting , delimiter", s, end - 1))
end = _w(s, end).end() end = _w(s, end).end()
nextchar = s[end:end + 1] nextchar = s[end:end + 1]
end += 1 end += 1
if nextchar != '"': if nextchar != '"':
raise ValueError(errmsg("Expecting property name", s, end - 1)) raise ValueError(errmsg("Expecting property name", s, end - 1))
object_pairs_hook = getattr(context, 'object_pairs_hook', None)
if object_pairs_hook is not None: if object_pairs_hook is not None:
result = object_pairs_hook(pairs) result = object_pairs_hook(pairs)
return result, end return result, end
pairs = dict(pairs) pairs = dict(pairs)
object_hook = getattr(context, 'object_hook', None)
if object_hook is not None: if object_hook is not None:
pairs = object_hook(pairs) pairs = object_hook(pairs)
return pairs, end return pairs, end
pattern(r'{')(JSONObject)
def JSONArray(match, context, _w=WHITESPACE.match): def JSONArray(s_and_end, scan_once, context, _w=WHITESPACE.match):
s, end = s_and_end
values = [] values = []
s = match.string
end = _w(s, match.end()).end()
# Look-ahead for trivial empty array
nextchar = s[end:end + 1] nextchar = s[end:end + 1]
if nextchar in _ws:
end = _w(s, end + 1).end()
nextchar = s[end:end + 1]
# Look-ahead for trivial empty array
if nextchar == ']': if nextchar == ']':
return values, end + 1 return values, end + 1
iterscan = JSONScanner.iterscan _append = values.append
while True: while True:
try: try:
value, end = next(iterscan(s, idx=end, context=context)) value, end = scan_once(s, end)
except StopIteration: except StopIteration:
raise ValueError(errmsg("Expecting object", s, end)) raise ValueError(errmsg("Expecting object", s, end))
values.append(value) _append(value)
end = _w(s, end).end()
nextchar = s[end:end + 1] nextchar = s[end:end + 1]
if nextchar in _ws:
end = _w(s, end + 1).end()
nextchar = s[end:end + 1]
end += 1 end += 1
if nextchar == ']': if nextchar == ']':
break break
if nextchar != ',': elif nextchar != ',':
raise ValueError(errmsg("Expecting , delimiter", s, end)) raise ValueError(errmsg("Expecting , delimiter", s, end))
end = _w(s, end).end() try:
return values, end if s[end] in _ws:
pattern(r'\[')(JSONArray) end += 1
if s[end] in _ws:
end = _w(s, end + 1).end()
ANYTHING = [ except IndexError:
JSONObject, pass
JSONArray,
JSONString,
JSONConstant,
JSONNumber,
]
JSONScanner = Scanner(ANYTHING) return values, end
class JSONDecoder(object): class JSONDecoder(object):
...@@ -278,23 +278,14 @@ class JSONDecoder(object): ...@@ -278,23 +278,14 @@ class JSONDecoder(object):
It also understands ``NaN``, ``Infinity``, and ``-Infinity`` as It also understands ``NaN``, ``Infinity``, and ``-Infinity`` as
their corresponding ``float`` values, which is outside the JSON spec. their corresponding ``float`` values, which is outside the JSON spec.
"""
_scanner = Scanner(ANYTHING) """
__all__ = ['__init__', 'decode', 'raw_decode']
def __init__(self, encoding=None, object_hook=None, parse_float=None, def __init__(self, object_hook=None, parse_float=None,
parse_int=None, parse_constant=None, strict=True, parse_int=None, parse_constant=None, strict=True,
object_pairs_hook=None): object_pairs_hook=None):
"""``encoding`` determines the encoding used to interpret any ``str`` """``object_hook``, if specified, will be called with the result
objects decoded by this instance (utf-8 by default). It has no of every JSON object decoded and its return value will be used in
effect when decoding ``unicode`` objects.
Note that currently only encodings that are a superset of ASCII work,
strings of other encodings should be passed in as ``unicode``.
``object_hook``, if specified, will be called with the result of
every JSON object decoded and its return value will be used in
place of the given ``dict``. This can be used to provide custom place of the given ``dict``. This can be used to provide custom
deserializations (e.g. to support JSON-RPC class hinting). deserializations (e.g. to support JSON-RPC class hinting).
...@@ -309,22 +300,25 @@ class JSONDecoder(object): ...@@ -309,22 +300,25 @@ class JSONDecoder(object):
for JSON integers (e.g. float). for JSON integers (e.g. float).
``parse_constant``, if specified, will be called with one of the ``parse_constant``, if specified, will be called with one of the
following strings: -Infinity, Infinity, NaN, null, true, false. following strings: -Infinity, Infinity, NaN.
This can be used to raise an exception if invalid JSON numbers This can be used to raise an exception if invalid JSON numbers
are encountered. are encountered.
""" """
self.encoding = encoding
self.object_hook = object_hook self.object_hook = object_hook
self.object_pairs_hook = object_pairs_hook self.parse_float = parse_float or float
self.parse_float = parse_float self.parse_int = parse_int or int
self.parse_int = parse_int self.parse_constant = parse_constant or _CONSTANTS.__getitem__
self.parse_constant = parse_constant
self.strict = strict self.strict = strict
self.object_pairs_hook = object_pairs_hook
self.parse_object = JSONObject
self.parse_array = JSONArray
self.parse_string = scanstring
self.scan_once = make_scanner(self)
def decode(self, s, _w=WHITESPACE.match): def decode(self, s, _w=WHITESPACE.match):
""" """Return the Python representation of ``s`` (a ``str`` or ``unicode``
Return the Python representation of ``s`` (a ``str`` or ``unicode``
instance containing a JSON document) instance containing a JSON document)
""" """
...@@ -334,18 +328,17 @@ class JSONDecoder(object): ...@@ -334,18 +328,17 @@ class JSONDecoder(object):
raise ValueError(errmsg("Extra data", s, end, len(s))) raise ValueError(errmsg("Extra data", s, end, len(s)))
return obj return obj
def raw_decode(self, s, **kw): def raw_decode(self, s, idx=0):
"""Decode a JSON document from ``s`` (a ``str`` or ``unicode`` beginning """Decode a JSON document from ``s`` (a ``str`` or ``unicode``
with a JSON document) and return a 2-tuple of the Python beginning with a JSON document) and return a 2-tuple of the Python
representation and the index in ``s`` where the document ended. representation and the index in ``s`` where the document ended.
This can be used to decode a JSON document from a string that may This can be used to decode a JSON document from a string that may
have extraneous data at the end. have extraneous data at the end.
""" """
kw.setdefault('context', self)
try: try:
obj, end = next(self._scanner.iterscan(s, **kw)) obj, end = self.scan_once(s, idx)
except StopIteration: except StopIteration:
raise ValueError("No JSON object could be decoded") raise ValueError("No JSON object could be decoded")
return obj, end return obj, end
"""Implementation of JSONEncoder """Implementation of JSONEncoder
""" """
import re import re
import math
try: try:
from _json import encode_basestring_ascii as c_encode_basestring_ascii from _json import encode_basestring_ascii as c_encode_basestring_ascii
except ImportError: except ImportError:
c_encode_basestring_ascii = None c_encode_basestring_ascii = None
try:
__all__ = ['JSONEncoder'] from _json import make_encoder as c_make_encoder
except ImportError:
c_make_encoder = None
ESCAPE = re.compile(r'[\x00-\x1f\\"\b\f\n\r\t]') ESCAPE = re.compile(r'[\x00-\x1f\\"\b\f\n\r\t]')
ESCAPE_ASCII = re.compile(r'([\\"]|[^\ -~])') ESCAPE_ASCII = re.compile(r'([\\"]|[^\ -~])')
HAS_UTF8 = re.compile(r'[\x80-\xff]') HAS_UTF8 = re.compile(b'[\x80-\xff]')
ESCAPE_DCT = { ESCAPE_DCT = {
'\\': '\\\\', '\\': '\\\\',
'"': '\\"', '"': '\\"',
...@@ -25,30 +25,12 @@ ESCAPE_DCT = { ...@@ -25,30 +25,12 @@ ESCAPE_DCT = {
} }
for i in range(0x20): for i in range(0x20):
ESCAPE_DCT.setdefault(chr(i), '\\u{0:04x}'.format(i)) ESCAPE_DCT.setdefault(chr(i), '\\u{0:04x}'.format(i))
#ESCAPE_DCT.setdefault(chr(i), '\\u%04x' % (i,))
# Assume this produces an infinity on all machines (probably not guaranteed)
INFINITY = float('1e66666')
FLOAT_REPR = repr FLOAT_REPR = repr
def floatstr(o, allow_nan=True):
# Check for specials. Note that this type of test is processor- and/or
# platform-specific, so do tests which don't depend on the internals.
if math.isnan(o):
text = 'NaN'
elif math.isinf(o):
if math.copysign(1., o) == 1.:
text = 'Infinity'
else:
text = '-Infinity'
else:
return FLOAT_REPR(o)
if not allow_nan:
msg = "Out of range float values are not JSON compliant: " + repr(o)
raise ValueError(msg)
return text
def encode_basestring(s): def encode_basestring(s):
"""Return a JSON representation of a Python string """Return a JSON representation of a Python string
...@@ -59,8 +41,9 @@ def encode_basestring(s): ...@@ -59,8 +41,9 @@ def encode_basestring(s):
def py_encode_basestring_ascii(s): def py_encode_basestring_ascii(s):
if isinstance(s, bytes): # and HAS_UTF8.search(s) is not None: """Return an ASCII-only JSON representation of a Python string
s = s.decode('utf-8')
"""
def replace(match): def replace(match):
s = match.group(0) s = match.group(0)
try: try:
...@@ -69,20 +52,18 @@ def py_encode_basestring_ascii(s): ...@@ -69,20 +52,18 @@ def py_encode_basestring_ascii(s):
n = ord(s) n = ord(s)
if n < 0x10000: if n < 0x10000:
return '\\u{0:04x}'.format(n) return '\\u{0:04x}'.format(n)
#return '\\u%04x' % (n,)
else: else:
# surrogate pair # surrogate pair
n -= 0x10000 n -= 0x10000
s1 = 0xd800 | ((n >> 10) & 0x3ff) s1 = 0xd800 | ((n >> 10) & 0x3ff)
s2 = 0xdc00 | (n & 0x3ff) s2 = 0xdc00 | (n & 0x3ff)
return '\\u{0:04x}\\u{1:04x}'.format(s1, s2) return '\\u{0:04x}\\u{1:04x}'.format(s1, s2)
return '"' + (ESCAPE_ASCII.sub(replace, s)) + '"' return '"' + ESCAPE_ASCII.sub(replace, s) + '"'
if c_encode_basestring_ascii is not None: encode_basestring_ascii = (
encode_basestring_ascii = c_encode_basestring_ascii c_encode_basestring_ascii or py_encode_basestring_ascii)
else:
encode_basestring_ascii = py_encode_basestring_ascii
class JSONEncoder(object): class JSONEncoder(object):
"""Extensible JSON <http://json.org> encoder for Python data structures. """Extensible JSON <http://json.org> encoder for Python data structures.
...@@ -113,33 +94,32 @@ class JSONEncoder(object): ...@@ -113,33 +94,32 @@ class JSONEncoder(object):
implementation (to raise ``TypeError``). implementation (to raise ``TypeError``).
""" """
__all__ = ['__init__', 'default', 'encode', 'iterencode']
item_separator = ', ' item_separator = ', '
key_separator = ': ' key_separator = ': '
def __init__(self, skipkeys=False, ensure_ascii=True, def __init__(self, skipkeys=False, ensure_ascii=True,
check_circular=True, allow_nan=True, sort_keys=False, check_circular=True, allow_nan=True, sort_keys=False,
indent=None, separators=None, encoding='utf-8', default=None): indent=None, separators=None, default=None):
"""Constructor for JSONEncoder, with sensible defaults. """Constructor for JSONEncoder, with sensible defaults.
If skipkeys is False, then it is a TypeError to attempt If skipkeys is false, then it is a TypeError to attempt
encoding of keys that are not str, int, long, float or None. If encoding of keys that are not str, int, long, float or None. If
skipkeys is True, such items are simply skipped. skipkeys is True, such items are simply skipped.
If ensure_ascii is True, the output is guaranteed to be str If ensure_ascii is true, the output is guaranteed to be str
objects with all incoming unicode characters escaped. If objects with all incoming unicode characters escaped. If
ensure_ascii is false, the output will be unicode object. ensure_ascii is false, the output will be unicode object.
If check_circular is True, then lists, dicts, and custom encoded If check_circular is true, then lists, dicts, and custom encoded
objects will be checked for circular references during encoding to objects will be checked for circular references during encoding to
prevent an infinite recursion (which would cause an OverflowError). prevent an infinite recursion (which would cause an OverflowError).
Otherwise, no such check takes place. Otherwise, no such check takes place.
If allow_nan is True, then NaN, Infinity, and -Infinity will be If allow_nan is true, then NaN, Infinity, and -Infinity will be
encoded as such. This behavior is not JSON specification compliant, encoded as such. This behavior is not JSON specification compliant,
but is consistent with most JavaScript based encoders and decoders. but is consistent with most JavaScript based encoders and decoders.
Otherwise, it will be a ValueError to encode such floats. Otherwise, it will be a ValueError to encode such floats.
If sort_keys is True, then the output of dictionaries will be If sort_keys is true, then the output of dictionaries will be
sorted by key; this is useful for regression tests to ensure sorted by key; this is useful for regression tests to ensure
that JSON serializations can be compared on a day-to-day basis. that JSON serializations can be compared on a day-to-day basis.
...@@ -156,28 +136,130 @@ class JSONEncoder(object): ...@@ -156,28 +136,130 @@ class JSONEncoder(object):
that can't otherwise be serialized. It should return a JSON encodable that can't otherwise be serialized. It should return a JSON encodable
version of the object or raise a ``TypeError``. version of the object or raise a ``TypeError``.
If encoding is not None, then all input strings will be
transformed into unicode using that encoding prior to JSON-encoding.
The default is UTF-8.
""" """
self.skipkeys = skipkeys self.skipkeys = skipkeys
self.ensure_ascii = ensure_ascii self.ensure_ascii = ensure_ascii
self.check_circular = check_circular self.check_circular = check_circular
self.allow_nan = allow_nan self.allow_nan = allow_nan
self.sort_keys = sort_keys self.sort_keys = sort_keys
self.indent = indent self.indent = indent
self.current_indent_level = 0
if separators is not None: if separators is not None:
self.item_separator, self.key_separator = separators self.item_separator, self.key_separator = separators
if default is not None: if default is not None:
self.default = default self.default = default
self.encoding = encoding
def _newline_indent(self): def default(self, o):
return '\n' + (' ' * (self.indent * self.current_indent_level)) """Implement this method in a subclass such that it returns
a serializable object for ``o``, or calls the base implementation
(to raise a ``TypeError``).
For example, to support arbitrary iterators, you could
implement default like this::
def default(self, o):
try:
iterable = iter(o)
except TypeError:
pass
else:
return list(iterable)
return JSONEncoder.default(self, o)
"""
raise TypeError(repr(o) + " is not JSON serializable")
def encode(self, o):
"""Return a JSON string representation of a Python data structure.
>>> JSONEncoder().encode({"foo": ["bar", "baz"]})
'{"foo": ["bar", "baz"]}'
"""
# This is for extremely simple cases and benchmarks.
if isinstance(o, str):
if self.ensure_ascii:
return encode_basestring_ascii(o)
else:
return encode_basestring(o)
# This doesn't pass the iterator directly to ''.join() because the
# exceptions aren't as detailed. The list call should be roughly
# equivalent to the PySequence_Fast that ''.join() would do.
chunks = self.iterencode(o, _one_shot=True)
if not isinstance(chunks, (list, tuple)):
chunks = list(chunks)
return ''.join(chunks)
def iterencode(self, o, _one_shot=False):
"""Encode the given object and yield each string
representation as available.
For example::
for chunk in JSONEncoder().iterencode(bigobject):
mysocket.write(chunk)
"""
if self.check_circular:
markers = {}
else:
markers = None
if self.ensure_ascii:
_encoder = encode_basestring_ascii
else:
_encoder = encode_basestring
def floatstr(o, allow_nan=self.allow_nan,
_repr=FLOAT_REPR, _inf=INFINITY, _neginf=-INFINITY):
# Check for specials. Note that this type of test is processor
# and/or platform-specific, so do tests which don't depend on the
# internals.
if o != o:
text = 'NaN'
elif o == _inf:
text = 'Infinity'
elif o == _neginf:
text = '-Infinity'
else:
return _repr(o)
if not allow_nan:
raise ValueError(
"Out of range float values are not JSON compliant: " +
repr(o))
return text
def _iterencode_list(self, lst, markers=None):
if (_one_shot and c_make_encoder is not None
and not self.indent and not self.sort_keys):
_iterencode = c_make_encoder(
markers, self.default, _encoder, self.indent,
self.key_separator, self.item_separator, self.sort_keys,
self.skipkeys, self.allow_nan)
else:
_iterencode = _make_iterencode(
markers, self.default, _encoder, self.indent, floatstr,
self.key_separator, self.item_separator, self.sort_keys,
self.skipkeys, _one_shot)
return _iterencode(o, 0)
def _make_iterencode(markers, _default, _encoder, _indent, _floatstr,
_key_separator, _item_separator, _sort_keys, _skipkeys, _one_shot,
## HACK: hand-optimized bytecode; turn globals into locals
ValueError=ValueError,
dict=dict,
float=float,
id=id,
int=int,
isinstance=isinstance,
list=list,
str=str,
tuple=tuple,
):
def _iterencode_list(lst, _current_indent_level):
if not lst: if not lst:
yield '[]' yield '[]'
return return
...@@ -186,31 +268,51 @@ class JSONEncoder(object): ...@@ -186,31 +268,51 @@ class JSONEncoder(object):
if markerid in markers: if markerid in markers:
raise ValueError("Circular reference detected") raise ValueError("Circular reference detected")
markers[markerid] = lst markers[markerid] = lst
yield '[' buf = '['
if self.indent is not None: if _indent is not None:
self.current_indent_level += 1 _current_indent_level += 1
newline_indent = self._newline_indent() newline_indent = '\n' + (' ' * (_indent * _current_indent_level))
separator = self.item_separator + newline_indent separator = _item_separator + newline_indent
yield newline_indent buf += newline_indent
else: else:
newline_indent = None newline_indent = None
separator = self.item_separator separator = _item_separator
first = True first = True
for value in lst: for value in lst:
if first: if first:
first = False first = False
else: else:
yield separator buf = separator
for chunk in self._iterencode(value, markers): if isinstance(value, str):
yield chunk yield buf + _encoder(value)
elif value is None:
yield buf + 'null'
elif value is True:
yield buf + 'true'
elif value is False:
yield buf + 'false'
elif isinstance(value, int):
yield buf + str(value)
elif isinstance(value, float):
yield buf + _floatstr(value)
else:
yield buf
if isinstance(value, (list, tuple)):
chunks = _iterencode_list(value, _current_indent_level)
elif isinstance(value, dict):
chunks = _iterencode_dict(value, _current_indent_level)
else:
chunks = _iterencode(value, _current_indent_level)
for chunk in chunks:
yield chunk
if newline_indent is not None: if newline_indent is not None:
self.current_indent_level -= 1 _current_indent_level -= 1
yield self._newline_indent() yield '\n' + (' ' * (_indent * _current_indent_level))
yield ']' yield ']'
if markers is not None: if markers is not None:
del markers[markerid] del markers[markerid]
def _iterencode_dict(self, dct, markers=None): def _iterencode_dict(dct, _current_indent_level):
if not dct: if not dct:
yield '{}' yield '{}'
return return
...@@ -220,78 +322,75 @@ class JSONEncoder(object): ...@@ -220,78 +322,75 @@ class JSONEncoder(object):
raise ValueError("Circular reference detected") raise ValueError("Circular reference detected")
markers[markerid] = dct markers[markerid] = dct
yield '{' yield '{'
key_separator = self.key_separator if _indent is not None:
if self.indent is not None: _current_indent_level += 1
self.current_indent_level += 1 newline_indent = '\n' + (' ' * (_indent * _current_indent_level))
newline_indent = self._newline_indent() item_separator = _item_separator + newline_indent
item_separator = self.item_separator + newline_indent
yield newline_indent yield newline_indent
else: else:
newline_indent = None newline_indent = None
item_separator = self.item_separator item_separator = _item_separator
first = True first = True
if self.ensure_ascii: if _sort_keys:
encoder = encode_basestring_ascii items = sorted(dct.items(), key=lambda kv: kv[0])
else:
encoder = encode_basestring
allow_nan = self.allow_nan
if self.sort_keys:
keys = list(dct.keys())
keys.sort()
items = [(k, dct[k]) for k in keys]
else: else:
items = iter(dct.items()) items = dct.items()
_encoding = self.encoding
_do_decode = (_encoding is not None
and not (_encoding == 'utf-8'))
for key, value in items: for key, value in items:
if isinstance(key, str): if isinstance(key, str):
if _do_decode:
key = key.decode(_encoding)
elif isinstance(key, str):
pass pass
# JavaScript is weakly typed for these, so it makes sense to # JavaScript is weakly typed for these, so it makes sense to
# also allow them. Many encoders seem to do something like this. # also allow them. Many encoders seem to do something like this.
elif isinstance(key, float): elif isinstance(key, float):
key = floatstr(key, allow_nan) key = _floatstr(key)
elif isinstance(key, (int, int)):
key = str(key)
elif key is True: elif key is True:
key = 'true' key = 'true'
elif key is False: elif key is False:
key = 'false' key = 'false'
elif key is None: elif key is None:
key = 'null' key = 'null'
elif self.skipkeys: elif isinstance(key, int):
key = str(key)
elif _skipkeys:
continue continue
else: else:
raise TypeError("key {0!r} is not a string".format(key)) raise TypeError("key " + repr(key) + " is not a string")
if first: if first:
first = False first = False
else: else:
yield item_separator yield item_separator
yield encoder(key) yield _encoder(key)
yield key_separator yield _key_separator
for chunk in self._iterencode(value, markers): if isinstance(value, str):
yield chunk yield _encoder(value)
elif value is None:
yield 'null'
elif value is True:
yield 'true'
elif value is False:
yield 'false'
elif isinstance(value, int):
yield str(value)
elif isinstance(value, float):
yield _floatstr(value)
else:
if isinstance(value, (list, tuple)):
chunks = _iterencode_list(value, _current_indent_level)
elif isinstance(value, dict):
chunks = _iterencode_dict(value, _current_indent_level)
else:
chunks = _iterencode(value, _current_indent_level)
for chunk in chunks:
yield chunk
if newline_indent is not None: if newline_indent is not None:
self.current_indent_level -= 1 _current_indent_level -= 1
yield self._newline_indent() yield '\n' + (' ' * (_indent * _current_indent_level))
yield '}' yield '}'
if markers is not None: if markers is not None:
del markers[markerid] del markers[markerid]
def _iterencode(self, o, markers=None): def _iterencode(o, _current_indent_level):
if isinstance(o, str): if isinstance(o, str):
if self.ensure_ascii: yield _encoder(o)
encoder = encode_basestring_ascii
else:
encoder = encode_basestring
_encoding = self.encoding
if (_encoding is not None and isinstance(o, str)
and not (_encoding == 'utf-8')):
o = o.decode(_encoding)
yield encoder(o)
elif o is None: elif o is None:
yield 'null' yield 'null'
elif o is True: elif o is True:
...@@ -301,12 +400,12 @@ class JSONEncoder(object): ...@@ -301,12 +400,12 @@ class JSONEncoder(object):
elif isinstance(o, (int, int)): elif isinstance(o, (int, int)):
yield str(o) yield str(o)
elif isinstance(o, float): elif isinstance(o, float):
yield floatstr(o, self.allow_nan) yield _floatstr(o)
elif isinstance(o, (list, tuple)): elif isinstance(o, (list, tuple)):
for chunk in self._iterencode_list(o, markers): for chunk in _iterencode_list(o, _current_indent_level):
yield chunk yield chunk
elif isinstance(o, dict): elif isinstance(o, dict):
for chunk in self._iterencode_dict(o, markers): for chunk in _iterencode_dict(o, _current_indent_level):
yield chunk yield chunk
else: else:
if markers is not None: if markers is not None:
...@@ -314,71 +413,9 @@ class JSONEncoder(object): ...@@ -314,71 +413,9 @@ class JSONEncoder(object):
if markerid in markers: if markerid in markers:
raise ValueError("Circular reference detected") raise ValueError("Circular reference detected")
markers[markerid] = o markers[markerid] = o
for chunk in self._iterencode_default(o, markers): o = _default(o)
for chunk in _iterencode(o, _current_indent_level):
yield chunk yield chunk
if markers is not None: if markers is not None:
del markers[markerid] del markers[markerid]
return _iterencode
def _iterencode_default(self, o, markers=None):
newobj = self.default(o)
return self._iterencode(newobj, markers)
def default(self, o):
"""Implement this method in a subclass such that it returns a serializable
object for ``o``, or calls the base implementation (to raise a
``TypeError``).
For example, to support arbitrary iterators, you could implement
default like this::
def default(self, o):
try:
iterable = iter(o)
except TypeError:
pass
else:
return list(iterable)
return JSONEncoder.default(self, o)
"""
raise TypeError(repr(o) + " is not JSON serializable")
def encode(self, o):
"""Return a JSON string representation of a Python data structure.
>>> JSONEncoder().encode({"foo": ["bar", "baz"]})
'{"foo": ["bar", "baz"]}'
"""
# This is for extremely simple cases and benchmarks.
if isinstance(o, (str, bytes)):
if isinstance(o, bytes):
_encoding = self.encoding
if (_encoding is not None
and not (_encoding == 'utf-8')):
o = o.decode(_encoding)
if self.ensure_ascii:
return encode_basestring_ascii(o)
else:
return encode_basestring(o)
# This doesn't pass the iterator directly to ''.join() because the
# exceptions aren't as detailed. The list call should be roughly
# equivalent to the PySequence_Fast that ''.join() would do.
chunks = list(self.iterencode(o))
return ''.join(chunks)
def iterencode(self, o):
"""Encode the given object and yield each string representation as
available.
For example::
for chunk in JSONEncoder().iterencode(bigobject):
mysocket.write(chunk)
"""
if self.check_circular:
markers = {}
else:
markers = None
return self._iterencode(o, markers)
"""Iterator based sre token scanner """JSON token scanner
""" """
import re import re
import sre_parse try:
import sre_compile from _json import make_scanner as c_make_scanner
import sre_constants except ImportError:
c_make_scanner = None
from re import VERBOSE, MULTILINE, DOTALL
from sre_constants import BRANCH, SUBPATTERN
__all__ = ['Scanner', 'pattern'] __all__ = ['make_scanner']
FLAGS = (VERBOSE | MULTILINE | DOTALL) NUMBER_RE = re.compile(
r'(-?(?:0|[1-9]\d*))(\.\d+)?([eE][-+]?\d+)?',
(re.VERBOSE | re.MULTILINE | re.DOTALL))
class Scanner(object): def py_make_scanner(context):
def __init__(self, lexicon, flags=FLAGS): parse_object = context.parse_object
self.actions = [None] parse_array = context.parse_array
# Combine phrases into a compound pattern parse_string = context.parse_string
s = sre_parse.Pattern() match_number = NUMBER_RE.match
s.flags = flags strict = context.strict
p = [] parse_float = context.parse_float
for idx, token in enumerate(lexicon): parse_int = context.parse_int
phrase = token.pattern parse_constant = context.parse_constant
try: object_hook = context.object_hook
subpattern = sre_parse.SubPattern(s,
[(SUBPATTERN, (idx + 1, sre_parse.parse(phrase, flags)))])
except sre_constants.error:
raise
p.append(subpattern)
self.actions.append(token)
s.groups = len(p) + 1 # NOTE(guido): Added to make SRE validation work def _scan_once(string, idx):
p = sre_parse.SubPattern(s, [(BRANCH, (None, p))]) try:
self.scanner = sre_compile.compile(p) nextchar = string[idx]
except IndexError:
raise StopIteration
def iterscan(self, string, idx=0, context=None): if nextchar == '"':
"""Yield match, end_idx for each match return parse_string(string, idx + 1, strict)
elif nextchar == '{':
return parse_object((string, idx + 1), strict,
_scan_once, object_hook, object_pairs_hook)
elif nextchar == '[':
return parse_array((string, idx + 1), _scan_once)
elif nextchar == 'n' and string[idx:idx + 4] == 'null':
return None, idx + 4
elif nextchar == 't' and string[idx:idx + 4] == 'true':
return True, idx + 4
elif nextchar == 'f' and string[idx:idx + 5] == 'false':
return False, idx + 5
""" m = match_number(string, idx)
match = self.scanner.scanner(string, idx).match if m is not None:
actions = self.actions integer, frac, exp = m.groups()
lastend = idx if frac or exp:
end = len(string) res = parse_float(integer + (frac or '') + (exp or ''))
while True: else:
m = match() res = parse_int(integer)
if m is None: return res, m.end()
break elif nextchar == 'N' and string[idx:idx + 3] == 'NaN':
matchbegin, matchend = m.span() return parse_constant('NaN'), idx + 3
if lastend == matchend: elif nextchar == 'I' and string[idx:idx + 8] == 'Infinity':
break return parse_constant('Infinity'), idx + 8
action = actions[m.lastindex] elif nextchar == '-' and string[idx:idx + 9] == '-Infinity':
if action is not None: return parse_constant('-Infinity'), idx + 9
rval, next_pos = action(m, context) else:
if next_pos is not None and next_pos != matchend: raise StopIteration
# "fast forward" the scanner
matchend = next_pos
match = self.scanner.scanner(string, matchend).match
yield rval, matchend
lastend = matchend
return _scan_once
def pattern(pattern, flags=FLAGS): make_scanner = c_make_scanner or py_make_scanner
def decorator(fn):
fn.pattern = pattern
fn.regex = re.compile(pattern, flags)
return fn
return decorator
...@@ -32,3 +32,10 @@ class TestDecode(TestCase): ...@@ -32,3 +32,10 @@ class TestDecode(TestCase):
object_pairs_hook = OrderedDict, object_pairs_hook = OrderedDict,
object_hook = lambda x: None), object_hook = lambda x: None),
OrderedDict(p)) OrderedDict(p))
def test_decoder_optimizations(self):
# Several optimizations were made that skip over calls to
# the whitespace regex, so this test is designed to try and
# exercise the uncommon cases. The array cases are already covered.
rval = json.loads('{ "key" : "value" , "k":"v" }')
self.assertEquals(rval, {"key":"value", "k":"v"})
...@@ -11,3 +11,11 @@ class TestDump(TestCase): ...@@ -11,3 +11,11 @@ class TestDump(TestCase):
def test_dumps(self): def test_dumps(self):
self.assertEquals(json.dumps({}), '{}') self.assertEquals(json.dumps({}), '{}')
def test_encode_truefalse(self):
self.assertEquals(json.dumps(
{True: False, False: True}, sort_keys=True),
'{"false": true, "true": false}')
self.assertEquals(json.dumps(
{2: 3.0, 4.0: 5, False: 1, 6: True}, sort_keys=True),
'{"false": 1, "2": 3.0, "4.0": 5, "6": true}')
...@@ -3,22 +3,20 @@ from unittest import TestCase ...@@ -3,22 +3,20 @@ from unittest import TestCase
import json.encoder import json.encoder
CASES = [ CASES = [
('/\\"\ucafe\ubabe\uab98\ufcde\ubcda\uef4a\x08\x0c\n\r\t`1~!@#$%^&*()_+-=[]{}|;:\',./<>?', b'"/\\\\\\"\\ucafe\\ubabe\\uab98\\ufcde\\ubcda\\uef4a\\b\\f\\n\\r\\t`1~!@#$%^&*()_+-=[]{}|;:\',./<>?"'), ('/\\"\ucafe\ubabe\uab98\ufcde\ubcda\uef4a\x08\x0c\n\r\t`1~!@#$%^&*()_+-=[]{}|;:\',./<>?', '"/\\\\\\"\\ucafe\\ubabe\\uab98\\ufcde\\ubcda\\uef4a\\b\\f\\n\\r\\t`1~!@#$%^&*()_+-=[]{}|;:\',./<>?"'),
('\u0123\u4567\u89ab\ucdef\uabcd\uef4a', b'"\\u0123\\u4567\\u89ab\\ucdef\\uabcd\\uef4a"'), ('\u0123\u4567\u89ab\ucdef\uabcd\uef4a', '"\\u0123\\u4567\\u89ab\\ucdef\\uabcd\\uef4a"'),
('controls', b'"controls"'), ('controls', '"controls"'),
('\x08\x0c\n\r\t', b'"\\b\\f\\n\\r\\t"'), ('\x08\x0c\n\r\t', '"\\b\\f\\n\\r\\t"'),
('{"object with 1 member":["array with 1 element"]}', b'"{\\"object with 1 member\\":[\\"array with 1 element\\"]}"'), ('{"object with 1 member":["array with 1 element"]}', '"{\\"object with 1 member\\":[\\"array with 1 element\\"]}"'),
(' s p a c e d ', b'" s p a c e d "'), (' s p a c e d ', '" s p a c e d "'),
('\U0001d120', b'"\\ud834\\udd20"'), ('\U0001d120', '"\\ud834\\udd20"'),
('\u03b1\u03a9', b'"\\u03b1\\u03a9"'), ('\u03b1\u03a9', '"\\u03b1\\u03a9"'),
(b'\xce\xb1\xce\xa9', b'"\\u03b1\\u03a9"'), ('\u03b1\u03a9', '"\\u03b1\\u03a9"'),
('\u03b1\u03a9', b'"\\u03b1\\u03a9"'), ('\u03b1\u03a9', '"\\u03b1\\u03a9"'),
(b'\xce\xb1\xce\xa9', b'"\\u03b1\\u03a9"'), ('\u03b1\u03a9', '"\\u03b1\\u03a9"'),
('\u03b1\u03a9', b'"\\u03b1\\u03a9"'), ("`1~!@#$%^&*()_+-={':[,]}|;.</>?", '"`1~!@#$%^&*()_+-={\':[,]}|;.</>?"'),
('\u03b1\u03a9', b'"\\u03b1\\u03a9"'), ('\x08\x0c\n\r\t', '"\\b\\f\\n\\r\\t"'),
("`1~!@#$%^&*()_+-={':[,]}|;.</>?", b'"`1~!@#$%^&*()_+-={\':[,]}|;.</>?"'), ('\u0123\u4567\u89ab\ucdef\uabcd\uef4a', '"\\u0123\\u4567\\u89ab\\ucdef\\uabcd\\uef4a"'),
('\x08\x0c\n\r\t', b'"\\b\\f\\n\\r\\t"'),
('\u0123\u4567\u89ab\ucdef\uabcd\uef4a', b'"\\u0123\\u4567\\u89ab\\ucdef\\uabcd\\uef4a"'),
] ]
class TestEncodeBaseStringAscii(TestCase): class TestEncodeBaseStringAscii(TestCase):
...@@ -26,12 +24,14 @@ class TestEncodeBaseStringAscii(TestCase): ...@@ -26,12 +24,14 @@ class TestEncodeBaseStringAscii(TestCase):
self._test_encode_basestring_ascii(json.encoder.py_encode_basestring_ascii) self._test_encode_basestring_ascii(json.encoder.py_encode_basestring_ascii)
def test_c_encode_basestring_ascii(self): def test_c_encode_basestring_ascii(self):
if json.encoder.c_encode_basestring_ascii is not None: if not json.encoder.c_encode_basestring_ascii:
self._test_encode_basestring_ascii(json.encoder.c_encode_basestring_ascii) return
self._test_encode_basestring_ascii(json.encoder.c_encode_basestring_ascii)
def _test_encode_basestring_ascii(self, encode_basestring_ascii): def _test_encode_basestring_ascii(self, encode_basestring_ascii):
fname = encode_basestring_ascii.__name__ fname = encode_basestring_ascii.__name__
for input_string, expect in CASES: for input_string, expect in CASES:
result = encode_basestring_ascii(input_string) result = encode_basestring_ascii(input_string)
result = result.encode("ascii") self.assertEquals(result, expect,
self.assertEquals(result, expect) '{0!r} != {1!r} for {2}({3!r})'.format(
result, expect, fname, input_string))
...@@ -73,4 +73,4 @@ class TestFail(TestCase): ...@@ -73,4 +73,4 @@ class TestFail(TestCase):
except ValueError: except ValueError:
pass pass
else: else:
self.fail("Expected failure for fail%d.json: %r" % (idx, doc)) self.fail("Expected failure for fail{0}.json: {1!r}".format(idx, doc))
...@@ -5,5 +5,11 @@ import json ...@@ -5,5 +5,11 @@ import json
class TestFloat(TestCase): class TestFloat(TestCase):
def test_floats(self): def test_floats(self):
for num in [1617161771.7650001, math.pi, math.pi**100, math.pi**-100]: for num in [1617161771.7650001, math.pi, math.pi**100, math.pi**-100, 3.1]:
self.assertEquals(float(json.dumps(num)), num) self.assertEquals(float(json.dumps(num)), num)
self.assertEquals(json.loads(json.dumps(num)), num)
def test_ints(self):
for num in [1, 1<<32, 1<<64]:
self.assertEquals(json.dumps(num), str(num))
self.assertEquals(int(json.dumps(num)), num)
...@@ -15,96 +15,90 @@ class TestScanString(TestCase): ...@@ -15,96 +15,90 @@ class TestScanString(TestCase):
def _test_scanstring(self, scanstring): def _test_scanstring(self, scanstring):
self.assertEquals( self.assertEquals(
scanstring('"z\\ud834\\udd20x"', 1, None, True), scanstring('"z\\ud834\\udd20x"', 1, True),
('z\U0001d120x', 16)) ('z\U0001d120x', 16))
if sys.maxunicode == 65535: if sys.maxunicode == 65535:
self.assertEquals( self.assertEquals(
scanstring('"z\U0001d120x"', 1, None, True), scanstring('"z\U0001d120x"', 1, True),
('z\U0001d120x', 6)) ('z\U0001d120x', 6))
else: else:
self.assertEquals( self.assertEquals(
scanstring('"z\U0001d120x"', 1, None, True), scanstring('"z\U0001d120x"', 1, True),
('z\U0001d120x', 5)) ('z\U0001d120x', 5))
self.assertEquals( self.assertEquals(
scanstring('"\\u007b"', 1, None, True), scanstring('"\\u007b"', 1, True),
('{', 8)) ('{', 8))
self.assertEquals( self.assertEquals(
scanstring('"A JSON payload should be an object or array, not a string."', 1, None, True), scanstring('"A JSON payload should be an object or array, not a string."', 1, True),
('A JSON payload should be an object or array, not a string.', 60)) ('A JSON payload should be an object or array, not a string.', 60))
self.assertEquals( self.assertEquals(
scanstring('["Unclosed array"', 2, None, True), scanstring('["Unclosed array"', 2, True),
('Unclosed array', 17)) ('Unclosed array', 17))
self.assertEquals( self.assertEquals(
scanstring('["extra comma",]', 2, None, True), scanstring('["extra comma",]', 2, True),
('extra comma', 14)) ('extra comma', 14))
self.assertEquals( self.assertEquals(
scanstring('["double extra comma",,]', 2, None, True), scanstring('["double extra comma",,]', 2, True),
('double extra comma', 21)) ('double extra comma', 21))
self.assertEquals( self.assertEquals(
scanstring('["Comma after the close"],', 2, None, True), scanstring('["Comma after the close"],', 2, True),
('Comma after the close', 24)) ('Comma after the close', 24))
self.assertEquals( self.assertEquals(
scanstring('["Extra close"]]', 2, None, True), scanstring('["Extra close"]]', 2, True),
('Extra close', 14)) ('Extra close', 14))
self.assertEquals( self.assertEquals(
scanstring('{"Extra comma": true,}', 2, None, True), scanstring('{"Extra comma": true,}', 2, True),
('Extra comma', 14)) ('Extra comma', 14))
self.assertEquals( self.assertEquals(
scanstring('{"Extra value after close": true} "misplaced quoted value"', 2, None, True), scanstring('{"Extra value after close": true} "misplaced quoted value"', 2, True),
('Extra value after close', 26)) ('Extra value after close', 26))
self.assertEquals( self.assertEquals(
scanstring('{"Illegal expression": 1 + 2}', 2, None, True), scanstring('{"Illegal expression": 1 + 2}', 2, True),
('Illegal expression', 21)) ('Illegal expression', 21))
self.assertEquals( self.assertEquals(
scanstring('{"Illegal invocation": alert()}', 2, None, True), scanstring('{"Illegal invocation": alert()}', 2, True),
('Illegal invocation', 21)) ('Illegal invocation', 21))
self.assertEquals( self.assertEquals(
scanstring('{"Numbers cannot have leading zeroes": 013}', 2, None, True), scanstring('{"Numbers cannot have leading zeroes": 013}', 2, True),
('Numbers cannot have leading zeroes', 37)) ('Numbers cannot have leading zeroes', 37))
self.assertEquals( self.assertEquals(
scanstring('{"Numbers cannot be hex": 0x14}', 2, None, True), scanstring('{"Numbers cannot be hex": 0x14}', 2, True),
('Numbers cannot be hex', 24)) ('Numbers cannot be hex', 24))
self.assertEquals( self.assertEquals(
scanstring('[[[[[[[[[[[[[[[[[[[["Too deep"]]]]]]]]]]]]]]]]]]]]', 21, None, True), scanstring('[[[[[[[[[[[[[[[[[[[["Too deep"]]]]]]]]]]]]]]]]]]]]', 21, True),
('Too deep', 30)) ('Too deep', 30))
self.assertEquals( self.assertEquals(
scanstring('{"Missing colon" null}', 2, None, True), scanstring('{"Missing colon" null}', 2, True),
('Missing colon', 16)) ('Missing colon', 16))
self.assertEquals( self.assertEquals(
scanstring('{"Double colon":: null}', 2, None, True), scanstring('{"Double colon":: null}', 2, True),
('Double colon', 15)) ('Double colon', 15))
self.assertEquals( self.assertEquals(
scanstring('{"Comma instead of colon", null}', 2, None, True), scanstring('{"Comma instead of colon", null}', 2, True),
('Comma instead of colon', 25)) ('Comma instead of colon', 25))
self.assertEquals( self.assertEquals(
scanstring('["Colon instead of comma": false]', 2, None, True), scanstring('["Colon instead of comma": false]', 2, True),
('Colon instead of comma', 25)) ('Colon instead of comma', 25))
self.assertEquals( self.assertEquals(
scanstring('["Bad value", truth]', 2, None, True), scanstring('["Bad value", truth]', 2, True),
('Bad value', 12)) ('Bad value', 12))
def test_issue3623(self):
self.assertRaises(ValueError, json.decoder.scanstring, b"xxx", 1,
"xxx")
self.assertRaises(UnicodeDecodeError,
json.encoder.encode_basestring_ascii, b"xx\xff")
...@@ -4,20 +4,8 @@ import json ...@@ -4,20 +4,8 @@ import json
from collections import OrderedDict from collections import OrderedDict
class TestUnicode(TestCase): class TestUnicode(TestCase):
def test_encoding1(self): # test_encoding1 and test_encoding2 from 2.x are irrelevant (only str
encoder = json.JSONEncoder(encoding='utf-8') # is supported as input, not bytes).
u = '\N{GREEK SMALL LETTER ALPHA}\N{GREEK CAPITAL LETTER OMEGA}'
s = u.encode('utf-8')
ju = encoder.encode(u)
js = encoder.encode(s)
self.assertEquals(ju, js)
def test_encoding2(self):
u = '\N{GREEK SMALL LETTER ALPHA}\N{GREEK CAPITAL LETTER OMEGA}'
s = u.encode('utf-8')
ju = json.dumps(u, encoding='utf-8')
js = json.dumps(s, encoding='utf-8')
self.assertEquals(ju, js)
def test_encoding3(self): def test_encoding3(self):
u = '\N{GREEK SMALL LETTER ALPHA}\N{GREEK CAPITAL LETTER OMEGA}' u = '\N{GREEK SMALL LETTER ALPHA}\N{GREEK CAPITAL LETTER OMEGA}'
...@@ -52,8 +40,22 @@ class TestUnicode(TestCase): ...@@ -52,8 +40,22 @@ class TestUnicode(TestCase):
def test_unicode_decode(self): def test_unicode_decode(self):
for i in range(0, 0xd7ff): for i in range(0, 0xd7ff):
u = chr(i) u = chr(i)
js = '"\\u{0:04x}"'.format(i) s = '"\\u{0:04x}"'.format(i)
self.assertEquals(json.loads(js), u) self.assertEquals(json.loads(s), u)
def test_unicode_preservation(self):
self.assertEquals(type(json.loads('""')), str)
self.assertEquals(type(json.loads('"a"')), str)
self.assertEquals(type(json.loads('["a"]')[0]), str)
def test_bytes_encode(self):
self.assertRaises(TypeError, json.dumps, b"hi")
self.assertRaises(TypeError, json.dumps, [b"hi"])
def test_bytes_decode(self):
self.assertRaises(TypeError, json.loads, b'"hi"')
self.assertRaises(TypeError, json.loads, b'["hi"]')
def test_object_pairs_hook_with_unicode(self): def test_object_pairs_hook_with_unicode(self):
s = '{"xkd":1, "kcw":2, "art":3, "hxm":4, "qrt":5, "pad":6, "hoy":7}' s = '{"xkd":1, "kcw":2, "art":3, "hxm":4, "qrt":5, "pad":6, "hoy":7}'
......
...@@ -2,11 +2,11 @@ r"""Command-line tool to validate and pretty-print JSON ...@@ -2,11 +2,11 @@ r"""Command-line tool to validate and pretty-print JSON
Usage:: Usage::
$ echo '{"json":"obj"}' | python -mjson.tool $ echo '{"json":"obj"}' | python -m json.tool
{ {
"json": "obj" "json": "obj"
} }
$ echo '{ 1.2:3.4}' | python -mjson.tool $ echo '{ 1.2:3.4}' | python -m json.tool
Expecting property name: line 1 column 2 (char 2) Expecting property name: line 1 column 2 (char 2)
""" """
...@@ -24,7 +24,7 @@ def main(): ...@@ -24,7 +24,7 @@ def main():
infile = open(sys.argv[1], 'rb') infile = open(sys.argv[1], 'rb')
outfile = open(sys.argv[2], 'wb') outfile = open(sys.argv[2], 'wb')
else: else:
raise SystemExit("{0} [infile [outfile]]".format(sys.argv[0])) raise SystemExit(sys.argv[0] + " [infile [outfile]]")
try: try:
obj = json.load(infile) obj = json.load(infile)
except ValueError as e: except ValueError as e:
......
...@@ -107,6 +107,8 @@ Installation ...@@ -107,6 +107,8 @@ Installation
Library Library
------- -------
- The json module now works exclusively with str and not bytes.
- Issue #3959: The ipaddr module has been added to the standard library. - Issue #3959: The ipaddr module has been added to the standard library.
Contributed by Google. Contributed by Google.
......
#include "Python.h" #include "Python.h"
#include "structmember.h"
#if PY_VERSION_HEX < 0x02060000 && !defined(Py_TYPE)
#define Py_TYPE(ob) (((PyObject*)(ob))->ob_type)
#endif
#if PY_VERSION_HEX < 0x02050000 && !defined(PY_SSIZE_T_MIN)
typedef int Py_ssize_t;
#define PY_SSIZE_T_MAX INT_MAX
#define PY_SSIZE_T_MIN INT_MIN
#define PyInt_FromSsize_t PyInt_FromLong
#define PyInt_AsSsize_t PyInt_AsLong
#endif
#ifndef Py_IS_FINITE
#define Py_IS_FINITE(X) (!Py_IS_INFINITY(X) && !Py_IS_NAN(X))
#endif
#ifdef __GNUC__
#define UNUSED __attribute__((__unused__))
#else
#define UNUSED
#endif
#define PyScanner_Check(op) PyObject_TypeCheck(op, &PyScannerType)
#define PyScanner_CheckExact(op) (Py_TYPE(op) == &PyScannerType)
#define PyEncoder_Check(op) PyObject_TypeCheck(op, &PyEncoderType)
#define PyEncoder_CheckExact(op) (Py_TYPE(op) == &PyEncoderType)
static PyTypeObject PyScannerType;
static PyTypeObject PyEncoderType;
typedef struct _PyScannerObject {
PyObject_HEAD
PyObject *strict;
PyObject *object_hook;
PyObject *object_pairs_hook;
PyObject *parse_float;
PyObject *parse_int;
PyObject *parse_constant;
} PyScannerObject;
static PyMemberDef scanner_members[] = {
{"strict", T_OBJECT, offsetof(PyScannerObject, strict), READONLY, "strict"},
{"object_hook", T_OBJECT, offsetof(PyScannerObject, object_hook), READONLY, "object_hook"},
{"object_pairs_hook", T_OBJECT, offsetof(PyScannerObject, object_pairs_hook), READONLY},
{"parse_float", T_OBJECT, offsetof(PyScannerObject, parse_float), READONLY, "parse_float"},
{"parse_int", T_OBJECT, offsetof(PyScannerObject, parse_int), READONLY, "parse_int"},
{"parse_constant", T_OBJECT, offsetof(PyScannerObject, parse_constant), READONLY, "parse_constant"},
{NULL}
};
typedef struct _PyEncoderObject {
PyObject_HEAD
PyObject *markers;
PyObject *defaultfn;
PyObject *encoder;
PyObject *indent;
PyObject *key_separator;
PyObject *item_separator;
PyObject *sort_keys;
PyObject *skipkeys;
int fast_encode;
int allow_nan;
} PyEncoderObject;
static PyMemberDef encoder_members[] = {
{"markers", T_OBJECT, offsetof(PyEncoderObject, markers), READONLY, "markers"},
{"default", T_OBJECT, offsetof(PyEncoderObject, defaultfn), READONLY, "default"},
{"encoder", T_OBJECT, offsetof(PyEncoderObject, encoder), READONLY, "encoder"},
{"indent", T_OBJECT, offsetof(PyEncoderObject, indent), READONLY, "indent"},
{"key_separator", T_OBJECT, offsetof(PyEncoderObject, key_separator), READONLY, "key_separator"},
{"item_separator", T_OBJECT, offsetof(PyEncoderObject, item_separator), READONLY, "item_separator"},
{"sort_keys", T_OBJECT, offsetof(PyEncoderObject, sort_keys), READONLY, "sort_keys"},
{"skipkeys", T_OBJECT, offsetof(PyEncoderObject, skipkeys), READONLY, "skipkeys"},
{NULL}
};
static PyObject *
ascii_escape_unicode(PyObject *pystr);
static PyObject *
py_encode_basestring_ascii(PyObject* self UNUSED, PyObject *pystr);
void init_json(void);
static PyObject *
scan_once_unicode(PyScannerObject *s, PyObject *pystr, Py_ssize_t idx, Py_ssize_t *next_idx_ptr);
static PyObject *
_build_rval_index_tuple(PyObject *rval, Py_ssize_t idx);
static PyObject *
scanner_new(PyTypeObject *type, PyObject *args, PyObject *kwds);
static int
scanner_init(PyObject *self, PyObject *args, PyObject *kwds);
static void
scanner_dealloc(PyObject *self);
static int
scanner_clear(PyObject *self);
static PyObject *
encoder_new(PyTypeObject *type, PyObject *args, PyObject *kwds);
static int
encoder_init(PyObject *self, PyObject *args, PyObject *kwds);
static void
encoder_dealloc(PyObject *self);
static int
encoder_clear(PyObject *self);
static int
encoder_listencode_list(PyEncoderObject *s, PyObject *rval, PyObject *seq, Py_ssize_t indent_level);
static int
encoder_listencode_obj(PyEncoderObject *s, PyObject *rval, PyObject *obj, Py_ssize_t indent_level);
static int
encoder_listencode_dict(PyEncoderObject *s, PyObject *rval, PyObject *dct, Py_ssize_t indent_level);
static PyObject *
_encoded_const(PyObject *const);
static void
raise_errmsg(char *msg, PyObject *s, Py_ssize_t end);
static PyObject *
encoder_encode_string(PyEncoderObject *s, PyObject *obj);
static int
_convertPyInt_AsSsize_t(PyObject *o, Py_ssize_t *size_ptr);
static PyObject *
_convertPyInt_FromSsize_t(Py_ssize_t *size_ptr);
static PyObject *
encoder_encode_float(PyEncoderObject *s, PyObject *obj);
#define DEFAULT_ENCODING "utf-8"
#define S_CHAR(c) (c >= ' ' && c <= '~' && c != '\\' && c != '"') #define S_CHAR(c) (c >= ' ' && c <= '~' && c != '\\' && c != '"')
#define MIN_EXPANSION 6 #define IS_WHITESPACE(c) (((c) == ' ') || ((c) == '\t') || ((c) == '\n') || ((c) == '\r'))
#define MIN_EXPANSION 6
#ifdef Py_UNICODE_WIDE #ifdef Py_UNICODE_WIDE
#define MAX_EXPANSION (2 * MIN_EXPANSION) #define MAX_EXPANSION (2 * MIN_EXPANSION)
#else #else
#define MAX_EXPANSION MIN_EXPANSION #define MAX_EXPANSION MIN_EXPANSION
#endif #endif
static int
_convertPyInt_AsSsize_t(PyObject *o, Py_ssize_t *size_ptr)
{
/* PyObject to Py_ssize_t converter */
*size_ptr = PyLong_AsSsize_t(o);
if (*size_ptr == -1 && PyErr_Occurred());
return 1;
return 0;
}
static PyObject *
_convertPyInt_FromSsize_t(Py_ssize_t *size_ptr)
{
/* Py_ssize_t to PyObject converter */
return PyLong_FromSsize_t(*size_ptr);
}
static Py_ssize_t static Py_ssize_t
ascii_escape_char(Py_UNICODE c, char *output, Py_ssize_t chars) ascii_escape_unichar(Py_UNICODE c, Py_UNICODE *output, Py_ssize_t chars)
{ {
Py_UNICODE x; /* Escape unicode code point c to ASCII escape sequences
in char *output. output must have at least 12 bytes unused to
accommodate an escaped surrogate pair "\uXXXX\uXXXX" */
output[chars++] = '\\'; output[chars++] = '\\';
switch (c) { switch (c) {
case '\\': output[chars++] = (char)c; break; case '\\': output[chars++] = c; break;
case '"': output[chars++] = (char)c; break; case '"': output[chars++] = c; break;
case '\b': output[chars++] = 'b'; break; case '\b': output[chars++] = 'b'; break;
case '\f': output[chars++] = 'f'; break; case '\f': output[chars++] = 'f'; break;
case '\n': output[chars++] = 'n'; break; case '\n': output[chars++] = 'n'; break;
...@@ -30,27 +167,19 @@ ascii_escape_char(Py_UNICODE c, char *output, Py_ssize_t chars) ...@@ -30,27 +167,19 @@ ascii_escape_char(Py_UNICODE c, char *output, Py_ssize_t chars)
Py_UNICODE v = c - 0x10000; Py_UNICODE v = c - 0x10000;
c = 0xd800 | ((v >> 10) & 0x3ff); c = 0xd800 | ((v >> 10) & 0x3ff);
output[chars++] = 'u'; output[chars++] = 'u';
x = (c & 0xf000) >> 12; output[chars++] = "0123456789abcdef"[(c >> 12) & 0xf];
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10); output[chars++] = "0123456789abcdef"[(c >> 8) & 0xf];
x = (c & 0x0f00) >> 8; output[chars++] = "0123456789abcdef"[(c >> 4) & 0xf];
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10); output[chars++] = "0123456789abcdef"[(c ) & 0xf];
x = (c & 0x00f0) >> 4;
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10);
x = (c & 0x000f);
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10);
c = 0xdc00 | (v & 0x3ff); c = 0xdc00 | (v & 0x3ff);
output[chars++] = '\\'; output[chars++] = '\\';
} }
#endif #endif
output[chars++] = 'u'; output[chars++] = 'u';
x = (c & 0xf000) >> 12; output[chars++] = "0123456789abcdef"[(c >> 12) & 0xf];
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10); output[chars++] = "0123456789abcdef"[(c >> 8) & 0xf];
x = (c & 0x0f00) >> 8; output[chars++] = "0123456789abcdef"[(c >> 4) & 0xf];
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10); output[chars++] = "0123456789abcdef"[(c ) & 0xf];
x = (c & 0x00f0) >> 4;
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10);
x = (c & 0x000f);
output[chars++] = (x < 10) ? '0' + x : 'a' + (x - 10);
} }
return chars; return chars;
} }
...@@ -58,118 +187,66 @@ ascii_escape_char(Py_UNICODE c, char *output, Py_ssize_t chars) ...@@ -58,118 +187,66 @@ ascii_escape_char(Py_UNICODE c, char *output, Py_ssize_t chars)
static PyObject * static PyObject *
ascii_escape_unicode(PyObject *pystr) ascii_escape_unicode(PyObject *pystr)
{ {
/* Take a PyUnicode pystr and return a new ASCII-only escaped PyUnicode */
Py_ssize_t i; Py_ssize_t i;
Py_ssize_t input_chars; Py_ssize_t input_chars;
Py_ssize_t output_size; Py_ssize_t output_size;
Py_ssize_t max_output_size;
Py_ssize_t chars; Py_ssize_t chars;
PyObject *rval; PyObject *rval;
char *output; Py_UNICODE *output;
Py_UNICODE *input_unicode; Py_UNICODE *input_unicode;
input_chars = PyUnicode_GET_SIZE(pystr); input_chars = PyUnicode_GET_SIZE(pystr);
input_unicode = PyUnicode_AS_UNICODE(pystr); input_unicode = PyUnicode_AS_UNICODE(pystr);
/* One char input can be up to 6 chars output, estimate 4 of these */ /* One char input can be up to 6 chars output, estimate 4 of these */
output_size = 2 + (MIN_EXPANSION * 4) + input_chars; output_size = 2 + (MIN_EXPANSION * 4) + input_chars;
rval = PyBytes_FromStringAndSize(NULL, output_size); max_output_size = 2 + (input_chars * MAX_EXPANSION);
rval = PyUnicode_FromStringAndSize(NULL, output_size);
if (rval == NULL) { if (rval == NULL) {
return NULL; return NULL;
} }
output = PyBytes_AS_STRING(rval); output = PyUnicode_AS_UNICODE(rval);
chars = 0; chars = 0;
output[chars++] = '"'; output[chars++] = '"';
for (i = 0; i < input_chars; i++) { for (i = 0; i < input_chars; i++) {
Py_UNICODE c = input_unicode[i]; Py_UNICODE c = input_unicode[i];
if (S_CHAR(c)) { if (S_CHAR(c)) {
output[chars++] = (char)c; output[chars++] = c;
} }
else { else {
chars = ascii_escape_char(c, output, chars); chars = ascii_escape_unichar(c, output, chars);
} }
if (output_size - chars < (1 + MAX_EXPANSION)) { if (output_size - chars < (1 + MAX_EXPANSION)) {
/* There's more than four, so let's resize by a lot */ /* There's more than four, so let's resize by a lot */
output_size *= 2; Py_ssize_t new_output_size = output_size * 2;
/* This is an upper bound */ /* This is an upper bound */
if (output_size > 2 + (input_chars * MAX_EXPANSION)) { if (new_output_size > max_output_size) {
output_size = 2 + (input_chars * MAX_EXPANSION); new_output_size = max_output_size;
}
if (_PyBytes_Resize(&rval, output_size) == -1) {
return NULL;
}
output = PyBytes_AS_STRING(rval);
}
}
output[chars++] = '"';
if (_PyBytes_Resize(&rval, chars) == -1) {
return NULL;
}
return rval;
}
static PyObject *
ascii_escape_str(PyObject *pystr)
{
Py_ssize_t i;
Py_ssize_t input_chars;
Py_ssize_t output_size;
Py_ssize_t chars;
PyObject *rval;
char *output;
char *input_str;
input_chars = PyBytes_GET_SIZE(pystr);
input_str = PyBytes_AS_STRING(pystr);
/* One char input can be up to 6 chars output, estimate 4 of these */
output_size = 2 + (MIN_EXPANSION * 4) + input_chars;
rval = PyBytes_FromStringAndSize(NULL, output_size);
if (rval == NULL) {
return NULL;
}
output = PyBytes_AS_STRING(rval);
chars = 0;
output[chars++] = '"';
for (i = 0; i < input_chars; i++) {
Py_UNICODE c = (Py_UNICODE)input_str[i];
if (S_CHAR(c)) {
output[chars++] = (char)c;
}
else if (c > 0x7F) {
/* We hit a non-ASCII character, bail to unicode mode */
PyObject *uni;
Py_DECREF(rval);
uni = PyUnicode_DecodeUTF8(input_str, input_chars, "strict");
if (uni == NULL) {
return NULL;
}
rval = ascii_escape_unicode(uni);
Py_DECREF(uni);
return rval;
}
else {
chars = ascii_escape_char(c, output, chars);
}
/* An ASCII char can't possibly expand to a surrogate! */
if (output_size - chars < (1 + MIN_EXPANSION)) {
/* There's more than four, so let's resize by a lot */
output_size *= 2;
if (output_size > 2 + (input_chars * MIN_EXPANSION)) {
output_size = 2 + (input_chars * MIN_EXPANSION);
} }
if (_PyBytes_Resize(&rval, output_size) == -1) { /* Make sure that the output size changed before resizing */
return NULL; if (new_output_size != output_size) {
output_size = new_output_size;
if (PyUnicode_Resize(&rval, output_size) == -1) {
return NULL;
}
output = PyUnicode_AS_UNICODE(rval);
} }
output = PyBytes_AS_STRING(rval);
} }
} }
output[chars++] = '"'; output[chars++] = '"';
if (_PyBytes_Resize(&rval, chars) == -1) { if (PyUnicode_Resize(&rval, chars) == -1) {
return NULL; return NULL;
} }
return rval; return rval;
} }
void static void
raise_errmsg(char *msg, PyObject *s, Py_ssize_t end) raise_errmsg(char *msg, PyObject *s, Py_ssize_t end)
{ {
/* Use the Python function json.decoder.errmsg to raise a nice
looking ValueError exception */
static PyObject *errmsg_fn = NULL; static PyObject *errmsg_fn = NULL;
PyObject *pymsg; PyObject *pymsg;
if (errmsg_fn == NULL) { if (errmsg_fn == NULL) {
...@@ -177,228 +254,68 @@ raise_errmsg(char *msg, PyObject *s, Py_ssize_t end) ...@@ -177,228 +254,68 @@ raise_errmsg(char *msg, PyObject *s, Py_ssize_t end)
if (decoder == NULL) if (decoder == NULL)
return; return;
errmsg_fn = PyObject_GetAttrString(decoder, "errmsg"); errmsg_fn = PyObject_GetAttrString(decoder, "errmsg");
Py_DECREF(decoder);
if (errmsg_fn == NULL) if (errmsg_fn == NULL)
return; return;
Py_DECREF(decoder);
} }
pymsg = PyObject_CallFunction(errmsg_fn, "(zOn)", msg, s, end); pymsg = PyObject_CallFunction(errmsg_fn, "(zOO&)", msg, s, _convertPyInt_FromSsize_t, &end);
if (pymsg) { if (pymsg) {
PyErr_SetObject(PyExc_ValueError, pymsg); PyErr_SetObject(PyExc_ValueError, pymsg);
Py_DECREF(pymsg); Py_DECREF(pymsg);
} }
/*
def linecol(doc, pos):
lineno = doc.count('\n', 0, pos) + 1
if lineno == 1:
colno = pos
else:
colno = pos - doc.rindex('\n', 0, pos)
return lineno, colno
def errmsg(msg, doc, pos, end=None):
lineno, colno = linecol(doc, pos)
if end is None:
return '%s: line %d column %d (char %d)' % (msg, lineno, colno, pos)
endlineno, endcolno = linecol(doc, end)
return '%s: line %d column %d - line %d column %d (char %d - %d)' % (
msg, lineno, colno, endlineno, endcolno, pos, end)
*/
} }
static PyObject * static PyObject *
join_list_unicode(PyObject *lst) join_list_unicode(PyObject *lst)
{ {
static PyObject *ustr = NULL; /* return u''.join(lst) */
static PyObject *joinstr = NULL; static PyObject *sep = NULL;
if (ustr == NULL) { if (sep == NULL) {
Py_UNICODE c = 0; sep = PyUnicode_FromStringAndSize("", 0);
ustr = PyUnicode_FromUnicode(&c, 0); if (sep == NULL)
} return NULL;
if (joinstr == NULL) {
joinstr = PyUnicode_InternFromString("join");
}
if (joinstr == NULL || ustr == NULL) {
return NULL;
} }
return PyObject_CallMethodObjArgs(ustr, joinstr, lst, NULL); return PyUnicode_Join(sep, lst);
} }
static PyObject * static PyObject *
scanstring_str(PyObject *pystr, Py_ssize_t end, char *encoding, int strict) _build_rval_index_tuple(PyObject *rval, Py_ssize_t idx) {
{ /* return (rval, idx) tuple, stealing reference to rval */
PyObject *rval; PyObject *tpl;
Py_ssize_t len = PyBytes_GET_SIZE(pystr); PyObject *pyidx;
Py_ssize_t begin = end - 1; /*
Py_ssize_t next = begin; steal a reference to rval, returns (rval, idx)
char *buf = PyBytes_AS_STRING(pystr); */
Py_buffer info; if (rval == NULL) {
PyObject *chunks = PyList_New(0); return NULL;
if (chunks == NULL) {
goto bail;
}
if (end < 0 || len <= end) {
PyErr_SetString(PyExc_ValueError, "end is out of bounds");
goto bail;
} }
while (1) { pyidx = PyLong_FromSsize_t(idx);
/* Find the end of the string or the next escape */ if (pyidx == NULL) {
Py_UNICODE c = 0; Py_DECREF(rval);
PyObject *chunk = NULL; return NULL;
for (next = end; next < len; next++) {
c = buf[next];
if (c == '"' || c == '\\') {
break;
}
else if (strict && c <= 0x1f) {
raise_errmsg("Invalid control character at", pystr, next);
goto bail;
}
}
if (!(c == '"' || c == '\\')) {
raise_errmsg("Unterminated string starting at", pystr, begin);
goto bail;
}
/* Pick up this chunk if it's not zero length */
if (next != end) {
PyObject *strchunk;
if (PyBuffer_FillInfo(&info, NULL, &buf[end], next - end, 1, 0) < 0) {
goto bail;
}
strchunk = PyMemoryView_FromBuffer(&info);
if (strchunk == NULL) {
goto bail;
}
chunk = PyUnicode_FromEncodedObject(strchunk, encoding, NULL);
Py_DECREF(strchunk);
if (chunk == NULL) {
goto bail;
}
if (PyList_Append(chunks, chunk)) {
Py_DECREF(chunk);
goto bail;
}
Py_DECREF(chunk);
}
next++;
if (c == '"') {
end = next;
break;
}
if (next == len) {
raise_errmsg("Unterminated string starting at", pystr, begin);
goto bail;
}
c = buf[next];
if (c != 'u') {
/* Non-unicode backslash escapes */
end = next + 1;
switch (c) {
case '"': break;
case '\\': break;
case '/': break;
case 'b': c = '\b'; break;
case 'f': c = '\f'; break;
case 'n': c = '\n'; break;
case 'r': c = '\r'; break;
case 't': c = '\t'; break;
default: c = 0;
}
if (c == 0) {
raise_errmsg("Invalid \\escape", pystr, end - 2);
goto bail;
}
}
else {
c = 0;
next++;
end = next + 4;
if (end >= len) {
raise_errmsg("Invalid \\uXXXX escape", pystr, next - 1);
goto bail;
}
/* Decode 4 hex digits */
for (; next < end; next++) {
Py_ssize_t shl = (end - next - 1) << 2;
Py_UNICODE digit = buf[next];
switch (digit) {
case '0': case '1': case '2': case '3': case '4':
case '5': case '6': case '7': case '8': case '9':
c |= (digit - '0') << shl; break;
case 'a': case 'b': case 'c': case 'd': case 'e':
case 'f':
c |= (digit - 'a' + 10) << shl; break;
case 'A': case 'B': case 'C': case 'D': case 'E':
case 'F':
c |= (digit - 'A' + 10) << shl; break;
default:
raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5);
goto bail;
}
}
#ifdef Py_UNICODE_WIDE
/* Surrogate pair */
if (c >= 0xd800 && c <= 0xdbff) {
Py_UNICODE c2 = 0;
if (end + 6 >= len) {
raise_errmsg("Invalid \\uXXXX\\uXXXX surrogate pair", pystr,
end - 5);
}
if (buf[next++] != '\\' || buf[next++] != 'u') {
raise_errmsg("Invalid \\uXXXX\\uXXXX surrogate pair", pystr,
end - 5);
}
end += 6;
/* Decode 4 hex digits */
for (; next < end; next++) {
Py_ssize_t shl = (end - next - 1) << 2;
Py_UNICODE digit = buf[next];
switch (digit) {
case '0': case '1': case '2': case '3': case '4':
case '5': case '6': case '7': case '8': case '9':
c2 |= (digit - '0') << shl; break;
case 'a': case 'b': case 'c': case 'd': case 'e':
case 'f':
c2 |= (digit - 'a' + 10) << shl; break;
case 'A': case 'B': case 'C': case 'D': case 'E':
case 'F':
c2 |= (digit - 'A' + 10) << shl; break;
default:
raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5);
goto bail;
}
}
c = 0x10000 + (((c - 0xd800) << 10) | (c2 - 0xdc00));
}
#endif
}
chunk = PyUnicode_FromUnicode(&c, 1);
if (chunk == NULL) {
goto bail;
}
if (PyList_Append(chunks, chunk)) {
Py_DECREF(chunk);
goto bail;
}
Py_DECREF(chunk);
} }
tpl = PyTuple_New(2);
rval = join_list_unicode(chunks); if (tpl == NULL) {
if (rval == NULL) { Py_DECREF(pyidx);
goto bail; Py_DECREF(rval);
return NULL;
} }
Py_CLEAR(chunks); PyTuple_SET_ITEM(tpl, 0, rval);
return Py_BuildValue("(Nn)", rval, end); PyTuple_SET_ITEM(tpl, 1, pyidx);
bail: return tpl;
Py_XDECREF(chunks);
return NULL;
} }
static PyObject * static PyObject *
scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict) scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict, Py_ssize_t *next_end_ptr)
{ {
/* Read the JSON string from PyUnicode pystr.
end is the index of the first character after the quote.
if strict is zero then literal control characters are allowed
*next_end_ptr is a return-by-reference index of the character
after the end quote
Return value is a new PyUnicode
*/
PyObject *rval; PyObject *rval;
Py_ssize_t len = PyUnicode_GET_SIZE(pystr); Py_ssize_t len = PyUnicode_GET_SIZE(pystr);
Py_ssize_t begin = end - 1; Py_ssize_t begin = end - 1;
...@@ -481,18 +398,18 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict) ...@@ -481,18 +398,18 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict)
} }
/* Decode 4 hex digits */ /* Decode 4 hex digits */
for (; next < end; next++) { for (; next < end; next++) {
Py_ssize_t shl = (end - next - 1) << 2;
Py_UNICODE digit = buf[next]; Py_UNICODE digit = buf[next];
c <<= 4;
switch (digit) { switch (digit) {
case '0': case '1': case '2': case '3': case '4': case '0': case '1': case '2': case '3': case '4':
case '5': case '6': case '7': case '8': case '9': case '5': case '6': case '7': case '8': case '9':
c |= (digit - '0') << shl; break; c |= (digit - '0'); break;
case 'a': case 'b': case 'c': case 'd': case 'e': case 'a': case 'b': case 'c': case 'd': case 'e':
case 'f': case 'f':
c |= (digit - 'a' + 10) << shl; break; c |= (digit - 'a' + 10); break;
case 'A': case 'B': case 'C': case 'D': case 'E': case 'A': case 'B': case 'C': case 'D': case 'E':
case 'F': case 'F':
c |= (digit - 'A' + 10) << shl; break; c |= (digit - 'A' + 10); break;
default: default:
raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5); raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5);
goto bail; goto bail;
...@@ -500,38 +417,46 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict) ...@@ -500,38 +417,46 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict)
} }
#ifdef Py_UNICODE_WIDE #ifdef Py_UNICODE_WIDE
/* Surrogate pair */ /* Surrogate pair */
if (c >= 0xd800 && c <= 0xdbff) { if ((c & 0xfc00) == 0xd800) {
Py_UNICODE c2 = 0; Py_UNICODE c2 = 0;
if (end + 6 >= len) { if (end + 6 >= len) {
raise_errmsg("Invalid \\uXXXX\\uXXXX surrogate pair", pystr, raise_errmsg("Unpaired high surrogate", pystr, end - 5);
end - 5); goto bail;
} }
if (buf[next++] != '\\' || buf[next++] != 'u') { if (buf[next++] != '\\' || buf[next++] != 'u') {
raise_errmsg("Invalid \\uXXXX\\uXXXX surrogate pair", pystr, raise_errmsg("Unpaired high surrogate", pystr, end - 5);
end - 5); goto bail;
} }
end += 6; end += 6;
/* Decode 4 hex digits */ /* Decode 4 hex digits */
for (; next < end; next++) { for (; next < end; next++) {
Py_ssize_t shl = (end - next - 1) << 2; c2 <<= 4;
Py_UNICODE digit = buf[next]; Py_UNICODE digit = buf[next];
switch (digit) { switch (digit) {
case '0': case '1': case '2': case '3': case '4': case '0': case '1': case '2': case '3': case '4':
case '5': case '6': case '7': case '8': case '9': case '5': case '6': case '7': case '8': case '9':
c2 |= (digit - '0') << shl; break; c2 |= (digit - '0'); break;
case 'a': case 'b': case 'c': case 'd': case 'e': case 'a': case 'b': case 'c': case 'd': case 'e':
case 'f': case 'f':
c2 |= (digit - 'a' + 10) << shl; break; c2 |= (digit - 'a' + 10); break;
case 'A': case 'B': case 'C': case 'D': case 'E': case 'A': case 'B': case 'C': case 'D': case 'E':
case 'F': case 'F':
c2 |= (digit - 'A' + 10) << shl; break; c2 |= (digit - 'A' + 10); break;
default: default:
raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5); raise_errmsg("Invalid \\uXXXX escape", pystr, end - 5);
goto bail; goto bail;
} }
} }
if ((c2 & 0xfc00) != 0xdc00) {
raise_errmsg("Unpaired high surrogate", pystr, end - 5);
goto bail;
}
c = 0x10000 + (((c - 0xd800) << 10) | (c2 - 0xdc00)); c = 0x10000 + (((c - 0xd800) << 10) | (c2 - 0xdc00));
} }
else if ((c & 0xfc00) == 0xdc00) {
raise_errmsg("Unpaired low surrogate", pystr, end - 5);
goto bail;
}
#endif #endif
} }
chunk = PyUnicode_FromUnicode(&c, 1); chunk = PyUnicode_FromUnicode(&c, 1);
...@@ -549,34 +474,41 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict) ...@@ -549,34 +474,41 @@ scanstring_unicode(PyObject *pystr, Py_ssize_t end, int strict)
if (rval == NULL) { if (rval == NULL) {
goto bail; goto bail;
} }
Py_CLEAR(chunks); Py_DECREF(chunks);
return Py_BuildValue("(Nn)", rval, end); *next_end_ptr = end;
return rval;
bail: bail:
*next_end_ptr = -1;
Py_XDECREF(chunks); Py_XDECREF(chunks);
return NULL; return NULL;
} }
PyDoc_STRVAR(pydoc_scanstring, PyDoc_STRVAR(pydoc_scanstring,
"scanstring(str_or_bytes, end, encoding) -> (bytes, end)\n"); "scanstring(basestring, end, strict=True) -> (bytes, end)\n"
"\n"
"Scan the string s for a JSON string. End is the index of the\n"
"character in s after the quote that started the JSON string.\n"
"Unescapes all valid JSON string escape sequences and raises ValueError\n"
"on attempt to decode an invalid string. If strict is False then literal\n"
"control characters are allowed in the string.\n"
"\n"
"Returns a tuple of the decoded string and the index of the character in s\n"
"after the end quote."
);
static PyObject * static PyObject *
py_scanstring(PyObject* self, PyObject *args) py_scanstring(PyObject* self UNUSED, PyObject *args)
{ {
PyObject *pystr; PyObject *pystr;
PyObject *rval;
Py_ssize_t end; Py_ssize_t end;
char *encoding = NULL; Py_ssize_t next_end = -1;
int strict = 0; int strict = 1;
if (!PyArg_ParseTuple(args, "On|zi:scanstring", &pystr, &end, &encoding, &strict)) { if (!PyArg_ParseTuple(args, "OO&|i:scanstring", &pystr, _convertPyInt_AsSsize_t, &end, &strict)) {
return NULL; return NULL;
} }
if (encoding == NULL) { if (PyUnicode_Check(pystr)) {
encoding = DEFAULT_ENCODING; rval = scanstring_unicode(pystr, end, strict, &next_end);
}
if (PyBytes_Check(pystr)) {
return scanstring_str(pystr, end, encoding, strict);
}
else if (PyUnicode_Check(pystr)) {
return scanstring_unicode(pystr, end, strict);
} }
else { else {
PyErr_Format(PyExc_TypeError, PyErr_Format(PyExc_TypeError,
...@@ -584,61 +516,1176 @@ py_scanstring(PyObject* self, PyObject *args) ...@@ -584,61 +516,1176 @@ py_scanstring(PyObject* self, PyObject *args)
Py_TYPE(pystr)->tp_name); Py_TYPE(pystr)->tp_name);
return NULL; return NULL;
} }
return _build_rval_index_tuple(rval, next_end);
} }
PyDoc_STRVAR(pydoc_encode_basestring_ascii, PyDoc_STRVAR(pydoc_encode_basestring_ascii,
"encode_basestring_ascii(str_or_bytes) -> bytes\n"); "encode_basestring_ascii(basestring) -> bytes\n"
"\n"
"Return an ASCII-only JSON representation of a Python string"
);
static PyObject * static PyObject *
py_encode_basestring_ascii(PyObject* self, PyObject *pystr) py_encode_basestring_ascii(PyObject* self UNUSED, PyObject *pystr)
{ {
PyObject *rval; PyObject *rval;
/* Return an ASCII-only JSON representation of a Python string */
/* METH_O */ /* METH_O */
if (PyBytes_Check(pystr)) { if (PyUnicode_Check(pystr)) {
rval = ascii_escape_str(pystr);
}
else if (PyUnicode_Check(pystr)) {
rval = ascii_escape_unicode(pystr); rval = ascii_escape_unicode(pystr);
} }
else { else {
PyErr_Format(PyExc_TypeError, PyErr_Format(PyExc_TypeError,
"first argument must be a string or unicode, not %.80s", "first argument must be a string, not %.80s",
Py_TYPE(pystr)->tp_name); Py_TYPE(pystr)->tp_name);
return NULL; return NULL;
} }
if (rval != NULL && PyBytes_Check(rval)) {
PyObject *urval = PyUnicode_DecodeASCII(PyBytes_AS_STRING(rval), PyBytes_GET_SIZE(rval), NULL);
Py_DECREF(rval);
return urval;
}
return rval; return rval;
} }
static PyMethodDef json_methods[] = { static void
{"encode_basestring_ascii", (PyCFunction)py_encode_basestring_ascii, scanner_dealloc(PyObject *self)
METH_O, pydoc_encode_basestring_ascii}, {
{"scanstring", (PyCFunction)py_scanstring, METH_VARARGS, /* Deallocate scanner object */
pydoc_scanstring}, scanner_clear(self);
{NULL, NULL, 0, NULL} Py_TYPE(self)->tp_free(self);
}; }
PyDoc_STRVAR(module_doc,
"json speedups\n");
static struct PyModuleDef jsonmodule = { static int
PyModuleDef_HEAD_INIT, scanner_traverse(PyObject *self, visitproc visit, void *arg)
"_json", {
module_doc, PyScannerObject *s;
-1, assert(PyScanner_Check(self));
json_methods, s = (PyScannerObject *)self;
NULL, Py_VISIT(s->strict);
NULL, Py_VISIT(s->object_hook);
NULL, Py_VISIT(s->object_pairs_hook);
NULL Py_VISIT(s->parse_float);
}; Py_VISIT(s->parse_int);
Py_VISIT(s->parse_constant);
return 0;
}
PyObject* static int
PyInit__json(void) scanner_clear(PyObject *self)
{ {
return PyModule_Create(&jsonmodule); PyScannerObject *s;
assert(PyScanner_Check(self));
s = (PyScannerObject *)self;
Py_CLEAR(s->strict);
Py_CLEAR(s->object_hook);
Py_CLEAR(s->object_pairs_hook);
Py_CLEAR(s->parse_float);
Py_CLEAR(s->parse_int);
Py_CLEAR(s->parse_constant);
return 0;
}
static PyObject *
_parse_object_unicode(PyScannerObject *s, PyObject *pystr, Py_ssize_t idx, Py_ssize_t *next_idx_ptr) {
/* Read a JSON object from PyUnicode pystr.
idx is the index of the first character after the opening curly brace.
*next_idx_ptr is a return-by-reference index to the first character after
the closing curly brace.
Returns a new PyObject (usually a dict, but object_hook can change that)
*/
Py_UNICODE *str = PyUnicode_AS_UNICODE(pystr);
Py_ssize_t end_idx = PyUnicode_GET_SIZE(pystr) - 1;
PyObject *val = NULL;
PyObject *rval = PyList_New(0);
PyObject *key = NULL;
int strict = PyObject_IsTrue(s->strict);
Py_ssize_t next_idx;
if (rval == NULL)
return NULL;
/* skip whitespace after { */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
/* only loop if the object is non-empty */
if (idx <= end_idx && str[idx] != '}') {
while (idx <= end_idx) {
/* read key */
if (str[idx] != '"') {
raise_errmsg("Expecting property name", pystr, idx);
goto bail;
}
key = scanstring_unicode(pystr, idx + 1, strict, &next_idx);
if (key == NULL)
goto bail;
idx = next_idx;
/* skip whitespace between key and : delimiter, read :, skip whitespace */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
if (idx > end_idx || str[idx] != ':') {
raise_errmsg("Expecting : delimiter", pystr, idx);
goto bail;
}
idx++;
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
/* read any JSON term */
val = scan_once_unicode(s, pystr, idx, &next_idx);
if (val == NULL)
goto bail;
{
PyObject *tuple = PyTuple_Pack(2, key, val);
if (tuple == NULL)
goto bail;
if (PyList_Append(rval, tuple) == -1) {
Py_DECREF(tuple);
goto bail;
}
Py_DECREF(tuple);
}
Py_CLEAR(key);
Py_CLEAR(val);
idx = next_idx;
/* skip whitespace before } or , */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
/* bail if the object is closed or we didn't get the , delimiter */
if (idx > end_idx) break;
if (str[idx] == '}') {
break;
}
else if (str[idx] != ',') {
raise_errmsg("Expecting , delimiter", pystr, idx);
goto bail;
}
idx++;
/* skip whitespace after , delimiter */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
}
}
/* verify that idx < end_idx, str[idx] should be '}' */
if (idx > end_idx || str[idx] != '}') {
raise_errmsg("Expecting object", pystr, end_idx);
goto bail;
}
*next_idx_ptr = idx + 1;
if (s->object_pairs_hook != Py_None) {
val = PyObject_CallFunctionObjArgs(s->object_pairs_hook, rval, NULL);
if (val == NULL)
goto bail;
Py_DECREF(rval);
return val;
}
val = PyDict_New();
if (val == NULL)
goto bail;
if (PyDict_MergeFromSeq2(val, rval, 1) == -1)
goto bail;
Py_DECREF(rval);
rval = val;
/* if object_hook is not None: rval = object_hook(rval) */
if (s->object_hook != Py_None) {
val = PyObject_CallFunctionObjArgs(s->object_hook, rval, NULL);
if (val == NULL)
goto bail;
Py_DECREF(rval);
rval = val;
val = NULL;
}
return rval;
bail:
Py_XDECREF(key);
Py_XDECREF(val);
Py_DECREF(rval);
return NULL;
}
static PyObject *
_parse_array_unicode(PyScannerObject *s, PyObject *pystr, Py_ssize_t idx, Py_ssize_t *next_idx_ptr) {
/* Read a JSON array from PyString pystr.
idx is the index of the first character after the opening brace.
*next_idx_ptr is a return-by-reference index to the first character after
the closing brace.
Returns a new PyList
*/
Py_UNICODE *str = PyUnicode_AS_UNICODE(pystr);
Py_ssize_t end_idx = PyUnicode_GET_SIZE(pystr) - 1;
PyObject *val = NULL;
PyObject *rval = PyList_New(0);
Py_ssize_t next_idx;
if (rval == NULL)
return NULL;
/* skip whitespace after [ */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
/* only loop if the array is non-empty */
if (idx <= end_idx && str[idx] != ']') {
while (idx <= end_idx) {
/* read any JSON term */
val = scan_once_unicode(s, pystr, idx, &next_idx);
if (val == NULL)
goto bail;
if (PyList_Append(rval, val) == -1)
goto bail;
Py_CLEAR(val);
idx = next_idx;
/* skip whitespace between term and , */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
/* bail if the array is closed or we didn't get the , delimiter */
if (idx > end_idx) break;
if (str[idx] == ']') {
break;
}
else if (str[idx] != ',') {
raise_errmsg("Expecting , delimiter", pystr, idx);
goto bail;
}
idx++;
/* skip whitespace after , */
while (idx <= end_idx && IS_WHITESPACE(str[idx])) idx++;
}
}
/* verify that idx < end_idx, str[idx] should be ']' */
if (idx > end_idx || str[idx] != ']') {
raise_errmsg("Expecting object", pystr, end_idx);
goto bail;
}
*next_idx_ptr = idx + 1;
return rval;
bail:
Py_XDECREF(val);
Py_DECREF(rval);
return NULL;
}
static PyObject *
_parse_constant(PyScannerObject *s, char *constant, Py_ssize_t idx, Py_ssize_t *next_idx_ptr) {
/* Read a JSON constant from PyString pystr.
constant is the constant string that was found
("NaN", "Infinity", "-Infinity").
idx is the index of the first character of the constant
*next_idx_ptr is a return-by-reference index to the first character after
the constant.
Returns the result of parse_constant
*/
PyObject *cstr;
PyObject *rval;
/* constant is "NaN", "Infinity", or "-Infinity" */
cstr = PyUnicode_InternFromString(constant);
if (cstr == NULL)
return NULL;
/* rval = parse_constant(constant) */
rval = PyObject_CallFunctionObjArgs(s->parse_constant, cstr, NULL);
idx += PyUnicode_GET_SIZE(cstr);
Py_DECREF(cstr);
*next_idx_ptr = idx;
return rval;
}
static PyObject *
_match_number_unicode(PyScannerObject *s, PyObject *pystr, Py_ssize_t start, Py_ssize_t *next_idx_ptr) {
/* Read a JSON number from PyUnicode pystr.
idx is the index of the first character of the number
*next_idx_ptr is a return-by-reference index to the first character after
the number.
Returns a new PyObject representation of that number:
PyInt, PyLong, or PyFloat.
May return other types if parse_int or parse_float are set
*/
Py_UNICODE *str = PyUnicode_AS_UNICODE(pystr);
Py_ssize_t end_idx = PyUnicode_GET_SIZE(pystr) - 1;
Py_ssize_t idx = start;
int is_float = 0;
PyObject *rval;
PyObject *numstr;
/* read a sign if it's there, make sure it's not the end of the string */
if (str[idx] == '-') {
idx++;
if (idx > end_idx) {
PyErr_SetNone(PyExc_StopIteration);
return NULL;
}
}
/* read as many integer digits as we find as long as it doesn't start with 0 */
if (str[idx] >= '1' && str[idx] <= '9') {
idx++;
while (idx <= end_idx && str[idx] >= '0' && str[idx] <= '9') idx++;
}
/* if it starts with 0 we only expect one integer digit */
else if (str[idx] == '0') {
idx++;
}
/* no integer digits, error */
else {
PyErr_SetNone(PyExc_StopIteration);
return NULL;
}
/* if the next char is '.' followed by a digit then read all float digits */
if (idx < end_idx && str[idx] == '.' && str[idx + 1] >= '0' && str[idx + 1] <= '9') {
is_float = 1;
idx += 2;
while (idx <= end_idx && str[idx] >= '0' && str[idx] <= '9') idx++;
}
/* if the next char is 'e' or 'E' then maybe read the exponent (or backtrack) */
if (idx < end_idx && (str[idx] == 'e' || str[idx] == 'E')) {
Py_ssize_t e_start = idx;
idx++;
/* read an exponent sign if present */
if (idx < end_idx && (str[idx] == '-' || str[idx] == '+')) idx++;
/* read all digits */
while (idx <= end_idx && str[idx] >= '0' && str[idx] <= '9') idx++;
/* if we got a digit, then parse as float. if not, backtrack */
if (str[idx - 1] >= '0' && str[idx - 1] <= '9') {
is_float = 1;
}
else {
idx = e_start;
}
}
/* copy the section we determined to be a number */
numstr = PyUnicode_FromUnicode(&str[start], idx - start);
if (numstr == NULL)
return NULL;
if (is_float) {
/* parse as a float using a fast path if available, otherwise call user defined method */
if (s->parse_float != (PyObject *)&PyFloat_Type) {
rval = PyObject_CallFunctionObjArgs(s->parse_float, numstr, NULL);
}
else {
rval = PyFloat_FromString(numstr);
}
}
else {
/* no fast path for unicode -> int, just call */
rval = PyObject_CallFunctionObjArgs(s->parse_int, numstr, NULL);
}
Py_DECREF(numstr);
*next_idx_ptr = idx;
return rval;
}
static PyObject *
scan_once_unicode(PyScannerObject *s, PyObject *pystr, Py_ssize_t idx, Py_ssize_t *next_idx_ptr)
{
/* Read one JSON term (of any kind) from PyUnicode pystr.
idx is the index of the first character of the term
*next_idx_ptr is a return-by-reference index to the first character after
the number.
Returns a new PyObject representation of the term.
*/
Py_UNICODE *str = PyUnicode_AS_UNICODE(pystr);
Py_ssize_t length = PyUnicode_GET_SIZE(pystr);
if (idx >= length) {
PyErr_SetNone(PyExc_StopIteration);
return NULL;
}
switch (str[idx]) {
case '"':
/* string */
return scanstring_unicode(pystr, idx + 1,
PyObject_IsTrue(s->strict),
next_idx_ptr);
case '{':
/* object */
return _parse_object_unicode(s, pystr, idx + 1, next_idx_ptr);
case '[':
/* array */
return _parse_array_unicode(s, pystr, idx + 1, next_idx_ptr);
case 'n':
/* null */
if ((idx + 3 < length) && str[idx + 1] == 'u' && str[idx + 2] == 'l' && str[idx + 3] == 'l') {
Py_INCREF(Py_None);
*next_idx_ptr = idx + 4;
return Py_None;
}
break;
case 't':
/* true */
if ((idx + 3 < length) && str[idx + 1] == 'r' && str[idx + 2] == 'u' && str[idx + 3] == 'e') {
Py_INCREF(Py_True);
*next_idx_ptr = idx + 4;
return Py_True;
}
break;
case 'f':
/* false */
if ((idx + 4 < length) && str[idx + 1] == 'a' && str[idx + 2] == 'l' && str[idx + 3] == 's' && str[idx + 4] == 'e') {
Py_INCREF(Py_False);
*next_idx_ptr = idx + 5;
return Py_False;
}
break;
case 'N':
/* NaN */
if ((idx + 2 < length) && str[idx + 1] == 'a' && str[idx + 2] == 'N') {
return _parse_constant(s, "NaN", idx, next_idx_ptr);
}
break;
case 'I':
/* Infinity */
if ((idx + 7 < length) && str[idx + 1] == 'n' && str[idx + 2] == 'f' && str[idx + 3] == 'i' && str[idx + 4] == 'n' && str[idx + 5] == 'i' && str[idx + 6] == 't' && str[idx + 7] == 'y') {
return _parse_constant(s, "Infinity", idx, next_idx_ptr);
}
break;
case '-':
/* -Infinity */
if ((idx + 8 < length) && str[idx + 1] == 'I' && str[idx + 2] == 'n' && str[idx + 3] == 'f' && str[idx + 4] == 'i' && str[idx + 5] == 'n' && str[idx + 6] == 'i' && str[idx + 7] == 't' && str[idx + 8] == 'y') {
return _parse_constant(s, "-Infinity", idx, next_idx_ptr);
}
break;
}
/* Didn't find a string, object, array, or named constant. Look for a number. */
return _match_number_unicode(s, pystr, idx, next_idx_ptr);
}
static PyObject *
scanner_call(PyObject *self, PyObject *args, PyObject *kwds)
{
/* Python callable interface to scan_once_{str,unicode} */
PyObject *pystr;
PyObject *rval;
Py_ssize_t idx;
Py_ssize_t next_idx = -1;
static char *kwlist[] = {"string", "idx", NULL};
PyScannerObject *s;
assert(PyScanner_Check(self));
s = (PyScannerObject *)self;
if (!PyArg_ParseTupleAndKeywords(args, kwds, "OO&:scan_once", kwlist, &pystr, _convertPyInt_AsSsize_t, &idx))
return NULL;
if (PyUnicode_Check(pystr)) {
rval = scan_once_unicode(s, pystr, idx, &next_idx);
}
else {
PyErr_Format(PyExc_TypeError,
"first argument must be a string, not %.80s",
Py_TYPE(pystr)->tp_name);
return NULL;
}
return _build_rval_index_tuple(rval, next_idx);
}
static PyObject *
scanner_new(PyTypeObject *type, PyObject *args, PyObject *kwds)
{
PyScannerObject *s;
s = (PyScannerObject *)type->tp_alloc(type, 0);
if (s != NULL) {
s->strict = NULL;
s->object_hook = NULL;
s->object_pairs_hook = NULL;
s->parse_float = NULL;
s->parse_int = NULL;
s->parse_constant = NULL;
}
return (PyObject *)s;
}
static int
scanner_init(PyObject *self, PyObject *args, PyObject *kwds)
{
/* Initialize Scanner object */
PyObject *ctx;
static char *kwlist[] = {"context", NULL};
PyScannerObject *s;
assert(PyScanner_Check(self));
s = (PyScannerObject *)self;
if (!PyArg_ParseTupleAndKeywords(args, kwds, "O:make_scanner", kwlist, &ctx))
return -1;
/* All of these will fail "gracefully" so we don't need to verify them */
s->strict = PyObject_GetAttrString(ctx, "strict");
if (s->strict == NULL)
goto bail;
s->object_hook = PyObject_GetAttrString(ctx, "object_hook");
if (s->object_hook == NULL)
goto bail;
s->object_pairs_hook = PyObject_GetAttrString(ctx, "object_pairs_hook");
if (s->object_pairs_hook == NULL)
goto bail;
s->parse_float = PyObject_GetAttrString(ctx, "parse_float");
if (s->parse_float == NULL)
goto bail;
s->parse_int = PyObject_GetAttrString(ctx, "parse_int");
if (s->parse_int == NULL)
goto bail;
s->parse_constant = PyObject_GetAttrString(ctx, "parse_constant");
if (s->parse_constant == NULL)
goto bail;
return 0;
bail:
Py_CLEAR(s->strict);
Py_CLEAR(s->object_hook);
Py_CLEAR(s->object_pairs_hook);
Py_CLEAR(s->parse_float);
Py_CLEAR(s->parse_int);
Py_CLEAR(s->parse_constant);
return -1;
}
PyDoc_STRVAR(scanner_doc, "JSON scanner object");
static
PyTypeObject PyScannerType = {
PyVarObject_HEAD_INIT(NULL, 0)
"_json.Scanner", /* tp_name */
sizeof(PyScannerObject), /* tp_basicsize */
0, /* tp_itemsize */
scanner_dealloc, /* tp_dealloc */
0, /* tp_print */
0, /* tp_getattr */
0, /* tp_setattr */
0, /* tp_compare */
0, /* tp_repr */
0, /* tp_as_number */
0, /* tp_as_sequence */
0, /* tp_as_mapping */
0, /* tp_hash */
scanner_call, /* tp_call */
0, /* tp_str */
0,/* PyObject_GenericGetAttr, */ /* tp_getattro */
0,/* PyObject_GenericSetAttr, */ /* tp_setattro */
0, /* tp_as_buffer */
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_HAVE_GC, /* tp_flags */
scanner_doc, /* tp_doc */
scanner_traverse, /* tp_traverse */
scanner_clear, /* tp_clear */
0, /* tp_richcompare */
0, /* tp_weaklistoffset */
0, /* tp_iter */
0, /* tp_iternext */
0, /* tp_methods */
scanner_members, /* tp_members */
0, /* tp_getset */
0, /* tp_base */
0, /* tp_dict */
0, /* tp_descr_get */
0, /* tp_descr_set */
0, /* tp_dictoffset */
scanner_init, /* tp_init */
0,/* PyType_GenericAlloc, */ /* tp_alloc */
scanner_new, /* tp_new */
0,/* PyObject_GC_Del, */ /* tp_free */
};
static PyObject *
encoder_new(PyTypeObject *type, PyObject *args, PyObject *kwds)
{
PyEncoderObject *s;
s = (PyEncoderObject *)type->tp_alloc(type, 0);
if (s != NULL) {
s->markers = NULL;
s->defaultfn = NULL;
s->encoder = NULL;
s->indent = NULL;
s->key_separator = NULL;
s->item_separator = NULL;
s->sort_keys = NULL;
s->skipkeys = NULL;
}
return (PyObject *)s;
}
static int
encoder_init(PyObject *self, PyObject *args, PyObject *kwds)
{
/* initialize Encoder object */
static char *kwlist[] = {"markers", "default", "encoder", "indent", "key_separator", "item_separator", "sort_keys", "skipkeys", "allow_nan", NULL};
PyEncoderObject *s;
PyObject *allow_nan;
assert(PyEncoder_Check(self));
s = (PyEncoderObject *)self;
if (!PyArg_ParseTupleAndKeywords(args, kwds, "OOOOOOOOO:make_encoder", kwlist,
&s->markers, &s->defaultfn, &s->encoder, &s->indent, &s->key_separator, &s->item_separator, &s->sort_keys, &s->skipkeys, &allow_nan))
return -1;
Py_INCREF(s->markers);
Py_INCREF(s->defaultfn);
Py_INCREF(s->encoder);
Py_INCREF(s->indent);
Py_INCREF(s->key_separator);
Py_INCREF(s->item_separator);
Py_INCREF(s->sort_keys);
Py_INCREF(s->skipkeys);
s->fast_encode = (PyCFunction_Check(s->encoder) && PyCFunction_GetFunction(s->encoder) == (PyCFunction)py_encode_basestring_ascii);
s->allow_nan = PyObject_IsTrue(allow_nan);
return 0;
}
static PyObject *
encoder_call(PyObject *self, PyObject *args, PyObject *kwds)
{
/* Python callable interface to encode_listencode_obj */
static char *kwlist[] = {"obj", "_current_indent_level", NULL};
PyObject *obj;
PyObject *rval;
Py_ssize_t indent_level;
PyEncoderObject *s;
assert(PyEncoder_Check(self));
s = (PyEncoderObject *)self;
if (!PyArg_ParseTupleAndKeywords(args, kwds, "OO&:_iterencode", kwlist,
&obj, _convertPyInt_AsSsize_t, &indent_level))
return NULL;
rval = PyList_New(0);
if (rval == NULL)
return NULL;
if (encoder_listencode_obj(s, rval, obj, indent_level)) {
Py_DECREF(rval);
return NULL;
}
return rval;
}
static PyObject *
_encoded_const(PyObject *obj)
{
/* Return the JSON string representation of None, True, False */
if (obj == Py_None) {
static PyObject *s_null = NULL;
if (s_null == NULL) {
s_null = PyUnicode_InternFromString("null");
}
Py_INCREF(s_null);
return s_null;
}
else if (obj == Py_True) {
static PyObject *s_true = NULL;
if (s_true == NULL) {
s_true = PyUnicode_InternFromString("true");
}
Py_INCREF(s_true);
return s_true;
}
else if (obj == Py_False) {
static PyObject *s_false = NULL;
if (s_false == NULL) {
s_false = PyUnicode_InternFromString("false");
}
Py_INCREF(s_false);
return s_false;
}
else {
PyErr_SetString(PyExc_ValueError, "not a const");
return NULL;
}
}
static PyObject *
encoder_encode_float(PyEncoderObject *s, PyObject *obj)
{
/* Return the JSON representation of a PyFloat */
double i = PyFloat_AS_DOUBLE(obj);
if (!Py_IS_FINITE(i)) {
if (!s->allow_nan) {
PyErr_SetString(PyExc_ValueError, "Out of range float values are not JSON compliant");
return NULL;
}
if (i > 0) {
return PyUnicode_FromString("Infinity");
}
else if (i < 0) {
return PyUnicode_FromString("-Infinity");
}
else {
return PyUnicode_FromString("NaN");
}
}
/* Use a better float format here? */
return PyObject_Repr(obj);
}
static PyObject *
encoder_encode_string(PyEncoderObject *s, PyObject *obj)
{
/* Return the JSON representation of a string */
if (s->fast_encode)
return py_encode_basestring_ascii(NULL, obj);
else
return PyObject_CallFunctionObjArgs(s->encoder, obj, NULL);
}
static int
_steal_list_append(PyObject *lst, PyObject *stolen)
{
/* Append stolen and then decrement its reference count */
int rval = PyList_Append(lst, stolen);
Py_DECREF(stolen);
return rval;
}
static int
encoder_listencode_obj(PyEncoderObject *s, PyObject *rval, PyObject *obj, Py_ssize_t indent_level)
{
/* Encode Python object obj to a JSON term, rval is a PyList */
PyObject *newobj;
int rv;
if (obj == Py_None || obj == Py_True || obj == Py_False) {
PyObject *cstr = _encoded_const(obj);
if (cstr == NULL)
return -1;
return _steal_list_append(rval, cstr);
}
else if (PyUnicode_Check(obj))
{
PyObject *encoded = encoder_encode_string(s, obj);
if (encoded == NULL)
return -1;
return _steal_list_append(rval, encoded);
}
else if (PyLong_Check(obj)) {
PyObject *encoded = PyObject_Str(obj);
if (encoded == NULL)
return -1;
return _steal_list_append(rval, encoded);
}
else if (PyFloat_Check(obj)) {
PyObject *encoded = encoder_encode_float(s, obj);
if (encoded == NULL)
return -1;
return _steal_list_append(rval, encoded);
}
else if (PyList_Check(obj) || PyTuple_Check(obj)) {
return encoder_listencode_list(s, rval, obj, indent_level);
}
else if (PyDict_Check(obj)) {
return encoder_listencode_dict(s, rval, obj, indent_level);
}
else {
PyObject *ident = NULL;
if (s->markers != Py_None) {
int has_key;
ident = PyLong_FromVoidPtr(obj);
if (ident == NULL)
return -1;
has_key = PyDict_Contains(s->markers, ident);
if (has_key) {
if (has_key != -1)
PyErr_SetString(PyExc_ValueError, "Circular reference detected");
Py_DECREF(ident);
return -1;
}
if (PyDict_SetItem(s->markers, ident, obj)) {
Py_DECREF(ident);
return -1;
}
}
newobj = PyObject_CallFunctionObjArgs(s->defaultfn, obj, NULL);
if (newobj == NULL) {
Py_XDECREF(ident);
return -1;
}
rv = encoder_listencode_obj(s, rval, newobj, indent_level);
Py_DECREF(newobj);
if (rv) {
Py_XDECREF(ident);
return -1;
}
if (ident != NULL) {
if (PyDict_DelItem(s->markers, ident)) {
Py_XDECREF(ident);
return -1;
}
Py_XDECREF(ident);
}
return rv;
}
}
static int
encoder_listencode_dict(PyEncoderObject *s, PyObject *rval, PyObject *dct, Py_ssize_t indent_level)
{
/* Encode Python dict dct a JSON term, rval is a PyList */
static PyObject *open_dict = NULL;
static PyObject *close_dict = NULL;
static PyObject *empty_dict = NULL;
PyObject *kstr = NULL;
PyObject *ident = NULL;
PyObject *key, *value;
Py_ssize_t pos;
int skipkeys;
Py_ssize_t idx;
if (open_dict == NULL || close_dict == NULL || empty_dict == NULL) {
open_dict = PyUnicode_InternFromString("{");
close_dict = PyUnicode_InternFromString("}");
empty_dict = PyUnicode_InternFromString("{}");
if (open_dict == NULL || close_dict == NULL || empty_dict == NULL)
return -1;
}
if (PyDict_Size(dct) == 0)
return PyList_Append(rval, empty_dict);
if (s->markers != Py_None) {
int has_key;
ident = PyLong_FromVoidPtr(dct);
if (ident == NULL)
goto bail;
has_key = PyDict_Contains(s->markers, ident);
if (has_key) {
if (has_key != -1)
PyErr_SetString(PyExc_ValueError, "Circular reference detected");
goto bail;
}
if (PyDict_SetItem(s->markers, ident, dct)) {
goto bail;
}
}
if (PyList_Append(rval, open_dict))
goto bail;
if (s->indent != Py_None) {
/* TODO: DOES NOT RUN */
indent_level += 1;
/*
newline_indent = '\n' + (' ' * (_indent * _current_indent_level))
separator = _item_separator + newline_indent
buf += newline_indent
*/
}
/* TODO: C speedup not implemented for sort_keys */
pos = 0;
skipkeys = PyObject_IsTrue(s->skipkeys);
idx = 0;
while (PyDict_Next(dct, &pos, &key, &value)) {
PyObject *encoded;
if (PyUnicode_Check(key)) {
Py_INCREF(key);
kstr = key;
}
else if (PyFloat_Check(key)) {
kstr = encoder_encode_float(s, key);
if (kstr == NULL)
goto bail;
}
else if (PyLong_Check(key)) {
kstr = PyObject_Str(key);
if (kstr == NULL)
goto bail;
}
else if (key == Py_True || key == Py_False || key == Py_None) {
kstr = _encoded_const(key);
if (kstr == NULL)
goto bail;
}
else if (skipkeys) {
continue;
}
else {
/* TODO: include repr of key */
PyErr_SetString(PyExc_ValueError, "keys must be a string");
goto bail;
}
if (idx) {
if (PyList_Append(rval, s->item_separator))
goto bail;
}
encoded = encoder_encode_string(s, kstr);
Py_CLEAR(kstr);
if (encoded == NULL)
goto bail;
if (PyList_Append(rval, encoded)) {
Py_DECREF(encoded);
goto bail;
}
Py_DECREF(encoded);
if (PyList_Append(rval, s->key_separator))
goto bail;
if (encoder_listencode_obj(s, rval, value, indent_level))
goto bail;
idx += 1;
}
if (ident != NULL) {
if (PyDict_DelItem(s->markers, ident))
goto bail;
Py_CLEAR(ident);
}
if (s->indent != Py_None) {
/* TODO: DOES NOT RUN */
indent_level -= 1;
/*
yield '\n' + (' ' * (_indent * _current_indent_level))
*/
}
if (PyList_Append(rval, close_dict))
goto bail;
return 0;
bail:
Py_XDECREF(kstr);
Py_XDECREF(ident);
return -1;
}
static int
encoder_listencode_list(PyEncoderObject *s, PyObject *rval, PyObject *seq, Py_ssize_t indent_level)
{
/* Encode Python list seq to a JSON term, rval is a PyList */
static PyObject *open_array = NULL;
static PyObject *close_array = NULL;
static PyObject *empty_array = NULL;
PyObject *ident = NULL;
PyObject *s_fast = NULL;
Py_ssize_t num_items;
PyObject **seq_items;
Py_ssize_t i;
if (open_array == NULL || close_array == NULL || empty_array == NULL) {
open_array = PyUnicode_InternFromString("[");
close_array = PyUnicode_InternFromString("]");
empty_array = PyUnicode_InternFromString("[]");
if (open_array == NULL || close_array == NULL || empty_array == NULL)
return -1;
}
ident = NULL;
s_fast = PySequence_Fast(seq, "_iterencode_list needs a sequence");
if (s_fast == NULL)
return -1;
num_items = PySequence_Fast_GET_SIZE(s_fast);
if (num_items == 0) {
Py_DECREF(s_fast);
return PyList_Append(rval, empty_array);
}
if (s->markers != Py_None) {
int has_key;
ident = PyLong_FromVoidPtr(seq);
if (ident == NULL)
goto bail;
has_key = PyDict_Contains(s->markers, ident);
if (has_key) {
if (has_key != -1)
PyErr_SetString(PyExc_ValueError, "Circular reference detected");
goto bail;
}
if (PyDict_SetItem(s->markers, ident, seq)) {
goto bail;
}
}
seq_items = PySequence_Fast_ITEMS(s_fast);
if (PyList_Append(rval, open_array))
goto bail;
if (s->indent != Py_None) {
/* TODO: DOES NOT RUN */
indent_level += 1;
/*
newline_indent = '\n' + (' ' * (_indent * _current_indent_level))
separator = _item_separator + newline_indent
buf += newline_indent
*/
}
for (i = 0; i < num_items; i++) {
PyObject *obj = seq_items[i];
if (i) {
if (PyList_Append(rval, s->item_separator))
goto bail;
}
if (encoder_listencode_obj(s, rval, obj, indent_level))
goto bail;
}
if (ident != NULL) {
if (PyDict_DelItem(s->markers, ident))
goto bail;
Py_CLEAR(ident);
}
if (s->indent != Py_None) {
/* TODO: DOES NOT RUN */
indent_level -= 1;
/*
yield '\n' + (' ' * (_indent * _current_indent_level))
*/
}
if (PyList_Append(rval, close_array))
goto bail;
Py_DECREF(s_fast);
return 0;
bail:
Py_XDECREF(ident);
Py_DECREF(s_fast);
return -1;
}
static void
encoder_dealloc(PyObject *self)
{
/* Deallocate Encoder */
encoder_clear(self);
Py_TYPE(self)->tp_free(self);
}
static int
encoder_traverse(PyObject *self, visitproc visit, void *arg)
{
PyEncoderObject *s;
assert(PyEncoder_Check(self));
s = (PyEncoderObject *)self;
Py_VISIT(s->markers);
Py_VISIT(s->defaultfn);
Py_VISIT(s->encoder);
Py_VISIT(s->indent);
Py_VISIT(s->key_separator);
Py_VISIT(s->item_separator);
Py_VISIT(s->sort_keys);
Py_VISIT(s->skipkeys);
return 0;
}
static int
encoder_clear(PyObject *self)
{
/* Deallocate Encoder */
PyEncoderObject *s;
assert(PyEncoder_Check(self));
s = (PyEncoderObject *)self;
Py_CLEAR(s->markers);
Py_CLEAR(s->defaultfn);
Py_CLEAR(s->encoder);
Py_CLEAR(s->indent);
Py_CLEAR(s->key_separator);
Py_CLEAR(s->item_separator);
Py_CLEAR(s->sort_keys);
Py_CLEAR(s->skipkeys);
return 0;
}
PyDoc_STRVAR(encoder_doc, "_iterencode(obj, _current_indent_level) -> iterable");
static
PyTypeObject PyEncoderType = {
PyVarObject_HEAD_INIT(NULL, 0)
"_json.Encoder", /* tp_name */
sizeof(PyEncoderObject), /* tp_basicsize */
0, /* tp_itemsize */
encoder_dealloc, /* tp_dealloc */
0, /* tp_print */
0, /* tp_getattr */
0, /* tp_setattr */
0, /* tp_compare */
0, /* tp_repr */
0, /* tp_as_number */
0, /* tp_as_sequence */
0, /* tp_as_mapping */
0, /* tp_hash */
encoder_call, /* tp_call */
0, /* tp_str */
0, /* tp_getattro */
0, /* tp_setattro */
0, /* tp_as_buffer */
Py_TPFLAGS_DEFAULT | Py_TPFLAGS_HAVE_GC, /* tp_flags */
encoder_doc, /* tp_doc */
encoder_traverse, /* tp_traverse */
encoder_clear, /* tp_clear */
0, /* tp_richcompare */
0, /* tp_weaklistoffset */
0, /* tp_iter */
0, /* tp_iternext */
0, /* tp_methods */
encoder_members, /* tp_members */
0, /* tp_getset */
0, /* tp_base */
0, /* tp_dict */
0, /* tp_descr_get */
0, /* tp_descr_set */
0, /* tp_dictoffset */
encoder_init, /* tp_init */
0, /* tp_alloc */
encoder_new, /* tp_new */
0, /* tp_free */
};
static PyMethodDef speedups_methods[] = {
{"encode_basestring_ascii",
(PyCFunction)py_encode_basestring_ascii,
METH_O,
pydoc_encode_basestring_ascii},
{"scanstring",
(PyCFunction)py_scanstring,
METH_VARARGS,
pydoc_scanstring},
{NULL, NULL, 0, NULL}
};
PyDoc_STRVAR(module_doc,
"json speedups\n");
static struct PyModuleDef jsonmodule = {
PyModuleDef_HEAD_INIT,
"_json",
module_doc,
-1,
speedups_methods,
NULL,
NULL,
NULL,
NULL
};
PyObject*
PyInit__json(void)
{
PyObject *m = PyModule_Create(&jsonmodule);
if (!m)
return NULL;
PyScannerType.tp_new = PyType_GenericNew;
if (PyType_Ready(&PyScannerType) < 0)
goto fail;
PyEncoderType.tp_new = PyType_GenericNew;
if (PyType_Ready(&PyEncoderType) < 0)
goto fail;
Py_INCREF((PyObject*)&PyScannerType);
if (PyModule_AddObject(m, "make_scanner", (PyObject*)&PyScannerType) < 0) {
Py_DECREF((PyObject*)&PyScannerType);
goto fail;
}
Py_INCREF((PyObject*)&PyEncoderType);
if (PyModule_AddObject(m, "make_encoder", (PyObject*)&PyEncoderType) < 0) {
Py_DECREF((PyObject*)&PyEncoderType);
goto fail;
}
return m;
fail:
Py_DECREF(m);
return NULL;
} }
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment