summaryrefslogtreecommitdiffstats
path: root/Python/pyhash.c
diff options
context:
space:
mode:
authorSerhiy Storchaka <storchaka@gmail.com>2015-03-24 17:49:23 (GMT)
committerSerhiy Storchaka <storchaka@gmail.com>2015-03-24 17:49:23 (GMT)
commit520e00dd9289a56d3af2f2700b1145f6c211b0db (patch)
treeb37f967345ba74743b319506d41965574ebbb6de /Python/pyhash.c
parent87eb482e30ec76976e7c04b1547faaab0df40261 (diff)
parent6c86fe2d4b3477eece6695604051273fb95b37b3 (diff)
downloadcpython-520e00dd9289a56d3af2f2700b1145f6c211b0db.zip
cpython-520e00dd9289a56d3af2f2700b1145f6c211b0db.tar.gz
cpython-520e00dd9289a56d3af2f2700b1145f6c211b0db.tar.bz2
Issue #23583: Added tests for standard IO streams in IDLE.
Diffstat (limited to 'Python/pyhash.c')
0 files changed, 0 insertions, 0 deletions
n class="hl opt">*readline, int extra_tokens, const char *encoding); static PyObject * tokenizeriter_new(PyTypeObject *type, PyObject *args, PyObject *kwargs) { PyObject *return_value = NULL; #if defined(Py_BUILD_CORE) && !defined(Py_BUILD_CORE_MODULE) #define NUM_KEYWORDS 2 static struct { PyGC_Head _this_is_not_used; PyObject_VAR_HEAD PyObject *ob_item[NUM_KEYWORDS]; } _kwtuple = { .ob_base = PyVarObject_HEAD_INIT(&PyTuple_Type, NUM_KEYWORDS) .ob_item = { &_Py_ID(extra_tokens), &_Py_ID(encoding), }, }; #undef NUM_KEYWORDS #define KWTUPLE (&_kwtuple.ob_base.ob_base) #else // !Py_BUILD_CORE # define KWTUPLE NULL #endif // !Py_BUILD_CORE static const char * const _keywords[] = {"", "extra_tokens", "encoding", NULL}; static _PyArg_Parser _parser = { .keywords = _keywords, .fname = "tokenizeriter", .kwtuple = KWTUPLE, }; #undef KWTUPLE PyObject *argsbuf[3]; PyObject * const *fastargs; Py_ssize_t nargs = PyTuple_GET_SIZE(args); Py_ssize_t noptargs = nargs + (kwargs ? PyDict_GET_SIZE(kwargs) : 0) - 2; PyObject *readline; int extra_tokens; const char *encoding = NULL; fastargs = _PyArg_UnpackKeywords(_PyTuple_CAST(args)->ob_item, nargs, kwargs, NULL, &_parser, 1, 1, 1, argsbuf); if (!fastargs) { goto exit; } readline = fastargs[0]; extra_tokens = PyObject_IsTrue(fastargs[1]); if (extra_tokens < 0) { goto exit; } if (!noptargs) { goto skip_optional_kwonly; } if (!PyUnicode_Check(fastargs[2])) { _PyArg_BadArgument("tokenizeriter", "argument 'encoding'", "str", fastargs[2]); goto exit; } Py_ssize_t encoding_length; encoding = PyUnicode_AsUTF8AndSize(fastargs[2], &encoding_length); if (encoding == NULL) { goto exit; } if (strlen(encoding) != (size_t)encoding_length) { PyErr_SetString(PyExc_ValueError, "embedded null character"); goto exit; } skip_optional_kwonly: return_value = tokenizeriter_new_impl(type, readline, extra_tokens, encoding); exit: return return_value; } /*[clinic end generated code: output=dcd6ec48f06a092e input=a9049054013a1b77]*/