/*[clinic input] preserve [clinic start generated code]*/ static PyObject * tokenizeriter_new_impl(PyTypeObject *type, const char *source); static PyObject * tokenizeriter_new(PyTypeObject *type, PyObject *args, PyObject *kwargs) { PyObject *return_value = NULL; static const char * const _keywords[] = {"source", NULL}; static _PyArg_Parser _parser = {NULL, _keywords, "tokenizeriter", 0}; PyObject *argsbuf[1]; PyObject * const *fastargs; Py_ssize_t nargs = PyTuple_GET_SIZE(args); const char *source; fastargs = _PyArg_UnpackKeywords(_PyTuple_CAST(args)->ob_item, nargs, kwargs, NULL, &_parser, 1, 1, 0, argsbuf); if (!fastargs) { goto exit; } if (!PyUnicode_Check(fastargs[0])) { _PyArg_BadArgument("tokenizeriter", "argument 'source'", "str", fastargs[0]); goto exit; } Py_ssize_t source_length; source = PyUnicode_AsUTF8AndSize(fastargs[0], &source_length); if (source == NULL) { goto exit; } if (strlen(source) != (size_t)source_length) { PyErr_SetString(PyExc_ValueError, "embedded null character"); goto exit; } return_value = tokenizeriter_new_impl(type, source); exit: return return_value; } /*[clinic end generated code: output=dfcd64774e01bfe6 input=a9049054013a1b77]*/