From 3fb2563d089c3380fabcc8c7563fa5fa9bbf00b5 Mon Sep 17 00:00:00 2001 From: Eli Schwartz Date: Mon, 17 Dec 2018 10:30:49 -0500 Subject: [PATCH] tokenizer: fix compile errors in merge of python3 port --- src/tinycss/tokenizer.c | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/src/tinycss/tokenizer.c b/src/tinycss/tokenizer.c index 5df9447a1b..be6e614ceb 100644 --- a/src/tinycss/tokenizer.c +++ b/src/tinycss/tokenizer.c @@ -192,7 +192,7 @@ tokenize_init(PyObject *self, PyObject *args) { int _kind = PyUnicode_KIND(unicode_object); \ void *_data = PyUnicode_DATA(unicode_object); \ for (Py_ssize_t iteridx = 0; iteridx < PyUnicode_GET_LENGTH(unicode_object); iteridx++) { \ - Py_UCS4 ch = PyUnicode_READ(kind, data, i); + Py_UCS4 ch = PyUnicode_READ(_kind, _data, iteridx); #else #define PyUnicode_GET_LENGTH PyUnicode_GET_SIZE #define ITER_CODE_PTS(unicode_object) { \ @@ -260,7 +260,7 @@ clone_unicode(const PyObject* src, Py_ssize_t start_offset, Py_ssize_t end_offse data = PyUnicode_4BYTE_DATA(src) + start_offset; break; } - return PyUnicode_FromKindAndData(kind, data, PyUnicode_GET_LENGTH(src) - start_offset - end_offset) + return PyUnicode_FromKindAndData(kind, data, PyUnicode_GET_LENGTH(src) - start_offset - end_offset); #else return PyUnicode_FromUnicode(PyUnicode_AS_UNICODE(src) + start_offset, PyUnicode_GET_LENGTH(src) - start_offset - end_offset); #endif @@ -308,7 +308,7 @@ tokenize_flat(PyObject *self, PyObject *args) { while (pos < source_len) { #if PY_VERSION_HEX >= 0x03030000 - c = PyUnicode_READ(css_kind, css_data, pos); + c = PyUnicode_READ(css_kind, css_source, pos); #else c = css_source[pos]; #endif