Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Commit 8a5f8ca

Browse files
committed
Run 2to3 on this library.
1 parent f733c60 commit 8a5f8ca

22 files changed

Lines changed: 196 additions & 194 deletions

Lib/lib2to3/fixes/basefix.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,7 @@ def new_name(self, template="xxx_todo_changeme"):
108108
"""
109109
name = template
110110
while name in self.used_names:
111-
name = template + str(self.numbers.next())
111+
name = template + str(next(self.numbers))
112112
self.used_names.add(name)
113113
return name
114114

Lib/lib2to3/fixes/fix_imports.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@
1010
# Local imports
1111
from . import basefix
1212
from .util import Name, attr_chain, any, set
13-
import __builtin__
14-
builtin_names = [name for name in dir(__builtin__)
13+
import builtins
14+
builtin_names = [name for name in dir(builtins)
1515
if name not in ("__name__", "__doc__")]
1616

1717
MAPPING = {"StringIO": ("io", ["StringIO"]),
@@ -26,7 +26,7 @@ def alternates(members):
2626

2727
def build_pattern():
2828
bare = set()
29-
for old_module, (new_module, members) in MAPPING.items():
29+
for old_module, (new_module, members) in list(MAPPING.items()):
3030
bare.add(old_module)
3131
bare.update(members)
3232
members = alternates(members)

Lib/lib2to3/fixes/fix_renames.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,8 +20,8 @@ def alternates(members):
2020

2121
def build_pattern():
2222
#bare = set()
23-
for module, replace in MAPPING.items():
24-
for old_attr, new_attr in replace.items():
23+
for module, replace in list(MAPPING.items()):
24+
for old_attr, new_attr in list(replace.items()):
2525
LOOKUP[(module, old_attr)] = new_attr
2626
#bare.add(module)
2727
#bare.add(old_attr)

Lib/lib2to3/fixes/util.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -323,7 +323,7 @@ def _is_import_binding(node, name, package=None):
323323
elif node.type == syms.import_from:
324324
# unicode(...) is used to make life easier here, because
325325
# from a.b import parses to ['import', ['a', '.', 'b'], ...]
326-
if package and unicode(node.children[1]).strip() != package:
326+
if package and str(node.children[1]).strip() != package:
327327
return None
328328
n = node.children[3]
329329
if package and _find('as', n):

Lib/lib2to3/patcomp.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
def tokenize_wrapper(input):
3232
"""Tokenizes a string suppressing significant whitespace."""
3333
skip = (token.NEWLINE, token.INDENT, token.DEDENT)
34-
tokens = tokenize.generate_tokens(driver.generate_lines(input).next)
34+
tokens = tokenize.generate_tokens(driver.generate_lines(input).__next__)
3535
for quintuple in tokens:
3636
type, value, start, end, line_text = quintuple
3737
if type not in skip:

Lib/lib2to3/pgen2/conv.py

Lines changed: 33 additions & 33 deletions
Original file line numberDiff line numberDiff line change
@@ -60,8 +60,8 @@ def parse_graminit_h(self, filename):
6060
"""
6161
try:
6262
f = open(filename)
63-
except IOError, err:
64-
print "Can't open %s: %s" % (filename, err)
63+
except IOError as err:
64+
print("Can't open %s: %s" % (filename, err))
6565
return False
6666
self.symbol2number = {}
6767
self.number2symbol = {}
@@ -70,8 +70,8 @@ def parse_graminit_h(self, filename):
7070
lineno += 1
7171
mo = re.match(r"^#define\s+(\w+)\s+(\d+)$", line)
7272
if not mo and line.strip():
73-
print "%s(%s): can't parse %s" % (filename, lineno,
74-
line.strip())
73+
print("%s(%s): can't parse %s" % (filename, lineno,
74+
line.strip()))
7575
else:
7676
symbol, number = mo.groups()
7777
number = int(number)
@@ -111,56 +111,56 @@ def parse_graminit_c(self, filename):
111111
"""
112112
try:
113113
f = open(filename)
114-
except IOError, err:
115-
print "Can't open %s: %s" % (filename, err)
114+
except IOError as err:
115+
print("Can't open %s: %s" % (filename, err))
116116
return False
117117
# The code below essentially uses f's iterator-ness!
118118
lineno = 0
119119

120120
# Expect the two #include lines
121-
lineno, line = lineno+1, f.next()
121+
lineno, line = lineno+1, next(f)
122122
assert line == '#include "pgenheaders.h"\n', (lineno, line)
123-
lineno, line = lineno+1, f.next()
123+
lineno, line = lineno+1, next(f)
124124
assert line == '#include "grammar.h"\n', (lineno, line)
125125

126126
# Parse the state definitions
127-
lineno, line = lineno+1, f.next()
127+
lineno, line = lineno+1, next(f)
128128
allarcs = {}
129129
states = []
130130
while line.startswith("static arc "):
131131
while line.startswith("static arc "):
132132
mo = re.match(r"static arc arcs_(\d+)_(\d+)\[(\d+)\] = {$",
133133
line)
134134
assert mo, (lineno, line)
135-
n, m, k = map(int, mo.groups())
135+
n, m, k = list(map(int, mo.groups()))
136136
arcs = []
137137
for _ in range(k):
138-
lineno, line = lineno+1, f.next()
138+
lineno, line = lineno+1, next(f)
139139
mo = re.match(r"\s+{(\d+), (\d+)},$", line)
140140
assert mo, (lineno, line)
141-
i, j = map(int, mo.groups())
141+
i, j = list(map(int, mo.groups()))
142142
arcs.append((i, j))
143-
lineno, line = lineno+1, f.next()
143+
lineno, line = lineno+1, next(f)
144144
assert line == "};\n", (lineno, line)
145145
allarcs[(n, m)] = arcs
146-
lineno, line = lineno+1, f.next()
146+
lineno, line = lineno+1, next(f)
147147
mo = re.match(r"static state states_(\d+)\[(\d+)\] = {$", line)
148148
assert mo, (lineno, line)
149-
s, t = map(int, mo.groups())
149+
s, t = list(map(int, mo.groups()))
150150
assert s == len(states), (lineno, line)
151151
state = []
152152
for _ in range(t):
153-
lineno, line = lineno+1, f.next()
153+
lineno, line = lineno+1, next(f)
154154
mo = re.match(r"\s+{(\d+), arcs_(\d+)_(\d+)},$", line)
155155
assert mo, (lineno, line)
156-
k, n, m = map(int, mo.groups())
156+
k, n, m = list(map(int, mo.groups()))
157157
arcs = allarcs[n, m]
158158
assert k == len(arcs), (lineno, line)
159159
state.append(arcs)
160160
states.append(state)
161-
lineno, line = lineno+1, f.next()
161+
lineno, line = lineno+1, next(f)
162162
assert line == "};\n", (lineno, line)
163-
lineno, line = lineno+1, f.next()
163+
lineno, line = lineno+1, next(f)
164164
self.states = states
165165

166166
# Parse the dfas
@@ -169,18 +169,18 @@ def parse_graminit_c(self, filename):
169169
assert mo, (lineno, line)
170170
ndfas = int(mo.group(1))
171171
for i in range(ndfas):
172-
lineno, line = lineno+1, f.next()
172+
lineno, line = lineno+1, next(f)
173173
mo = re.match(r'\s+{(\d+), "(\w+)", (\d+), (\d+), states_(\d+),$',
174174
line)
175175
assert mo, (lineno, line)
176176
symbol = mo.group(2)
177-
number, x, y, z = map(int, mo.group(1, 3, 4, 5))
177+
number, x, y, z = list(map(int, mo.group(1, 3, 4, 5)))
178178
assert self.symbol2number[symbol] == number, (lineno, line)
179179
assert self.number2symbol[number] == symbol, (lineno, line)
180180
assert x == 0, (lineno, line)
181181
state = states[z]
182182
assert y == len(state), (lineno, line)
183-
lineno, line = lineno+1, f.next()
183+
lineno, line = lineno+1, next(f)
184184
mo = re.match(r'\s+("(?:\\\d\d\d)*")},$', line)
185185
assert mo, (lineno, line)
186186
first = {}
@@ -191,18 +191,18 @@ def parse_graminit_c(self, filename):
191191
if byte & (1<<j):
192192
first[i*8 + j] = 1
193193
dfas[number] = (state, first)
194-
lineno, line = lineno+1, f.next()
194+
lineno, line = lineno+1, next(f)
195195
assert line == "};\n", (lineno, line)
196196
self.dfas = dfas
197197

198198
# Parse the labels
199199
labels = []
200-
lineno, line = lineno+1, f.next()
200+
lineno, line = lineno+1, next(f)
201201
mo = re.match(r"static label labels\[(\d+)\] = {$", line)
202202
assert mo, (lineno, line)
203203
nlabels = int(mo.group(1))
204204
for i in range(nlabels):
205-
lineno, line = lineno+1, f.next()
205+
lineno, line = lineno+1, next(f)
206206
mo = re.match(r'\s+{(\d+), (0|"\w+")},$', line)
207207
assert mo, (lineno, line)
208208
x, y = mo.groups()
@@ -212,35 +212,35 @@ def parse_graminit_c(self, filename):
212212
else:
213213
y = eval(y)
214214
labels.append((x, y))
215-
lineno, line = lineno+1, f.next()
215+
lineno, line = lineno+1, next(f)
216216
assert line == "};\n", (lineno, line)
217217
self.labels = labels
218218

219219
# Parse the grammar struct
220-
lineno, line = lineno+1, f.next()
220+
lineno, line = lineno+1, next(f)
221221
assert line == "grammar _PyParser_Grammar = {\n", (lineno, line)
222-
lineno, line = lineno+1, f.next()
222+
lineno, line = lineno+1, next(f)
223223
mo = re.match(r"\s+(\d+),$", line)
224224
assert mo, (lineno, line)
225225
ndfas = int(mo.group(1))
226226
assert ndfas == len(self.dfas)
227-
lineno, line = lineno+1, f.next()
227+
lineno, line = lineno+1, next(f)
228228
assert line == "\tdfas,\n", (lineno, line)
229-
lineno, line = lineno+1, f.next()
229+
lineno, line = lineno+1, next(f)
230230
mo = re.match(r"\s+{(\d+), labels},$", line)
231231
assert mo, (lineno, line)
232232
nlabels = int(mo.group(1))
233233
assert nlabels == len(self.labels), (lineno, line)
234-
lineno, line = lineno+1, f.next()
234+
lineno, line = lineno+1, next(f)
235235
mo = re.match(r"\s+(\d+)$", line)
236236
assert mo, (lineno, line)
237237
start = int(mo.group(1))
238238
assert start in self.number2symbol, (lineno, line)
239239
self.start = start
240-
lineno, line = lineno+1, f.next()
240+
lineno, line = lineno+1, next(f)
241241
assert line == "};\n", (lineno, line)
242242
try:
243-
lineno, line = lineno+1, f.next()
243+
lineno, line = lineno+1, next(f)
244244
except StopIteration:
245245
pass
246246
else:

Lib/lib2to3/pgen2/driver.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -99,7 +99,7 @@ def parse_file(self, filename, debug=False):
9999

100100
def parse_string(self, text, debug=False):
101101
"""Parse a string and return the syntax tree."""
102-
tokens = tokenize.generate_tokens(generate_lines(text).next)
102+
tokens = tokenize.generate_tokens(generate_lines(text).__next__)
103103
return self.parse_tokens(tokens, debug)
104104

105105

Lib/lib2to3/pgen2/grammar.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -100,17 +100,17 @@ def load(self, filename):
100100
def report(self):
101101
"""Dump the grammar tables to standard output, for debugging."""
102102
from pprint import pprint
103-
print "s2n"
103+
print("s2n")
104104
pprint(self.symbol2number)
105-
print "n2s"
105+
print("n2s")
106106
pprint(self.number2symbol)
107-
print "states"
107+
print("states")
108108
pprint(self.states)
109-
print "dfas"
109+
print("dfas")
110110
pprint(self.dfas)
111-
print "labels"
111+
print("labels")
112112
pprint(self.labels)
113-
print "start", self.start
113+
print("start", self.start)
114114

115115

116116
# Map from operator to number (since tokenize doesn't do this)

Lib/lib2to3/pgen2/literals.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -53,7 +53,7 @@ def test():
5353
s = repr(c)
5454
e = evalString(s)
5555
if e != c:
56-
print i, c, s, e
56+
print(i, c, s, e)
5757

5858

5959
if __name__ == "__main__":

0 commit comments

Comments
 (0)