Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Commit a5dca7e

Browse files
committed
fix: fix content based on the suggestion
1 parent 5b290a3 commit a5dca7e

File tree

1 file changed

+22
-18
lines changed

1 file changed

+22
-18
lines changed

library/urllib.robotparser.po

Lines changed: 22 additions & 18 deletions
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ msgstr ""
1010
"Project-Id-Version: Python 3.10\n"
1111
"Report-Msgid-Bugs-To: \n"
1212
"POT-Creation-Date: 2021-10-26 16:47+0000\n"
13-
"PO-Revision-Date: 2022-01-23 14:36+0800\n"
13+
"PO-Revision-Date: 2022-01-27 13:40+0800\n"
1414
"Last-Translator: Phil Lin <[email protected]>\n"
1515
"Language-Team: Chinese - TAIWAN (https://github.com/python/python-docs-zh-"
1616
"tw)\n"
@@ -23,7 +23,7 @@ msgstr ""
2323

2424
#: ../../library/urllib.robotparser.rst:2
2525
msgid ":mod:`urllib.robotparser` --- Parser for robots.txt"
26-
msgstr ":mod:`urllib.robotparser` --- robots.txt 的解析器"
26+
msgstr ":mod:`urllib.robotparser` --- robots.txt 的剖析器"
2727

2828
#: ../../library/urllib.robotparser.rst:10
2929
msgid "**Source code:** :source:`Lib/urllib/robotparser.py`"
@@ -37,17 +37,17 @@ msgid ""
3737
"on the structure of :file:`robots.txt` files, see http://www.robotstxt.org/"
3838
"orig.html."
3939
msgstr ""
40-
"此模組提供了一個單獨的類別 :class:`RobotFileParser`, 它可以知道某個特定 user "
41-
"agent (使用者代理)是否能在有發布 :file:`robots.txt` 文件的網站 fetch(擷"
42-
"取)此網站特定的 URL。有關 :file:`robots.txt` 文件結構的更多細節,請參閱 "
43-
"http://www.robotstxt.org/orig.html 。"
40+
"此模組 (module) 提供了一個單獨的類別 (class) \\ :class:`RobotFileParser`\\"
41+
"它可以知道某個特定 user agent(使用者代理)是否能在有發布 :file:`robots.txt` "
42+
"文件的網站 fetch(擷取)特定 URL。有關 :file:`robots.txt` 文件結構的更多細"
43+
"節,請參閱 http://www.robotstxt.org/orig.html。"
4444

4545
#: ../../library/urllib.robotparser.rst:28
4646
msgid ""
4747
"This class provides methods to read, parse and answer questions about the :"
4848
"file:`robots.txt` file at *url*."
4949
msgstr ""
50-
"此類別提供了一些方法可以讀取、解析和回答關於 *url* 上的 :file:`robots.txt` 文"
50+
"此類別提供了一些方法可以讀取、剖析和回答關於 *url* 上的 :file:`robots.txt` 文"
5151
"件的問題。"
5252

5353
#: ../../library/urllib.robotparser.rst:33
@@ -56,19 +56,19 @@ msgstr "設置指向 :file:`robots.txt` 文件的 URL。"
5656

5757
#: ../../library/urllib.robotparser.rst:37
5858
msgid "Reads the :file:`robots.txt` URL and feeds it to the parser."
59-
msgstr "讀取 :file:`robots.txt` URL 並將其輸入到解析器。"
59+
msgstr "讀取 :file:`robots.txt` URL 並將其輸入到剖析器。"
6060

6161
#: ../../library/urllib.robotparser.rst:41
6262
msgid "Parses the lines argument."
63-
msgstr "解析行參數(此參數為 ``robots.txt`` 文件裡的行)。"
63+
msgstr "剖析 lines 引數。"
6464

6565
#: ../../library/urllib.robotparser.rst:45
6666
msgid ""
6767
"Returns ``True`` if the *useragent* is allowed to fetch the *url* according "
6868
"to the rules contained in the parsed :file:`robots.txt` file."
6969
msgstr ""
70-
"如果根據被解析的 :file:`robots.txt` 文件中的規則,*useragent* 被允許 fetch "
71-
"*url* 的話,則返回 ``True``。"
70+
"根據從 :file:`robots.txt` 文件中剖析出的規則,如果 *useragent* 被允許 fetch "
71+
"*url* 的話,則回傳 ``True``。"
7272

7373
#: ../../library/urllib.robotparser.rst:51
7474
msgid ""
@@ -82,7 +82,7 @@ msgstr ""
8282
#: ../../library/urllib.robotparser.rst:57
8383
msgid ""
8484
"Sets the time the ``robots.txt`` file was last fetched to the current time."
85-
msgstr "將最後一次獲取 ``robots.txt`` 文件的時間設置為當前時間。"
85+
msgstr "將最近一次 fetch ``robots.txt`` 文件的時間設置為當前時間。"
8686

8787
#: ../../library/urllib.robotparser.rst:62
8888
msgid ""
@@ -92,8 +92,8 @@ msgid ""
9292
"parameter has invalid syntax, return ``None``."
9393
msgstr ""
9494
"針對指定的 *useragent* 從 ``robots.txt`` 返回 ``Crawl-delay`` 參數的值。如果"
95-
"此參數不存在或不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 存在語"
96-
"法錯誤,則返回 ``None``。"
95+
"此參數不存在、不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 中所指"
96+
"的條目含有無效語法,則回傳 ``None``。"
9797

9898
#: ../../library/urllib.robotparser.rst:71
9999
msgid ""
@@ -103,8 +103,9 @@ msgid ""
103103
"``robots.txt`` entry for this parameter has invalid syntax, return ``None``."
104104
msgstr ""
105105
"以 :term:`named tuple` ``RequestRate(requests, seconds)`` 的形式從 ``robots."
106-
"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在或不適用於指定的 "
107-
"*useragent* ,或是此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。"
106+
"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在、不適用於指定的 "
107+
"*useragent* ,或是此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 "
108+
"``None``。"
108109

109110
#: ../../library/urllib.robotparser.rst:81
110111
msgid ""
@@ -113,10 +114,13 @@ msgid ""
113114
"entry for this parameter has invalid syntax, return ``None``."
114115
msgstr ""
115116
"以 :func:`list` 的形式從 ``robots.txt`` 返回 ``Sitemap`` 參數的內容。如果此參"
116-
"數不存在或此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。"
117+
"數不存在或此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 ``None``。"
117118

118119
#: ../../library/urllib.robotparser.rst:89
119120
msgid ""
120121
"The following example demonstrates basic use of the :class:`RobotFileParser` "
121122
"class::"
122-
msgstr "下面的範例展示了 :class:`RobotFileParser` 類別的基本用法::"
123+
msgstr ""
124+
"下面的範例展示了 :class:`RobotFileParser` 類別的基本用法:\n"
125+
"\n"
126+
"::"

0 commit comments

Comments
 (0)