From 5b290a3b54365346a25fd856c8557f971f167f45 Mon Sep 17 00:00:00 2001 From: linooohon Date: Sun, 23 Jan 2022 14:57:20 +0800 Subject: [PATCH 1/3] translate library/urllib.robotparser --- library/urllib.robotparser.po | 37 +++++++++++++++++++++++++++-------- 1 file changed, 29 insertions(+), 8 deletions(-) diff --git a/library/urllib.robotparser.po b/library/urllib.robotparser.po index 6c325ebd6c..019caea519 100644 --- a/library/urllib.robotparser.po +++ b/library/urllib.robotparser.po @@ -3,13 +3,15 @@ # This file is distributed under the same license as the Python package. # # Translators: +# Adrian Liaw , 2018 +# Phil Lin , 2022 msgid "" msgstr "" "Project-Id-Version: Python 3.10\n" "Report-Msgid-Bugs-To: \n" "POT-Creation-Date: 2021-10-26 16:47+0000\n" -"PO-Revision-Date: 2018-05-23 16:14+0000\n" -"Last-Translator: Adrian Liaw \n" +"PO-Revision-Date: 2022-01-23 14:36+0800\n" +"Last-Translator: Phil Lin \n" "Language-Team: Chinese - TAIWAN (https://github.com/python/python-docs-zh-" "tw)\n" "Language: zh_TW\n" @@ -17,10 +19,11 @@ msgstr "" "Content-Type: text/plain; charset=UTF-8\n" "Content-Transfer-Encoding: 8bit\n" "Plural-Forms: nplurals=1; plural=0;\n" +"X-Generator: Poedit 3.0.1\n" #: ../../library/urllib.robotparser.rst:2 msgid ":mod:`urllib.robotparser` --- Parser for robots.txt" -msgstr "" +msgstr ":mod:`urllib.robotparser` --- robots.txt 的解析器" #: ../../library/urllib.robotparser.rst:10 msgid "**Source code:** :source:`Lib/urllib/robotparser.py`" @@ -34,30 +37,38 @@ msgid "" "on the structure of :file:`robots.txt` files, see http://www.robotstxt.org/" "orig.html." msgstr "" +"此模組提供了一個單獨的類別 :class:`RobotFileParser`, 它可以知道某個特定 user " +"agent (使用者代理)是否能在有發布 :file:`robots.txt` 文件的網站 fetch(擷" +"取)此網站特定的 URL。有關 :file:`robots.txt` 文件結構的更多細節,請參閱 " +"http://www.robotstxt.org/orig.html 。" #: ../../library/urllib.robotparser.rst:28 msgid "" "This class provides methods to read, parse and answer questions about the :" "file:`robots.txt` file at *url*." msgstr "" +"此類別提供了一些方法可以讀取、解析和回答關於 *url* 上的 :file:`robots.txt` 文" +"件的問題。" #: ../../library/urllib.robotparser.rst:33 msgid "Sets the URL referring to a :file:`robots.txt` file." -msgstr "" +msgstr "設置指向 :file:`robots.txt` 文件的 URL。" #: ../../library/urllib.robotparser.rst:37 msgid "Reads the :file:`robots.txt` URL and feeds it to the parser." -msgstr "" +msgstr "讀取 :file:`robots.txt` URL 並將其輸入到解析器。" #: ../../library/urllib.robotparser.rst:41 msgid "Parses the lines argument." -msgstr "" +msgstr "解析行參數(此參數為 ``robots.txt`` 文件裡的行)。" #: ../../library/urllib.robotparser.rst:45 msgid "" "Returns ``True`` if the *useragent* is allowed to fetch the *url* according " "to the rules contained in the parsed :file:`robots.txt` file." msgstr "" +"如果根據被解析的 :file:`robots.txt` 文件中的規則,*useragent* 被允許 fetch " +"*url* 的話,則返回 ``True``。" #: ../../library/urllib.robotparser.rst:51 msgid "" @@ -65,11 +76,13 @@ msgid "" "for long-running web spiders that need to check for new ``robots.txt`` files " "periodically." msgstr "" +"返回最近一次 fetch ``robots.txt`` 文件的時間。這適用於需要定期檢查 ``robots." +"txt`` 文件更新情況的長時間運行網頁爬蟲。" #: ../../library/urllib.robotparser.rst:57 msgid "" "Sets the time the ``robots.txt`` file was last fetched to the current time." -msgstr "" +msgstr "將最後一次獲取 ``robots.txt`` 文件的時間設置為當前時間。" #: ../../library/urllib.robotparser.rst:62 msgid "" @@ -78,6 +91,9 @@ msgid "" "apply to the *useragent* specified or the ``robots.txt`` entry for this " "parameter has invalid syntax, return ``None``." msgstr "" +"針對指定的 *useragent* 從 ``robots.txt`` 返回 ``Crawl-delay`` 參數的值。如果" +"此參數不存在或不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 存在語" +"法錯誤,則返回 ``None``。" #: ../../library/urllib.robotparser.rst:71 msgid "" @@ -86,6 +102,9 @@ msgid "" "such parameter or it doesn't apply to the *useragent* specified or the " "``robots.txt`` entry for this parameter has invalid syntax, return ``None``." msgstr "" +"以 :term:`named tuple` ``RequestRate(requests, seconds)`` 的形式從 ``robots." +"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在或不適用於指定的 " +"*useragent* ,或是此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。" #: ../../library/urllib.robotparser.rst:81 msgid "" @@ -93,9 +112,11 @@ msgid "" "form of a :func:`list`. If there is no such parameter or the ``robots.txt`` " "entry for this parameter has invalid syntax, return ``None``." msgstr "" +"以 :func:`list` 的形式從 ``robots.txt`` 返回 ``Sitemap`` 參數的內容。如果此參" +"數不存在或此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。" #: ../../library/urllib.robotparser.rst:89 msgid "" "The following example demonstrates basic use of the :class:`RobotFileParser` " "class::" -msgstr "" +msgstr "下面的範例展示了 :class:`RobotFileParser` 類別的基本用法::" From a5dca7e5d05cdfc35811d601523b9a9fb3f34f2e Mon Sep 17 00:00:00 2001 From: linooohon Date: Wed, 26 Jan 2022 23:12:58 +0800 Subject: [PATCH 2/3] fix: fix content based on the suggestion --- library/urllib.robotparser.po | 40 +++++++++++++++++++---------------- 1 file changed, 22 insertions(+), 18 deletions(-) diff --git a/library/urllib.robotparser.po b/library/urllib.robotparser.po index 019caea519..99d6363046 100644 --- a/library/urllib.robotparser.po +++ b/library/urllib.robotparser.po @@ -10,7 +10,7 @@ msgstr "" "Project-Id-Version: Python 3.10\n" "Report-Msgid-Bugs-To: \n" "POT-Creation-Date: 2021-10-26 16:47+0000\n" -"PO-Revision-Date: 2022-01-23 14:36+0800\n" +"PO-Revision-Date: 2022-01-27 13:40+0800\n" "Last-Translator: Phil Lin \n" "Language-Team: Chinese - TAIWAN (https://github.com/python/python-docs-zh-" "tw)\n" @@ -23,7 +23,7 @@ msgstr "" #: ../../library/urllib.robotparser.rst:2 msgid ":mod:`urllib.robotparser` --- Parser for robots.txt" -msgstr ":mod:`urllib.robotparser` --- robots.txt 的解析器" +msgstr ":mod:`urllib.robotparser` --- robots.txt 的剖析器" #: ../../library/urllib.robotparser.rst:10 msgid "**Source code:** :source:`Lib/urllib/robotparser.py`" @@ -37,17 +37,17 @@ msgid "" "on the structure of :file:`robots.txt` files, see http://www.robotstxt.org/" "orig.html." msgstr "" -"此模組提供了一個單獨的類別 :class:`RobotFileParser`, 它可以知道某個特定 user " -"agent (使用者代理)是否能在有發布 :file:`robots.txt` 文件的網站 fetch(擷" -"取)此網站特定的 URL。有關 :file:`robots.txt` 文件結構的更多細節,請參閱 " -"http://www.robotstxt.org/orig.html 。" +"此模組 (module) 提供了一個單獨的類別 (class) \\ :class:`RobotFileParser`\\ ," +"它可以知道某個特定 user agent(使用者代理)是否能在有發布 :file:`robots.txt` " +"文件的網站 fetch(擷取)特定 URL。有關 :file:`robots.txt` 文件結構的更多細" +"節,請參閱 http://www.robotstxt.org/orig.html。" #: ../../library/urllib.robotparser.rst:28 msgid "" "This class provides methods to read, parse and answer questions about the :" "file:`robots.txt` file at *url*." msgstr "" -"此類別提供了一些方法可以讀取、解析和回答關於 *url* 上的 :file:`robots.txt` 文" +"此類別提供了一些方法可以讀取、剖析和回答關於 *url* 上的 :file:`robots.txt` 文" "件的問題。" #: ../../library/urllib.robotparser.rst:33 @@ -56,19 +56,19 @@ msgstr "設置指向 :file:`robots.txt` 文件的 URL。" #: ../../library/urllib.robotparser.rst:37 msgid "Reads the :file:`robots.txt` URL and feeds it to the parser." -msgstr "讀取 :file:`robots.txt` URL 並將其輸入到解析器。" +msgstr "讀取 :file:`robots.txt` URL 並將其輸入到剖析器。" #: ../../library/urllib.robotparser.rst:41 msgid "Parses the lines argument." -msgstr "解析行參數(此參數為 ``robots.txt`` 文件裡的行)。" +msgstr "剖析 lines 引數。" #: ../../library/urllib.robotparser.rst:45 msgid "" "Returns ``True`` if the *useragent* is allowed to fetch the *url* according " "to the rules contained in the parsed :file:`robots.txt` file." msgstr "" -"如果根據被解析的 :file:`robots.txt` 文件中的規則,*useragent* 被允許 fetch " -"*url* 的話,則返回 ``True``。" +"根據從 :file:`robots.txt` 文件中剖析出的規則,如果 *useragent* 被允許 fetch " +"*url* 的話,則回傳 ``True``。" #: ../../library/urllib.robotparser.rst:51 msgid "" @@ -82,7 +82,7 @@ msgstr "" #: ../../library/urllib.robotparser.rst:57 msgid "" "Sets the time the ``robots.txt`` file was last fetched to the current time." -msgstr "將最後一次獲取 ``robots.txt`` 文件的時間設置為當前時間。" +msgstr "將最近一次 fetch ``robots.txt`` 文件的時間設置為當前時間。" #: ../../library/urllib.robotparser.rst:62 msgid "" @@ -92,8 +92,8 @@ msgid "" "parameter has invalid syntax, return ``None``." msgstr "" "針對指定的 *useragent* 從 ``robots.txt`` 返回 ``Crawl-delay`` 參數的值。如果" -"此參數不存在或不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 存在語" -"法錯誤,則返回 ``None``。" +"此參數不存在、不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 中所指" +"的條目含有無效語法,則回傳 ``None``。" #: ../../library/urllib.robotparser.rst:71 msgid "" @@ -103,8 +103,9 @@ msgid "" "``robots.txt`` entry for this parameter has invalid syntax, return ``None``." msgstr "" "以 :term:`named tuple` ``RequestRate(requests, seconds)`` 的形式從 ``robots." -"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在或不適用於指定的 " -"*useragent* ,或是此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。" +"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在、不適用於指定的 " +"*useragent* ,或是此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 " +"``None``。" #: ../../library/urllib.robotparser.rst:81 msgid "" @@ -113,10 +114,13 @@ msgid "" "entry for this parameter has invalid syntax, return ``None``." msgstr "" "以 :func:`list` 的形式從 ``robots.txt`` 返回 ``Sitemap`` 參數的內容。如果此參" -"數不存在或此參數在 ``robots.txt`` 存在語法錯誤,則返回 ``None``。" +"數不存在或此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 ``None``。" #: ../../library/urllib.robotparser.rst:89 msgid "" "The following example demonstrates basic use of the :class:`RobotFileParser` " "class::" -msgstr "下面的範例展示了 :class:`RobotFileParser` 類別的基本用法::" +msgstr "" +"下面的範例展示了 :class:`RobotFileParser` 類別的基本用法:\n" +"\n" +"::" From fbc54f9d232c990b8c9f043f81b29cd5680aa8cb Mon Sep 17 00:00:00 2001 From: "Matt.Wang" Date: Thu, 27 Jan 2022 15:47:58 +0800 Subject: [PATCH 3/3] =?UTF-8?q?fix:=20`=E8=BF=94=E5=9B=9E`=20->=20`?= =?UTF-8?q?=E5=9B=9E=E5=82=B3`?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- library/urllib.robotparser.po | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/library/urllib.robotparser.po b/library/urllib.robotparser.po index 99d6363046..0195c033d9 100644 --- a/library/urllib.robotparser.po +++ b/library/urllib.robotparser.po @@ -76,7 +76,7 @@ msgid "" "for long-running web spiders that need to check for new ``robots.txt`` files " "periodically." msgstr "" -"返回最近一次 fetch ``robots.txt`` 文件的時間。這適用於需要定期檢查 ``robots." +"回傳最近一次 fetch ``robots.txt`` 文件的時間。這適用於需要定期檢查 ``robots." "txt`` 文件更新情況的長時間運行網頁爬蟲。" #: ../../library/urllib.robotparser.rst:57 @@ -91,7 +91,7 @@ msgid "" "apply to the *useragent* specified or the ``robots.txt`` entry for this " "parameter has invalid syntax, return ``None``." msgstr "" -"針對指定的 *useragent* 從 ``robots.txt`` 返回 ``Crawl-delay`` 參數的值。如果" +"針對指定的 *useragent* 從 ``robots.txt`` 回傳 ``Crawl-delay`` 參數的值。如果" "此參數不存在、不適用於指定的 *useragent* ,或是此參數在 ``robots.txt`` 中所指" "的條目含有無效語法,則回傳 ``None``。" @@ -103,7 +103,7 @@ msgid "" "``robots.txt`` entry for this parameter has invalid syntax, return ``None``." msgstr "" "以 :term:`named tuple` ``RequestRate(requests, seconds)`` 的形式從 ``robots." -"txt`` 返回 ``Request-rate`` 參數的內容。如果此參數不存在、不適用於指定的 " +"txt`` 回傳 ``Request-rate`` 參數的內容。如果此參數不存在、不適用於指定的 " "*useragent* ,或是此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 " "``None``。" @@ -113,7 +113,7 @@ msgid "" "form of a :func:`list`. If there is no such parameter or the ``robots.txt`` " "entry for this parameter has invalid syntax, return ``None``." msgstr "" -"以 :func:`list` 的形式從 ``robots.txt`` 返回 ``Sitemap`` 參數的內容。如果此參" +"以 :func:`list` 的形式從 ``robots.txt`` 回傳 ``Sitemap`` 參數的內容。如果此參" "數不存在或此參數在 ``robots.txt`` 中所指的條目含有無效語法,則回傳 ``None``。" #: ../../library/urllib.robotparser.rst:89