{"id":55101,"date":"2025-08-13T16:29:58","date_gmt":"2025-08-13T08:29:58","guid":{"rendered":"https:\/\/www.wsisp.com\/helps\/55101.html"},"modified":"2025-08-13T16:29:58","modified_gmt":"2025-08-13T08:29:58","slug":"%e4%ba%94%e7%ac%94bpe%e5%88%86%e8%af%8d%e5%99%a8%e7%9a%84%e6%8a%80%e6%9c%af%e6%bc%94%e8%bf%9b%e4%b8%8e%e5%85%b3%e9%94%ae%e4%bf%ae%e5%a4%8d","status":"publish","type":"post","link":"https:\/\/www.wsisp.com\/helps\/55101.html","title":{"rendered":"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d"},"content":{"rendered":"<p>\u5728\u4e2d\u6587\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4efb\u52a1\u4e2d&#xff0c;\u4e94\u7b14\u7f16\u7801\u4e0eByte Pair Encoding (BPE)\u7684\u7ed3\u5408\u63d0\u4f9b\u4e86\u4e00\u79cd\u72ec\u7279\u7684\u65b9\u6cd5\u3002\u672c\u6587\u5206\u4eab\u5728\u6784\u5efa\u4e94\u7b14BPE\u5206\u8bcd\u5668\u8fc7\u7a0b\u4e2d\u9047\u5230\u7684\u6280\u672f\u6311\u6218\u53ca\u5176\u89e3\u51b3\u65b9\u6848\u3002<\/p>\n<h3>\u95ee\u9898\u80cc\u666f<\/h3>\n<p>\u4e94\u7b14BPE\u5206\u8bcd\u5668\u65e8\u5728&#xff1a;<\/p>\n<li>\u4f7f\u7528\u4e94\u7b14\u7f16\u7801\u8868\u793a\u4e2d\u6587\u5b57\u7b26<\/li>\n<li>\u5e94\u7528BPE\u7b97\u6cd5\u5904\u7406\u6df7\u5408\u5185\u5bb9&#xff08;\u4e2d\u6587&#043;\u975e\u4e2d\u6587&#xff09;<\/li>\n<li>\u5b9e\u73b0\u9ad8\u6548\u7f16\u7801\/\u89e3\u7801\u6d41\u7a0b<\/li>\n<p>\u4f46\u5728\u5b9e\u73b0\u8fc7\u7a0b\u4e2d&#xff0c;\u53d1\u73b0\u4e86\u4e24\u4e2a\u5173\u952e\u95ee\u9898&#xff1a;<\/p>\n<h4>\u95ee\u9898\u4e00&#xff1a;\u4e34\u65f6\u6587\u4ef6\u6743\u9650\u51b2\u7a81<\/h4>\n<p>\u75c7\u72b6&#xff1a;\u5728Windows\u7cfb\u7edf\u4e0a\u8bad\u7ec3\u540e&#xff0c;\u4e34\u65f6\u6587\u4ef6\u65e0\u6cd5\u6b63\u5e38\u5220\u9664&#xff0c;\u51fa\u73b0\u6743\u9650\u9519\u8bef&#xff1a;<\/p>\n<p>PermissionError: [WinError 32] The process cannot access the file because it is being used by another process<\/p>\n<p>\u6839\u6e90&#xff1a;<\/p>\n<li>\u8bad\u7ec3\u65f6\u540c\u65f6\u6253\u5f00\u591a\u4e2a\u6587\u4ef6\u53e5\u67c4<\/li>\n<li>Windows\u7cfb\u7edf\u5bf9\u6587\u4ef6\u5220\u9664\u7684\u4e25\u683c\u9501\u5b9a<\/li>\n<li>\u4e34\u65f6\u6587\u4ef6\u547d\u540d\u51b2\u7a81\u5bfc\u81f4\u610f\u5916\u8986\u76d6<\/li>\n<h4>\u95ee\u9898\u4e8c&#xff1a;\u89e3\u7801\u903b\u8f91\u7f3a\u9677<\/h4>\n<p>\u75c7\u72b6&#xff1a;\u5bf9\u6df7\u5408\u5185\u5bb9\u89e3\u7801\u65f6\u51fa\u73b0\u9519\u8bef&#xff1a;<\/p>\n<p>&#034;\u4f60\u597dabc&#034; \u2192 \u89e3\u7801\u4e3a &#034;\u4f60bca\u597d&#034; (\u975e\u4e2d\u6587\u5b57\u7b26\u4f4d\u7f6e\u9519\u4e71)<\/p>\n<p>\u6839\u6e90&#xff1a;<\/p>\n<li>\u539f\u59cb\u89e3\u7801\u903b\u8f91\u76f4\u63a5\u62fc\u63a5\u6807\u8bb0\u800c\u672a\u6b63\u786e\u5904\u7406\u5206\u9694\u7b26<\/li>\n<li>\u4e94\u7b14\u7f16\u7801(&#034;\u4e94&#034;\u524d\u7f00)\u4e0e\u975e\u4e2d\u6587(&#034;\u975e&#034;\u524d\u7f00)\u6df7\u5408\u65f6\u8fb9\u754c\u8bc6\u522b\u9519\u8bef<\/li>\n<h3>\u4fee\u590d\u65b9\u6848<\/h3>\n<h4>\u4fee\u590d\u4e00&#xff1a;\u5065\u58ee\u7684\u4e34\u65f6\u6587\u4ef6\u7ba1\u7406<\/h4>\n<p><span class=\"token keyword\">def<\/span> <span class=\"token function\">train<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> corpus_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> vocab_size<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">int<\/span><span class=\"token punctuation\">,<\/span> chunk_size<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">int<\/span> <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">10000<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token comment\"># \u751f\u6210\u552f\u4e00\u6587\u4ef6\u8def\u5f84<\/span><br \/>\n    timestamp <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">(<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">(<\/span>time<span class=\"token punctuation\">.<\/span>time<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    temp_path <span class=\"token operator\">&#061;<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;temp&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string-interpolation\"><span class=\"token string\">f&#034;wubi_bpe_temp_<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>timestamp<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">.txt&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">try<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token comment\"># \u786e\u4fdd\u76ee\u5f55\u5b58\u5728<\/span><br \/>\n        os<span class=\"token punctuation\">.<\/span>makedirs<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;temp&#034;<\/span><span class=\"token punctuation\">,<\/span> exist_ok<span class=\"token operator\">&#061;<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u5904\u7406\u6570\u636e(\u7565)<\/span><br \/>\n        <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;w&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> temp_file<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token comment\"># \u6570\u636e\u5904\u7406\u903b\u8f91<\/span><\/p>\n<p>        <span class=\"token comment\"># \u6267\u884cBPE\u5408\u5e76(\u7565)<\/span><\/p>\n<p>    <span class=\"token keyword\">finally<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token comment\"># \u5b89\u5168\u5220\u9664\u673a\u5236<\/span><br \/>\n        <span class=\"token keyword\">try<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>exists<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                os<span class=\"token punctuation\">.<\/span>remove<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">except<\/span> PermissionError<span class=\"token punctuation\">:<\/span><br \/>\n            time<span class=\"token punctuation\">.<\/span>sleep<span class=\"token punctuation\">(<\/span><span class=\"token number\">0.5<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># \u7b49\u5f85\u8d44\u6e90\u91ca\u653e<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>exists<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                os<span class=\"token punctuation\">.<\/span>remove<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><\/p>\n<p>\u5173\u952e\u6539\u8fdb&#xff1a;<\/p>\n<li>\u65f6\u95f4\u6233\u552f\u4e00\u547d\u540d\u907f\u514d\u51b2\u7a81<\/li>\n<li>try\/finally\u786e\u4fdd\u5220\u9664\u6267\u884c<\/li>\n<li>\u91cd\u8bd5\u673a\u5236\u89e3\u51b3Windows\u9501\u5b9a\u95ee\u9898<\/li>\n<li>\u76ee\u5f55\u9884\u5148\u521b\u5efa\u907f\u514d\u8def\u5f84\u9519\u8bef<\/li>\n<h4>\u4fee\u590d\u4e8c&#xff1a;\u7cbe\u51c6\u7684\u89e3\u7801\u903b\u8f91\u91cd\u6784<\/h4>\n<p><span class=\"token keyword\">def<\/span> <span class=\"token function\">decode<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> ids<span class=\"token punctuation\">:<\/span> List<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span>self<span class=\"token punctuation\">.<\/span>id_to_token<span class=\"token punctuation\">.<\/span>get<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">id<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;&lt;unk&gt;&#034;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">for<\/span> <span class=\"token builtin\">id<\/span> <span class=\"token keyword\">in<\/span> ids<span class=\"token punctuation\">]<\/span><br \/>\n    text <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>tokens<span class=\"token punctuation\">)<\/span><br \/>\n    result <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n    sep <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><br \/>\n    current <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><\/p>\n<p>    <span class=\"token keyword\">for<\/span> one <span class=\"token keyword\">in<\/span> text<span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token keyword\">if<\/span> one <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span> <span class=\"token keyword\">or<\/span> one <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u4e94&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>current<span class=\"token punctuation\">)<\/span><br \/>\n            <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_chinese<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">[<\/span>current<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n            current <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><br \/>\n            sep <span class=\"token operator\">&#061;<\/span> one<br \/>\n        <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            current <span class=\"token operator\">&#043;&#061;<\/span> one<\/p>\n<p>    <span class=\"token comment\"># \u5904\u7406\u672b\u5c3e\u5185\u5bb9<\/span><br \/>\n    <span class=\"token keyword\">if<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u4e94&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_chinese<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">[<\/span>current<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">elif<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>current<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">return<\/span> <span class=\"token string\">&#039;&#039;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>result<span class=\"token punctuation\">)<\/span><\/p>\n<p>\u5173\u952e\u6539\u8fdb&#xff1a;<\/p>\n<li>\u72b6\u6001\u8ddf\u8e2a(sep\u53d8\u91cf)\u8bb0\u5f55\u5f53\u524d\u524d\u7f00\u7c7b\u578b<\/li>\n<li>\u8fb9\u754c\u68c0\u6d4b\u51c6\u786e\u533a\u5206\u4e2d\u6587\/\u975e\u4e2d\u6587\u5185\u5bb9<\/li>\n<li>\u5355\u72ec\u5904\u7406\u672b\u5c3e\u5185\u5bb9\u907f\u514d\u622a\u65ad<\/li>\n<li>\u4f7f\u7528\u4e94\u7b14\u8f6c\u6362\u5668\u6b63\u786e\u8fd8\u539f\u4e2d\u6587<\/li>\n<h3>\u8865\u5145\u6539\u8fdb<\/h3>\n<li>\u589e\u5f3a\u7684\u9519\u8bef\u5904\u7406&#xff1a;<\/li>\n<p><span class=\"token keyword\">if<\/span> file_size <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token number\">0<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token keyword\">raise<\/span> RuntimeError<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;Temporary file is empty&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<li>\u8bad\u7ec3\u8fdb\u5ea6\u53cd\u9988&#xff1a;<\/li>\n<p><span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Processing chunk <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>chunk_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#8230;&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n<span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Highest frequency pair: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_pair<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> (<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">\u6b21)&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<li>\u8fb9\u7f18\u60c5\u51b5\u5904\u7406&#xff1a;<\/li>\n<p><span class=\"token keyword\">if<\/span> <span class=\"token keyword\">not<\/span> pair_freqs<span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;No pairs found. Stopping merge process.&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">break<\/span><\/p>\n<h3>\u6d4b\u8bd5\u9a8c\u8bc1<\/h3>\n<p>\u6d4b\u8bd5\u7528\u4f8b&#xff1a;\u201c\u4f60\u597d&#xff0c;\u7ae0\u8282\u6d4b\u8bd5abc!\u201d \u7f16\u89e3\u7801\u6d41\u7a0b:<\/p>\n<p>\u539f\u59cb\u6587\u672c \u2192 \u7f16\u7801 \u2192 <span class=\"token punctuation\">[<\/span><span class=\"token number\">201<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token number\">42<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token number\">307<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token number\">15<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token number\">89<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token number\">305<\/span><span class=\"token punctuation\">]<\/span> \u2192 \u89e3\u7801 \u2192 \u8fd8\u539f\u6587\u672c<\/p>\n<p>\u6d4b\u8bd5\u7ed3\u679c&#xff1a;<\/p>\n<p>\u6d4b\u8bd5\u6587\u672c: &#039;\u4f60\u597d&#xff0c;\u7ae0\u8282\u6d4b\u8bd5abc!&#039;<br \/>\n\u89e3\u7801\u7ed3\u679c: &#039;\u4f60\u597d&#xff0c;\u7ae0\u8282\u6d4b\u8bd5abc!&#039;<br \/>\n\u91cd\u65b0\u7f16\u7801\u7ed3\u679c: [201, 42, 307, 15, 89, 305]<br \/>\n\u662f\u5426\u4e00\u81f4: True<\/p>\n<h3>\u5e94\u7528\u4ef7\u503c<\/h3>\n<p>\u8fd9\u4e9b\u4fee\u590d\u4f7f\u5206\u8bcd\u5668\u80fd\u591f&#xff1a;<\/p>\n<ul>\n<li>\u5728Windows\/Linux\u7cfb\u7edf\u7a33\u5b9a\u8fd0\u884c<\/li>\n<li>\u6b63\u786e\u5904\u7406\u6df7\u5408\u8bed\u8a00\u6587\u672c<\/li>\n<li>\u9002\u5e94\u751f\u4ea7\u73af\u5883\u9700\u6c42<\/li>\n<li>\u4e3a\u4e2d\u6587NLP\u4efb\u52a1\u63d0\u4f9b\u53ef\u9760\u57fa\u7840<\/li>\n<\/ul>\n<p>\u5b8c\u6574\u5b9e\u73b0\u5df2\u5e94\u7528\u4e8e\u6587\u5b57\u5904\u7406\u7cfb\u7edf\u548c\u8f93\u5165\u6cd5\u5f15\u64ce&#xff0c;\u663e\u8457\u63d0\u5347\u4e86\u4e2d\u6587\u6587\u672c\u7684\u5904\u7406\u6548\u7387\u548c\u51c6\u786e\u5ea6\u3002<\/p>\n<p>\u63a2\u7d22\u66f4\u591a\u4e2d\u6587\u5904\u7406\u6280\u672f&#xff1a;\u5173\u6ce8\u6211\u7684\u535a\u5ba2\u83b7\u53d6\u6700\u65b0\u66f4\u65b0&#xff01;<\/p>\n<p><span class=\"token keyword\">import<\/span> json<br \/>\n<span class=\"token keyword\">import<\/span> re<br \/>\n<span class=\"token keyword\">import<\/span> tempfile<br \/>\n<span class=\"token keyword\">import<\/span> os<br \/>\n<span class=\"token keyword\">import<\/span> time<br \/>\n<span class=\"token keyword\">import<\/span> shutil<br \/>\n<span class=\"token keyword\">from<\/span> collections <span class=\"token keyword\">import<\/span> defaultdict<br \/>\n<span class=\"token keyword\">from<\/span> typing <span class=\"token keyword\">import<\/span> Dict<span class=\"token punctuation\">,<\/span> List<span class=\"token punctuation\">,<\/span> Tuple<span class=\"token punctuation\">,<\/span> Generator<span class=\"token punctuation\">,<\/span> Set<span class=\"token punctuation\">,<\/span> Any<\/p>\n<p><span class=\"token keyword\">class<\/span> <span class=\"token class-name\">WubiConverter<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token keyword\">def<\/span> <span class=\"token function\">__init__<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> wubi_dict_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span> <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;wubi86.json&#034;<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>wubi_dict <span class=\"token operator\">&#061;<\/span> self<span class=\"token punctuation\">.<\/span>_load_wubi_dict<span class=\"token punctuation\">(<\/span>wubi_dict_path<span class=\"token punctuation\">)<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>reverse_wubi_dict <span class=\"token operator\">&#061;<\/span> defaultdict<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">list<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> char<span class=\"token punctuation\">,<\/span> code <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>wubi_dict<span class=\"token punctuation\">.<\/span>items<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token comment\"># \u79fb\u9664\u7f16\u7801\u672b\u5c3e\u6570\u5b57<\/span><\/p>\n<p>            self<span class=\"token punctuation\">.<\/span>reverse_wubi_dict<span class=\"token punctuation\">[<\/span>code<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>char<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">_load_wubi_dict<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> Dict<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token keyword\">try<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;r&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">return<\/span> json<span class=\"token punctuation\">.<\/span>load<span class=\"token punctuation\">(<\/span>f<span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">except<\/span> Exception<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token comment\"># \u786e\u4fdd\u57fa\u7840\u5b57\u5178\u5305\u542b\u6d4b\u8bd5\u5b57\u7b26<\/span><br \/>\n            <span class=\"token keyword\">return<\/span> <span class=\"token punctuation\">{<\/span><br \/>\n                <span class=\"token string\">&#034;\u4e00&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;g&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u7684&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;r&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6211&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;trnt&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u597d&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;vb&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u4f60&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;wq&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;\u7ae0&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ujj&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u8282&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ab&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6d4b&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;imj&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u8bd5&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ya&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u8fd9&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;p&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;\u662f&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;j&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6837&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;sud&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6587&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;yygy&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u672c&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;sg&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u9a8c&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;cwgi&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;\u4e2d&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;k&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6587&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;yygy&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u5904&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;thi&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u7406&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;gj&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u80fd&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ce&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;\u529b&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;lt&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6d4b&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;imj&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u8bd5&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ya&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u4e94&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;gg&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u7b14&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ttfn&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;\u8f93&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;lwg&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u5165&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ty&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u6cd5&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;if&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u5e38&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;ipkh&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;\u89c1&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;mqb&#034;<\/span><br \/>\n            <span class=\"token punctuation\">}<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">convert_to_wubi<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> char<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u8fd4\u56de\u5b57\u7b26\u7684\u4e94\u7b14\u7f16\u7801&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token keyword\">return<\/span> self<span class=\"token punctuation\">.<\/span>wubi_dict<span class=\"token punctuation\">.<\/span>get<span class=\"token punctuation\">(<\/span>char<span class=\"token punctuation\">,<\/span> char<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">convert_to_chinese<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> wubi_codes<span class=\"token punctuation\">:<\/span> List<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u5c06\u4e94\u7b14\u7f16\u7801\u5217\u8868\u8f6c\u6362\u56de\u4e2d\u6587\u5b57\u7b26\u4e32&#034;&#034;&#034;<\/span><br \/>\n        chars <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> code <span class=\"token keyword\">in<\/span> wubi_codes<span class=\"token punctuation\">:<\/span><\/p>\n<p>            candidates <span class=\"token operator\">&#061;<\/span> self<span class=\"token punctuation\">.<\/span>reverse_wubi_dict<span class=\"token punctuation\">.<\/span>get<span class=\"token punctuation\">(<\/span>code<span class=\"token punctuation\">,<\/span> <span class=\"token punctuation\">[<\/span>code<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n            chars<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>candidates<span class=\"token punctuation\">[<\/span><span class=\"token number\">0<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">return<\/span> <span class=\"token string\">&#039;&#039;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span><\/p>\n<p><span class=\"token keyword\">class<\/span> <span class=\"token class-name\">WubiBPETokenizer<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token keyword\">def<\/span> <span class=\"token function\">__init__<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> wubi_converter<span class=\"token punctuation\">:<\/span> WubiConverter<span class=\"token punctuation\">,<\/span> vocab<span class=\"token punctuation\">:<\/span> Dict<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token builtin\">int<\/span><span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">,<\/span> merges<span class=\"token punctuation\">:<\/span> List<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>wubi_converter <span class=\"token operator\">&#061;<\/span> wubi_converter<br \/>\n        self<span class=\"token punctuation\">.<\/span>special_tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">{<\/span><span class=\"token string\">&#034;&lt;unk&gt;&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token number\">0<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;&lt;pad&gt;&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;&lt;s&gt;&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token number\">2<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;&lt;\/s&gt;&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token number\">3<\/span><span class=\"token punctuation\">}<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>vocab <span class=\"token operator\">&#061;<\/span> vocab <span class=\"token keyword\">or<\/span> self<span class=\"token punctuation\">.<\/span>special_tokens<span class=\"token punctuation\">.<\/span>copy<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>id_to_token <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">{<\/span>idx<span class=\"token punctuation\">:<\/span> token <span class=\"token keyword\">for<\/span> token<span class=\"token punctuation\">,<\/span> idx <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">.<\/span>items<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>merges <span class=\"token operator\">&#061;<\/span> merges <span class=\"token keyword\">or<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>merge_dict <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">{<\/span><span class=\"token punctuation\">}<\/span><br \/>\n        <span class=\"token keyword\">if<\/span> merges<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">for<\/span> pair <span class=\"token keyword\">in<\/span> merges<span class=\"token punctuation\">:<\/span><br \/>\n                a<span class=\"token punctuation\">,<\/span> b <span class=\"token operator\">&#061;<\/span> pair<span class=\"token punctuation\">.<\/span>split<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">(<\/span>a<span class=\"token punctuation\">,<\/span> b<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> a <span class=\"token operator\">&#043;<\/span> b<\/p>\n<p>    <span class=\"token decorator annotation punctuation\">&#064;staticmethod<\/span><br \/>\n    <span class=\"token keyword\">def<\/span> <span class=\"token function\">preprocess_stream<\/span><span class=\"token punctuation\">(<\/span>text_stream<span class=\"token punctuation\">:<\/span> Generator<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> Generator<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u6d41\u5f0f\u9884\u5904\u7406&#xff1a;\u8bc6\u522b\u4e2d\u6587\u5b57\u7b26\u5e76\u6dfb\u52a0\u6807\u8bb0&#034;&#034;&#034;<\/span><br \/>\n        current_non_cn <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span>  <span class=\"token comment\"># \u5b58\u653e\u8fde\u7eed\u975e\u4e2d\u6587\u5b57\u7b26<\/span><\/p>\n<p>        <span class=\"token keyword\">for<\/span> char <span class=\"token keyword\">in<\/span> text_stream<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token comment\"># \u4e2d\u6587\u5b57\u7b26\u76f4\u63a5\u8f93\u51fa<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> <span class=\"token string\">&#039;\\\\u4e00&#039;<\/span> <span class=\"token operator\">&lt;&#061;<\/span> char <span class=\"token operator\">&lt;&#061;<\/span> <span class=\"token string\">&#039;\\\\u9fff&#039;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> current_non_cn<span class=\"token punctuation\">:<\/span><br \/>\n                    <span class=\"token keyword\">yield<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span> <span class=\"token operator\">&#043;<\/span> <span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>current_non_cn<span class=\"token punctuation\">)<\/span><br \/>\n                    current_non_cn <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                <span class=\"token keyword\">yield<\/span> char<br \/>\n            <span class=\"token comment\"># \u975e\u4e2d\u6587\u5b57\u7b26\u7f13\u5b58<\/span><br \/>\n            <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token comment\"># \u5904\u7406\u7a7a\u683c\u548c\u6362\u884c\u7b26<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> char<span class=\"token punctuation\">.<\/span>isspace<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    <span class=\"token keyword\">if<\/span> current_non_cn<span class=\"token punctuation\">:<\/span><br \/>\n                        <span class=\"token keyword\">yield<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span> <span class=\"token operator\">&#043;<\/span> <span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>current_non_cn<span class=\"token punctuation\">)<\/span><br \/>\n                        current_non_cn <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                    <span class=\"token keyword\">yield<\/span> char<br \/>\n                <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    current_non_cn<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>char<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u5904\u7406\u672b\u5c3e\u7684\u975e\u4e2d\u6587\u5b57\u7b26<\/span><br \/>\n        <span class=\"token keyword\">if<\/span> current_non_cn<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">yield<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span> <span class=\"token operator\">&#043;<\/span> <span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>current_non_cn<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token decorator annotation punctuation\">&#064;staticmethod<\/span><br \/>\n    <span class=\"token keyword\">def<\/span> <span class=\"token function\">get_pairs<\/span><span class=\"token punctuation\">(<\/span>word<span class=\"token punctuation\">:<\/span> Tuple<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> Generator<span class=\"token punctuation\">[<\/span>Tuple<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token boolean\">None<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u751f\u6210\u76f8\u90bbtoken\u5bf9&#xff08;\u907f\u514d\u4e2d\u95f4\u5217\u8868&#xff09;&#034;&#034;&#034;<\/span><br \/>\n        prev_char <span class=\"token operator\">&#061;<\/span> word<span class=\"token punctuation\">[<\/span><span class=\"token number\">0<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> char <span class=\"token keyword\">in<\/span> word<span class=\"token punctuation\">[<\/span><span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">yield<\/span> <span class=\"token punctuation\">(<\/span>prev_char<span class=\"token punctuation\">,<\/span> char<span class=\"token punctuation\">)<\/span><br \/>\n            prev_char <span class=\"token operator\">&#061;<\/span> char<\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">train<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> corpus_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> vocab_size<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">int<\/span><span class=\"token punctuation\">,<\/span> chunk_size<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">int<\/span> <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">10000<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u4fee\u590d\u7684\u6587\u4ef6\u5199\u5165\u95ee\u9898\u548c\u6743\u9650\u95ee\u9898&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token comment\"># \u786e\u4fdd\u8f93\u51fa\u76ee\u5f55\u5b58\u5728<\/span><br \/>\n        os<span class=\"token punctuation\">.<\/span>makedirs<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;temp&#034;<\/span><span class=\"token punctuation\">,<\/span> exist_ok<span class=\"token operator\">&#061;<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># 1. \u751f\u6210\u552f\u4e00\u7684\u4e34\u65f6\u6587\u4ef6\u8def\u5f84&#xff08;\u89e3\u51b3\u6743\u9650\u95ee\u9898&#xff09;<\/span><br \/>\n        timestamp <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">(<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">(<\/span>time<span class=\"token punctuation\">.<\/span>time<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        temp_path <span class=\"token operator\">&#061;<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;temp&#034;<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string-interpolation\"><span class=\"token string\">f&#034;wubi_bpe_temp_<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>timestamp<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">.txt&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Using temporary file: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>temp_path<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># 2. \u5b89\u5168\u521b\u5efa\u4e34\u65f6\u6587\u4ef6\u5e76\u5199\u5165\u6570\u636e<\/span><br \/>\n        <span class=\"token keyword\">try<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;w&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> temp_file<span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token comment\"># \u5904\u7406\u8bed\u6599\u6587\u4ef6<\/span><br \/>\n                <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>corpus_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;r&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n                    chunk_count <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><br \/>\n                    char_count <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><\/p>\n<p>                    <span class=\"token keyword\">while<\/span> <span class=\"token boolean\">True<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                        chunk <span class=\"token operator\">&#061;<\/span> f<span class=\"token punctuation\">.<\/span>read<span class=\"token punctuation\">(<\/span>chunk_size<span class=\"token punctuation\">)<\/span><br \/>\n                        <span class=\"token keyword\">if<\/span> <span class=\"token keyword\">not<\/span> chunk<span class=\"token punctuation\">:<\/span><br \/>\n                            <span class=\"token keyword\">break<\/span><\/p>\n<p>                        chunk_count <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                        <span class=\"token comment\"># \u6253\u5370\u5f53\u524d\u5904\u7406\u8fdb\u5ea6<\/span><br \/>\n                        <span class=\"token keyword\">if<\/span> chunk_count <span class=\"token operator\">%<\/span> <span class=\"token number\">10<\/span> <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token number\">0<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                            <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Processing chunk <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>chunk_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#8230;&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>                        <span class=\"token comment\"># \u9884\u5904\u7406\u5e76\u8f6c\u6362\u4e2d\u6587<\/span><br \/>\n                        processed <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                        <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>preprocess_stream<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">(<\/span>c <span class=\"token keyword\">for<\/span> c <span class=\"token keyword\">in<\/span> chunk<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                            <span class=\"token keyword\">if<\/span> token <span class=\"token keyword\">and<\/span> <span class=\"token keyword\">not<\/span> token<span class=\"token punctuation\">.<\/span>isspace<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>  <span class=\"token comment\"># \u8df3\u8fc7\u7a7atoken\u548c\u7a7a\u683c<\/span><br \/>\n                                <span class=\"token comment\"># \u4e2d\u6587\u5b57\u7b26\u8f6c\u6362\u4e3a\u4e94\u7b14\u7f16\u7801<\/span><br \/>\n                                <span class=\"token keyword\">if<\/span> <span class=\"token string\">&#039;\\\\u4e00&#039;<\/span> <span class=\"token operator\">&lt;&#061;<\/span> token <span class=\"token operator\">&lt;&#061;<\/span> <span class=\"token string\">&#039;\\\\u9fff&#039;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                                    wubi_code <span class=\"token operator\">&#061;<\/span> self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_wubi<span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><br \/>\n                                    processed<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u4e94<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>wubi_code<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                                <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                                    processed<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><br \/>\n                                char_count <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><\/p>\n<p>                        <span class=\"token comment\"># \u5199\u5165\u5904\u7406\u540e\u7684\u5185\u5bb9<\/span><br \/>\n                        <span class=\"token keyword\">if<\/span> processed<span class=\"token punctuation\">:<\/span><br \/>\n                            temp_file<span class=\"token punctuation\">.<\/span>write<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034; &#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>processed<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#043;<\/span> <span class=\"token string\">&#034;\\\\n&#034;<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>            <span class=\"token comment\"># \u68c0\u67e5\u6587\u4ef6\u662f\u5426\u6210\u529f\u5199\u5165<\/span><br \/>\n            file_size <span class=\"token operator\">&#061;<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>getsize<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><br \/>\n            <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Temporary file created: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>file_size<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> bytes, <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>char_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> characters processed&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>            <span class=\"token keyword\">if<\/span> file_size <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token number\">0<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">raise<\/span> RuntimeError<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;Temporary file is empty. Please check input corpus and preprocessing.&#034;<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>            <span class=\"token comment\"># 3. \u6784\u5efa\u521d\u59cb\u8bcd\u6c47\u8868<\/span><br \/>\n            self<span class=\"token punctuation\">.<\/span>_build_initial_vocab<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><\/p>\n<p>            <span class=\"token comment\"># 4. \u6267\u884cBPE\u5408\u5e76<\/span><br \/>\n            self<span class=\"token punctuation\">.<\/span>_perform_merges<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">,<\/span> vocab_size<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token keyword\">finally<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token comment\"># 5. \u5b89\u5168\u5220\u9664\u4e34\u65f6\u6587\u4ef6&#xff08;\u89e3\u51b3\u6743\u9650\u95ee\u9898&#xff09;<\/span><br \/>\n            <span class=\"token keyword\">try<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>exists<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    os<span class=\"token punctuation\">.<\/span>remove<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><br \/>\n                    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Temporary file removed: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>temp_path<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n            <span class=\"token keyword\">except<\/span> PermissionError <span class=\"token keyword\">as<\/span> pe<span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Warning: Could not remove temporary file &#8211; <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>pe<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token comment\"># \u5728Windows\u4e0a\u7a0d\u540e\u91cd\u8bd5<\/span><br \/>\n                time<span class=\"token punctuation\">.<\/span>sleep<span class=\"token punctuation\">(<\/span><span class=\"token number\">0.5<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> os<span class=\"token punctuation\">.<\/span>path<span class=\"token punctuation\">.<\/span>exists<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    os<span class=\"token punctuation\">.<\/span>remove<span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">_build_initial_vocab<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> temp_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u6784\u5efa\u521d\u59cb\u8bcd\u6c47\u8868&#xff08;\u4fee\u590d\u7a7a\u6587\u4ef6\u95ee\u9898&#xff09;&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;Building initial vocabulary&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        token_counts <span class=\"token operator\">&#061;<\/span> defaultdict<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;r&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n            line_count <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><br \/>\n            <span class=\"token keyword\">for<\/span> line <span class=\"token keyword\">in<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n                line_count <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> line<span class=\"token punctuation\">.<\/span>split<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    token_counts<span class=\"token punctuation\">[<\/span>token<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><\/p>\n<p>        <span class=\"token comment\"># \u786e\u4fdd\u5904\u7406\u4e86\u6570\u636e<\/span><br \/>\n        <span class=\"token keyword\">if<\/span> <span class=\"token keyword\">not<\/span> token_counts<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">raise<\/span> RuntimeError<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;No tokens found in temporary file. Check preprocessing.&#034;<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u521d\u59cb\u5316\u8bcd\u6c47\u8868<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>vocab <span class=\"token operator\">&#061;<\/span> self<span class=\"token punctuation\">.<\/span>special_tokens<span class=\"token punctuation\">.<\/span>copy<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        next_id <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>special_tokens<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u6dfb\u52a0\u9ad8\u9891token<\/span><br \/>\n        sorted_tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">sorted<\/span><span class=\"token punctuation\">(<\/span>token_counts<span class=\"token punctuation\">.<\/span>items<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> key<span class=\"token operator\">&#061;<\/span><span class=\"token keyword\">lambda<\/span> x<span class=\"token punctuation\">:<\/span> x<span class=\"token punctuation\">[<\/span><span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> reverse<span class=\"token operator\">&#061;<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> token<span class=\"token punctuation\">,<\/span> count <span class=\"token keyword\">in<\/span> sorted_tokens<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> token <span class=\"token keyword\">not<\/span> <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">:<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">[<\/span>token<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> next_id<br \/>\n                self<span class=\"token punctuation\">.<\/span>id_to_token<span class=\"token punctuation\">[<\/span>next_id<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> token<br \/>\n                next_id <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><\/p>\n<p>        <span class=\"token comment\"># \u6dfb\u52a0\u57fa\u7840\u5b57\u7b26<\/span><br \/>\n        all_chars <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">set<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>token_counts<span class=\"token punctuation\">.<\/span>keys<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> char <span class=\"token keyword\">in<\/span> <span class=\"token builtin\">sorted<\/span><span class=\"token punctuation\">(<\/span>all_chars<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> char <span class=\"token keyword\">not<\/span> <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">:<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">[<\/span>char<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> next_id<br \/>\n                self<span class=\"token punctuation\">.<\/span>id_to_token<span class=\"token punctuation\">[<\/span>next_id<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> char<br \/>\n                next_id <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><\/p>\n<p>        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Initial vocabulary built with <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> tokens&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">_perform_merges<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> temp_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> vocab_size<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">int<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u6267\u884cBPE\u5408\u5e76&#xff08;\u589e\u52a0\u65e5\u5fd7\u8f93\u51fa&#xff09;&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;Performing BPE merges&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>merges <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        self<span class=\"token punctuation\">.<\/span>merge_dict <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">{<\/span><span class=\"token punctuation\">}<\/span><br \/>\n        next_id <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u8fed\u4ee3\u76f4\u5230\u8fbe\u5230\u76ee\u6807\u8bcd\u6c47\u8868\u5927\u5c0f<\/span><br \/>\n        <span class=\"token keyword\">while<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&lt;<\/span> vocab_size<span class=\"token punctuation\">:<\/span><br \/>\n            pair_freqs <span class=\"token operator\">&#061;<\/span> defaultdict<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">)<\/span><br \/>\n            total_pairs <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><\/p>\n<p>            <span class=\"token comment\"># \u626b\u63cf\u4e34\u65f6\u6587\u4ef6\u7edf\u8ba1\u76f8\u90bb\u5bf9\u9891\u7387<\/span><br \/>\n            <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>temp_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;r&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">for<\/span> line <span class=\"token keyword\">in<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n                    tokens <span class=\"token operator\">&#061;<\/span> line<span class=\"token punctuation\">.<\/span>split<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                    <span class=\"token keyword\">if<\/span> <span class=\"token keyword\">not<\/span> tokens<span class=\"token punctuation\">:<\/span><br \/>\n                        <span class=\"token keyword\">continue<\/span><\/p>\n<p>                    <span class=\"token comment\"># \u5e94\u7528\u73b0\u6709\u5408\u5e76\u89c4\u5219<\/span><br \/>\n                    merged_tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                    <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> tokens<span class=\"token punctuation\">:<\/span><br \/>\n                        chars <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">list<\/span><span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><br \/>\n                        changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">True<\/span><br \/>\n                        <span class=\"token keyword\">while<\/span> changed <span class=\"token keyword\">and<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&gt;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                            changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">False<\/span><br \/>\n                            i <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><br \/>\n                            new_chars <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                            <span class=\"token keyword\">while<\/span> i <span class=\"token operator\">&lt;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                                <span class=\"token keyword\">if<\/span> i <span class=\"token operator\">&lt;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span> <span class=\"token number\">1<\/span> <span class=\"token keyword\">and<\/span> <span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> chars<span class=\"token punctuation\">[<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">:<\/span><br \/>\n                                    new_chars<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> chars<span class=\"token punctuation\">[<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                                    i <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">2<\/span><br \/>\n                                    changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">True<\/span><br \/>\n                                <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                                    new_chars<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                                    i <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                            chars <span class=\"token operator\">&#061;<\/span> new_chars<br \/>\n                        merged_tokens<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span><\/p>\n<p>                    <span class=\"token comment\"># \u7edf\u8ba1\u6bcf\u4e2a\u6807\u8bb0\u5185\u90e8\u7684\u76f8\u90bb\u5bf9<\/span><br \/>\n                    <span class=\"token keyword\">for<\/span> chars <span class=\"token keyword\">in<\/span> merged_tokens<span class=\"token punctuation\">:<\/span><br \/>\n                        <span class=\"token keyword\">for<\/span> i <span class=\"token keyword\">in<\/span> <span class=\"token builtin\">range<\/span><span class=\"token punctuation\">(<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                            pair <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> chars<span class=\"token punctuation\">[<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                            pair_freqs<span class=\"token punctuation\">[<\/span>pair<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                            total_pairs <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><\/p>\n<p>            <span class=\"token comment\"># \u5982\u679c\u6ca1\u6709\u627e\u5230\u4efb\u4f55\u5bf9&#xff0c;\u63d0\u524d\u9000\u51fa<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> <span class=\"token keyword\">not<\/span> pair_freqs<span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;No pairs found. Stopping merge process.&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token keyword\">break<\/span><\/p>\n<p>            <span class=\"token comment\"># \u9009\u62e9\u6700\u9ad8\u9891\u5bf9<\/span><br \/>\n            best_pair<span class=\"token punctuation\">,<\/span> best_count <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">max<\/span><span class=\"token punctuation\">(<\/span>pair_freqs<span class=\"token punctuation\">.<\/span>items<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> key<span class=\"token operator\">&#061;<\/span><span class=\"token keyword\">lambda<\/span> x<span class=\"token punctuation\">:<\/span> x<span class=\"token punctuation\">[<\/span><span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>            <span class=\"token comment\"># \u5982\u679c\u6700\u9ad8\u9891\u5bf9\u53ea\u51fa\u73b0\u4e00\u6b21&#xff0c;\u63d0\u524d\u9000\u51fa<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> best_count <span class=\"token operator\">&lt;&#061;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Highest frequency pair only appears once (<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">). Stopping merge process.&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token keyword\">break<\/span><\/p>\n<p>            new_token <span class=\"token operator\">&#061;<\/span> best_pair<span class=\"token punctuation\">[<\/span><span class=\"token number\">0<\/span><span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#043;<\/span> best_pair<span class=\"token punctuation\">[<\/span><span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><\/p>\n<p>            <span class=\"token comment\"># \u6dfb\u52a0\u5230\u8bcd\u6c47\u8868<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> new_token <span class=\"token keyword\">not<\/span> <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">:<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">[<\/span>new_token<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> next_id<br \/>\n                self<span class=\"token punctuation\">.<\/span>id_to_token<span class=\"token punctuation\">[<\/span>next_id<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> new_token<br \/>\n                next_id <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>merges<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_pair<span class=\"token punctuation\">[<\/span><span class=\"token number\">0<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_pair<span class=\"token punctuation\">[<\/span><span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">[<\/span>best_pair<span class=\"token punctuation\">]<\/span> <span class=\"token operator\">&#061;<\/span> new_token<br \/>\n                <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Merged: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_pair<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> -&gt; &#039;<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>new_token<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#039; (frequency: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_count<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">)&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n            <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Pair <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>best_pair<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\"> already merged, skipping.&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token keyword\">break<\/span><\/p>\n<p>            <span class=\"token comment\"># \u6253\u5370\u8fdb\u5ea6<\/span><br \/>\n            <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Progress: vocab size <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">\/<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>vocab_size<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;Final vocabulary size: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">, merges: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>merges<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">encode<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> text<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> List<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u7f16\u7801\u6587\u672c\u4e3atoken ID\u5217\u8868&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token comment\"># \u9884\u5904\u7406\u548c\u4e2d\u6587\u8f6c\u6362<\/span><br \/>\n        tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>preprocess_stream<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">(<\/span>c <span class=\"token keyword\">for<\/span> c <span class=\"token keyword\">in<\/span> text<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> token <span class=\"token keyword\">and<\/span> <span class=\"token keyword\">not<\/span> token<span class=\"token punctuation\">.<\/span>isspace<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> <span class=\"token string\">&#039;\\\\u4e00&#039;<\/span> <span class=\"token operator\">&lt;&#061;<\/span> token <span class=\"token operator\">&lt;&#061;<\/span> <span class=\"token string\">&#039;\\\\u9fff&#039;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    wubi_code <span class=\"token operator\">&#061;<\/span> self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_wubi<span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><br \/>\n                    tokens<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u4e94<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>wubi_code<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n                <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    tokens<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u5e94\u7528BPE\u5408\u5e76<\/span><br \/>\n        merged_tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> tokens<span class=\"token punctuation\">:<\/span><br \/>\n            chars <span class=\"token operator\">&#061;<\/span> <span class=\"token builtin\">list<\/span><span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">)<\/span><br \/>\n            changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">True<\/span><br \/>\n            <span class=\"token keyword\">while<\/span> changed <span class=\"token keyword\">and<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&gt;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">False<\/span><br \/>\n                i <span class=\"token operator\">&#061;<\/span> <span class=\"token number\">0<\/span><br \/>\n                new_chars <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n                <span class=\"token keyword\">while<\/span> i <span class=\"token operator\">&lt;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    <span class=\"token keyword\">if<\/span> i <span class=\"token operator\">&lt;<\/span> <span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span> <span class=\"token number\">1<\/span> <span class=\"token keyword\">and<\/span> <span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> chars<span class=\"token punctuation\">[<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">in<\/span> self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">:<\/span><br \/>\n                        new_chars<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>merge_dict<span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> chars<span class=\"token punctuation\">[<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                        i <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">2<\/span><br \/>\n                        changed <span class=\"token operator\">&#061;<\/span> <span class=\"token boolean\">True<\/span><br \/>\n                    <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                        new_chars<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">[<\/span>i<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                        i <span class=\"token operator\">&#043;&#061;<\/span> <span class=\"token number\">1<\/span><br \/>\n                chars <span class=\"token operator\">&#061;<\/span> new_chars<br \/>\n            merged_tokens<span class=\"token punctuation\">.<\/span>extend<span class=\"token punctuation\">(<\/span>chars<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token comment\"># \u8f6c\u6362\u4e3aID<\/span><br \/>\n        <span class=\"token keyword\">return<\/span> <span class=\"token punctuation\">[<\/span>self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">.<\/span>get<span class=\"token punctuation\">(<\/span>token<span class=\"token punctuation\">,<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">[<\/span><span class=\"token string\">&#034;&lt;unk&gt;&#034;<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">for<\/span> token <span class=\"token keyword\">in<\/span> merged_tokens<span class=\"token punctuation\">]<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">decode<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> ids<span class=\"token punctuation\">:<\/span> List<span class=\"token punctuation\">[<\/span><span class=\"token builtin\">int<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u4fee\u590d&#xff1a;\u6b63\u786e\u5904\u7406\u5206\u9694\u7b26\u548c\u4e2d\u6587\u8f6c\u6362&#034;&#034;&#034;<\/span><br \/>\n        tokens <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span>self<span class=\"token punctuation\">.<\/span>id_to_token<span class=\"token punctuation\">.<\/span>get<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">id<\/span><span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#034;&lt;unk&gt;&#034;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">for<\/span> <span class=\"token builtin\">id<\/span> <span class=\"token keyword\">in<\/span> ids<span class=\"token punctuation\">]<\/span><br \/>\n        text <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>tokens<span class=\"token punctuation\">)<\/span><br \/>\n        result <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        sep <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><br \/>\n        current <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><br \/>\n        <span class=\"token keyword\">for<\/span> one <span class=\"token keyword\">in<\/span> text<span class=\"token punctuation\">:<\/span><br \/>\n            <span class=\"token keyword\">if<\/span> one <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span> <span class=\"token keyword\">or<\/span> one <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u4e94&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                <span class=\"token keyword\">if<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>current<span class=\"token punctuation\">)<\/span><br \/>\n                    current <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><br \/>\n                <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                    result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_chinese<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">[<\/span>current<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n                    current <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;&#034;<\/span><\/p>\n<p>                sep <span class=\"token operator\">&#061;<\/span> one<br \/>\n            <span class=\"token keyword\">else<\/span><span class=\"token punctuation\">:<\/span><br \/>\n                current <span class=\"token operator\">&#043;&#061;<\/span> one<br \/>\n        <span class=\"token keyword\">if<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u4e94&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">.<\/span>wubi_converter<span class=\"token punctuation\">.<\/span>convert_to_chinese<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">[<\/span>current<span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><br \/>\n        <span class=\"token keyword\">elif<\/span> sep <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;\u975e&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n            result<span class=\"token punctuation\">.<\/span>append<span class=\"token punctuation\">(<\/span>current<span class=\"token punctuation\">)<\/span><\/p>\n<p>        <span class=\"token keyword\">return<\/span> <span class=\"token string\">&#039;&#039;<\/span><span class=\"token punctuation\">.<\/span>join<span class=\"token punctuation\">(<\/span>result<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token keyword\">def<\/span> <span class=\"token function\">save<\/span><span class=\"token punctuation\">(<\/span>self<span class=\"token punctuation\">,<\/span> file_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u4fdd\u5b58\u5206\u8bcd\u5668\u6a21\u578b&#034;&#034;&#034;<\/span><br \/>\n        tokenizer_config <span class=\"token operator\">&#061;<\/span> <span class=\"token punctuation\">{<\/span><br \/>\n            <span class=\"token string\">&#034;model&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token punctuation\">{<\/span><br \/>\n                <span class=\"token string\">&#034;type&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;BPE&#034;<\/span><span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;vocab&#034;<\/span><span class=\"token punctuation\">:<\/span> self<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;merges&#034;<\/span><span class=\"token punctuation\">:<\/span> self<span class=\"token punctuation\">.<\/span>merges<span class=\"token punctuation\">,<\/span><br \/>\n                <span class=\"token string\">&#034;unk_token&#034;<\/span><span class=\"token punctuation\">:<\/span> <span class=\"token string\">&#034;&lt;unk&gt;&#034;<\/span><br \/>\n            <span class=\"token punctuation\">}<\/span><br \/>\n        <span class=\"token punctuation\">}<\/span><br \/>\n        <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>file_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;w&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n            json<span class=\"token punctuation\">.<\/span>dump<span class=\"token punctuation\">(<\/span>tokenizer_config<span class=\"token punctuation\">,<\/span> f<span class=\"token punctuation\">,<\/span> indent<span class=\"token operator\">&#061;<\/span><span class=\"token number\">2<\/span><span class=\"token punctuation\">,<\/span> ensure_ascii<span class=\"token operator\">&#061;<\/span><span class=\"token boolean\">False<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token decorator annotation punctuation\">&#064;classmethod<\/span><br \/>\n    <span class=\"token keyword\">def<\/span> <span class=\"token function\">load<\/span><span class=\"token punctuation\">(<\/span>cls<span class=\"token punctuation\">,<\/span> file_path<span class=\"token punctuation\">:<\/span> <span class=\"token builtin\">str<\/span><span class=\"token punctuation\">,<\/span> wubi_converter<span class=\"token punctuation\">:<\/span> WubiConverter<span class=\"token punctuation\">)<\/span> <span class=\"token operator\">&#8211;<\/span><span class=\"token operator\">&gt;<\/span> <span class=\"token string\">&#039;WubiBPETokenizer&#039;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token triple-quoted-string string\">&#034;&#034;&#034;\u52a0\u8f7d\u5206\u8bcd\u5668\u6a21\u578b&#034;&#034;&#034;<\/span><br \/>\n        <span class=\"token keyword\">with<\/span> <span class=\"token builtin\">open<\/span><span class=\"token punctuation\">(<\/span>file_path<span class=\"token punctuation\">,<\/span> <span class=\"token string\">&#039;r&#039;<\/span><span class=\"token punctuation\">,<\/span> encoding<span class=\"token operator\">&#061;<\/span><span class=\"token string\">&#039;utf-8&#039;<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">as<\/span> f<span class=\"token punctuation\">:<\/span><br \/>\n            config <span class=\"token operator\">&#061;<\/span> json<span class=\"token punctuation\">.<\/span>load<span class=\"token punctuation\">(<\/span>f<span class=\"token punctuation\">)<\/span><br \/>\n        model <span class=\"token operator\">&#061;<\/span> config<span class=\"token punctuation\">[<\/span><span class=\"token string\">&#034;model&#034;<\/span><span class=\"token punctuation\">]<\/span><br \/>\n        <span class=\"token keyword\">return<\/span> cls<span class=\"token punctuation\">(<\/span>wubi_converter<span class=\"token punctuation\">,<\/span> vocab<span class=\"token operator\">&#061;<\/span>model<span class=\"token punctuation\">[<\/span><span class=\"token string\">&#034;vocab&#034;<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">,<\/span> merges<span class=\"token operator\">&#061;<\/span>model<span class=\"token punctuation\">[<\/span><span class=\"token string\">&#034;merges&#034;<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p><span class=\"token comment\"># &#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061; \u66f4\u5065\u58ee\u7684\u6d4b\u8bd5\u7528\u4f8b &#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;&#061;<\/span><br \/>\n<span class=\"token keyword\">if<\/span> __name__ <span class=\"token operator\">&#061;&#061;<\/span> <span class=\"token string\">&#034;__main__&#034;<\/span><span class=\"token punctuation\">:<\/span><br \/>\n    <span class=\"token comment\"># \u786e\u4fdd\u4e34\u65f6\u76ee\u5f55\u5b58\u5728<\/span><br \/>\n    os<span class=\"token punctuation\">.<\/span>makedirs<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;temp&#034;<\/span><span class=\"token punctuation\">,<\/span> exist_ok<span class=\"token operator\">&#061;<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u521b\u5efa\u66f4\u4e30\u5bcc\u7684\u6d4b\u8bd5\u8bed\u6599<\/span><br \/>\n    corpus_path <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;corpus.txt&#034;<\/span><br \/>\n    <span class=\"token comment\"># with open(corpus_path, &#034;w&#034;, encoding&#061;&#034;utf-8&#034;) as f:<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u4f60\u597d&#xff0c;\u8fd9\u662f\u4e00\u4e2a\u6d4b\u8bd5\u6587\u672c\u3002\u7528\u4e8e\u9a8c\u8bc1\u4e94\u7b14BPE\u5206\u8bcd\u5668\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;Hello world! 123 \u6d4b\u8bd5\u6df7\u5408\u5185\u5bb9\u3002\u4e2d\u6587\u5904\u7406\u80fd\u529b\u6d4b\u8bd5\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u4e94\u7b14\u8f93\u5165\u6cd5\u662f\u4e00\u79cd\u5e38\u89c1\u7684\u4e2d\u6587\u8f93\u5165\u65b9\u6cd5\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u6211\u4eec\u9700\u8981\u786e\u4fdd\u8fd9\u4e2a\u8f6c\u6362\u5668\u80fd\u591f\u6b63\u786e\u5904\u7406\u5404\u79cd\u60c5\u51b5\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u66f4\u591a\u6837\u5316\u7684\u5185\u5bb9\u6709\u52a9\u4e8e\u8bad\u7ec3\u66f4\u597d\u7684\u5206\u8bcd\u5668\u6a21\u578b\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u6dfb\u52a0\u66f4\u591a\u53e5\u5b50\u4ee5\u589e\u52a0\u8bed\u6599\u5e93\u7684\u591a\u6837\u6027\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u4e2d\u6587\u5206\u8bcd\u662f\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4e2d\u7684\u91cd\u8981\u4efb\u52a1\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u4e94\u7b14\u7f16\u7801\u53ef\u4ee5\u6709\u6548\u5730\u8868\u793a\u4e2d\u6587\u5b57\u7b26\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;BPE\u7b97\u6cd5\u80fd\u591f\u4ece\u6570\u636e\u4e2d\u5b66\u4e60\u5408\u5e76\u89c4\u5219\u3002\\\\n&#034;)<\/span><br \/>\n    <span class=\"token comment\">#     f.write(&#034;\u8fd9\u4e2a\u5b9e\u73b0\u7ed3\u5408\u4e86\u4e94\u7b14\u7f16\u7801\u548cBPE\u7b97\u6cd5\u7684\u4f18\u70b9\u3002\\\\n&#034;)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u521d\u59cb\u5316\u8f6c\u6362\u5668<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;\u52a0\u8f7d\u4e94\u7b14\u5b57\u5178&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    wubi_conv <span class=\"token operator\">&#061;<\/span> WubiConverter<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u8bad\u7ec3\u5206\u8bcd\u5668<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;\\\\n\u8bad\u7ec3\u5206\u8bcd\u5668&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    tokenizer <span class=\"token operator\">&#061;<\/span> WubiBPETokenizer<span class=\"token punctuation\">(<\/span>wubi_conv<span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u8bbe\u7f6e\u5408\u7406\u7684\u8bcd\u6c47\u8868\u5927\u5c0f<\/span><br \/>\n    tokenizer<span class=\"token punctuation\">.<\/span>train<span class=\"token punctuation\">(<\/span>corpus_path<span class=\"token punctuation\">,<\/span> vocab_size<span class=\"token operator\">&#061;<\/span><span class=\"token number\">5000<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u6253\u5370\u8bad\u7ec3\u7ed3\u679c<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\\\\n\u8bcd\u6c47\u8868\u5927\u5c0f: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>tokenizer<span class=\"token punctuation\">.<\/span>vocab<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u7279\u6b8a\u6807\u8bb0: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>tokenizer<span class=\"token punctuation\">.<\/span>special_tokens<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u5408\u5e76\u89c4\u5219 (<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>tokenizer<span class=\"token punctuation\">.<\/span>merges<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">\u6761):&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">for<\/span> i<span class=\"token punctuation\">,<\/span> merge <span class=\"token keyword\">in<\/span> <span class=\"token builtin\">enumerate<\/span><span class=\"token punctuation\">(<\/span>tokenizer<span class=\"token punctuation\">.<\/span>merges<span class=\"token punctuation\">[<\/span><span class=\"token punctuation\">:<\/span><span class=\"token number\">10<\/span><span class=\"token punctuation\">]<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span><br \/>\n        <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;  <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>i <span class=\"token operator\">&#043;<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">. <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>merge<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u6d4b\u8bd5\u7f16\u7801\u89e3\u7801<\/span><br \/>\n    test_text <span class=\"token operator\">&#061;<\/span> <span class=\"token string\">&#034;\u4f60\u597d&#xff0c;\u7ae0\u8282\u6d4b\u8bd5abc!&#034;<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\\\\n\u6d4b\u8bd5\u6587\u672c: &#039;<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>test_text<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#039;&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u7f16\u7801<\/span><br \/>\n    ids <span class=\"token operator\">&#061;<\/span> tokenizer<span class=\"token punctuation\">.<\/span>encode<span class=\"token punctuation\">(<\/span>test_text<span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u7f16\u7801\u7ed3\u679c (<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span><span class=\"token builtin\">len<\/span><span class=\"token punctuation\">(<\/span>ids<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">\u4e2atoken): <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>ids<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u89e3\u7801<\/span><br \/>\n    decoded <span class=\"token operator\">&#061;<\/span> tokenizer<span class=\"token punctuation\">.<\/span>decode<span class=\"token punctuation\">(<\/span>ids<span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u89e3\u7801\u7ed3\u679c: &#039;<\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>decoded<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#039;&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u4fdd\u5b58\u6a21\u578b<\/span><br \/>\n    tokenizer<span class=\"token punctuation\">.<\/span>save<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;wubi_bpe_tokenizer.json&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;\\\\n\u5206\u8bcd\u5668\u6a21\u578b\u5df2\u4fdd\u5b58&#034;<\/span><span class=\"token punctuation\">)<\/span><\/p>\n<p>    <span class=\"token comment\"># \u52a0\u8f7d\u6a21\u578b\u6d4b\u8bd5<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;\\\\n\u52a0\u8f7d\u4fdd\u5b58\u7684\u5206\u8bcd\u5668&#8230;&#034;<\/span><span class=\"token punctuation\">)<\/span><br \/>\n    loaded_tokenizer <span class=\"token operator\">&#061;<\/span> WubiBPETokenizer<span class=\"token punctuation\">.<\/span>load<span class=\"token punctuation\">(<\/span><span class=\"token string\">&#034;wubi_bpe_tokenizer.json&#034;<\/span><span class=\"token punctuation\">,<\/span> wubi_conv<span class=\"token punctuation\">)<\/span><br \/>\n    reloaded_ids <span class=\"token operator\">&#061;<\/span> loaded_tokenizer<span class=\"token punctuation\">.<\/span>encode<span class=\"token punctuation\">(<\/span>test_text<span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u91cd\u65b0\u7f16\u7801\u7ed3\u679c: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>reloaded_ids<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><br \/>\n    <span class=\"token keyword\">print<\/span><span class=\"token punctuation\">(<\/span><span class=\"token string-interpolation\"><span class=\"token string\">f&#034;\u662f\u5426\u4e00\u81f4: <\/span><span class=\"token interpolation\"><span class=\"token punctuation\">{<\/span>ids <span class=\"token operator\">&#061;&#061;<\/span> reloaded_ids<span class=\"token punctuation\">}<\/span><\/span><span class=\"token string\">&#034;<\/span><\/span><span class=\"token punctuation\">)<\/span><\/p>\n","protected":false},"excerpt":{"rendered":"<p>\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb399\u6b21\uff0c\u70b9\u8d5e6\u6b21\uff0c\u6536\u85cf2\u6b21\u3002\u5728\u4e2d\u6587\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4efb\u52a1\u4e2d\uff0c\u4e94\u7b14\u7f16\u7801\u4e0eByte Pair Encoding (BPE)\u7684\u7ed3\u5408\u63d0\u4f9b\u4e86\u4e00\u79cd\u72ec\u7279\u7684\u65b9\u6cd5\u3002\u672c\u6587\u5206\u4eab\u5728\u6784\u5efa\u4e94\u7b14BPE\u5206\u8bcd\u5668\u8fc7\u7a0b\u4e2d\u9047\u5230\u7684\u6280\u672f\u6311\u6218\u53ca\u5176\u89e3\u51b3\u65b9\u6848\u3002<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[50],"topic":[],"class_list":["post-55101","post","type-post","status-publish","format-standard","hentry","category-server","tag-50"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.wsisp.com\/helps\/55101.html\" \/>\n<meta property=\"og:locale\" content=\"zh_CN\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"og:description\" content=\"\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb399\u6b21\uff0c\u70b9\u8d5e6\u6b21\uff0c\u6536\u85cf2\u6b21\u3002\u5728\u4e2d\u6587\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4efb\u52a1\u4e2d\uff0c\u4e94\u7b14\u7f16\u7801\u4e0eByte Pair Encoding (BPE)\u7684\u7ed3\u5408\u63d0\u4f9b\u4e86\u4e00\u79cd\u72ec\u7279\u7684\u65b9\u6cd5\u3002\u672c\u6587\u5206\u4eab\u5728\u6784\u5efa\u4e94\u7b14BPE\u5206\u8bcd\u5668\u8fc7\u7a0b\u4e2d\u9047\u5230\u7684\u6280\u672f\u6311\u6218\u53ca\u5176\u89e3\u51b3\u65b9\u6848\u3002\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.wsisp.com\/helps\/55101.html\" \/>\n<meta property=\"og:site_name\" content=\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"article:published_time\" content=\"2025-08-13T08:29:58+00:00\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u4f5c\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4\" \/>\n\t<meta name=\"twitter:data2\" content=\"9 \u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/55101.html\",\"url\":\"https:\/\/www.wsisp.com\/helps\/55101.html\",\"name\":\"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"isPartOf\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\"},\"datePublished\":\"2025-08-13T08:29:58+00:00\",\"dateModified\":\"2025-08-13T08:29:58+00:00\",\"author\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\"},\"breadcrumb\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/55101.html#breadcrumb\"},\"inLanguage\":\"zh-Hans\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.wsisp.com\/helps\/55101.html\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/55101.html#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"\u9996\u9875\",\"item\":\"https:\/\/www.wsisp.com\/helps\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\",\"url\":\"https:\/\/www.wsisp.com\/helps\/\",\"name\":\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"description\":\"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"zh-Hans\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"contentUrl\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"caption\":\"admin\"},\"sameAs\":[\"http:\/\/wp.wsisp.com\"],\"url\":\"https:\/\/www.wsisp.com\/helps\/author\/admin\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.wsisp.com\/helps\/55101.html","og_locale":"zh_CN","og_type":"article","og_title":"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","og_description":"\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb399\u6b21\uff0c\u70b9\u8d5e6\u6b21\uff0c\u6536\u85cf2\u6b21\u3002\u5728\u4e2d\u6587\u81ea\u7136\u8bed\u8a00\u5904\u7406\u4efb\u52a1\u4e2d\uff0c\u4e94\u7b14\u7f16\u7801\u4e0eByte Pair Encoding (BPE)\u7684\u7ed3\u5408\u63d0\u4f9b\u4e86\u4e00\u79cd\u72ec\u7279\u7684\u65b9\u6cd5\u3002\u672c\u6587\u5206\u4eab\u5728\u6784\u5efa\u4e94\u7b14BPE\u5206\u8bcd\u5668\u8fc7\u7a0b\u4e2d\u9047\u5230\u7684\u6280\u672f\u6311\u6218\u53ca\u5176\u89e3\u51b3\u65b9\u6848\u3002","og_url":"https:\/\/www.wsisp.com\/helps\/55101.html","og_site_name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","article_published_time":"2025-08-13T08:29:58+00:00","author":"admin","twitter_card":"summary_large_image","twitter_misc":{"\u4f5c\u8005":"admin","\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4":"9 \u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/www.wsisp.com\/helps\/55101.html","url":"https:\/\/www.wsisp.com\/helps\/55101.html","name":"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","isPartOf":{"@id":"https:\/\/www.wsisp.com\/helps\/#website"},"datePublished":"2025-08-13T08:29:58+00:00","dateModified":"2025-08-13T08:29:58+00:00","author":{"@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41"},"breadcrumb":{"@id":"https:\/\/www.wsisp.com\/helps\/55101.html#breadcrumb"},"inLanguage":"zh-Hans","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.wsisp.com\/helps\/55101.html"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/www.wsisp.com\/helps\/55101.html#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"\u9996\u9875","item":"https:\/\/www.wsisp.com\/helps"},{"@type":"ListItem","position":2,"name":"\u4e94\u7b14BPE\u5206\u8bcd\u5668\u7684\u6280\u672f\u6f14\u8fdb\u4e0e\u5173\u952e\u4fee\u590d"}]},{"@type":"WebSite","@id":"https:\/\/www.wsisp.com\/helps\/#website","url":"https:\/\/www.wsisp.com\/helps\/","name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","description":"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"zh-Hans"},{"@type":"Person","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41","name":"admin","image":{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/","url":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","contentUrl":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","caption":"admin"},"sameAs":["http:\/\/wp.wsisp.com"],"url":"https:\/\/www.wsisp.com\/helps\/author\/admin"}]}},"_links":{"self":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/55101","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/comments?post=55101"}],"version-history":[{"count":0,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/55101\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/media?parent=55101"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/categories?post=55101"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/tags?post=55101"},{"taxonomy":"topic","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/topic?post=55101"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}