<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>分词 on Answer</title>
    <link>https://answer.freetools.me/tags/%E5%88%86%E8%AF%8D/</link>
    <description>Recent content in 分词 on Answer</description>
    <generator>Hugo -- 0.152.2</generator>
    <language>zh-cn</language>
    <lastBuildDate>Thu, 12 Mar 2026 04:10:51 +0800</lastBuildDate>
    <atom:link href="https://answer.freetools.me/tags/%E5%88%86%E8%AF%8D/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>从输入文本到输出：大模型推理的完整流程解析</title>
      <link>https://answer.freetools.me/%E4%BB%8E%E8%BE%93%E5%85%A5%E6%96%87%E6%9C%AC%E5%88%B0%E8%BE%93%E5%87%BA%E5%A4%A7%E6%A8%A1%E5%9E%8B%E6%8E%A8%E7%90%86%E7%9A%84%E5%AE%8C%E6%95%B4%E6%B5%81%E7%A8%8B%E8%A7%A3%E6%9E%90/</link>
      <pubDate>Thu, 12 Mar 2026 04:10:51 +0800</pubDate>
      <guid>https://answer.freetools.me/%E4%BB%8E%E8%BE%93%E5%85%A5%E6%96%87%E6%9C%AC%E5%88%B0%E8%BE%93%E5%87%BA%E5%A4%A7%E6%A8%A1%E5%9E%8B%E6%8E%A8%E7%90%86%E7%9A%84%E5%AE%8C%E6%95%B4%E6%B5%81%E7%A8%8B%E8%A7%A3%E6%9E%90/</guid>
      <description>深入解析大语言模型推理的完整技术链路，从分词、嵌入、位置编码、注意力计算到自回归生成，揭示模型如何将输入文本转化为输出响应的每一步。</description>
    </item>
    <item>
      <title>Tokenizer 如何塑造大语言模型的世界观：从 BPE 到 Byte Latent Transformer 的三十年技术演进</title>
      <link>https://answer.freetools.me/tokenizer-%E5%A6%82%E4%BD%95%E5%A1%91%E9%80%A0%E5%A4%A7%E8%AF%AD%E8%A8%80%E6%A8%A1%E5%9E%8B%E7%9A%84%E4%B8%96%E7%95%8C%E8%A7%82%E4%BB%8E-bpe-%E5%88%B0-byte-latent-transformer-%E7%9A%84%E4%B8%89%E5%8D%81%E5%B9%B4%E6%8A%80%E6%9C%AF%E6%BC%94%E8%BF%9B/</link>
      <pubDate>Wed, 11 Mar 2026 14:01:29 +0800</pubDate>
      <guid>https://answer.freetools.me/tokenizer-%E5%A6%82%E4%BD%95%E5%A1%91%E9%80%A0%E5%A4%A7%E8%AF%AD%E8%A8%80%E6%A8%A1%E5%9E%8B%E7%9A%84%E4%B8%96%E7%95%8C%E8%A7%82%E4%BB%8E-bpe-%E5%88%B0-byte-latent-transformer-%E7%9A%84%E4%B8%89%E5%8D%81%E5%B9%B4%E6%8A%80%E6%9C%AF%E6%BC%94%E8%BF%9B/</guid>
      <description>深入解析大语言模型 Tokenizer 的工作原理：从 BPE、WordPiece 到 Unigram 三种主流算法的技术差异，到 tokenization 对算术推理、多语言处理、字符级任务的深层影响，以及 Byte Latent Transformer 等无 tokenizer 架构的未来探索。</description>
    </item>
  </channel>
</rss>
