<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>大模型原理 on Answer</title>
    <link>https://answer.freetools.me/categories/%E5%A4%A7%E6%A8%A1%E5%9E%8B%E5%8E%9F%E7%90%86/</link>
    <description>Recent content in 大模型原理 on Answer</description>
    <generator>Hugo -- 0.152.2</generator>
    <language>zh-cn</language>
    <lastBuildDate>Fri, 13 Mar 2026 01:08:27 +0800</lastBuildDate>
    <atom:link href="https://answer.freetools.me/categories/%E5%A4%A7%E6%A8%A1%E5%9E%8B%E5%8E%9F%E7%90%86/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>Exposure Bias的数学本质：为什么误差会以二次方速度累积</title>
      <link>https://answer.freetools.me/exposure-bias%E7%9A%84%E6%95%B0%E5%AD%A6%E6%9C%AC%E8%B4%A8%E4%B8%BA%E4%BB%80%E4%B9%88%E8%AF%AF%E5%B7%AE%E4%BC%9A%E4%BB%A5%E4%BA%8C%E6%AC%A1%E6%96%B9%E9%80%9F%E5%BA%A6%E7%B4%AF%E7%A7%AF/</link>
      <pubDate>Fri, 13 Mar 2026 01:08:27 +0800</pubDate>
      <guid>https://answer.freetools.me/exposure-bias%E7%9A%84%E6%95%B0%E5%AD%A6%E6%9C%AC%E8%B4%A8%E4%B8%BA%E4%BB%80%E4%B9%88%E8%AF%AF%E5%B7%AE%E4%BC%9A%E4%BB%A5%E4%BA%8C%E6%AC%A1%E6%96%B9%E9%80%9F%E5%BA%A6%E7%B4%AF%E7%A7%AF/</guid>
      <description>Exposure Bias的数学本质：为什么误差会以二次方速度累积</description>
    </item>
    <item>
      <title>为什么大模型每次回答都不一样：从温度参数到批次不变性的完整技术解析</title>
      <link>https://answer.freetools.me/%E4%B8%BA%E4%BB%80%E4%B9%88%E5%A4%A7%E6%A8%A1%E5%9E%8B%E6%AF%8F%E6%AC%A1%E5%9B%9E%E7%AD%94%E9%83%BD%E4%B8%8D%E4%B8%80%E6%A0%B7%E4%BB%8E%E6%B8%A9%E5%BA%A6%E5%8F%82%E6%95%B0%E5%88%B0%E6%89%B9%E6%AC%A1%E4%B8%8D%E5%8F%98%E6%80%A7%E7%9A%84%E5%AE%8C%E6%95%B4%E6%8A%80%E6%9C%AF%E8%A7%A3%E6%9E%90/</link>
      <pubDate>Thu, 12 Mar 2026 14:00:43 +0800</pubDate>
      <guid>https://answer.freetools.me/%E4%B8%BA%E4%BB%80%E4%B9%88%E5%A4%A7%E6%A8%A1%E5%9E%8B%E6%AF%8F%E6%AC%A1%E5%9B%9E%E7%AD%94%E9%83%BD%E4%B8%8D%E4%B8%80%E6%A0%B7%E4%BB%8E%E6%B8%A9%E5%BA%A6%E5%8F%82%E6%95%B0%E5%88%B0%E6%89%B9%E6%AC%A1%E4%B8%8D%E5%8F%98%E6%80%A7%E7%9A%84%E5%AE%8C%E6%95%B4%E6%8A%80%E6%9C%AF%E8%A7%A3%E6%9E%90/</guid>
      <description>深入分析LLM输出随机性的技术根源，从温度参数的数学原理到batch invariance这一被忽视的真正原因，以及如何在生产环境中实现可复现输出</description>
    </item>
    <item>
      <title>FFN如何成为大模型的&#34;知识仓库&#34;：从键值存储到知识编辑的技术真相</title>
      <link>https://answer.freetools.me/ffn%E5%A6%82%E4%BD%95%E6%88%90%E4%B8%BA%E5%A4%A7%E6%A8%A1%E5%9E%8B%E7%9A%84%E7%9F%A5%E8%AF%86%E4%BB%93%E5%BA%93%E4%BB%8E%E9%94%AE%E5%80%BC%E5%AD%98%E5%82%A8%E5%88%B0%E7%9F%A5%E8%AF%86%E7%BC%96%E8%BE%91%E7%9A%84%E6%8A%80%E6%9C%AF%E7%9C%9F%E7%9B%B8/</link>
      <pubDate>Thu, 12 Mar 2026 04:20:22 +0800</pubDate>
      <guid>https://answer.freetools.me/ffn%E5%A6%82%E4%BD%95%E6%88%90%E4%B8%BA%E5%A4%A7%E6%A8%A1%E5%9E%8B%E7%9A%84%E7%9F%A5%E8%AF%86%E4%BB%93%E5%BA%93%E4%BB%8E%E9%94%AE%E5%80%BC%E5%AD%98%E5%82%A8%E5%88%B0%E7%9F%A5%E8%AF%86%E7%BC%96%E8%BE%91%E7%9A%84%E6%8A%80%E6%9C%AF%E7%9C%9F%E7%9B%B8/</guid>
      <description>FFN如何成为大模型的&amp;#34;知识仓库&amp;#34;：从键值存储到知识编辑的技术真相</description>
    </item>
    <item>
      <title>大模型解码策略全景解析：从贪婪搜索到动态阈值采样的二十年演进</title>
      <link>https://answer.freetools.me/%E5%A4%A7%E6%A8%A1%E5%9E%8B%E8%A7%A3%E7%A0%81%E7%AD%96%E7%95%A5%E5%85%A8%E6%99%AF%E8%A7%A3%E6%9E%90%E4%BB%8E%E8%B4%AA%E5%A9%AA%E6%90%9C%E7%B4%A2%E5%88%B0%E5%8A%A8%E6%80%81%E9%98%88%E5%80%BC%E9%87%87%E6%A0%B7%E7%9A%84%E4%BA%8C%E5%8D%81%E5%B9%B4%E6%BC%94%E8%BF%9B/</link>
      <pubDate>Wed, 11 Mar 2026 23:31:16 +0800</pubDate>
      <guid>https://answer.freetools.me/%E5%A4%A7%E6%A8%A1%E5%9E%8B%E8%A7%A3%E7%A0%81%E7%AD%96%E7%95%A5%E5%85%A8%E6%99%AF%E8%A7%A3%E6%9E%90%E4%BB%8E%E8%B4%AA%E5%A9%AA%E6%90%9C%E7%B4%A2%E5%88%B0%E5%8A%A8%E6%80%81%E9%98%88%E5%80%BC%E9%87%87%E6%A0%B7%E7%9A%84%E4%BA%8C%E5%8D%81%E5%B9%B4%E6%BC%94%E8%BF%9B/</guid>
      <description>深入解析大语言模型文本生成的核心机制——解码策略。从基础的贪婪解码、束搜索，到温度采样、Top-k、Top-p，再到前沿的Min-p、Top-n-sigma等动态阈值方法，全面梳理这一领域二十年的技术演进，并深入探讨ICLR 2025热门论文Min-p采样的学术争议。</description>
    </item>
    <item>
      <title>为什么大模型连简单的加法都会算错？从分词陷阱到算术推理的技术困境</title>
      <link>https://answer.freetools.me/%E4%B8%BA%E4%BB%80%E4%B9%88%E5%A4%A7%E6%A8%A1%E5%9E%8B%E8%BF%9E%E7%AE%80%E5%8D%95%E7%9A%84%E5%8A%A0%E6%B3%95%E9%83%BD%E4%BC%9A%E7%AE%97%E9%94%99%E4%BB%8E%E5%88%86%E8%AF%8D%E9%99%B7%E9%98%B1%E5%88%B0%E7%AE%97%E6%9C%AF%E6%8E%A8%E7%90%86%E7%9A%84%E6%8A%80%E6%9C%AF%E5%9B%B0%E5%A2%83/</link>
      <pubDate>Mon, 09 Mar 2026 02:47:33 +0800</pubDate>
      <guid>https://answer.freetools.me/%E4%B8%BA%E4%BB%80%E4%B9%88%E5%A4%A7%E6%A8%A1%E5%9E%8B%E8%BF%9E%E7%AE%80%E5%8D%95%E7%9A%84%E5%8A%A0%E6%B3%95%E9%83%BD%E4%BC%9A%E7%AE%97%E9%94%99%E4%BB%8E%E5%88%86%E8%AF%8D%E9%99%B7%E9%98%B1%E5%88%B0%E7%AE%97%E6%9C%AF%E6%8E%A8%E7%90%86%E7%9A%84%E6%8A%80%E6%9C%AF%E5%9B%B0%E5%A2%83/</guid>
      <description>为什么大模型连简单的加法都会算错？从分词陷阱到算术推理的技术困境</description>
    </item>
  </channel>
</rss>
