<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>正则化 on Answer</title>
    <link>https://answer.freetools.me/tags/%E6%AD%A3%E5%88%99%E5%8C%96/</link>
    <description>Recent content in 正则化 on Answer</description>
    <generator>Hugo -- 0.152.2</generator>
    <language>zh-cn</language>
    <lastBuildDate>Thu, 12 Mar 2026 16:31:58 +0800</lastBuildDate>
    <atom:link href="https://answer.freetools.me/tags/%E6%AD%A3%E5%88%99%E5%8C%96/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>数据增强技术：为何简单的变换能显著提升模型泛化能力</title>
      <link>https://answer.freetools.me/%E6%95%B0%E6%8D%AE%E5%A2%9E%E5%BC%BA%E6%8A%80%E6%9C%AF%E4%B8%BA%E4%BD%95%E7%AE%80%E5%8D%95%E7%9A%84%E5%8F%98%E6%8D%A2%E8%83%BD%E6%98%BE%E8%91%97%E6%8F%90%E5%8D%87%E6%A8%A1%E5%9E%8B%E6%B3%9B%E5%8C%96%E8%83%BD%E5%8A%9B/</link>
      <pubDate>Thu, 12 Mar 2026 16:31:58 +0800</pubDate>
      <guid>https://answer.freetools.me/%E6%95%B0%E6%8D%AE%E5%A2%9E%E5%BC%BA%E6%8A%80%E6%9C%AF%E4%B8%BA%E4%BD%95%E7%AE%80%E5%8D%95%E7%9A%84%E5%8F%98%E6%8D%A2%E8%83%BD%E6%98%BE%E8%91%97%E6%8F%90%E5%8D%87%E6%A8%A1%E5%9E%8B%E6%B3%9B%E5%8C%96%E8%83%BD%E5%8A%9B/</guid>
      <description>深入解析数据增强技术的理论原理、核心方法和实践指南，从图像到文本再到音频，全面覆盖深度学习中最重要的正则化技术之一。</description>
    </item>
    <item>
      <title>过拟合、欠拟合与偏差-方差权衡：机器学习最核心困境的完整解析</title>
      <link>https://answer.freetools.me/%E8%BF%87%E6%8B%9F%E5%90%88%E6%AC%A0%E6%8B%9F%E5%90%88%E4%B8%8E%E5%81%8F%E5%B7%AE-%E6%96%B9%E5%B7%AE%E6%9D%83%E8%A1%A1%E6%9C%BA%E5%99%A8%E5%AD%A6%E4%B9%A0%E6%9C%80%E6%A0%B8%E5%BF%83%E5%9B%B0%E5%A2%83%E7%9A%84%E5%AE%8C%E6%95%B4%E8%A7%A3%E6%9E%90/</link>
      <pubDate>Thu, 12 Mar 2026 10:28:48 +0800</pubDate>
      <guid>https://answer.freetools.me/%E8%BF%87%E6%8B%9F%E5%90%88%E6%AC%A0%E6%8B%9F%E5%90%88%E4%B8%8E%E5%81%8F%E5%B7%AE-%E6%96%B9%E5%B7%AE%E6%9D%83%E8%A1%A1%E6%9C%BA%E5%99%A8%E5%AD%A6%E4%B9%A0%E6%9C%80%E6%A0%B8%E5%BF%83%E5%9B%B0%E5%A2%83%E7%9A%84%E5%AE%8C%E6%95%B4%E8%A7%A3%E6%9E%90/</guid>
      <description>从偏差-方差分解的数学推导到双下降现象的现代理解，深入解析过拟合与欠拟合的本质、诊断方法与缓解策略</description>
    </item>
    <item>
      <title>Early Stopping：为什么&#34;提前终止&#34;能拯救你的模型免于过拟合</title>
      <link>https://answer.freetools.me/early-stopping%E4%B8%BA%E4%BB%80%E4%B9%88%E6%8F%90%E5%89%8D%E7%BB%88%E6%AD%A2%E8%83%BD%E6%8B%AF%E6%95%91%E4%BD%A0%E7%9A%84%E6%A8%A1%E5%9E%8B%E5%85%8D%E4%BA%8E%E8%BF%87%E6%8B%9F%E5%90%88/</link>
      <pubDate>Thu, 12 Mar 2026 06:43:40 +0800</pubDate>
      <guid>https://answer.freetools.me/early-stopping%E4%B8%BA%E4%BB%80%E4%B9%88%E6%8F%90%E5%89%8D%E7%BB%88%E6%AD%A2%E8%83%BD%E6%8B%AF%E6%95%91%E4%BD%A0%E7%9A%84%E6%A8%A1%E5%9E%8B%E5%85%8D%E4%BA%8E%E8%BF%87%E6%8B%9F%E5%90%88/</guid>
      <description>从Prechelt的经典停止准则到LLM微调的实践指南，深入解析早停法如何通过监控验证集性能在模型学习到噪声之前及时终止训练，揭示其与L2正则化的理论等价性以及在现代大模型微调中的应用。</description>
    </item>
    <item>
      <title>标签平滑的默认值为何是0.1：从训练稳定性到收敛理论的数学解析</title>
      <link>https://answer.freetools.me/%E6%A0%87%E7%AD%BE%E5%B9%B3%E6%BB%91%E7%9A%84%E9%BB%98%E8%AE%A4%E5%80%BC%E4%B8%BA%E4%BD%95%E6%98%AF0.1%E4%BB%8E%E8%AE%AD%E7%BB%83%E7%A8%B3%E5%AE%9A%E6%80%A7%E5%88%B0%E6%94%B6%E6%95%9B%E7%90%86%E8%AE%BA%E7%9A%84%E6%95%B0%E5%AD%A6%E8%A7%A3%E6%9E%90/</link>
      <pubDate>Thu, 12 Mar 2026 00:08:10 +0800</pubDate>
      <guid>https://answer.freetools.me/%E6%A0%87%E7%AD%BE%E5%B9%B3%E6%BB%91%E7%9A%84%E9%BB%98%E8%AE%A4%E5%80%BC%E4%B8%BA%E4%BD%95%E6%98%AF0.1%E4%BB%8E%E8%AE%AD%E7%BB%83%E7%A8%B3%E5%AE%9A%E6%80%A7%E5%88%B0%E6%94%B6%E6%95%9B%E7%90%86%E8%AE%BA%E7%9A%84%E6%95%B0%E5%AD%A6%E8%A7%A3%E6%9E%90/</guid>
      <description>深度解析标签平滑技术：为何ε=0.1成为默认值？从Szegedy的Inception到Transformer的训练技巧，揭示其正则化机制、模型校准改进、与知识蒸馏的复杂关系，以及在噪声标签处理中的意外效果。</description>
    </item>
    <item>
      <title>Dropout机制：为什么随机丢弃神经元反而能提升泛化能力</title>
      <link>https://answer.freetools.me/dropout%E6%9C%BA%E5%88%B6%E4%B8%BA%E4%BB%80%E4%B9%88%E9%9A%8F%E6%9C%BA%E4%B8%A2%E5%BC%83%E7%A5%9E%E7%BB%8F%E5%85%83%E5%8F%8D%E8%80%8C%E8%83%BD%E6%8F%90%E5%8D%87%E6%B3%9B%E5%8C%96%E8%83%BD%E5%8A%9B/</link>
      <pubDate>Wed, 11 Mar 2026 21:31:43 +0800</pubDate>
      <guid>https://answer.freetools.me/dropout%E6%9C%BA%E5%88%B6%E4%B8%BA%E4%BB%80%E4%B9%88%E9%9A%8F%E6%9C%BA%E4%B8%A2%E5%BC%83%E7%A5%9E%E7%BB%8F%E5%85%83%E5%8F%8D%E8%80%8C%E8%83%BD%E6%8F%90%E5%8D%87%E6%B3%9B%E5%8C%96%E8%83%BD%E5%8A%9B/</guid>
      <description>深入解析Dropout正则化技术的核心原理：从神经元共适应问题到集成学习视角，从贝叶斯推断到Transformer中的实际应用，揭示这个看似简单却深刻影响深度学习的技术本质。</description>
    </item>
  </channel>
</rss>
