<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:media="http://search.yahoo.com/mrss/"><channel><title>AI on k4i's blog</title><link>https://k4i.top/zh/categories/ai/</link><description>Recent content in AI on k4i's blog</description><generator>Hugo -- gohugo.io</generator><language>zh</language><managingEditor>sky_io@outlook.com (K4i)</managingEditor><webMaster>sky_io@outlook.com (K4i)</webMaster><copyright>All content is subject to the license of &lt;a rel="license noopener" href="https://creativecommons.org/licenses/by-nc-sa/4.0/" target="_blank"&gt;CC BY-NC-SA 4.0&lt;/a&gt; .</copyright><lastBuildDate>Mon, 16 Feb 2026 23:30:00 +0800</lastBuildDate><atom:link href="https://k4i.top/zh/categories/ai/index.xml" rel="self" type="application/rss+xml"/><item><title>批量梯度下降与随机梯度下降</title><link>https://k4i.top/zh/posts/batch-vs-stochastic-gradient-descent/</link><pubDate>Mon, 16 Feb 2026 23:30:00 +0800</pubDate><author>sky_io@outlook.com (K4i)</author><atom:modified>Sun, 19 Apr 2026 23:18:51 +0800</atom:modified><guid>https://k4i.top/zh/posts/batch-vs-stochastic-gradient-descent/</guid><description>&lt;h2 id="introduction"&gt;引言&lt;/h2&gt;
&lt;p&gt;从上一篇前向传播与反向传播的文章中，我们已知梯度下降的更新公式：&lt;/p&gt;
&lt;p&gt;\[&lt;br /&gt;
\theta \leftarrow \theta - \eta \nabla C&lt;br /&gt;
\]&lt;/p&gt;
&lt;p&gt;其中 \(\nabla C\) 是&lt;strong&gt;整个训练数据集&lt;/strong&gt;上的损失梯度。但计算 \(\nabla C\) 意味着要对每一个训练样本运行前向传播和反向传播，然后取平均。当数据集有数百万样本时，这是极其昂贵的。&lt;/p&gt;</description><dc:creator>K4i</dc:creator><media:content url="https://k4i.top//images/icons/gradient-descent.png" medium="image"><media:title type="html">featured image</media:title></media:content><category>gradient-descent</category><category>AI</category></item><item><title>前向传播与反向传播</title><link>https://k4i.top/zh/posts/forward-and-backward-propagation/</link><pubDate>Mon, 16 Feb 2026 23:26:00 +0800</pubDate><author>sky_io@outlook.com (K4i)</author><atom:modified>Sun, 19 Apr 2026 23:18:51 +0800</atom:modified><guid>https://k4i.top/zh/posts/forward-and-backward-propagation/</guid><description>&lt;h2 id="deep-neural-network"&gt;深度神经网络&lt;/h2&gt;
&lt;figure&gt;&lt;img src="https://k4i.top/images/posts/forward-backward-propagation/neural_network_example.svg"
alt="图 1： 神经网络示例" width="70%"&gt;&lt;figcaption&gt;
&lt;p&gt;&lt;span class="figure-number"&gt;图 1： &lt;/span&gt;神经网络示例&lt;/p&gt;
&lt;/figcaption&gt;
&lt;/figure&gt;
&lt;p&gt;每个层可以用数学形式描述如下，从输入层逐层计算到最终输出层的过程称为&lt;strong&gt;前向传播（forward propagation）&lt;/strong&gt;：&lt;/p&gt;</description><dc:creator>K4i</dc:creator><media:content url="https://k4i.top//images/icons/gradient-descent.png" medium="image"><media:title type="html">featured image</media:title></media:content><category>gradient-descent</category><category>AI</category></item></channel></rss>