<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/">
  <channel>
    <title>Zhiyuan Li</title>
    <link>https://zhiyuan1i.github.io/en/</link>
    <description>Recent content on Zhiyuan Li</description>
    <image>
      <title>Zhiyuan Li</title>
      <url>https://zhiyuan1i.github.io/images/papermod-cover.png</url>
      <link>https://zhiyuan1i.github.io/images/papermod-cover.png</link>
    </image>
    <generator>Hugo -- 0.155.3</generator>
    <language>en-US</language>
    <lastBuildDate>Sat, 21 Feb 2026 10:44:23 +0000</lastBuildDate>
    <atom:link href="https://zhiyuan1i.github.io/en/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>The Mathematics of DPLR (Diagonal Plus Low Rank): Parallel Computing with Explicit Transition Matrices</title>
      <link>https://zhiyuan1i.github.io/en/posts/dplr-mathematics/</link>
      <pubDate>Sat, 21 Feb 2026 10:44:23 +0000</pubDate>
      <guid>https://zhiyuan1i.github.io/en/posts/dplr-mathematics/</guid>
      <description>A deep dive into the chunk-wise parallel algorithm for DPLR, understanding the WY representation of explicit diagonal-plus-low-rank transition matrices, and exploring the unified framework with KDA/IPLR</description>
    </item>
    <item>
      <title>KDA (Kimi Delta Attention): From Matrix Multiplication to Affine Transformation</title>
      <link>https://zhiyuan1i.github.io/en/posts/kda-mathematics/</link>
      <pubDate>Tue, 17 Feb 2026 03:00:00 +0000</pubDate>
      <guid>https://zhiyuan1i.github.io/en/posts/kda-mathematics/</guid>
      <description>A deep dive into the chunk-wise parallel algorithm of KDA, establishing the theoretical framework of Affine transformations from basic matrix multiplication lemmas</description>
    </item>
    <item>
      <title>About Me</title>
      <link>https://zhiyuan1i.github.io/en/about/</link>
      <pubDate>Mon, 16 Feb 2026 00:00:00 +0000</pubDate>
      <guid>https://zhiyuan1i.github.io/en/about/</guid>
      <description>&lt;h2 id=&#34;zhiyuan-li&#34;&gt;Zhiyuan Li&lt;/h2&gt;
&lt;p&gt;AI Infra Engineer at &lt;a href=&#34;https://www.moonshot.cn/&#34;&gt;Moonshot AI&lt;/a&gt;.&lt;/p&gt;
&lt;p&gt;Focusing on efficient implementation and optimization of &lt;strong&gt;Linear Attention&lt;/strong&gt;. Honored to have contributed to the development of &lt;a href=&#34;https://github.com/MoonshotAI/Kimi-Linear&#34;&gt;Kimi Linear&lt;/a&gt; and &lt;strong&gt;Kimi Delta Attention (KDA)&lt;/strong&gt;, learning a lot from the excellent colleagues on the team.&lt;/p&gt;
&lt;hr&gt;
&lt;h3 id=&#34;-research-interests&#34;&gt;🔬 Research Interests&lt;/h3&gt;
&lt;ul&gt;
&lt;li&gt;&lt;strong&gt;Linear Attention&lt;/strong&gt;: Exploring sub-quadratic sequence modeling methods for more efficient long sequences&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Efficient Inference Optimization&lt;/strong&gt;: CUDA kernel optimization, memory bandwidth optimization, Tensor Core acceleration&lt;/li&gt;
&lt;li&gt;&lt;strong&gt;Model Architectures&lt;/strong&gt;: RWKV-6/7, Gated DeltaNet, and other novel attention mechanisms&lt;/li&gt;
&lt;/ul&gt;
&lt;h3 id=&#34;-open-source-contributions&#34;&gt;🚀 Open Source Contributions&lt;/h3&gt;
&lt;ul&gt;
&lt;li&gt;Contributed to &lt;a href=&#34;https://github.com/fla-org/flash-linear-attention&#34;&gt;&lt;strong&gt;flash-linear-attention&lt;/strong&gt;&lt;/a&gt; community project - Efficient implementations of state-of-the-art linear attention models&lt;/li&gt;
&lt;/ul&gt;
&lt;h3 id=&#34;-articles&#34;&gt;📝 Articles&lt;/h3&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://zhuanlan.zhihu.com/p/1989809041849988324&#34;&gt;Learning KDA from Scratch - Part 1&lt;/a&gt; - Understanding KDA parallelization from an Infra perspective (Chinese)&lt;/li&gt;
&lt;/ul&gt;
&lt;hr&gt;
&lt;h3 id=&#34;-about-this-site&#34;&gt;💬 About This Site&lt;/h3&gt;
&lt;p&gt;This site documents my learning notes, technical articles, and some immature thoughts in the AI Infra field. I&amp;rsquo;m still learning, so please feel free to point out any mistakes. Looking forward to exchanging ideas with you.&lt;/p&gt;</description>
    </item>
    <item>
      <title>Tech Stack</title>
      <link>https://zhiyuan1i.github.io/en/posts/tech-stack/</link>
      <pubDate>Mon, 16 Feb 2026 00:00:00 +0000</pubDate>
      <guid>https://zhiyuan1i.github.io/en/posts/tech-stack/</guid>
      <description>Introduction to the tech stack used to build this site</description>
    </item>
  </channel>
</rss>
