<?xml version="1.0" encoding="UTF-8"?>
<rss  xmlns:atom="http://www.w3.org/2005/Atom" 
      xmlns:media="http://search.yahoo.com/mrss/" 
      xmlns:content="http://purl.org/rss/1.0/modules/content/" 
      xmlns:dc="http://purl.org/dc/elements/1.1/" 
      version="2.0">
<channel>
<title>ML Notes — English</title>
<link>https://notes.iwase.dev/en/</link>
<atom:link href="https://notes.iwase.dev/en/index.xml" rel="self" type="application/rss+xml"/>
<description>Machine learning and deep learning research notes (English edition)</description>
<generator>quarto-1.9.37</generator>
<lastBuildDate>Fri, 15 May 2026 00:00:00 GMT</lastBuildDate>
<item>
  <title>Diffusion Language Models</title>
  <dc:creator>Naoto Iwase</dc:creator>
  <link>https://notes.iwase.dev/en/dllm/</link>
  <description><![CDATA[ 





<p>Diffusion Language Models (DLLM) bring the ideas behind the diffusion models that succeeded in image generation into language modeling, with recent large-scale implementations such as LLaDA and Dream. This book organizes the key references needed to understand modern DLLMs, covering their formulation, sampling strategies, and the correspondence with continuous diffusion models.</p>



 ]]></description>
  <category>LLM</category>
  <category>Generative Models</category>
  <guid>https://notes.iwase.dev/en/dllm/</guid>
  <pubDate>Fri, 15 May 2026 00:00:00 GMT</pubDate>
  <media:content url="https://notes.iwase.dev/en/dllm/images/dllm.png" medium="image" type="image/png" height="76" width="144"/>
</item>
<item>
  <title>One-Step Generation</title>
  <dc:creator>Naoto Iwase</dc:creator>
  <link>https://notes.iwase.dev/en/one-step-generation/</link>
  <description><![CDATA[ 





<p>Between 2025 and 2026, methods that overcome the multi-step inference of diffusion models and Flow Matching to <strong>generate high-quality images with a single network evaluation (1-NFE)</strong> have been rapidly advancing. This series curates four papers driving this field, tracing the technical evolution from extensions of Flow Matching to entirely new paradigms.</p>



 ]]></description>
  <category>Deep Learning</category>
  <category>Generative Models</category>
  <guid>https://notes.iwase.dev/en/one-step-generation/</guid>
  <pubDate>Wed, 11 Feb 2026 00:00:00 GMT</pubDate>
  <media:content url="https://notes.iwase.dev/en/one-step-generation/images/one-step-generation.png" medium="image" type="image/png" height="76" width="144"/>
</item>
<item>
  <title>Molmo2</title>
  <dc:creator>Naoto Iwase</dc:creator>
  <link>https://notes.iwase.dev/en/molmo2/</link>
  <description><![CDATA[ 





<p>Molmo2 (Multimodal Open Language Model 2) is a fully open Vision-Language Model (VLM) family developed by the Allen Institute for AI (AI2) and the University of Washington. Its key distinguishing feature is <strong>video grounding</strong> capability, which enables the model to precisely indicate “when and where” specific events or objects occur within a video.</p>
<p>Using 9 new datasets (constructed entirely without relying on proprietary models), Molmo2 achieves state-of-the-art performance among open-source models. In particular, it surpasses proprietary models such as Gemini 3 Pro in video pointing and tracking.</p>
<p><strong>Paper</strong>: <a href="https://arxiv.org/abs/2601.10611">arXiv:2601.10611</a></p>
<p><strong>Code</strong>: <a href="https://github.com/allenai/molmo2">github.com/allenai/molmo2</a></p>
<p><strong>Demo</strong>: <a href="https://playground.allenai.org">playground.allenai.org</a></p>



 ]]></description>
  <category>VLM</category>
  <category>Multimodal</category>
  <guid>https://notes.iwase.dev/en/molmo2/</guid>
  <pubDate>Tue, 03 Feb 2026 00:00:00 GMT</pubDate>
  <media:content url="https://notes.iwase.dev/en/molmo2/images/molmo2.png" medium="image" type="image/png" height="76" width="144"/>
</item>
<item>
  <title>Olmo 3</title>
  <dc:creator>Naoto Iwase</dc:creator>
  <link>https://notes.iwase.dev/en/olmo-3/</link>
  <description><![CDATA[ 





<p>Olmo 3 is a family of state-of-the-art, fully-open language models at the 7B and 32B parameter scales developed by the Allen Institute for AI (AI2). This release includes the entire Model Flow, i.e., the full lifecycle of the family of models, including every stage, checkpoint, data point, and dependency used to build it.</p>
<p><strong>Paper</strong>: <a href="https://arxiv.org/abs/2512.13961">arXiv:2512.13961</a></p>



 ]]></description>
  <category>LLM</category>
  <category>Reasoning</category>
  <guid>https://notes.iwase.dev/en/olmo-3/</guid>
  <pubDate>Mon, 02 Feb 2026 00:00:00 GMT</pubDate>
  <media:content url="https://notes.iwase.dev/en/olmo-3/images/olmo-3.png" medium="image" type="image/png" height="76" width="144"/>
</item>
</channel>
</rss>
