<rss version="2.0">
  <channel>
    <title>omlx on Honeypot.net</title>
    <link>https://honeypot.net/categories/omlx/</link>
    <description></description>
    
    <language>en</language>
    
    <lastBuildDate>Sat, 02 May 2026 08:53:32 -0700</lastBuildDate>
    
    <item>
      <title></title>
      <link>https://honeypot.net/2026/05/02/ive-been-running-ollama-on.html</link>
      <pubDate>Sat, 02 May 2026 08:53:32 -0700</pubDate>
      
      <guid>http://kirk.micro.blog/2026/05/02/ive-been-running-ollama-on.html</guid>
      <description>&lt;p&gt;I&amp;rsquo;ve been running Ollama on my Mac Studio for local AI experiments. I followed advice to try oMLX instead and it&amp;rsquo;s ludicrously faster, like maybe 5-10x for both time to first token and completing the response. I haven&amp;rsquo;t benchmarked it, but it subjectively feels like when I replaced a hard drive with an SSD.&lt;/p&gt;
</description>
    </item>
    
  </channel>
</rss>