<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Developers Digest - Open Source AI</title>
    <link>https://www.developersdigest.tech/tags/open-source-ai</link>
    <description>2 items tagged Open Source AI on Developers Digest - blog posts, tools, guides, and tutorials.</description>
    <language>en</language>
    <lastBuildDate>Sat, 25 Apr 2026 08:44:23 GMT</lastBuildDate>
    <atom:link href="https://www.developersdigest.tech/tags/open-source-ai/feed.xml" rel="self" type="application/rss+xml" />
    <image>
      <url>https://avatars.githubusercontent.com/u/124798203?v=4</url>
      <title>Developers Digest - Open Source AI</title>
      <link>https://www.developersdigest.tech/tags/open-source-ai</link>
    </image>
    <item>
      <title><![CDATA[Microsoft PHI-4: A 14B Parameter Model That Rivals Models 5x Its Size]]></title>
      <link>https://www.developersdigest.tech/blog/microsoft-phi-4-guide</link>
      <guid isPermaLink="true">https://www.developersdigest.tech/blog/microsoft-phi-4-guide</guid>
      <description><![CDATA[Microsoft's PHI-4 is an MIT-licensed 14 billion parameter model that matches Llama 3.3 70B and Qwen 2.5 72B on key benchmarks. Here is what makes it special, how to run it locally, and why small language models are increasingly practical for real development work.]]></description>
      <pubDate>Thu, 09 Jan 2025 00:00:00 GMT</pubDate>
      <category>Microsoft</category>
      <category>PHI-4</category>
      <category>Open Source AI</category>
      <category>LLM</category>
      <category>Ollama</category>
      <category>Local AI</category>
    </item>
    <item>
      <title><![CDATA[Llama 3.3 70B: Meta's Cost-Effective Frontier Model]]></title>
      <link>https://www.developersdigest.tech/blog/llama-3-3-70b-guide</link>
      <guid isPermaLink="true">https://www.developersdigest.tech/blog/llama-3-3-70b-guide</guid>
      <description><![CDATA[Meta surprised the AI community with Llama 3.3, a 70 billion parameter model that delivers 405B-class performance at a fraction of the cost. Here is what the benchmarks show, where to run it, and why this release matters for developers building with open-source models.]]></description>
      <pubDate>Sat, 07 Dec 2024 00:00:00 GMT</pubDate>
      <category>Meta</category>
      <category>Llama</category>
      <category>Open Source AI</category>
      <category>LLM</category>
      <category>Ollama</category>
    </item>
  </channel>
</rss>