<rss
      xmlns:atom="http://www.w3.org/2005/Atom"
      xmlns:media="http://search.yahoo.com/mrss/"
      xmlns:content="http://purl.org/rss/1.0/modules/content/"
      xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd"
      xmlns:dc="http://purl.org/dc/elements/1.1/"
      version="2.0"
    >
      <channel>
        <title><![CDATA[Melvin Carvalho]]></title>
        <description><![CDATA[Mathematician and Web Developer]]></description>
        <link>https://melvincarvalho.npub.pro/tag/3/</link>
        <atom:link href="https://melvincarvalho.npub.pro/tag/3/rss/" rel="self" type="application/rss+xml"/>
        <itunes:new-feed-url>https://melvincarvalho.npub.pro/tag/3/rss/</itunes:new-feed-url>
        <itunes:author><![CDATA[Melvin Carvalho]]></itunes:author>
        <itunes:subtitle><![CDATA[Mathematician and Web Developer]]></itunes:subtitle>
        <itunes:type>episodic</itunes:type>
        <itunes:owner>
          <itunes:name><![CDATA[Melvin Carvalho]]></itunes:name>
          <itunes:email><![CDATA[Melvin Carvalho]]></itunes:email>
        </itunes:owner>
            
      <pubDate>Mon, 22 Jul 2024 06:20:59 GMT</pubDate>
      <lastBuildDate>Mon, 22 Jul 2024 06:20:59 GMT</lastBuildDate>
      
      <itunes:image href="https://void.cat/d/Eon6A2onyvTWGSseNkwQ4v.webp" />
      <image>
        <title><![CDATA[Melvin Carvalho]]></title>
        <link>https://melvincarvalho.npub.pro/tag/3/</link>
        <url>https://void.cat/d/Eon6A2onyvTWGSseNkwQ4v.webp</url>
      </image>
      <item>
      <title><![CDATA[Probably the craziest week in Open…]]></title>
      <description><![CDATA[Probably the craziest week in Open Source AI (yet):


Mistral (in collaboration with Nvidia) dropped Apache 2.0 licensed NeMo 12B LLM, better than L3 8B and Gemma 2 9B. Models are multilingual with 128K context and a highly efficient tokenizer - tekken.
Apple released DCLM 7B - truly open source LLM,…]]></description>
             <itunes:subtitle><![CDATA[Probably the craziest week in Open Source AI (yet):


Mistral (in collaboration with Nvidia) dropped Apache 2.0 licensed NeMo 12B LLM, better than L3 8B and Gemma 2 9B. Models are multilingual with 128K context and a highly efficient tokenizer - tekken.
Apple released DCLM 7B - truly open source LLM,…]]></itunes:subtitle>
      <pubDate>Mon, 22 Jul 2024 06:20:59 GMT</pubDate>
      <link>https://melvincarvalho.npub.pro/post/note1agyt4d0nrq2e9spa903ggyt29zge4l253dfc8y68sqedqhfk3r7qr2gjvl/</link>
      <comments>https://melvincarvalho.npub.pro/post/note1agyt4d0nrq2e9spa903ggyt29zge4l253dfc8y68sqedqhfk3r7qr2gjvl/</comments>
      <guid isPermaLink="false">note1agyt4d0nrq2e9spa903ggyt29zge4l253dfc8y68sqedqhfk3r7qr2gjvl</guid>
      <category>3</category>
      
      <noteId>note1agyt4d0nrq2e9spa903ggyt29zge4l253dfc8y68sqedqhfk3r7qr2gjvl</noteId>
      <npub>npub1melv683fw6n2mvhl5h6dhqd8mqfv3wmxnz4qph83ua4dk4006ezsrt5c24</npub>
      <dc:creator><![CDATA[Melvin Carvalho]]></dc:creator>
      <content:encoded><![CDATA[<p>Probably the craziest week in Open Source AI (yet):<br><br>1. Mistral (in collaboration with Nvidia) dropped Apache 2.0 licensed NeMo 12B LLM, better than L3 8B and Gemma 2 9B. Models are multilingual with 128K context and a highly efficient tokenizer - tekken.<br> <br>2. Apple released DCLM 7B - truly open source LLM, based on OpenELM, trained on 2.5T tokens with 63.72 MMLU (better than Mistral 7B)<br><br>3. HF shared SmolLM - 135M, 360M, &amp; 1.7B Smol LMs capable of running directly in the browser; they beat Qwen 1.5B, Phi 1.5B and more. Trained on just 650B tokens. <br><br>4. Groq put out Llama 3 8B &amp; 70B tool use &amp; function calling model checkpoints - achieves 90.76% accuracy on Berkely Function Calling Leaderboard (BFCL).  Excels at API usage &amp; structured data manipulation!<br><br>5.  Salesforce released xLAM 1.35B &amp; 7B Large Action Models along with 60K instruction fine-tuning dataset. The 7B model scores 88.24% on BFCL &amp; 2B 78.94%<br><br>6. Deepseek changed the game with  v2 chat 0628 - The best open LLM on LYMSYS arena right now - 236B parameter model with 21B active parameters. It also excels at coding (rank <a href='/tag/3/'>#3</a>) and arena hard problems (rank <a href='/tag/3/'>#3</a>)<br><br>There's a lot more; Arcee (mergekit) released a series of LLMs, each better than the other, and Numina and HF Numina 72B (based on Qwen 2) and Math datasets, Mixbread with embedding models (english + german) and a lot more!<br><br>It's fun to see so many releases next week with L3 405B </p>
]]></content:encoded>
      <itunes:author><![CDATA[Melvin Carvalho]]></itunes:author>
      <itunes:summary><![CDATA[<p>Probably the craziest week in Open Source AI (yet):<br><br>1. Mistral (in collaboration with Nvidia) dropped Apache 2.0 licensed NeMo 12B LLM, better than L3 8B and Gemma 2 9B. Models are multilingual with 128K context and a highly efficient tokenizer - tekken.<br> <br>2. Apple released DCLM 7B - truly open source LLM, based on OpenELM, trained on 2.5T tokens with 63.72 MMLU (better than Mistral 7B)<br><br>3. HF shared SmolLM - 135M, 360M, &amp; 1.7B Smol LMs capable of running directly in the browser; they beat Qwen 1.5B, Phi 1.5B and more. Trained on just 650B tokens. <br><br>4. Groq put out Llama 3 8B &amp; 70B tool use &amp; function calling model checkpoints - achieves 90.76% accuracy on Berkely Function Calling Leaderboard (BFCL).  Excels at API usage &amp; structured data manipulation!<br><br>5.  Salesforce released xLAM 1.35B &amp; 7B Large Action Models along with 60K instruction fine-tuning dataset. The 7B model scores 88.24% on BFCL &amp; 2B 78.94%<br><br>6. Deepseek changed the game with  v2 chat 0628 - The best open LLM on LYMSYS arena right now - 236B parameter model with 21B active parameters. It also excels at coding (rank <a href='/tag/3/'>#3</a>) and arena hard problems (rank <a href='/tag/3/'>#3</a>)<br><br>There's a lot more; Arcee (mergekit) released a series of LLMs, each better than the other, and Numina and HF Numina 72B (based on Qwen 2) and Math datasets, Mixbread with embedding models (english + german) and a lot more!<br><br>It's fun to see so many releases next week with L3 405B </p>
]]></itunes:summary>
      
      </item>
      
      <item>
      <title><![CDATA[Prag is playing well.  2 good…]]></title>
      <description><![CDATA[Prag is playing well.  2 good wins vs the #3.  Surprise b4 in the English today.  Playing confidently.  But Magnus with white tomorrow, will probably play a standard London to see if Prag can cope.  He'll struggle to get a draw I think.…]]></description>
             <itunes:subtitle><![CDATA[Prag is playing well.  2 good wins vs the #3.  Surprise b4 in the English today.  Playing confidently.  But Magnus with white tomorrow, will probably play a standard London to see if Prag can cope.  He'll struggle to get a draw I think.…]]></itunes:subtitle>
      <pubDate>Tue, 22 Aug 2023 18:12:38 GMT</pubDate>
      <link>https://melvincarvalho.npub.pro/post/note107vu7fd5wzvtl7ye9ms927qs4dzjateemuvkvan6s95gau3jwf9sxyjsed/</link>
      <comments>https://melvincarvalho.npub.pro/post/note107vu7fd5wzvtl7ye9ms927qs4dzjateemuvkvan6s95gau3jwf9sxyjsed/</comments>
      <guid isPermaLink="false">note107vu7fd5wzvtl7ye9ms927qs4dzjateemuvkvan6s95gau3jwf9sxyjsed</guid>
      <category>3</category>
      
      <noteId>note107vu7fd5wzvtl7ye9ms927qs4dzjateemuvkvan6s95gau3jwf9sxyjsed</noteId>
      <npub>npub1melv683fw6n2mvhl5h6dhqd8mqfv3wmxnz4qph83ua4dk4006ezsrt5c24</npub>
      <dc:creator><![CDATA[Melvin Carvalho]]></dc:creator>
      <content:encoded><![CDATA[<p>Prag is playing well.  2 good wins vs the <a href='/tag/3/'>#3</a>.  Surprise b4 in the English today.  Playing confidently.  But Magnus with white tomorrow, will probably play a standard London to see if Prag can cope.  He'll struggle to get a draw I think.</p>
]]></content:encoded>
      <itunes:author><![CDATA[Melvin Carvalho]]></itunes:author>
      <itunes:summary><![CDATA[<p>Prag is playing well.  2 good wins vs the <a href='/tag/3/'>#3</a>.  Surprise b4 in the English today.  Playing confidently.  But Magnus with white tomorrow, will probably play a standard London to see if Prag can cope.  He'll struggle to get a draw I think.</p>
]]></itunes:summary>
      
      </item>
      
      </channel>
      </rss>
    