<?xml version="1.0" encoding="UTF-8"?>
<rss  xmlns:atom="http://www.w3.org/2005/Atom" 
      xmlns:media="http://search.yahoo.com/mrss/" 
      xmlns:content="http://purl.org/rss/1.0/modules/content/" 
      xmlns:dc="http://purl.org/dc/elements/1.1/" 
      version="2.0">
<channel>
<title>Operating Intelligence - Cedric Vidal</title>
<link>https://vidal.biz/posts.html</link>
<atom:link href="https://vidal.biz/posts.xml" rel="self" type="application/rss+xml"/>
<description>Operating Intelligence by Cedric Vidal. Writing about building, evaluating, and scaling intelligent systems in production.</description>
<image>
<url>https://vidal.biz/profile.jpeg</url>
<title>Operating Intelligence - Cedric Vidal</title>
<link>https://vidal.biz/posts.html</link>
</image>
<generator>quarto-1.8.27</generator>
<lastBuildDate>Thu, 31 Oct 2024 00:00:00 GMT</lastBuildDate>
<item>
  <title>LLM Distillation just got easier - Part 4 - Maximize your fine-tuned model performance with the new Azure AI Evaluation SDK</title>
  <link>https://vidal.biz/</link>
  <description>Discover how to enhance AI model performance with the Azure AI Evaluation SDK. Learn the evaluation process of fine-tuned models, compare distilled and baseline models using advanced metrics.</description>
  <category>Evaluation</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Thu, 31 Oct 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/raft-4.png" medium="image" type="image/png" height="82" width="144"/>
</item>
<item>
  <title>LLM Distillation just got easier - Part 3 - Deploying your LoRA Fine-tuned Llama 3.1 8B model, why it&#39;s a breeze!</title>
  <link>https://vidal.biz/</link>
  <description>Learn how Azure AI makes it effortless to deploy your LoRA fine-tuned models using Azure AI.</description>
  <category>Infrastructure</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Thu, 24 Oct 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/raft-3.png" medium="image" type="image/png" height="82" width="144"/>
</item>
<item>
  <title>Thoughts about OpenAI’s new Distillation feature</title>
  <dc:creator>Cedric Vidal</dc:creator>
  <link>https://vidal.biz/posts/2024-10-02-openai-distillation-thoughts/</link>
  <description><![CDATA[ OpenAI has recently introduced a compelling new distillation feature, setting itself apart from traditional methods. Unlike <strong>Synthetic Data Distillation</strong>—on which I <a href="https://aka.ms/raft-recipe-blog-datagen">wrote previously</a> and coded a <a href="https://aka.ms/raft-recipe">⁠raft-distillation-recipe repository</a> Github repository—this approach leverages <strong>Production Data driven Distillation</strong>. This innovative method harnesses real-time production data, offering a dynamic and practical alternative for optimizing AI models. ]]></description>
  <category>Strategy</category>
  <guid>https://vidal.biz/posts/2024-10-02-openai-distillation-thoughts/</guid>
  <pubDate>Wed, 02 Oct 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/2024-10-02-openai-distillation-thoughts/header.webp" medium="image" type="image/webp"/>
</item>
<item>
  <title>LLM Distillation just got easier - Part 2 - Fine-Tuning Llama 3.1 8B on Azure AI Serverless</title>
  <link>https://vidal.biz/</link>
  <description>How Azure AI Serverless Fine-tuning, LoRA, RAFT and the AI Python SDK are streamlining fine-tuning of domain specific models.</description>
  <category>Infrastructure</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Wed, 18 Sep 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/raft-2.png" medium="image" type="image/png" height="82" width="144"/>
</item>
<item>
  <title>LLM Distillation just got easier - Part 1 - Synthetic Data Gen with Llama 3.1 405B &amp; RAFT</title>
  <link>https://vidal.biz/</link>
  <description>How Llama 405B and RAFT on Azure AI are changing the landscape of synthetic dataset generation and making model distillation much more approachable</description>
  <category>Infrastructure</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Wed, 04 Sep 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/raft-1.png" medium="image" type="image/png" height="82" width="144"/>
</item>
<item>
  <title>GoEX: a safer way to build autonomous Agentic AI applications</title>
  <link>https://vidal.biz/</link>
  <description>The Gorilla Execution Engine, from a paper by the UC Berkeley researchers behind Gorilla LLM and RAFT, helps developers create safer and more private Agentic AI applications</description>
  <category>Systems</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Wed, 24 Apr 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/goex.jpeg" medium="image" type="image/jpeg"/>
</item>
<item>
  <title>RAFT:  A new way to teach LLMs to be better at RAG</title>
  <link>https://vidal.biz/</link>
  <description>&#39;Retrieval-Augmented Fine-Tuning&#39; combines the benefits of Retrieval-Augmented Generation and Fine-Tuning for better domain adaptation</description>
  <category>Infrastructure</category>
  <guid>https://vidal.biz/</guid>
  <pubDate>Fri, 15 Mar 2024 00:00:00 GMT</pubDate>
  <media:content url="https://vidal.biz/posts/thumbnails/raft-student-gorilla.jpeg" medium="image" type="image/jpeg"/>
</item>
</channel>
</rss>
