<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Code Sport I/O</title>
	<atom:link href="https://codesport.io/feed/" rel="self" type="application/rss+xml" />
	<link>https://codesport.io</link>
	<description>Weekend Code School in North Carolina for Busy Adults</description>
	<lastBuildDate>Thu, 02 Apr 2026 17:40:43 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.8.1</generator>
	<item>
		<title>Comparing Embedding Models</title>
		<link>https://codesport.io/artificial-intelligence/comparing-embedding-models/</link>
					<comments>https://codesport.io/artificial-intelligence/comparing-embedding-models/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Thu, 05 Feb 2026 19:26:43 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7808</guid>

					<description><![CDATA[Embedding Model Comparison Feature bge-m3 qwen3-embedding-0.6b qwen3-embedding-4b all-MiniLM-L6-v2 nomic-embed-text-v1.5 Context Window 8,192 tokens 32,768 tokens 32,768 tokens 256 word pieces (~190 tokens) 8,192 tokens Dimensions 1,024 1,024 2,560 384 768 (Flexible) Model Size 567M Parameters 600M Parameters 4B Parameters 22.7M Parameters 137M Parameters MTEB Score / Best For 64.6 / Multilingual &#038; Hybrid 61.8 / [&#8230;]]]></description>
										<content:encoded><![CDATA[<script src="https://cdn.tailwindcss.com"></script>
<script>
  tailwind.config = {
    corePlugins: {
      preflight: false, // Disables Tailwind base styles (resets)
    },
    // Optional: Only apply styles within this selector
    important: '#tailwind-blog-content', 
  }
</script>
    <style>
        @import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;600;700&display=swap');
        body {
            font-family: 'Inter', sans-serif;
            background-color: #f7f9fb;
        }
        .responsive-table th, .responsive-table td {
            border-bottom: 1px solid #e2e8f0;
        }
        @media (max-width: 768px) {
            .responsive-table th, .responsive-table td {
                padding: 0.75rem 0.5rem;
                font-size: 0.8125rem;
            }
        }
    </style>

    <div  id="tailwind-blog-content" class="max-w-full mx-auto space-y-10">

        <!-- Main Comparison Table -->
        <div class="space-y-4">
            <h1 class="text-2xl font-extrabold text-slate-900 tracking-tight border-b-4 border-indigo-600 pb-2">Embedding Model Comparison</h1>
            <div class="overflow-x-auto rounded-xl shadow-lg bg-white border border-slate-200">
                <table class="w-full text-left border-collapse responsive-table">
                    <thead>
                        <tr class="bg-slate-800 text-white uppercase text-lg tracking-wider">
                            <th class="p-4">Feature</th>
                            <th class="p-4">bge-m3</th>
                            <th class="p-4">qwen3-embedding-0.6b</th>
                            <th class="p-4">qwen3-embedding-4b</th>
                            <th class="p-4">all-MiniLM-L6-v2</th>
                            <th class="p-4">nomic-embed-text-v1.5</th>
                        </tr>
                    </thead>
                    <tbody class="text-slate-700">
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Context Window</td>
                            <td class="p-4">8,192 tokens</td>
                            <td class="p-4">32,768 tokens</td>
                            <td class="p-4">32,768 tokens</td>
                            <td class="p-4">256 word pieces (~190 tokens)</td>
                            <td class="p-4">8,192 tokens</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Dimensions</td>
                            <td class="p-4">1,024</td>
                            <td class="p-4">1,024</td>
                            <td class="p-4">2,560</td>
                            <td class="p-4">384</td>
                            <td class="p-4">768 (Flexible)</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Model Size</td>
                            <td class="p-4">567M Parameters</td>
                            <td class="p-4">600M Parameters</td>
                            <td class="p-4">4B Parameters</td>
                            <td class="p-4">22.7M Parameters</td>
                            <td class="p-4">137M Parameters</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">MTEB Score / Best For</td>
                            <td class="p-4">64.6 / Multilingual &#038; Hybrid</td>
                            <td class="p-4">61.8 / Fast Multilingual</td>
                            <td class="p-4">~67.5 / High-Precision RAG</td>
                            <td class="p-4">56.0 / Speed &#038; Similarity</td>
                            <td class="p-4">62.3 / Long-Context English</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Primary Use</td>
                            <td class="p-4">Multilingual Search</td>
                            <td class="p-4">Lightweight local search</td>
                            <td class="p-4">Enterprise-grade retrieval</td>
                            <td class="p-4">Mobile/Edge tasks</td>
                            <td class="p-4">Processing long docs</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Size (VRAM)</td>
                            <td class="p-4">~1.2 GB</td>
                            <td class="p-4">~1.2 GB</td>
                            <td class="p-4">~2.5 GB (Quantized)</td>
                            <td class="p-4">~45 MB</td>
                            <td class="p-4">~274 MB</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900">Accuracy</td>
                            <td class="p-4">High (Top Multilingual)</td>
                            <td class="p-4">High</td>
                            <td class="p-4">Very High (SOTA)</td>
                            <td class="p-4">Moderate</td>
                            <td class="p-4">High (Beats OpenAI Small)</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold bg-slate-50 text-slate-900 border-b-0">Popularity</td>
                            <td class="p-4 text-green-600 font-bold border-b-0"><a href="https://ollama.com/library/bge-m3" class="underline hover:no-underline">3.2M+</a></td>
                            <td class="p-4 text-green-600 font-bold border-b-0""><a href="https://ollama.com/library/qwen3-embedding:0.6b" class="underline hover:no-underline">548K+ (Series Total)</a></td>
                            <td class="p-4 text-green-600 font-bold border-b-0""><a href="https://ollama.com/library/qwen3-embedding:4b" class="underline hover:no-underline">548K+ (Series Total)</a></td>
                            <td class="p-4 text-green-600 font-bold border-b-0"><a href="https://ollama.com/library/all-minilm" class="underline hover:no-underline">2.3M+</a></td>
                            <td class="p-4 text-green-600 font-bold border-b-0"><a href="https://ollama.com/library/nomic-embed-text" class="underline hover:no-underline">52.4M+</a></td>
                        </tr>
                    </tbody>
                </table>
            </div>
        </div>

        <!-- Strategic Breakdown -->
        <div class="grid grid-cols-1 md:grid-cols-2 gap-6">
            <div class="p-6 bg-white rounded-xl shadow-md border-t-4 border-blue-500">
                <h2 class="text-lg font-bold text-slate-800 mb-2"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f4c4.png" alt="📄" class="wp-smiley" style="height: 1em; max-height: 1em;" /> For Long Documents</h2>
                <p class="text-slate-600 text-sm">The <strong>Qwen3 series</strong> and <strong>nomic-embed-text</strong> are the clear winners with context windows of 32k and 8k tokens respectively. Most standard models (like MiniLM) will truncate text after just a few paragraphs.</p>
            </div>
            <div class="p-6 bg-white rounded-xl shadow-md border-t-4 border-purple-500">
                <h2 class="text-lg font-bold text-slate-800 mb-2"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f310.png" alt="🌐" class="wp-smiley" style="height: 1em; max-height: 1em;" /> For Multilingual Tasks</h2>
                <p class="text-slate-600 text-sm"><strong>bge-m3</strong> is specifically designed for cross-lingual tasks and supports over 100 languages with high efficiency and hybrid retrieval capabilities.</p>
            </div>
            <div class="p-6 bg-white rounded-xl shadow-md border-t-4 border-orange-500">
                <h2 class="text-lg font-bold text-slate-800 mb-2"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f3c6.png" alt="🏆" class="wp-smiley" style="height: 1em; max-height: 1em;" /> For Maximum Performance</h2>
                <p class="text-slate-600 text-sm"><strong>qwen3-embedding-4b</strong> currently leads open-weight benchmarks, offering retrieval accuracy that rivals or exceeds proprietary models like OpenAI&#8217;s text-embedding-3-large.</p>
            </div>
            <div class="p-6 bg-white rounded-xl shadow-md border-t-4 border-green-500">
                <h2 class="text-lg font-bold text-slate-800 mb-2"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/26a1.png" alt="⚡" class="wp-smiley" style="height: 1em; max-height: 1em;" /> For Speed / Low Resource</h2>
                <p class="text-slate-600 text-sm"><strong>all-MiniLM-L6-v2</strong> is incredibly small and fast (~45MB VRAM), making it ideal if you are running on a CPU or a very low-end device.</p>
            </div>
        </div>

    </div>]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/comparing-embedding-models/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Enterprise-Ready GenAI: Determinism is the Priority</title>
		<link>https://codesport.io/artificial-intelligence/llm-determinism-in-enterprise-systems/</link>
					<comments>https://codesport.io/artificial-intelligence/llm-determinism-in-enterprise-systems/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Thu, 05 Feb 2026 15:31:37 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7774</guid>

					<description><![CDATA[Enterprise GenAI and the Role of Deterministic Outputs In many enterprise environments, determinism is not merely a technical preference, it is a fundamental requirement for operational stability. With the rise of modern Generative AI, deterministic systems are critical in industries where &#8220;hallucinations&#8221; (making things up) could be dangerous or illegal. Defining Deterministic Systems The ability [&#8230;]]]></description>
										<content:encoded><![CDATA[    <script src="https://cdn.tailwindcss.com"></script>
<script>
  tailwind.config = {
    corePlugins: {
      preflight: false, // Disables Tailwind base styles (resets)
    },
    // Optional: Only apply styles within this selector
    important: '#tailwind-blog-content', 
  }
</script>
    <style>
        @import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;600;700&display=swap');
        body {
            font-family: 'Inter', sans-serif;
           /* background-color: #f7f9fb;*/
            color: #1e293b;
        }
        h2 { 
            padding-bottom: 0.5rem; 
            border-bottom: 2px solid #94a3b8; 
            margin-top: 2.5rem !important; 
            font-weight: 700;
            color: #0f172a;
        }
        .bench-table tbody tr:nth-child(even) { background-color: #f9fafb; }
        .bench-table tbody tr:nth-child(odd) { background-color: #ffffff; }
        .bench-table tbody tr:hover { background-color: #eff6ff !important; }
        
        .math-tex {
            font-family: "Times New Roman", Times, serif;
            font-style: italic;
            font-weight: 600;
             /*color: black #4338ca;*/}
 .blueBoxtext{font-size: 1.3em !important;}
.p-6{padding: .2rem 1.5rem 1.5rem 1.5rem}
     
    </style>

    <div id="tailwind-blog-content" class="max-w-full mx-auto space-y-8">
        
        <header class="border-b-4 border-indigo-600 pb-6">
            <h2 class="text-4xl text-slate-900 tracking-tight">Enterprise GenAI and the Role of Deterministic Outputs</h1>
        </header>

        <section class="leading-relaxed text-slate-700 space-y-4 text-lg">
            <p>In many enterprise environments, determinism is not merely a technical preference, it is a fundamental requirement
             for operational stability.</p>

              <p>With the rise of modern Generative AI, deterministic systems are critical in industries where <strong>&#8220;hallucinations&#8221; (making things up) could be dangerous or illegal</strong>.</p>   
   <!--    <blockquote>While GenAI thrives on "creative" randomness, it is possible to "locked it down" and  ensure it provides the nearly the same output for the same input.</blockquote> 
       <p>Generative AI is probabilistic by nature, but it can be made mostly deterministic through specific settings and architectural choices. </p>
-->

        <div class="p-6 bg-slate-100 rounded-xl border-l-4 border-slate-400">
            <h2 class="font-bold text-slate-700 uppercase mb-2 tracking-wide text-2xl">Defining Deterministic Systems</h2>
            <span class="text-slate-600 leading-relaxed text-2xl">
                  The ability to generate the exact same output, <span class="math-tex">y</span>, for the exact same prompt input <span class="math-tex">x</span> sent to an LLM, <span class="math-tex">f(x)</span>, 100% of the time. As such, the goal of <em>Deterministic AI</em> is to produce repeatable, consistent outputs for a given prompt fed into an Large Language Learning Model (LLM)</span>
        </div>


       <h2 class="text-3xl">Use Cases: Generative AI for Expert Systems</h2>

        <div class="p-6 bg-slate-100 rounded-xl border-l-4 border-slate-400">
            <h2 class="font-bold text-slate-700 uppercase mb-2 tracking-wide text-2xl">Defining Expert Systems</h2>
            <span class="text-slate-600 leading-relaxed text-2xl">
      	Expert Systems are active reasoning tools which  use Retrieval Augmented Generation (RAG) as a supplemental knowledge base. It then adds an inference engine to process those facts and solve a specific problem (e.g., diagnosing a disease or calculating tax returns).</span>
        </div>


<p>A growing use case for GenAI in the Enterprise is exposing LLMs to their internal documents and knowledge bases. <!--These are digital libraries.--> These are repositories (like a wiki or SharePoint or dedicated document server) stores documents, facts, and FAQs. They are largely seen as a passive tool where humans go to search for information.</p>
  
<section class="bg-white p-6 rounded-xl shadow-md border border-slate-200  blueBoxtext">
    <h3 class="text-2xl font-bold text-indigo-700 mb-4">Use Cases for Deterministic AI</h3>
            <ul class="space-y-3 text-slate-600">
               <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span><span><strong class="text-slate-800">Finance:</strong> For processing transactions or enforcing strict compliance rules.
      
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span><span><strong class="text-slate-800">Healthcare:</strong> For clinical decision support systems that must follow proven medical protocols.
        
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span><span><strong class="text-slate-800">Industrial Automation:</strong> For controlling robotic arms on assembly lines where every move must be exact and repeatable
            <!--
            Automation: IT workflows, such as CI/CD, Terraform, or Kubernetes management.
Process Guidance: Systems that must follow strict safety procedures.-->
            
            </ul>
    </section>

       
            <p>
When deploying Generative AI for Enterprise Knowledge Bases, stakeholders expect &#8220;The Canonical Answer&#8221; — a single, verified response
 to a specific query that does not fluctuate based on the model&#8217;s random state. Without determinism, two employees asking the same policy question could receive different nuances, leading to internal misalignment.
            </p>

     <h2 class="text-3xl">Auditability: Consistent, Deterministic Outputs Supports Governance Risk, and Compliance </h2>

        <p>
                For <strong>Corporate Training and Compliance</strong>, determinism serves as the backbone of auditability. If an AI tutor or compliance assistant is used to certify staff, the logic used to evaluate their performance must be repeatable. High-variance responses in training scenarios introduce &#8220;drift,&#8221; where the educational material may slowly deviate from the intended curriculum over thousands of user sessions.
            </p>

        
        </section>


        <section class="bg-white p-6 rounded-xl shadow-md border border-slate-200">
            <h3 class="text-2xl font-bold text-indigo-700 mb-4">Key Benefits for Organizations:</h3>
            <ul class="space-y-3 text-slate-600">
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span> <span><strong class="text-slate-800">Regulatory Compliance:</strong> Ensures that automated advice meets legal standards consistently.</span></li>
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span> <span><strong class="text-slate-800">Debugging &#038; QA:</strong> Allows engineers to replicate errors by using the same prompt and seed.</span></li>
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span> <span><strong class="text-slate-800">User Trust:</strong> Increases reliability perception when the system provides stable, predictable outputs.</span></li>
                <li class="flex items-start"><span class="text-indigo-500 mr-2">•</span> <span><strong class="text-slate-800">Cost Control:</strong> Deterministic paths are easier to cache, reducing API token costs.</span></li>
            </ul>   
        </section>

        <section>
            <h2 class="text-2xl">Optimal Model Parameters to Enforce Deterministic AI</h2>
            <p>
                To ensure the highest degree of determinism, LLM parameters should be locked to a &#8220;ground state.&#8221; Any value that allows the model to choose from a distribution of tokens rather than the single highest-probability token will introduce variance.</p>

            <p>The below parameters move the LLM away from &#8220;creative sampling&#8221; and toward Greedy Decoding. In Greedy Decoding, the model picks the most likely next token at every step. Randomess is diminished.</p>
            
            <div class="overflow-hidden rounded-xl border border-slate-200 bg-white shadow-lg mt-6">
                <table class="w-full text-left border-collapse bench-table">
                    <thead class="text-white uppercase tracking-wider bg-emerald-800"> <!--text-xs--> 
                        <tr>
                            <th class="p-4">Parameter</th>
                            <th class="p-4">Optimum Value</th>
                            <th class="p-4">Determinism Impact</th>
                            <th class="p-4">Explanation</th>
                        </tr>
                    </thead>
                    <tbody> <!--class="text-sm"-->
                        <tr>
                            <td class="p-4 font-bold">Temperature</td>
                            <td class="p-4">0</td>
                            <td class="p-4">90%</td>
                            <td class="p-4">Disables random sampling; model always picks the most likely token.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Seed</td>
                            <td class="p-4">Fixed Integer</td>
                            <td class="p-4">5%</td>
                            <td class="p-4">Ensures the random number generator starts at the same point. non-deterministic noise in GPU floating-point calculations (cuBLAS/flash-attention)</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Top_P</td>
                            <td class="p-4">1.0</td>
                            <td class="p-4">1%</td>
                            <td class="p-4">When Temp is 0, Top_P becomes irrelevant, but keeping it at 1.0 prevents it from interfering with the top choice.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Top_K</td>
                            <td class="p-4">0 (or 1)</td>
                            <td class="p-4">1%</td>
                            <td class="p-4"> Setting this to 1 forces the model to only look at the single most likely token.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Frequency_Penalty</td>
                            <td class="p-4">0</td>
                            <td class="p-4">1%</td>
                            <td class="p-4">Ensures token scores and therefore token selection are not modified by historical usage.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Presence_Penalty</td>
                            <td class="p-4">0</td>
                            <td class="p-4">1%</td>
                            <td class="p-4">Prevents token scores from being modified by existence in text.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Repeat_Penalty</td>
                            <td class="p-4">1.0</td>
                            <td class="p-4">1%</td>
                            <td class="p-4">In most systems, 1.0 means &#8220;no change.&#8221; Anything else modifies the logic scores.</td>
                        </tr>
                        <tr>
                            <td class="p-4 font-bold">Min_P</td>
                            <td class="p-4">0</td>
                            <td class="p-4">1%</td>
                            <td class="p-4">This is a newer sampling technique; keeping it at 0 ensures it doesn&#8217;t prune the top-ranked token. Thus, disabling threshold-based token pruning.</td>
                        </tr>
                    </tbody>
                </table>
            </div>
<BR>
<p><strong>NB:</strong> These parameters are not stored in the model file itself. They are injected at runtime when the user sends a prompt.</p>
        </section>

        <section>
            <h2 class="text-2xl">Critical Nuance: The &#8220;Hidden&#8221; 100%</h2>
            <p class="mt-4 text-slate-700 leading-relaxed">
                Even with the settings above, achieving 100% determinism in LLMs is notoriously difficult due to hardware non-determinism.
            </p>
            <ol class="list-decimal list-inside mt-4 space-y-4 text-slate-700">
                <li class="pl-2">
                    <strong class="text-slate-900">Atomic Operations:</strong> Modern GPUs perform calculations in parallel. Because of the way floating-point numbers are added across thousands of cores (the order of operations can vary slightly), you might get a tiny difference in the 10th decimal place.
                </li>
                <li class="pl-2">
                    <strong class="text-slate-900">Logit Shifting:</strong> That tiny decimal difference can occasionally cause token A (score 10.000001) and token B (score 10.000002) to swap places.
                </li>
                <li class="pl-2">
                    <strong class="text-slate-900">The Fix:</strong> This is why the Seed and Temperature = 0 are used together. On some enterprise APIs (like OpenAI), you can also look for the <code class="bg-slate-100 px-1 rounded border border-slate-300">system_fingerprint</code> in the response to ensure the underlying hardware/software version hasn&#8217;t changed. Monitor this to ensure the provider didn&#8217;t update the model or hardware.
                </li>
            </ol>
        </section>

        <section>
            <h2 class="text-2xl">Why Penalties are &#8220;0&#8221; for Determinism</h2>
            <p class="mt-4 text-slate-700 leading-relaxed">
                While you can have a deterministic output with a <code class="text-indigo-600 font-semibold">frequency_penalty</code> of 0.5, it makes the output harder to predict and debug. If your goal is <span class="math-tex">f(x)=y</span>, you want the purest path from the model&#8217;s training to the output. Penalties add a layer of &#8220;stateful&#8221; math (where the next word depends on every word before it in a modified way) which increases the chance of a floating-point error causing a divergence.
            </p>
        </section>

        <!-- Technical Note -->
        <div class="p-6 bg-slate-100 rounded-xl border-l-4 border-slate-400">
            <h2 class="font-bold text-slate-700 uppercase mb-2 tracking-wide">Quality Assurance Note</h2>
            <p class="text-slate-600 leading-relaxed">
                Even with deterministic parameters set to zero, floating-point variance across parallel GPU cores can introduce non-determinism at high token counts. Use the <strong>System Fingerprint</strong> to verify environment stability during longitudinal testing.
            </p>
        </div>

<hr>
Referecnes
<ul>
	<li><a href="https://www.ebsco.com/research-starters/computer-science/expert-system-artificial-intelligence">Expert System (artificial intelligence)</a></li>

	<li><a href="https://www.geeksforgeeks.org/artificial-intelligence/expert-systems/">Expert Systems in AI</a></li>
</ul>
 <!-- 
        <div class="mt-8 p-6 bg-white rounded-xl shadow-lg border-l-4 border-indigo-500">
            <h2 class="text-xl font-semibold text-indigo-700 mb-4 font-bold">Footnotes & Technical References</h2>
            <ol class="list-decimal list-inside space-y-3 text-sm text-gray-600">
                <li>
                    <a href="https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html" target="_blank" class="text-indigo-600 hover:underline font-medium">NVIDIA Container Toolkit Documentation</a>: Official guide confirming that while the kernel-mode driver must be on the host, the library-mode CUDA components should be containerized to avoid version conflicts.
                </li>
                <li>
                    <a href="https://www.docker.com/blog/how-to-deploy-ai-applications-with-docker-compose/" target="_blank" class="text-indigo-600 hover:underline font-medium">Docker Engineering Blog</a>: Highlighting best practices for using Compose to orchestrate multi-service AI applications (Ollama + Vector DBs).
                </li>
                <li>
                    <a href="https://cheatsheetseries.owasp.org/cheatsheets/Docker_Security_Cheat_Sheet.html" target="_blank" class="text-indigo-600 hover:underline font-medium">OWASP Docker Security Guide</a>: Verification of the isolation benefits of cgroups and namespaces when running potentially untrusted agent code (e.g., Code Interpreter agents).
                </li>
            </ol>
        </div>
              <div class="bg-indigo-50 border-l-4 border-indigo-500 p-6 rounded-r-lg">
                    <h4 class="font-bold text-indigo-900 mb-2">Key Benefits for Organizations:</h4>
                    <ul class="list-disc list-inside space-y-2 text-indigo-800 text-sm">
                        <li><strong>Regulatory Compliance:</strong> Ensures that automated advice meets legal standards consistently.</li>
                        <li><strong>Debugging & QA:</strong> Allows engineers to replicate errors by using the same prompt and seed.</li>
                        <li><strong>User Trust:</strong> Increases reliability perception when the system provides stable, predictable outputs.</li>
                        <li><strong>Cost Control:</strong> Deterministic paths are easier to cache, reducing API token costs for repeated queries.</li>
                    </ul>
                </div>        
-->

    </div>
]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/llm-determinism-in-enterprise-systems/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Inference Parameter Optimization for Qwen3-30B-A3B-2507-instruct</title>
		<link>https://codesport.io/artificial-intelligence/inference-parameter-optimization-for-qwen3-30b-a3b-2507-instruct/</link>
					<comments>https://codesport.io/artificial-intelligence/inference-parameter-optimization-for-qwen3-30b-a3b-2507-instruct/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Thu, 05 Feb 2026 15:29:28 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7757</guid>

					<description><![CDATA[LLM Parameter Management &#038; Determinism Technical Training Manual: Precedence, Overrides, and Enterprise Configuration. 1. Diagnostic Verification &#038; Inspection Before adjusting settings in the WebUI, engineers must verify the model&#8217;s native configuration. Use the following commands to inspect the Modelfile and the parameter defaults assigned to the local instance. # Inspect parameters for the specific Qwen3 [&#8230;]]]></description>
										<content:encoded><![CDATA[    <script src="https://cdn.tailwindcss.com"></script>

<script>
  tailwind.config = {
    corePlugins: {
      preflight: false, // Disables Tailwind base styles (resets)
    },
    // Optional: Only apply styles within this selector
    important: '#tailwind-blog-content', 
  }
</script>
    <style>
        @import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;600;700&display=swap');

   /*     body {
            font-family: 'Inter', sans-serif;
            color: #1e293b;
            background-color: #f7f9fb;
        }
*/
        .bench-table tbody tr:nth-child(even) { background-color: #f9fafb; }
        .bench-table tbody tr:nth-child(odd) { background-color: #ffffff; }
        .bench-table tbody tr:hover { background-color: #eff6ff !important; }

        .command-box {
            background-color: #1e293b;
            color: #e2e8f0;
            padding: 1.25rem;
            border-radius: 0.5rem;
            font-family: 'Courier New', Courier, monospace;
/*            font-size: 0.85rem;*/
            line-height: 1.6;
            overflow-x: auto;
            border-left: 4px solid #6366f1;
        }

.table-caption {
    display: block;
    width: 100%;
    font-size: 1.3em;
    font-weight: 600;
    color: #64748b;
    margin-top: 1.5em;
    margin-bottom: 2em;
    text-align: center;
    clear: both;
        }



        /* Updated Heading Styles with Darker Borders */
        h2 { 
            padding-bottom: 0.5rem; 
            border-bottom: 2px solid #94a3b8; 
            margin-top: 2.5rem !important; 
        }
        h3 { color: #4338ca; margin-top: 1.5rem !important; font-weight: 700; }
    </style>


    <div id="tailwind-blog-content" class="mx-auto space-y-12">

        <!-- Header -->
        <header class="border-b-4 border-indigo-600 pb-6">
            <h1 class="text-4xl font-extrabold text-slate-900 tracking-tight">LLM Parameter Management &#038; Determinism</h1>
            <p class="text-lg text-slate-600 mt-3">Technical Training Manual: Precedence, Overrides, and Enterprise Configuration.</p>
        </header>

        <!-- Section 1 -->
        <section>
            <h2 class="text-3xl font-bold text-slate-800">1. Diagnostic Verification &#038; Inspection</h2>
            <p class="mt-4 text-slate-700 leading-relaxed">
                Before adjusting settings in the WebUI, engineers must verify the model&#8217;s native configuration. Use the following commands to inspect the Modelfile and the parameter defaults assigned to the local instance.</p>
            <div class="command-box mt-4">
                # Inspect parameters for the specific Qwen3 tag<br>
                docker exec -it ollama ollama show &#8211;parameters qwen3:30b-a3b-instruct-2507-q4_K_M<br><br>
                
                # View the full Modelfile including System Prompts and Stop Sequences<br>
                docker exec -it ollama ollama show &#8211;modelfile qwen3:30b-a3b-instruct-2507-q4_K_M<br><br>

                # Verification for Unsloth-specific GGUF exports<br>
                docker exec -it ollama ollama show &#8211;parameters hf.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF:UD-Q4_K_XL
            </div>
        </section>

        <!-- Section 2 -->
        <section>
            <h2 class="text-3xl font-bold text-slate-800">2. The Three Sources of Parameter Data</h2>
            
            <div class="space-y-8 mt-6">
                <div>
                    <h3>2.1 The Ollama Modelfile (Baseline)</h3>
                    <p class="text-slate-700 leading-relaxed">
                        The Modelfile represents the foundational &#8220;hard-coded&#8221; configuration of the model. It defines the system prompt and default sampling parameters. These values serve as the fallback logic for the model; they are used only if the inference request does not specify its own parameters. This is considered the &#8220;floor&#8221; of the configuration stack.
                    </p>
                </div>

                <div>
                    <h3>2.2 Open-WebUI Settings (Master Interface)</h3>
                    <p class="text-slate-700 leading-relaxed">
                        Open-WebUI serves as the master orchestration layer. It maintains its own internal state for model parameters which frequently differs from the Ollama defaults. When a user submits a prompt, <strong>the WebUI explicitly injects its settings into the API request body</strong>. This payload override ensures that the UI settings take precedence over whatever is defined in the Modelfile.
                    </p>
                </div>

                <div>
                    <h3>2.3 Qwen3&#8217;s Recommendations (Target)</h3>
                    <p class="text-slate-700 leading-relaxed">
                        Optimization guidelines from Qwen3 (and reported by <a href="https://unsloth.ai/docs/models/qwen3-how-to-run-and-fine-tune/qwen3-2507#best-practices">Unsloth</a>) provide the mathematical targets for high-performance inference. These recommendations are based on rigorous testing of specific model architectures. Engineers should aim to mirror these targets within Open-WebUI to ensure the model operates within its intended reasoning parameters.</p>

<ul>
<li><strong>Output Length:</strong> Use an output length of 32,768 tokens for most queries, which is adequate for most queries</li>

<li><strong>presence_penalty:</strong> 0.0 to 2.0 (llama.cpp default turns it off, but to reduce repetitions, you may use 1.0)</li></ul>


                    </p>
                </div>
            </div>
        </section>

        <!-- Table 1 -->
        <section>
            <div class="overflow-hidden rounded-xl border border-slate-200 bg-white shadow-md">
                <table class="caption-bottom w-full text-left border-collapse bench-table">
     <caption class="table-caption">Table 1: Qwen3-30B-AB-2507-Instruct Value Comparison</caption>
                     <thead class="uppercase bg-blue-600 text-white sticky top-0">
                        <tr>
                            <th class="p-4">Parameter</th>
                            <th class="p-4">CLI (ollama show)</th>
                            <th class="p-4">Open-WebUI Settings</th>
                            <th class="p-4">Qwen3 Recommended</th>
                        </tr>
                    </thead>
                     <tbody class="divide-y divide-gray-200 text-gray-700">
                        <tr><td class="p-4 font-bold">temperature</td><td class="p-4">0.8</td><td class="p-4">0.8</td><td class="p-4">0.7</td></tr>
                        <tr><td class="p-4 font-bold">top_p</td><td class="p-4">0.9</td><td class="p-4">0.9</td><td class="p-4">0.8</td></tr>
                        <tr><td class="p-4 font-bold">min_p</td><td class="p-4">0.0</td><td class="p-4">0.05</td><td class="p-4">0.00</td></tr>
                        <tr><td class="p-4 font-bold">top_k</td><td class="p-4">40</td><td class="p-4">40</td><td class="p-4">20</td></tr>
                        <tr><td class="p-4 font-bold">presence_penalty</td><td class="p-4">0.0</td><td class="p-4">0.0</td><td class="p-4">0.0 to 2.0</td></tr>
                    </tbody>

                </table>
            </div>
           
        </section>

   
        <!-- Section 3: Subtle Influences -->
        <section>
            <h2 class="text-3xl font-bold text-slate-800">3. Subtle Influences of Sampling Logic</h2>
            <div class="space-y-8 mt-6">
                <div>
                    <h3>3.1 Temperature Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">Higher = more “creative” (risky) word choices. This is the primary &#8220;randomness&#8221; dial. At 0.0, the model stops &#8220;rolling the dice. At 1 it always rolls!</p>

                       <p><strong>Technical Deep-dive:</strong>  Temperature modifies the logits before the final softmax layer. Decreasing it sharpens the probability distribution. At 0.0, the &#8220;soft&#8221; max becomes a &#8220;hard&#8221; max, preventing the model from sampling from the &#8220;tail&#8221; of the distribution.
                    </p>
                </div>
                <div>
                    <h3>3.2 Top_P (Nucleus) Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">Limits choices to a cumulative probability “mass&#8221;. Anything less than 1 increases creativity and randomness. The LLM will exclude the top tokens if less than 1. For enterprise determinism, set it to 1.</p>

<p><strong>Technical Deep-dive:</strong> Top_P sampling dynamically adjusts the vocabulary size based on confidence. In confident predictions, it considers very few tokens. Setting this to 1.0 ensures that no candidates are removed based solely on cumulative probability.
                    </p>
                </div>


                <div>
                    <h3>3.3 Top_K Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">
A higher number increases token selection randomness by expanding choices to the top K most likely words.


    
                    </p>
                </div>
                <div>

                <div>
                    <h3>3.4 Min_P Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">
                        Min_P ensures a token is only considered if its probability is at least a specified percentage (e.g., 5%) of the most likely token&#8217;s probability. This preserves diverse choices when the model is genuinely uncertain while scaling more naturally than Top_P.
                    </p>
                </div>
                <div>





                    <h3>3.4 Frequency Penalty Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">High values force the AI to use “new” words to avoid repetition. High values encourage the model to avoid words it just used. This can cause semantic and meaning drift. For enterprise determinism, set to 0.0</p>



                        <p><strong>Technical Deep-dive:</strong> High values prevents repetition by subtracting a value from the logits of tokens that have already appeared. For enterprise determinism, it must be 0.0, as technical accuracy often requires repeating specific terminology that the penalty might otherwise suppress.
                    </p>
                </div>
                <div>
                    <h3>3.5 Presence Penalty Nuance</h3>
                    <p class="text-slate-700 leading-relaxed">
                        Presence Penalty applies a one-time penalty to tokens that have already appeared in the text at least once, regardless of their frequency. This encourages the model to introduce completely new topics or words into the conversation, which is generally avoided in deterministic retrieval.
                    </p>
                </div>
            </div>
        </section>

    </div>

]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/inference-parameter-optimization-for-qwen3-30b-a3b-2507-instruct/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Agentic Ai Server Build Tutorial Part I: The MVP</title>
		<link>https://codesport.io/artificial-intelligence/agentic-ai-server-build-tutorial-part-i-the-mvp/</link>
					<comments>https://codesport.io/artificial-intelligence/agentic-ai-server-build-tutorial-part-i-the-mvp/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Tue, 03 Feb 2026 17:36:35 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[advanced]]></category>
		<category><![CDATA[classes]]></category>
		<category><![CDATA[courses]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7700</guid>

					<description><![CDATA[Project Overview and Introduction This tutorial assumes you have Ubuntu 24.04 LTS server installed along with an NVIDIA GPU. It serves as Part I of an on-premises Agentic AI server buildout series. Part I, is the minimal viable product. It creates a headless and GPU-accelerated AI server with a client-accesible UI, LLM Runner, LLM, and [&#8230;]]]></description>
										<content:encoded><![CDATA[<h2 class="page-header">Project Overview and Introduction</h2>

<p>This tutorial assumes you have Ubuntu 24.04 LTS server installed along with an NVIDIA GPU. It serves as Part I of an on-premises Agentic AI server buildout series.  Part I, is the minimal viable product.  It creates a headless and GPU-accelerated AI server with a client-accesible UI, LLM Runner, LLM, and RAG capabilities via manual document uploads.</p>

 
        <ol class="indent">
	<li>Part I: Minimum Viable Product &#8211; Proof-of-concept (open-WebUI + Ollama)</li>
	<li>Part II: Pilot &#8211; Production buildout with 3-concurrent users (N8N + Dify + vLLM + image generation and recognition + Canvas generation  + 2-way voice communication)</li>
	<li>Part III: Enterprise &#8211; Production build-out for a medium-sized business with 100 to 500 concurrent users</li>
</ol>

    
<p>As we advance through the stages we&#8217;ll deploy enterprise tools. These include deployment and cybersecurity tools such as Ansible, VMs, Wireguard and a commercial AI-powered Next Generation Firewall (NGFW).</p>


<h2 class="page-header">Phase 1: Update Packages and Nvidia GPU Drivers</h2>
<pre>
# 1. update OS & Kernel
sudo apt update && sudo apt upgrade -y
sudo reboot

# 2. update nvidia drivers for headless, AI workloads
sudo ubuntu-drivers install --gpgpu
sudo reboot

# 3. manually install nvidia utility package
major_ver=$(modinfo -F version nvidia | cut -d'.' -f1)
sudo apt install nvidia-utils-${major_ver}-server 

sudo reboot
</pre>

<h2 class="page-header">Phase 2: Set-up Containerization</h2>

<p>In this phase we achieve software isolation, dependency management, and configuration portability.  Install Docker &#038; Nvidia Toolkit. We use Docker so that tools like Dify don&#8217;t affect system files. Nvidia&#8217;s toolkit allows docker  to &#8220;talk&#8221; to your GPU. </p>

<pre>
# Install Docker using the official convenience script
# curl -fsSL https://get.docker.com -o get-docker.sh
# sudo sh get-docker.sh
# Or install docker engine and docker compose from official ubuntu packages
sudo apt install docker.io -y # https://stackoverflow.com/a/57678382
sudo apt install docker-compose-v2 -y # 

#Download and install Nvidia's Docker container kit
curl -fsSL https://nvidia.github.io/libnvidia-container/gpgkey | sudo gpg --dearmor -o /usr/share/keyrings/nvidia-container-toolkit-keyring.gpg
curl -s -L https://nvidia.github.io/libnvidia-container/stable/deb/nvidia-container-toolkit.list | \
  sed 's#deb https://#deb [signed-by=/usr/share/keyrings/nvidia-container-toolkit-keyring.gpg] https://#g' | \
  sudo tee /etc/apt/sources.list.d/nvidia-container-toolkit.list
#  Fix the Architecture Variable
sudo sed -i "s/\$(ARCH)/$(dpkg --print-architecture)/g" /etc/apt/sources.list.d/nvidia-container-toolkit.list


# Enable GPU access for Docker
sudo apt update && sudo apt install -y nvidia-container-toolkit
sudo nvidia-ctk runtime configure --runtime=docker
sudo systemctl restart docker # sudo systemctl enable --now docker

# Add user to docker group (avoid sudo for docker commands)
sudo usermod -aG docker $USER
newgrp docker # "shortcut" that refreshes group memberships for current session so you don't have to log out
</pre>


<h2 class="page-header">Phase 3: Dockerizing the AI Stack</h2>

<p>Create a folder called <code>ai-server</code> and create a file named <code>docker-compose.yml</code>.</p>
<p>This is boss-mode. Use this for maximum flexibility:</p>

<pre>
# Directory Structure
mkdir -p ~/ai-stack/mvp/data/ollama ~/ai-stack/mvp/data/webui
cd ~/ai-stack/mvp && nano docker-compose.yml
</pre>

<p>Next create and save the following docker file:</p>
<pre>
# File path: ~/ai-stack/mvp/docker-compose.yml
services:
  ollama:
    image: ollama/ollama:latest
    container_name: ollama
#  Useful for external access (e.g., for API calls from outside the Docker network).
#  When used internally (e.g., by Open WebUI), not exposing the port is more secure.
#    ports:
#      - "11434:11434"
    pull_policy: always # disable for production
    restart: unless-stopped
    tty: true # for add'l debug info
    volumes:
      - ./ollama_data:/root/.ollama
    deploy:
      resources:
        reservations:
          devices:
            - driver: nvidia
              count: 1  # may also set to 'all' too
              capabilities: [gpu]

  open-webui:
    image: ghcr.io/open-webui/open-webui:main
    container_name: open-webui
    ports:
      - "3000:8080"
    environment:
      - OLLAMA_BASE_URL=http://ollama:11434
      #- RAG_EMBEDDING_ENGINE=ollama
      #- RAG_EMBEDDING_MODEL=nomic-embed-text:latest # Consider changing this for Part 2 or 3
      #- ENABLE_RAG_HYBRID_SEARCH=True
# If you need super strong security.  
#     - WEBUI_SECRET_KEY=change_me_to_random_string
# bypass login page for a single-user setup, set the WEBUI_AUTH environment variable to False
      - WEBUI_AUTH=False
    volumes:
      - ./webui_data:/app/backend/data
    depends_on:
      - ollama
    restart: unless-stopped
</pre>


<p>Run: <code>docker compose up -d</code> then download Qwen3 with: <pre>docker exec -it ollama ollama pull qwen3:30b-a3b-instruct-2507-q4_K_M</pre> and the embedding model with <pre>docker exec ollama ollama pull nomic-embed-text</pre></p> 

<small><strong>NB</strong>: Docker compose commands like <code>docker compose up</code> or <code>docker compose down</code> must be run from the directory where your docker-compose.yml file is located.</small>

<h4 class="page-header">What This Command Does</h4>
 
        <ol class="indent">
	<li><code>docker exec -it ollama</code>: Runs a command inside the ollama container.<code>-it</code> spawns an interactive session that shows download progress.</li>

	<li><code>ollama pull qwen3:30b-q4_K_M</code>: Downloads the Qwen3 30B model with q4_K_M quantization</li>
</ol>

    

<h3 class="page-header">Quantized Alternatives of the Same LLM: Lower VRAM Usage with More Accuracy</h3>

<p>Alternatively you may download Unsloth&#8217;s version: <code>sudo docker exec -it ollama ollama pull hf.co/unsloth/Qwen3-30B-A3B-Instruct-2507-GGUF:UD-Q4_K_XL</code>.</p>

<p>The disadvantage is that Unsloth does not update its quants for new updates of the same model.   For example Quen3-30b-A3-instrcut was updated in September 2025. But, Unsloth still uses the July 2025 version for its quant.</p>

<p>Unsloth quantization algo (UD-Q4_K_XL) results in smaller (17GB vs 18GB) but more accurate model than the standard (e.g., Q4_K_M). See Daniel Han&#8217;s reply <a href="https://www.reddit.com/r/LocalLLaMA/comments/1f92brm/comment/llkr7ia">here</a>.</p>

<small><strong>NB:</strong> The open source LLM community often uses the word &#8220;quant&#8221; as shorthand for <em>quantization</em> (the process) and <em>quantization algorithm</em> (the method or algo type: Q4_K_M, Q5_1, UD-Q4_K_XL). For example: <em>I’m using the Q4_K_M quant for this model.</em> or <em>Unsloth’s UD quants are optimized for accuracy.</em></small>

<p>GLM-4.7-flash will be used in part 2. However, You may download Unsloth&#8217;s version with: <code>docker exec -it ollama ollama pull hf.co/unsloth/GLM-4.7-Flash-GGUF:UD-Q4_K_XL</code>.</p>


<h3 class="page-header">Model Housekeeping</h3>

<h4 class="page-header">Docker Commands</h4>
 
        <ol class="indent">

	<li><strong>View downloaded models</strong>: <pre>docker exec -it ollama ollama list</pre></li>

	<li><strong>List downloaded models:</strong> <pre>docker exec -it ollama ollama list</pre></li>

	<li><strong>Delete a model:</strong> <pre>docker exec -it ollama ollama rm <model-name></pre></li>

	<li>check logs: docker compose log</li>

</ol>

    
<!--

docker exec -it ollama ollama ps
docker logs open-webui
docker stats
docker exec -it open-webui sh
docker logs open-webui | grep -i "upload\|rag\|embedding"

watch -n 1 nvidia-smi
WebUI Python packages: docker exec -it open-webui pip list
Files uploaded via WebUI: ls -al ai-stack/mvp/web*/up*
Loaded models: ls -al ai*/mvp/ollama_data/models/manifests/r*/l*/q*

find / -name "*MiniLM*" 2>/dev/null
ls -al /app/backend/data/cache/embedding/models/
-->



<h4 class="page-header">Security Notes</h4>
 
        <ol class="indent">
	<li><strong>Docker security:</strong> Running containers as root (default) is a security risk. Consider adding a non-root user in the Dockerfile or using user: &#8220;1000:1000&#8221; in docker-compose.yml</li>

	<li><strong>Open WebUI auth:</strong> For production, strongly recommend enabling WEBUI_AUTH and setting a WEBUI_SECRET_KEY</li>

</ol>

    


<h2 class="page-header">Phase 4: MVP of RAG On Uploaded Documents</h3>

<p>RAG functionality is built into webUI.  No additional action needed for now.</p>

<!--

(the process of mapping a large set of values to a smaller set, often used in the context of reducing the precision of model weights)
<p>Ollama uses uniform quantization (e.g., all layers at Q4_0, Q5_1, etc.). On the other hand, Unsloth uses a custom quantization algo call <em>Dynamic 2.0</em> (UD-Q4_K_XL). It uses selective, layer-by-layer quantization, optimizing each layer differently producing higher accuracy per layer for any given layer size.</p>

Unsloth's "Dynamic 2.0 GGUF" approach is different from standard quantization methods like those used by Ollama. 
**Key Differences:**
- **Selective, Layer-by-Layer Quantization:** Unlike Ollama, which typically applies a uniform quantization (e.g., Q4_0) across all layers of a model, Unsloth’s Dynamic 2.0 GGUFs use a "selective layer quantization" approach. This means each layer of the model is analyzed and quantized differently—some layers may use more aggressive quantization (e.g., 2-bit or even 1-bit), while others use less aggressive (e.g., 4-bit or 5-bit), depending on their importance and sensitivity to quantization errors. This is done to maximize efficiency and accuracy simultaneously.
- **Dynamic Calibration:** Unsloth uses a custom calibration dataset and framework to determine the optimal quantization for each layer. This process is model-specific, meaning the quantization profile for Llama 4 will differ from Gemma 3, for example.
- **Performance and Compatibility:** The resulting GGUF files are compatible with standard inference engines like llama.cpp and Ollama, but they are optimized for better accuracy and smaller size compared to uniform quantization. Unsloth’s method aims to minimize the performance gap between full-precision and quantized models, often outperforming other quantization methods on benchmarks like MMLU and KL Divergence.

**In Summary:**
- Ollama’s GGUFs are typically uniformly quantized (e.g., all layers at Q4_0).
- Unsloth’s Dynamic 2.0 GGUFs are selectively quantized per layer, using a mix of quantization levels (e.g., 1-bit, 2-bit, 4-bit, etc.) to balance size, speed, and accuracy.
- Both produce GGUF files, but Unsloth’s approach is more nuanced and often more performant for the same or smaller file size.


2. The "Two Locations" Mystery
the documents existing in two places. This is actually how Docker Volumes work:

Physical Server Folder: This is your "source of truth" on your 3090 rig (where your docker-compose.yml lives).

Container App Folder: This is the "window" through which the Docker container sees those same files.

The Reality: They aren't actually two different sets of files; they are the same data being shared through a "portal". If you delete a file in one, it disappears from the other.


--->




<!--
<h3 class="page-header">Phase 4: Orchestration (Master Control)</h3>

<p>Dify is the UI and "Agent" manager. </p>
<pre>
mkdir -p ~/ai-stack/production && cd ~/ai-stack/production
git clone https://github.com/langgenius/dify.git
cd dify/docker
cp .env.example .env
docker compose up -d
</pre>
-->
<hr>
<h3>Resources and References</h3>
https://huggingface.co/collections/Qwen/qwen3
https://huggingface.co/Qwen/Qwen3-Omni-30B-A3B-Instruct
https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507
https://ollama.com/library/qwen3/tags
https://ollama.com/library/qwen3:30b-a3b-instruct-2507-q4_K_M

https://documentation.ubuntu.com/server/how-to/graphics/install-nvidia-drivers/#installing-the-drivers-on-servers-and-or-for-computing-purposes
https://docs.ollama.com/linux#install

https://docs.openwebui.com/getting-started/quick-start/
https://docs.openwebui.com/getting-started/quick-start/starting-with-ollama/
https://community.home-assistant.io/t/my-journey-to-a-reliable-and-enjoyable-locally-hosted-voice-assistant/944860


<!--
<h3 class="page-header">Phase 3: The LLM Engine (Neural Network)</h3>


<p>This is suboptimal and we suggest you install using docker for modularity</p>
<pre>
curl -fsSL https://ollama.com/install.sh | sh
ollama run qwen3:30b-a3b-instruct-2507-q4_K_M
</pre>

 <code>docker compose -f ~/ai-stack/docker-compose.yml up -d</code></p>

<pre>
# 1. Create the project directory
mkdir ~/ai-stack

# 2. Move into that directory
cd ~/ai-stack

# 3. Create the file here
nano docker-compose.yml</pre>
-->]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/agentic-ai-server-build-tutorial-part-i-the-mvp/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>LLM Quantization</title>
		<link>https://codesport.io/artificial-intelligence/llm-quantization/</link>
					<comments>https://codesport.io/artificial-intelligence/llm-quantization/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Mon, 02 Feb 2026 04:00:18 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7732</guid>

					<description><![CDATA[The purpose of quantization is reducing VRAM requirements while maintaining LLM accuracy and inference speed. For example, the compression algorithm NF4 (Q4-NF4) is a model best suited for models with 14B or more parameters. In LLM quantization, the &#8220;bits&#8221; (e.g., 4-bit, 8-bit) represent the precision used to store model weights, determining memory usage and computational [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>The purpose of quantization is reducing VRAM requirements while maintaining LLM accuracy and inference speed. For example, the compression algorithm <em>NF4 (Q4-NF4)</em> is a model  best suited for models with 14B or more parameters. </p>

<p>In LLM quantization, the &#8220;bits&#8221; (e.g., 4-bit, 8-bit) represent the precision used to store model weights, determining memory usage and computational efficiency. 4-bit quantization reduces memory by ~75% compared to 16-bit (roughly 0.5 bytes per parameter vs 2 bytes), enabling smaller hardware to run large models with minimal accuracy loss (<1–2%).</p>

<p>Common options for GGUF are in the table below.</p>

<style>
  .modern-table-container {
    font-family: 'Segoe UI', Tahoma, Geneva, Verdana, sans-serif;
    margin: 20px 0;
    overflow-x: auto;
    border-radius: 8px;
    box-shadow: 0 4px 15px rgba(0, 0, 0, 0.1);
  }
  .modern-table {
    width: 100%;
    border-collapse: collapse;
    background-color: #ffffff;
    text-align: left;
  }
  .modern-table th {
    background-color: #2c3e50;
    color: #ffffff;
    padding: 12px 15px;
    font-weight: 600;
    text-transform: uppercase;
    font-size: 1em;
    letter-spacing: 0.05em;
  }
  .modern-table td {
    padding: 12px 15px;
    border-bottom: 1px solid #edf2f7;
    color: #4a5568;
    font-size: 1em;
  }
  .modern-table tbody tr:nth-of-type(even) {
    background-color: #f8fafc;
  }
  .modern-table tbody tr:hover {
    background-color: #edf2f7;
    transition: background-color 0.2s ease;
  }
  .badge {
    display: inline-block;
    padding: 4px 8px;
    border-radius: 4px;
    font-weight: bold;
    font-size: .9em;
    background-color: #e2e8f0;
    color: #2d3748;
  }
</style>

<div class="modern-table-container">
  <table class="modern-table">
    <thead>
      <tr>
        <th>Algorithm</th>
        <th>Bits</th>
        <th>Pros</th>
        <th>Cons</th>
        <th>Notes</th>
      </tr>
    </thead>
    <tbody>
      <tr>
        <td><strong>Q4_0</strong></td>
        <td><span class="badge">4-bit</span></td>
        <td>Very small VRAM footprint, fast</td>
        <td>Slightly lower accuracy on reasoning / math</td>
        <td>Default for most local GPU setups</td>
      </tr>
      <tr>
        <td><strong>Q4_1</strong></td>
        <td><span class="badge">4-bit</span></td>
        <td>Slightly better accuracy than Q4_0</td>
        <td>Slightly more VRAM</td>
        <td>Good for agentic reasoning</td>
      </tr>
      <tr>
        <td><strong>NF4 (Q4-NF4)</strong></td>
        <td><span class="badge">4-bit</span></td>
        <td>Optimized for FP16-like performance</td>
        <td>Requires GGUF model converted for NF4</td>
        <td>Recommended for larger models (14B)</td>
      </tr>
      <tr>
        <td><strong>Q5_0 / Q5_1</strong></td>
        <td><span class="badge">5-bit</span></td>
        <td>Better accuracy; still small</td>
        <td>Higher VRAM</td>
        <td>Only if you want slightly higher quality reasoning</td>
      </tr>
      <tr>
        <td><strong>FP16 / FP32</strong></td>
        <td><span class="badge">16/32-bit</span></td>
        <td>Max quality</td>
        <td>VRAM-heavy</td>
        <td>Not practical for 14B+ on RTX 3090</td>
      </tr>
    </tbody>
  </table>
</div>]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/llm-quantization/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Apple Silicon Security Architecture &#038; Physical Layout</title>
		<link>https://codesport.io/security/apple-silicon-security-architecture-physical-layout/</link>
					<comments>https://codesport.io/security/apple-silicon-security-architecture-physical-layout/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Sat, 10 Jan 2026 23:39:48 +0000</pubDate>
				<category><![CDATA[Security]]></category>
		<category><![CDATA[security]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7674</guid>

					<description><![CDATA[🏗️ The SoC vs. SSD Relationship The diagram below clarifies that the Operating System (Kernel) is stored on the physical NAND (SSD), but is managed by a controller inside the SoC. +-----------------------------------------------------------+ +----------------+ &#124; APPLE SILICON (SoC) &#124; &#124; INTERNAL SSD &#124; &#124; &#124; &#124; (NAND) &#124; &#124; +-----------------------+ +---------------------+ &#124; &#124; &#124; &#124; &#124; [&#8230;]]]></description>
										<content:encoded><![CDATA[<h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f3d7.png" alt="🏗" class="wp-smiley" style="height: 1em; max-height: 1em;" /> The SoC vs. SSD Relationship</h2>
<p>The diagram below clarifies that the Operating System (Kernel) is stored on the physical NAND (SSD), but is managed by a controller inside the SoC.</p>
<pre>
+-----------------------------------------------------------+          +----------------+
|                   APPLE SILICON (SoC)                     |          |  INTERNAL SSD  |
|                                                           |          |     (NAND)     |
|  +-----------------------+       +---------------------+  |          |                |
|  | Application Processor |       |   Secure Enclave    |  |          | +------------+ |
|  |    (The Main CPU)     |       |      (SEP)          |  |          | | iBoot (S2) | |
|  |                       |       |                     |  |  Secure  | +------------+ |
|  |  +-----------------+  |Mailbox|  +---------------+  |  |  Channel | | XNU Kernel | |
|  |  |  XNU (In RAM)   | &lt;=======&gt;|  |    sepOS      |  | &lt;---------&gt; | +------------+ |
|  |  +-----------------+  |       |  +---------------+  |  |          | | User Data  | |
|  |  |  App Sandbox    |  |       |  | AES Crypto Eng|  |  |          | +------------+ |
|  +-----------------------+       +---------------------+  |          +----------------+
|              ^                              ^             |           Physical NAND
|              |                              |             |         (Encrypted at Rest)
|      +---------------+              +---------------+     |
|      |   Boot ROM    |              |   UID Key     |     |
|      | (Root/Trust)  |              | (Silicon Fuse)|     |
|      +---------------+              +---------------+     |
+-----------------------------------------------------------+
</pre>
<center><p class="center"><strong>Figure 1:</strong> Apple Silicon Security Architecture &amp; Physical Layout</p></center>

<h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f511.png" alt="🔑" class="wp-smiley" style="height: 1em; max-height: 1em;" /> The &#8220;UID&#8221; Confusion Resolved</h2>
<p>The UID (Unique ID) is a physical AES-256 key fused into the SoC during manufacturing. It is the &#8220;Master Secret&#8221; of the chip.</p>

<p>Why is it &#8220;randomly&#8221; there? It provides Hardware-Bound Encryption. Because the SSD data is encrypted using a key derived from this specific UID, you cannot desolder the SSD chips and read them on another Mac. The data is cryptographically married to that specific processor.</p>

<h3 class="page-header">The Workflow</h3>

 
        <ol class="indent">

 	<li>The SEP takes the UID.</li>
 	<li>It combines it with your Passcode.</li>
 	<li>It generates a KEK (Key Encryption Key).</li>
 	<li>This KEK is sent to the AES Engine to decrypt the SSD.</li>
</ol>

    
<h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f6e1.png" alt="🛡" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Physical Boot Sequence (The &#8220;Hand-off&#8221;)</h2>
 
        <ol class="indent">
 	<li><strong>Stage 1 (Hardware):</strong> The Boot ROM (inside the chip) starts. It knows nothing about the SSD yet except how to talk to the NVMe controller.</li>
 	<li><strong>Stage 2 (Loading)</strong>: Boot ROM reads iBoot from the SSD. It checks the signature. If valid, it loads iBoot into the SoC&#8217;s internal RAM.</li>
 	<li><strong>Stage 3 (Kernel): </strong>iBoot now looks for the XNU Kernel on the SSD.</li>
 	<li><strong>Stage 4 (Execution):</strong> iBoot verifies the Kernel&#8217;s signature, loads it into the system&#8217;s main RAM, and the OS begins to run.</li>
</ol>

    
<h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f4c2.png" alt="📂" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Summary of Storage Locations</h2>


<div style="overflow-x:auto; font-family: -apple-system, BlinkMacSystemFont, 'Segoe UI', Roboto, Helvetica, Arial, sans-serif;">
    <table style="width:100%; border-collapse: collapse; background-color: #1e1e1e; color: #ffffff; border-radius: 8px; overflow: hidden; box-shadow: 0 4px 15px rgba(0,0,0,0.3);">
        <thead>
            <tr style="background-color: #2d2d2d; text-align: left;">
                <th style="padding: 15px; border-bottom: 2px solid #3d3d3d; font-weight: 600; color: #ffcc00;">Component</th>
                <th style="padding: 15px; border-bottom: 2px solid #3d3d3d; font-weight: 600; color: #ffcc00;">Physical Location</th>
                <th style="padding: 15px; border-bottom: 2px solid #3d3d3d; font-weight: 600; color: #ffcc00;">Identity / Role</th>
            </tr>
        </thead>
        <tbody>
            <tr style="border-bottom: 1px solid #333;">
                <td style="padding: 15px; font-weight: 500;">Boot ROM</td>
                <td style="padding: 15px; color: #cccccc;">Inside SoC (Silicon)</td>
                <td style="padding: 15px; color: #cccccc;">Immutable Root of Trust.</td>
            </tr>
            <tr style="border-bottom: 1px solid #333; background-color: #252525;">
                <td style="padding: 15px; font-weight: 500;">UID Key</td>
                <td style="padding: 15px; color: #cccccc;">Inside SoC (Fuses)</td>
                <td style="padding: 15px; color: #cccccc;">The silicon&#8217;s unique cryptographic thumbprint.</td>
            </tr>
            <tr style="border-bottom: 1px solid #333;">
                <td style="padding: 15px; font-weight: 500;">Application Processor</td>
                <td style="padding: 15px; color: #cccccc;">Inside SoC (Logic)</td>
                <td style="padding: 15px; color: #cccccc;"><strong>The Main CPU</strong> (Runs macOS and apps).</td>
            </tr>
            <tr style="border-bottom: 1px solid #333; background-color: #252525;">
                <td style="padding: 15px; font-weight: 500;">iBoot / Kernel</td>
                <td style="padding: 15px; color: #cccccc;">On the SSD</td>
                <td style="padding: 15px; color: #cccccc;">The software that manages the hardware.</td>
            </tr>
            <tr>
                <td style="padding: 15px; font-weight: 500;">Biometrics</td>
                <td style="padding: 15px; color: #cccccc;">SEP Protected RAM</td>
                <td style="padding: 15px; color: #cccccc;">Isolated fingerprint/face data storage.</td>
            </tr>
        </tbody>
    </table>
</div>]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/security/apple-silicon-security-architecture-physical-layout/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Protected: The AI Agent Stack</title>
		<link>https://codesport.io/artificial-intelligence/the-ai-agent-stack/</link>
					<comments>https://codesport.io/artificial-intelligence/the-ai-agent-stack/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Tue, 28 Oct 2025 02:26:56 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7496</guid>

					<description><![CDATA[There is no excerpt because this is a protected post.]]></description>
										<content:encoded><![CDATA[<form action="https://codesport.io/wp-login.php?action=postpass" class="post-password-form" method="post"><input type="hidden" name="redirect_to" value="https://codesport.io/artificial-intelligence/the-ai-agent-stack/" />
	<p>This content is password protected. To view it please enter your password below:</p>
	<p><label for="pwbox-7496">Password: <input name="post_password" id="pwbox-7496" type="password" spellcheck="false" required size="20" /></label> <input type="submit" name="Submit" value="Enter" /></p></form>
	]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/the-ai-agent-stack/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Quantitative Finance and Numerical Methods</title>
		<link>https://codesport.io/quantitative-finance/quantitative-finance-and-numerical-methods/</link>
					<comments>https://codesport.io/quantitative-finance/quantitative-finance-and-numerical-methods/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Tue, 28 Oct 2025 01:08:06 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Quantitative Finance]]></category>
		<category><![CDATA[python]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7651</guid>

					<description><![CDATA[Time series analysis, regression, and PCA can be considered types of machine learning, while hypothesis testing and cohort analysis are statistical methods that support and inform machine learning processes. The distinction lies in whether the technique is primarily used to train a predictive model (ML) or to analyze and interpret data (statistical) Types of machine [&#8230;]]]></description>
										<content:encoded><![CDATA[<p>Time series analysis, regression, and PCA can be considered types of machine learning, while hypothesis testing and cohort analysis are statistical methods that support and inform machine learning processes. The distinction lies in whether the technique is primarily used to train a predictive model (ML) or to analyze and interpret data (statistical)</p>

<h2 class="page-header">Types of machine learning</h2>
 
        <ul class="indent">
<li><b>Time Series Analysis:</b> When used for forecasting future values based on historical, time-ordered data, it is a type of supervised machine learning. Some of the most effective methods for time series forecasting are machine learning models, including neural networks and gradient boosting frameworks.</li>
<li><b>Principal Component Analysis (PCA):</b> This is a key technique in unsupervised machine learning. PCA is a dimensionality reduction method that finds new, uncorrelated variables (principal components) that capture the most variance in a dataset. It is often used to preprocess data for other machine learning algorithms by reducing model complexity and preventing overfitting.</li>
<li><b>Regression:</b> This is a fundamental supervised machine learning technique. It predicts a continuous numerical output based on input features and works by estimating the relationship between variables. A regression algorithm is used to train a machine learning model, which learns to make predictions based on labeled data.</li>
</ul>

    

<h2 class="page-header">Statistical methods that support machine learning</h2>
 
        <ul class="indent">
<li><b>Hypothesis Testing:</b> This is a statistical method used to evaluate assumptions about a population based on sample data. In machine learning, hypothesis testing can be used to compare different models, evaluate the significance of features, and validate assumptions about data distribution, but it is not a machine learning technique itself.</li>
<li><b>Cohort Analysis:</b> This is a statistical and analytical technique for studying the behavior of different groups (cohorts) over time. While it provides insights that can inform machine learning models (e.g., identifying valuable customer segments), it is a form of descriptive or diagnostic analysis, not a machine learning technique.</li>
</ul>

    

<h2 class="page-header">Quantitative Finance and Numerical Methods</h2>
 
        <ul class="indent">
<li><b>Used diffusive stochastic processes (including geometric Brownian motion with local volatility) and numerical methods including: Monte Carlo, finite difference methods) to estimate the fair value of equity derivatives</b></li>
<ul>
<li><b>Diffusive stochastic processes:</b> These are mathematical models used in quantitative finance to simulate the random movement of asset prices over time.</li>
<li><b>Geometric Brownian Motion (GBM):</b> A specific type of diffusive stochastic process commonly used in the Black-Scholes model for option pricing.</li>
<li><b>Local Volatility:</b> A more advanced model that allows the volatility of an asset to change over time and with respect to the asset&#8217;s price level.</li>
<li><b>Monte Carlo Simulation:</b> A numerical technique that uses random sampling to model the behavior of financial assets and find approximate solutions to complex problems, like valuing options.</li>
<li><b>Finite Difference Methods:</b> Numerical methods for solving differential equations, which are fundamental to pricing derivatives.</li>
</ul>
<li><b>Analyzed risk on a portfolio of options using stochastic volatility models (Local Volatility, Local-Stochastic Volatility)</b></li>
<ul>
<li><b>Stochastic Volatility Models:</b> These are used to model the fact that the volatility of an asset, like an option, is not constant but changes over time in a random way.</li>
</ul>
</ul>

    

<h2 class="page-header">Machine Learning</h2>
 
        <ul class="indent">
<li><b>Machine learning methods (regularization, clustering algorithms) applied to financial data sets</b></li>
<ul>
<li><b>Regularization:</b> Techniques like L1 (Lasso) and L2 (Ridge) that are used to prevent overfitting in a machine learning model by penalizing large coefficients.</li>
<li><b>Clustering Algorithms:</b> Unsupervised machine learning methods, such as K-means or DBSCAN, used to group similar data points together based on their features.</li>
</ul>
<li><b>Predictive modeling:</b> The process of using statistical and machine learning techniques to forecast future outcomes.</li>
<li><b>Classification:</b> A supervised machine learning task for predicting a discrete class label (e.g., spam or not spam) for a given data point.</li>
<li><b>Feature Engineering:</b> The process of creating new input features from existing ones to improve the performance of machine learning models.</li>
</ul>

    

<h2 class="page-header">Statistical Methods</h2>
 
        <ul class="indent">
<li><b>Applied statistical methods (including Principal Component Analysis (PCA), linear and logistic regression, time series analysis) to financial data sets</b></li>
<ul>
<li><b>Principal Component Analysis (PCA):</b> A statistical and unsupervised machine learning technique used for dimensionality reduction.</li>
<li><b>Linear Regression:</b> A supervised learning method used to model the relationship between a dependent variable and one or more independent variables.</li>
<li><b>Logistic Regression:</b> A supervised learning method used for classification problems where the goal is to predict a categorical outcome.</li>
<li><b>Time Series Analysis:</b> The analysis of time-stamped data to identify patterns, trends, and seasonality, often for forecasting.</li>
</ul>
<li><b>Used std deviation (Ỽ), variance (Ỽ2), co-variance, correlation (r-value), linear regression, Q-Q Plots, hypothesis tests, p-value, cohort analysis to inform investment decisions from data</b></li>
<ul>
<li><b>Descriptive Statistics:</b>
<ul>
<li><b>Standard Deviation, Variance, Co-variance, and Correlation:</b> Measures of data dispersion and the relationship between variables.</li>
</ul>
</li>
<li><b>Inferential Statistics:</b>
<ul>
<li><b>Hypothesis Tests and p-value:</b> Used to make inferences about a population from sample data and determine the statistical significance of results.</li>
<li><b>Q-Q Plots (Quantile-Quantile Plots):</b> Used to check if a dataset follows a particular theoretical distribution, such as a normal distribution.</li>
</ul>
</li>
<li><b>Analytical Technique:</b>
<ul>
<li><b>Cohort Analysis:</b> A method for studying the behavior of a group of people (cohort) over a period of time.</li>
</ul>
</li>
</ul>
</ul>

    

<h2 class="page-header">Programming and Data Analysis Tools</h2>
 
        <ul class="indent">
<li><b>Python (Pandas, NumPy, SciPy, Matplotlib, Seaborn, PyTorch), SQL, R, Excel VBA, Bash</b></li>
<ul>
<li><b>Python Libraries (Pandas, NumPy, SciPy, etc.):</b> Packages used for data manipulation, scientific computing, visualization, and machine learning.</li>
<li><b>SQL:</b> A language used for managing data in relational databases.</li>
<li><b>R:</b> A programming language and environment specifically designed for statistical computing and graphics.</li>
<li><b>Excel VBA:</b> A programming language for creating macros in Microsoft Excel.</li>
<li><b>Bash:</b> A command-line shell used for interacting with an operating system.</li>
</ul>
</ul>

    
]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/quantitative-finance/quantitative-finance-and-numerical-methods/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Python Frameworks &#038; Libraries for RAG and Agentic AI</title>
		<link>https://codesport.io/artificial-intelligence/python-frameworks-libraries-for-rag-and-agentic-ai/</link>
					<comments>https://codesport.io/artificial-intelligence/python-frameworks-libraries-for-rag-and-agentic-ai/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Mon, 27 Oct 2025 23:38:08 +0000</pubDate>
				<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[python]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7648</guid>

					<description><![CDATA[Here&#8217;s what&#8217;s actually being used in production right now: 🤖 Agentic AI Frameworks (Orchestration &#038; Multi-Agent Systems) Tier 1 &#8211; Production-Ready (Most Common) LangChain / LangGraph (Industry Standard) CrewAI (Industry Standard) AutoGen (Microsoft) (Production-Ready) Haystack (Production-Ready) Tier 2 &#8211; Emerging/Specialized smolagents (Hugging Face) (Emerging) Google ADK (Agent Development Kit) (Emerging) Semantic Kernel (Microsoft) (Emerging) LlamaIndex [&#8230;]]]></description>
										<content:encoded><![CDATA[<article>
 
  <p>Here&#8217;s what&#8217;s actually being used in production right now:</p>

  <h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f916.png" alt="🤖" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Agentic AI Frameworks (Orchestration &#038; Multi-Agent Systems)</h2>

  <h3 class="page-header">Tier 1 &#8211; Production-Ready (Most Common)</h3>

  <h4 class="page-header">LangChain / LangGraph (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Most popular framework for building AI agents</li>
    <li>LangChain = basic chains and agents</li>
    <li>LangGraph = advanced multi-agent workflows with state management</li>
  </ul>

    

  <h4 class="page-header">CrewAI (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Specialized for multi-agent collaboration</li>
    <li>Agents work together like a &#8220;crew&#8221; with roles</li>
  </ul>

    

  <h4 class="page-header">AutoGen (Microsoft) (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Multi-agent conversations</li>
    <li>Good for code generation tasks</li>
    <li>Rising popularity</li>
  </ul>

    

  <h4 class="page-header">Haystack (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Originally for search, now strong in RAG + agents</li>
    <li>Good for production deployments</li>
  </ul>

    

  <h3 class="page-header">Tier 2 &#8211; Emerging/Specialized</h3>

  <h4 class="page-header">smolagents (Hugging Face) (Emerging)</h4>
   
        <ul class="indent">
    <li>Lightweight agent framework</li>
  </ul>

    

  <h4 class="page-header">Google ADK (Agent Development Kit) (Emerging)</h4>
   
        <ul class="indent">
    <li>Google&#8217;s agent framework</li>
  </ul>

    

  <h4 class="page-header">Semantic Kernel (Microsoft) (Emerging)</h4>
   
        <ul class="indent">
    <li>Enterprise-focused agent framework</li>
    <li>C# and Python support</li>
  </ul>

    

  <h4 class="page-header">LlamaIndex Workflows</h4>
   
        <ul class="indent">
    <li>Evolved from RAG-only to include agentic capabilities</li>
  </ul>

    

  <h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f4da.png" alt="📚" class="wp-smiley" style="height: 1em; max-height: 1em;" /> RAG Frameworks &#038; Libraries</h2>

  <h3 class="page-header">Core RAG Orchestration</h3>

  <h4 class="page-header">LangChain (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Most popular for RAG pipelines</li>
    <li>Document loaders, splitters, retrievers</li>
  </ul>

    

  <h4 class="page-header">LlamaIndex (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Purpose-built for RAG</li>
    <li>Better for complex document structures</li>
  </ul>

    

  <h4 class="page-header">Haystack (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Enterprise RAG pipelines</li>
    <li>Good for production deployments</li>
  </ul>

    

  <h3 class="page-header">Vector Databases (Storage &#038; Retrieval)</h3>

  <h4 class="page-header">ChromaDB (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Most popular embedded vector DB</li>
  </ul>

    

  <h4 class="page-header">pgvector (PostgreSQL extension) (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Vector search in PostgreSQL</li>
  </ul>

    

  <h4 class="page-header">Pinecone (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Managed vector DB (cloud)</li>
    <li>Popular in production</li>
  </ul>

    

  <h4 class="page-header">Weaviate (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Open-source vector DB</li>
    <li>Good for hybrid search</li>
  </ul>

    

  <h4 class="page-header">Milvus (Emerging)</h4>
   
        <ul class="indent">
    <li>Scalable vector DB</li>
    <li>Good for large deployments</li>
  </ul>

    

  <h4 class="page-header">FAISS (Facebook AI Similarity Search) (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Vector similarity search library</li>
    <li>Not a full DB, but very fast</li>
  </ul>

    

  <h3 class="page-header">Document Processing &#038; Chunking</h3>

  <h4 class="page-header">LangChain Document Loaders (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Handles PDFs, Word, HTML, etc.</li>
  </ul>

    

  <h4 class="page-header">Unstructured (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Advanced document parsing</li>
    <li>Handles complex layouts</li>
  </ul>

    

  <h4 class="page-header">PyPDF2 / pypdf (Production-Ready)</h4>
   
        <ul class="indent">
    <li>PDF text extraction</li>
  </ul>

    

  <h4 class="page-header">python-docx (Emerging)</h4>
   
        <ul class="indent">
    <li>Word document processing</li>
  </ul>

    

  <h4 class="page-header">BeautifulSoup4 (Production-Ready)</h4>
   
        <ul class="indent">
    <li>HTML parsing (for web scraping)</li>
  </ul>

    

  <h3 class="page-header">Embedding Models</h3>


<p>Embedding models are a specialized type of machine learning algorithm designed to translate complex, unstructured data into numerical vectors. These &#8220;vectors&#8221; act as a list of numbers that represent the underlying meaning and relationships of the input data</p>

  <h4 class="page-header">sentence-transformers (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Create embeddings locally</li>
    <li>Hugging Face models</li>
  </ul>

    

  <h4 class="page-header">OpenAI embeddings (Industry Standard)</h4>
   
        <ul class="indent">
    <li>Via API (text-embedding-ada-002, text-embedding-3-small/large)</li>
  </ul>

    

  <h4 class="page-header">Cohere embeddings (Production-Ready)</h4>
   
        <ul class="indent">
    <li>Via API</li>
  </ul>

    

  <h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f50d.png" alt="🔍" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Evaluation &#038; Monitoring</h2>

  <h4 class="page-header">LangSmith (Industry Standard)</h4>
   
        <ul class="indent">
    <li>LangChain&#8217;s evaluation platform</li>
  </ul>

    

  <h4 class="page-header">TruLens (Production-Ready)</h4>
   
        <ul class="indent">
    <li>RAG evaluation and observability</li>
  </ul>

    

  <h4 class="page-header">Ragas (Production-Ready)</h4>
   
        <ul class="indent">
    <li>RAG evaluation framework</li>
    <li>Measures faithfulness, relevance, etc.</li>
  </ul>

    

  <h4 class="page-header">Phoenix (Arize AI) (Emerging)</h4>
   
        <ul class="indent">
    <li>LLM observability</li>
  </ul>

    

  <h4 class="page-header">Weights &#038; Biases (Production-Ready)</h4>
   
        <ul class="indent">
    <li>ML experiment tracking</li>
  </ul>

    

  <h2 class="page-header"><img src="https://s.w.org/images/core/emoji/15.1.0/72x72/1f6e0.png" alt="🛠" class="wp-smiley" style="height: 1em; max-height: 1em;" /> Supporting Libraries</h2>

  <h3 class="page-header">Core Python for AI</h3>
   
        <ul class="indent">
    <li><strong>NumPy</strong> &#8211; Array operations</li>
    <li><strong>Pandas</strong> &#8211; Data manipulation</li>
    <li><strong>Pydantic</strong> &#8211; Data validation (important for agents!)</li>
    <li><strong>asyncio</strong> &#8211; Async operations (important for agents)</li>
  </ul>

    

  <h3 class="page-header">API &#038; Integration</h3>
   
        <ul class="indent">
    <li><strong>requests</strong> &#8211; HTTP requests</li>
    <li><strong>httpx</strong> &#8211; Async HTTP</li>
    <li><strong>FastAPI</strong> &#8211; Building APIs for agents</li>
    <li><strong>Gradio / Streamlit</strong> &#8211; Quick UIs for demos</li>
  </ul>

    

  <h3 class="page-header">Prompt Engineering</h3>
   
        <ul class="indent">
    <li><strong>promptify</strong> &#8211; Prompt templates</li>
    <li><strong>guidance</strong> (Microsoft) &#8211; Structured LLM outputs</li>
  </ul>

    
</article>]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/artificial-intelligence/python-frameworks-libraries-for-rag-and-agentic-ai/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
		<item>
		<title>Biotech Foundations: Modality, Delivery, Disease Target</title>
		<link>https://codesport.io/biotech/biotech-foundational-concepts-modality-delivery-disease-target/</link>
					<comments>https://codesport.io/biotech/biotech-foundational-concepts-modality-delivery-disease-target/#respond</comments>
		
		<dc:creator><![CDATA[Code Sport]]></dc:creator>
		<pubDate>Tue, 21 Oct 2025 07:51:33 +0000</pubDate>
				<category><![CDATA[Biotech]]></category>
		<guid isPermaLink="false">https://codesport.io/?p=7610</guid>

					<description><![CDATA[Glossary of Terms in Clinical Research These three terms are fundamental to understanding how a drug is conceived and developed: Term Definition Context/Role in Drug Development Disease Target The specific molecular entity (e.g., a protein, enzyme, receptor, or gene) in the body that the drug is designed to physically interact with to produce a therapeutic [&#8230;]]]></description>
										<content:encoded><![CDATA[   <script src="https://cdn.tailwindcss.com"></script>

<script>
  tailwind.config = {
    corePlugins: {
     preflight: false, // Disables Tailwind base styles (resets)
    },
    // Optional: Only apply styles within this selector
    important: '#tailwind-blog-content', 
  }
</script>
    <style>
        /* Custom font import for a clean look */
        @import url('https://fonts.googleapis.com/css2?family=Inter:wght@400;600;700&display=swap');
        body {
            font-family: 'Inter', sans-serif;
            background-color: #f7f9fb; /* Light background for contrast */
        }
        /* Custom utility class to make the table look great on small screens */
        @media (max-width: 768px) {
            .responsive-table th, .responsive-table td {
                padding: 0.75rem 0.5rem; /* Reduced padding on mobile */
                font-size: 0.875rem; /* Slightly smaller text on mobile */
            }
        }

*, ::before, ::after {
  border-width: 0;
  border-style: solid; /* This is what you lose with preflight: false */
  border-color: theme('borderColor.DEFAULT', currentColor);
}
    </style>


<!-- I deleted 6xl and changed to full -->
    <div id="tailwind-blog-content"><div class="max-w-full mx-auto space-y-12"><!-- delete or mod to tweak width see: 
www.google.com/search?client=opera&q=what+does+this+do+in+tailwind%3F+max-w-6xl+mx-auto+space-y-12

-->

       <!-- Table 1: The Distinction: Modality vs. Drug Delivery -->
<div class="space-y-4">
    <h1 class="text-3xl font-bold text-gray-800 border-b-2 border-blue-500 pb-2">
        Glossary of Terms in Clinical Research
    </h1>
    <p class="text-gray-600">
        These three terms are fundamental to understanding how a drug is conceived and developed:
    </p>

    <div class="bg-white shadow-2xl rounded-xl overflow-x-auto">
        <table class="w-full text-left responsive-table">
            <thead class="uppercase bg-blue-600 text-white sticky top-0">
                <tr>
                    <th scope="col" class="p-4">Term</th>
                    <th scope="col" class="p-4">Definition</th>
                    <th scope="col" class="p-4">Context/Role in Drug Development</th>
                </tr>
            </thead>
            <!-- The 'text-sm' class ensures consistent font size with the other tables -->
            <tbody class="divide-y divide-gray-200 text-gray-700">
                <!-- Row 1: Disease Target -->
                <tr class="hover:bg-blue-50 transition duration-150">
                    <td class="p-4 font-semibold text-blue-700"> <!--- text-lg removed -->
                        Disease Target
                    </td>
                    <td class="p-4">
                        The specific molecular entity (e.g., a protein, enzyme, receptor, or gene) in the body that the drug is designed to physically interact with to produce a therapeutic effect.
                    </td>
                    <td class="p-4">
                        This is the scientific hypothesis. For the drug we&#8217;ve been discussing, the target is IL-17A and IL-17F (two specific inflammatory cytokine proteins).
                    </td>
                </tr>
                <!-- Row 2: Modality -->
                <tr class="hover:bg-blue-50 transition duration-150">
                    <td class="p-4 font-semibold text-blue-700"> <!--- text-lg removed -->
                        Modality
                    </td>
                    <td class="p-4">
                        <!--The physical form or category of the therapeutic agent itself. It describes what the drug is.--->
<ol class="list-decimal">
	<li>In Biotech all modalities are biological molecules</li>
	<li>A biological molecule is used to create therapeutic drugs</li>
	<li>The drugs are created using living organisms or biological processes
<li>These are called Large Molecules drugs. And differ from traditional, small-molecule, pharmaceutical drugs that are made through chemical synthesis </li></ol>

                    </td>
                    <td class="p-4">Describes the drug&#8217;s structure. Examples include:  large-molecule proteins or antibodies, or smaller, more complex molecules like nucleic acids, cell therapies, and gene editing technologies, Small Molecule, Monoclonal Antibody (mAb), Cell Therapy, Gene Therapy, or, a Nanobody® (a small, single-domain antibody).
                    </td>
                </tr>
                <!-- Row 3: Indication -->
                <tr class="hover:bg-blue-50 transition duration-150">
                    <td class="p-4 font-semibold text-blue-700"> <!--- text-lg removed -->
                        Indication
                    </td>
                    <td class="p-4">
                        The specific disease, condition, or manifestation for which a drug has been, or is seeking to be, officially approved by a regulatory agency (like the FDA).
                    </td>
                    <td class="p-4">
                        The disease or symptoms the drug treats. Often the endpoint in clinical trials. A single drug may have multiple approved indications
                    </td>
                </tr>
            </tbody>
        </table>
    </div>
</div>


       <!-- Table 2: The Distinction: Modality vs. Drug Delivery -->
        <div class="space-y-4">
            <h1 class="text-3xl font-bold text-gray-800 border-b-2 border-green-500 pb-2">
                The Distinction: Modality vs. Drug Delivery
            </h1>
            <p class="text-gray-600">
                Understanding the difference between *what* the drug is (modality) and *how* it gets where it needs to go (delivery) is crucial in modern biotech.
            </p>

            <div class="bg-white shadow-2xl rounded-xl overflow-x-auto">
                <table class="w-full text-left responsive-table">
                   <!-- removed "text-sm" from thead -->
                    <thead class="uppercase bg-green-600 text-white sticky top-0">
                        <tr>
                            <th scope="col" class="p-4">Feature</th>
                            <th scope="col" class="p-4">Drug Modality</th>
                            <th scope="col" class="p-4">Drug Delivery System (DDS)</th>
                        </tr>
                    </thead>
                    <tbody class="divide-y divide-gray-200 text-gray-700">
                        <!-- Row 1: Core Definition -->
                        <tr class="hover:bg-green-50 transition duration-150">
                            <td class="p-4 font-semibold text-green-700">Core Definition</td>
                            <td class="p-4">
                                <span class="inline-block bg-green-100 text-green-800 text-xs font-semibold px-2.5 py-0.5 rounded-full mr-2">What It Is</span>
                                The active therapeutic agent itself—the type of molecule (e.g., antibody, mRNA, cell).
                            </td>
                            <td class="p-4">
                                <span class="inline-block bg-gray-100 text-gray-800 text-xs font-semibold px-2.5 py-0.5 rounded-full mr-2">How It Gets There</span>
                                The mechanism or vehicle used to transport, protect, and release the drug at the target site.
                            </td>
                        </tr>
                        <!-- Row 2: Analogy -->
                        <tr class="hover:bg-green-50 transition duration-150">
                            <td class="p-4 font-semibold text-green-700">Analogy</td>
                            <td class="p-4">
                                The **&#8221;cargo&#8221;** or the therapeutic instruction within the package.
                            </td>
                            <td class="p-4">
                                The **&#8221;truck&#8221;** or **&#8221;vessel&#8221;** responsible for transporting the cargo safely to its destination.
                            </td>
                        </tr>
                        <!-- Row 3: Examples -->
                        <tr class="hover:bg-green-50 transition duration-150">
                            <td class="p-4 font-semibold text-green-700">Examples</td>
                            <td class="p-4">
                                Monoclonal Antibodies (mAbs), Small Molecules, Antisense Oligonucleotides (ASOs), CAR T-Cells.
                            </td>
                            <td class="p-4">
                                Lipid Nanoparticles (LNPs), Viral Vectors (AAV), Pegylation, Transdermal Patches, Orally Ingestible Devices.
                            </td>
                        </tr>
                        <!-- Row 4: Primary Goal -->
                        <tr class="hover:bg-green-50 transition duration-150">
                            <td class="p-4 font-semibold text-green-700">Primary Goal</td>
                            <td class="p-4">
                                To interact with a specific biological target to achieve the therapeutic effect (e.g., block a receptor, replace a protein).
                            </td>
                            <td class="p-4">
                                To improve stability, enhance bioavailability, increase targeting specificity, and control release kinetics.
                            </td>
                        </tr>
                    </tbody>
                </table>
            </div>
        </div>

       <!-- Table 2b: The Distinction: Modality, Target, Drug Delivery, Indication -->
   <div class="bg-white shadow-2xl rounded-xl overflow-x-auto">
        <table class="w-full text-left responsive-table">
            <!-- Unified Header Color (bg-blue-600) -->
            <thead class="uppercase bg-green-600 text-white sticky top-0">
                <tr>
                    <th scope="col" class="p-4">Feature</th>
                    <th scope="col" class="p-4">Drug Modality (The Tool)</th>
                    <th scope="col" class="p-4">Disease Target (The Lock)</th>
                    <th scope="col" class="p-4">Drug Delivery System (DDS) (The Package)</th>
                    <th scope="col" class="p-4">Indication (The Goal)</th>
                </tr>
            </thead>
            <tbody class="divide-y divide-gray-200 text-gray-700">
                <!-- Row 1: Core Definition -->
                <tr class="hover:bg-green-50 transition duration-150">
                    <td class="p-4 font-semibold text-green-700">Core Definition</td>
                    <td class="p-4">The <strong>molecular or cellular class</strong> of the therapeutic tool. Defines the drug&#8217;s fundamental <strong>structure</strong> and <strong>mechanism of action</strong>.</td>
                    <td class="p-4">The specific <strong>biological component</strong> the Modality interacts with to initiate a therapeutic effect.</td>
                    <td class="p-4">The protective vehicle or formulation designed to transport, stabilize, and maximize the Modality&#8217;s exposure at the site of action.</td>
                    <td class="p-4">The specific disease or condition the drug is approved or being developed to treat.</td>
                </tr>
                <!-- Row 2: Analogy -->
                <tr class="hover:bg-green-50 transition duration-150">
                    <td class="p-4 font-semibold text-green-700">Analogy</td>
                    <td class="p-4">The <strong>Specific Tool</strong> (Wrench, Scalpel, Blueprint).</td>
                    <td class="p-4">The <strong>Specific Nut, Protein, or Cell</strong> that needs to be fixed or blocked.</td>
                    <td class="p-4">The <strong>Container or Transport Vehicle</strong> (Pill bottle, LNP truck, Syringe).</td>
                    <td class="p-4">The <strong>Patient&#8217;s Illness</strong> (e.g., A Fever, a Cancer, a Genetic Defect).</td>
                </tr>
                <!-- Row 3: Examples -->
                <tr class="hover:bg-green-50 transition duration-150">
                    <td class="p-4 font-semibold text-green-700">Key Examples</td>
                    <td class="p-4">>Small Molecule (Tylenol), mAb, Nanobody®, mRNA, CAR-T Cell Therapy
<!--
                        <span class="bg-indigo-100 text-indigo-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">Small Molecule</span> (Tylenol), 
                        <span class="bg-green-100 text-green-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">mAb</span>, 
                        <span class="bg-yellow-100 text-yellow-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">Nanobody®</span>, 
                        <span class="bg-red-100 text-red-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">mRNA</span>, 
                        <span class="bg-purple-100 text-purple-800 text-xs font-medium px-2.5 py-0.5 rounded-full">Cell Therapy</span> (CAR T).
-->
                    </td>
                    <td class="p-4">An <strong>Enzyme</strong> (COX), a <strong>Cytokine</strong> (IL-17), a <strong>Viral Protein</strong> (Spike), or a <strong>Defective Gene</strong>.</td>
                    <td class="p-4">
                        <span class="bg-gray-100 text-gray-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">LNP</span>, 
                        <span class="bg-gray-100 text-gray-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">Viral Vector</span> (AAV), 
                        <span class="bg-gray-100 text-gray-800 text-xs font-medium mr-2 px-2.5 py-0.5 rounded-full">Tablet Coating</span>, 
                        Injectable Depot Formulations.
                    </td>
                    <td class="p-4">Hidradenitis Suppurativa (HS), COVID-19, Acute Myeloid Leukemia (AML).</td>
                </tr>
            </tbody>
        </table>
    </div



        <!-- Spacer -->
        <div class="border-t border-gray-300 my-10"></div>

        <!-- Table 3: Relevance to Clinical Trials and the Biotech Industry -->
        <div class="space-y-4">
            <h1 class="text-3xl font-bold text-gray-800 border-b-2 border-purple-500 pb-2">
                Relevance to Clinical Trials and the Biotech Industry
            </h1>
            <p class="text-gray-600">
                The choice of modality and delivery system fundamentally dictates the risks, costs, and timeline of a biopharma program.
            </p>

            <div class="bg-white shadow-2xl rounded-xl overflow-x-auto">
                <table class="w-full text-left responsive-table">
                    <!-- removed "text-sm" from thead -->
                    <thead class="uppercase bg-purple-600 text-white sticky top-0">
                        <tr>
                            <th scope="col" class="p-4">Area</th>
                            <th scope="col" class="p-4">Impact of Drug Modality</th>
                            <th scope="col" class="p-4">Impact of Drug Delivery System (DDS)</th>
                        </tr>
                    </thead>
                    <tbody class="divide-y divide-gray-200 text-gray-700">
                        <!-- Row 1: Clinical Trial Design -->
                        <tr class="hover:bg-purple-50 transition duration-150">
                            <td class="p-4 font-semibold text-purple-700">Clinical Trial Design</td>
                            <td class="p-4">
                                Dictates primary endpoints, necessary safety monitoring (e.g., immunogenicity checks), and potential for curative (one-and-done) or chronic dosing.
                            </td>
                            <td class="p-4">
                                Affects dosing frequency, patient compliance (oral vs. injection), and the need for specialized administration protocols (e.g., surgery for implanted devices).
                            </td>
                        </tr>
                        <!-- Row 2: Regulatory Strategy -->
                        <tr class="hover:bg-purple-50 transition duration-150">
                            <td class="p-4 font-semibold text-purple-700">Regulatory Strategy</td>
                            <td class="p-4">
                                Newer modalities (Cell/Gene Therapy) often lack regulatory precedence, requiring adaptive strategies and extensive CMC (Chemistry, Manufacturing, and Controls) packages.
                            </td>
                            <td class="p-4">
                                Must prove the safety and long-term biocompatibility of the delivery vehicle. Can enable accelerated approval pathways by significantly reducing off-target toxicity.
                            </td>
                        </tr>
                        <!-- Row 3: Biotech Industry Focus -->
                        <tr class="hover:bg-purple-50 transition duration-150">
                            <td class="p-4 font-semibold text-purple-700">Biotech Industry Focus</td>
                            <td class="p-4">
                                Defines a company&#8217;s core platform (e.g., &#8220;a protein degrader company&#8221;). High IP value focused on the mechanism of action.
                            </td>
                            <td class="p-4">
                                Creates specialized platform technology focused on overcoming natural biological barriers (e.g., blood-brain barrier). High value in improving bioavailability of existing modalities.
                            </td>
                        </tr>
                        <!-- Row 4: Manufacturing -->
                        <tr class="hover:bg-purple-50 transition duration-150">
                            <td class="p-4 font-semibold text-purple-700">Manufacturing</td>
                            <td class="p-4">
                                Requires highly specialized and often extremely complex processes (e.g., live cell culture, viral vector purity/titer). High production cost per dose.
                            </td>
                            <td class="p-4">
                                Focuses on large-scale, consistent, and sterile production of the formulation components (e.g., LNP formulation consistency and stability). Often easier to scale than the modality itself.
                            </td>
                        </tr>
                    </tbody>
                </table>
            </div>
        </div>


        <!-- Table 4: Application: COVID-19 Vaccine Technologies -->
        <h1 class="text-3xl font-bold text-gray-800 mb-6 border-b-2 border-indigo-400 pb-2">
            COVID-19 Vaccine Technologies: A Modality Comparison
        </h1>
        <p class="text-gray-600 mb-8">
These three vaccines illustrate two different modality and delivery approaches to achieving the same goal: immunity against the SARS-CoV-2 Spike (S) protein.
        </p>

        <!-- Main Comparison Table Card -->
        <div class="bg-white shadow-2xl rounded-xl overflow-x-auto">
            <table class="w-full text-left responsive-table">
                <!-- removed "text-sm" from thead -->
                <thead class="uppercase bg-indigo-600 text-white sticky top-0">
                    <tr>
                        <th scope="col" class="p-4">Vaccine (Company)</th>
                        <th scope="col" class="p-4">Modality (What is the drug?)</th>
                        <th scope="col" class="p-4">Delivery System/Vehicle</th>
                        <th scope="col" class="p-4">Route of Administration</th>
                        <th scope="col" class="p-4">Drug Target/Mechanism</th>
                    </tr>
                </thead>
                <tbody class="divide-y divide-gray-200 text-gray-700">
                    <!-- Pfizer/BioNTech Row -->
                    <tr class="hover:bg-indigo-50 transition duration-150">
                        <td class="p-4 font-semibold text-indigo-700">
                            Comirnaty <span class="text-sm font-normal text-gray-500">(Pfizer/BioNTech)</span>
                        </td>
                        <td class="p-4 font-medium">
                            <span class="inline-block bg-blue-100 text-blue-800 text-xs font-semibold px-2.5 py-0.5 rounded-full">mRNA</span>
                            Nucleoside-Modified Messenger RNA
                        </td>
                        <td class="p-4">
                            Lipid Nanoparticle (LNP)
                        </td>
                        <td class="p-4">
                            Intramuscular (IM) Injection
                        </td>
                        <td class="p-4">
                            Instructions for human cells to produce the SARS-CoV-2 Spike (S) protein.
                        </td>
                    </tr>
                    <!-- Moderna Row -->
                    <tr class="hover:bg-indigo-50 transition duration-150">
                        <td class="p-4 font-semibold text-indigo-700">
                            Spikevax <span class="text-sm font-normal text-gray-500">(Moderna)</span>
                        </td>
                        <td class="p-4 font-medium">
                            <span class="inline-block bg-blue-100 text-blue-800 text-xs font-semibold px-2.5 py-0.5 rounded-full">mRNA</span>
                            Nucleoside-Modified Messenger RNA
                        </td>
                        <td class="p-4">
                            Lipid Nanoparticle (LNP)
                        </td>
                        <td class="p-4">
                            Intramuscular (IM) Injection
                        </td>
                        <td class="p-4">
                            Instructions for human cells to produce the SARS-CoV-2 Spike (S) protein.
                        </td>
                    </tr>
                    <!-- Novavax Row -->
                    <tr class="hover:bg-indigo-50 transition duration-150">
                        <td class="p-4 font-semibold text-indigo-700">
                            Nuvaxovid <span class="text-sm font-normal text-gray-500">(Novavax)</span>
                        </td>
                        <td class="p-4 font-medium">
                            <span class="inline-block bg-green-100 text-green-800 text-xs font-semibold px-2.5 py-0.5 rounded-full">Protein Subunit</span>
                            Recombinant Protein Subunit
                        </td>
                        <td class="p-4">
                            Nanoparticles with a Saponin-based Adjuvant (Matrix-M)
                        </td>
                        <td class="p-4">
                            Intramuscular (IM) Injection
                        </td>
                        <td class="p-4">
                            The actual, pre-made SARS-CoV-2 Spike (S) protein (antigen) is delivered directly to trigger an immune response.
                        </td>
                    </tr>
                </tbody>
            </table>
        </div>

        <!-- Key Takeaways -->
        <div class="mt-8 p-6 bg-white rounded-xl shadow-lg border-l-4 border-indigo-500">
            <h2 class="text-xl font-semibold text-indigo-700 mb-3">Key Takeaways on Modality</h2>
            <ul class="space-y-2 text-gray-600 list-disc list-inside">
                <li><strong class="text-gray-800">mRNA Vaccines (Pfizer/Moderna):</strong> Use genetic instructions (mRNA) wrapped in fat bubbles (LNP) to teach the body&#8217;s cells how to make the antigen.</li>
                <li><strong class="text-gray-800">Protein Subunit Vaccine (Novavax):</strong> Uses traditional vaccine technology by injecting a non-infectious, pre-made piece of the virus (the spike protein), boosted by an adjuvant.</li>
                <li><strong class="text-gray-800">Route of Delivery:</strong> All three utilize the same common route: an intramuscular injection.</li>
            </ul>
        </div>

    </div>

</div>]]></content:encoded>
					
					<wfw:commentRss>https://codesport.io/biotech/biotech-foundational-concepts-modality-delivery-disease-target/feed/</wfw:commentRss>
			<slash:comments>0</slash:comments>
		
		
			</item>
	</channel>
</rss>
