<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Edge Computing on Carles Abarca</title><link>https://carlesabarca.com/tags/edge-computing/</link><description>Recent content in Edge Computing on Carles Abarca</description><generator>Hugo -- gohugo.io</generator><language>en</language><copyright>© 2026 Carles Abarca</copyright><lastBuildDate>Wed, 23 Oct 2024 00:00:00 +0000</lastBuildDate><atom:link href="https://carlesabarca.com/tags/edge-computing/index.xml" rel="self" type="application/rss+xml"/><item><title>Small LLMs: Powerful Alternatives for Business</title><link>https://carlesabarca.com/posts/small-llms-powerful-alternatives/</link><pubDate>Wed, 23 Oct 2024 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/small-llms-powerful-alternatives/</guid><description>Smaller LLMs like DistilBERT, TinyBERT, and ALBERT are proving to be efficient and powerful alternatives for businesses.</description><content:encoded>&lt;p&gt;In the world of AI, Large Language Models like Claude and GPT-4 often grab the headlines, but &lt;strong&gt;smaller LLMs are proving to be efficient and powerful alternatives&lt;/strong&gt; for businesses. Here is why models like DistilBERT, TinyBERT, ALBERT, MiniLM, MobileBERT, and ELECTRA-Small deserve your attention:&lt;/p&gt;

&lt;h2 class="relative group"&gt;Cost Efficiency
 &lt;div id="cost-efficiency" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#cost-efficiency" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Models such as DistilBERT and MobileBERT are significantly smaller than their larger counterparts but retain nearly the same language understanding capabilities. This means reduced computational power and lower costs, making AI more accessible to businesses of all sizes.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Speed and Performance
 &lt;div id="speed-and-performance" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#speed-and-performance" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Lightweight architectures like TinyBERT and MiniLM offer faster responses, improving user experiences in real-time applications such as chatbots, virtual assistants, and automated customer support. Quick inference speeds make them ideal for low-latency environments.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Data Privacy and Customization
 &lt;div id="data-privacy-and-customization" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#data-privacy-and-customization" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Open-source models like ALBERT and ELECTRA-Small provide the flexibility to fine-tune on localized data. This ensures sensitive data stays on-premises or in private cloud instances, boosting security while also enabling businesses to tailor AI models to specific industry needs with minimal data.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Tailored Solutions for Niche Markets
 &lt;div id="tailored-solutions-for-niche-markets" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#tailored-solutions-for-niche-markets" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;With models like ALBERT, businesses can deploy AI that is finely tuned for specialized tasks or sectors, allowing them to innovate in niche markets without sacrificing performance.&lt;/p&gt;
&lt;p&gt;As AI becomes more deeply integrated into every industry, these smaller LLMs bring flexibility, cost savings, and targeted results &amp;ndash; proving that sometimes, less is more when it comes to AI.&lt;/p&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/small-llms-powerful-alternatives/featured.png"/></item><item><title>Unlocking AI Efficiency with LoRA and Quantization</title><link>https://carlesabarca.com/posts/lora-quantization-ai-efficiency/</link><pubDate>Mon, 06 May 2024 00:00:00 +0000</pubDate><guid>https://carlesabarca.com/posts/lora-quantization-ai-efficiency/</guid><description>Two pivotal techniques &amp;ndash; LoRA and Quantization &amp;ndash; are shaping the future of lean and efficient AI systems.</description><content:encoded>&lt;p&gt;As we push the boundaries of what AI can achieve, the need for optimized models that perform at scale while conserving resources becomes paramount. Two pivotal techniques that are shaping the future of lean and efficient AI are Low Rank Adaptation (LoRA) and Quantization.&lt;/p&gt;

&lt;h2 class="relative group"&gt;What is LoRA?
 &lt;div id="what-is-lora" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#what-is-lora" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Low Rank Adaptation is a novel technique that allows for the efficient tuning of large pre-trained models. LoRA works by inserting trainable low-rank matrices into the model, enabling significant updates to model behavior without altering the majority of the pre-trained weights. This approach not only preserves the strengths of the original model but also reduces the computational overhead typically associated with training large models.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Why Quantization Matters
 &lt;div id="why-quantization-matters" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#why-quantization-matters" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Quantization reduces the precision of the numbers used within an AI model from floating-point to integers, which are less computationally intensive. This process dramatically decreases the model size and speeds up inference time, making it ideal for deployment on edge devices where resources are limited.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Combining LoRA and Quantization
 &lt;div id="combining-lora-and-quantization" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#combining-lora-and-quantization" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;When used together, LoRA and Quantization offer a powerful synergy that boosts model performance and efficiency. This combination allows for deploying state-of-the-art models on platforms with strict memory and processing constraints, such as mobile phones and IoT devices.&lt;/p&gt;

&lt;h2 class="relative group"&gt;Real-World Impact
 &lt;div id="real-world-impact" class="anchor"&gt;&lt;/div&gt;
 
 &lt;span
 class="absolute top-0 w-6 transition-opacity opacity-0 -start-6 not-prose group-hover:opacity-100 select-none"&gt;
 &lt;a class="text-primary-300 dark:text-neutral-700 !no-underline" href="#real-world-impact" aria-label="Anchor"&gt;#&lt;/a&gt;
 &lt;/span&gt;
 
&lt;/h2&gt;
&lt;p&gt;Industries ranging from telecommunications to healthcare are already reaping the benefits of these technologies. By integrating LoRA and Quantization, businesses are able to deploy advanced AI solutions more broadly and at a lower cost.&lt;/p&gt;</content:encoded><media:content xmlns:media="http://search.yahoo.com/mrss/" url="https://carlesabarca.com/posts/lora-quantization-ai-efficiency/featured.png"/></item></channel></rss>