<?xml version="1.0"?>
<oembed><version>1.0</version><provider_name>Lightning AI</provider_name><provider_url>https://lightning.ai/pages</provider_url><author_name>JP Hennessy</author_name><author_url>https://lightning.ai/pages/author/jplightning-ai/</author_url><title>Parameter-Efficient LLM Finetuning With Low-Rank Adaptation (LoRA) - Lightning AI</title><type>rich</type><width>600</width><height>338</height><html>&lt;blockquote class="wp-embedded-content" data-secret="ddPBjcuGE6"&gt;&lt;a href="https://lightning.ai/pages/community/tutorial/lora-llm/"&gt;Parameter-Efficient LLM Finetuning With Low-Rank Adaptation (LoRA)&lt;/a&gt;&lt;/blockquote&gt;&lt;iframe sandbox="allow-scripts" security="restricted" src="https://lightning.ai/pages/community/tutorial/lora-llm/embed/#?secret=ddPBjcuGE6" width="600" height="338" title="&#x201C;Parameter-Efficient LLM Finetuning With Low-Rank Adaptation (LoRA)&#x201D; &#x2014; Lightning AI" data-secret="ddPBjcuGE6" frameborder="0" marginwidth="0" marginheight="0" scrolling="no" class="wp-embedded-content"&gt;&lt;/iframe&gt;&lt;script type="text/javascript"&gt;
/* &lt;![CDATA[ */
/*! This file is auto-generated */
!function(d,l){"use strict";l.querySelector&amp;&amp;d.addEventListener&amp;&amp;"undefined"!=typeof URL&amp;&amp;(d.wp=d.wp||{},d.wp.receiveEmbedMessage||(d.wp.receiveEmbedMessage=function(e){var t=e.data;if((t||t.secret||t.message||t.value)&amp;&amp;!/[^a-zA-Z0-9]/.test(t.secret)){for(var s,r,n,a=l.querySelectorAll('iframe[data-secret="'+t.secret+'"]'),o=l.querySelectorAll('blockquote[data-secret="'+t.secret+'"]'),c=new RegExp("^https?:$","i"),i=0;i&lt;o.length;i++)o[i].style.display="none";for(i=0;i&lt;a.length;i++)s=a[i],e.source===s.contentWindow&amp;&amp;(s.removeAttribute("style"),"height"===t.message?(1e3&lt;(r=parseInt(t.value,10))?r=1e3:~~r&lt;200&amp;&amp;(r=200),s.height=r):"link"===t.message&amp;&amp;(r=new URL(s.getAttribute("src")),n=new URL(t.value),c.test(n.protocol))&amp;&amp;n.host===r.host&amp;&amp;l.activeElement===s&amp;&amp;(d.top.location.href=t.value))}},d.addEventListener("message",d.wp.receiveEmbedMessage,!1),l.addEventListener("DOMContentLoaded",function(){for(var e,t,s=l.querySelectorAll("iframe.wp-embedded-content"),r=0;r&lt;s.length;r++)(t=(e=s[r]).getAttribute("data-secret"))||(t=Math.random().toString(36).substring(2,12),e.src+="#?secret="+t,e.setAttribute("data-secret",t)),e.contentWindow.postMessage({message:"ready",secret:t},"*")},!1)))}(window,document);
/* ]]&gt; */
&lt;/script&gt;
</html><thumbnail_url>https://lightningaidev.wpengine.com/wp-content/uploads/2023/04/lora-thumbnail.png</thumbnail_url><thumbnail_width>1902</thumbnail_width><thumbnail_height>886</thumbnail_height><description>In the rapidly evolving field of AI, using large language models in an efficient and effective manner is becoming more and more important. In this article, you will learn how to tune an LLM with Low-Rank Adaptation (LoRA) computationally efficiently!</description></oembed>
