{"id":598515,"date":"2024-07-26T11:11:25","date_gmt":"2024-07-26T05:41:25","guid":{"rendered":"https:\/\/www.digit.in\/?p=598515"},"modified":"2024-08-01T11:33:12","modified_gmt":"2024-08-01T06:03:12","slug":"llm-to-rag-decoding-ai-jargon-matters-heres-why","status":"publish","type":"post","link":"https:\/\/www.digit.in\/features\/general\/llm-to-rag-why-decoding-ai-jargon-matters.html","title":{"rendered":"LLM to RAG: Decoding AI jargon matters, here\u2019s why"},"content":{"rendered":"\n<p>Whenever a new facet of technology debuts and gathers momentum, it introduces lots of new terminology into the lexicon. From <em>tuning<\/em> into FM radio to <em>streaming<\/em> music on-the-go, <em>STD<\/em> phone calls in the past to emoji and <em>GIF<\/em>-enabled chats in your favourite app, for example, haven\u2019t we all learned, unlearned, and relearned quite a few tech terms over the years? Now it\u2019s AI\u2019s turn to introduce a whole new wave of words for us to throw around. <\/p>\n\n\n\n<p>Naturally, as the use of AI proliferates, so does its related jargon. It\u2019s important to address some of that jargon to avoid unnecessary confusion and to save our own sanity. I think there have been too many acronyms floating around and key concepts unexplained as far as AI is concerned. I\u2019ve been guilty of not paying enough attention to it, something that I\u2019m trying to correct through this article. Better late than never, right?<\/p>\n\n\n\n<p>Also read: <a href=\"https:\/\/www.digit.in\/features\/general\/how-rag-boosts-llm-accuracy-to-limit-ai-hallucination.html\">How RAG boosts LLM accuracy and reduces AI hallucination<\/a><\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/static.digit.in\/AI_jargon_buster.png\"><img decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/static.digit.in\/AI_jargon_buster-1024x576.png\" alt=\"\" class=\"wp-image-598540\" srcset=\"https:\/\/static.digit.in\/AI_jargon_buster-1024x576.png 1024w, https:\/\/static.digit.in\/AI_jargon_buster-300x169.png 300w, https:\/\/static.digit.in\/AI_jargon_buster-768x432.png 768w, https:\/\/static.digit.in\/AI_jargon_buster-1536x864.png 1536w, https:\/\/static.digit.in\/AI_jargon_buster-2048x1152.png 2048w, https:\/\/static.digit.in\/AI_jargon_buster-267x150.png 267w, https:\/\/static.digit.in\/AI_jargon_buster-100x56.png 100w, https:\/\/static.digit.in\/AI_jargon_buster-622x350.png 622w, https:\/\/static.digit.in\/AI_jargon_buster-788x443.png 788w, https:\/\/static.digit.in\/AI_jargon_buster-599x337.png 599w, https:\/\/static.digit.in\/AI_jargon_buster-150x84.png 150w, https:\/\/static.digit.in\/AI_jargon_buster.png 1280w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>Take LLM or Large Language Models, for instance \u2013 the reason behind the current AI renaissance we are all experiencing. It\u2019s the backbone of AI chatbots like OpenAI\u2019s ChatGPT, Microsoft Copilot, or Google Gemini, but how many of us had even heard of such a term called LLM even five years ago? Simply put, it\u2019s an AI algorithm trained on vast amounts of text and natural language-related data to generate textual output similar to a human would. But what about text-to-image generation services like Midjourney, Stable Diffusion, or DALL-E? They use LLMs for language demystification, but they\u2019re strictly AI diffusion models \u2013 machine learning algorithms that progressively add noise to data and learn to reverse the process, especially useful for generating images, videos, and sound. LLMs and diffusion models make up most of the Generative AI products and services you\u2019ve seen and used since 2022. It\u2019s not even been two years since ChatGPT 3 blew the AI floodgates open for the general public, and we\u2019ve already abbreviated Generative AI to GenAI. Keeping it snappy!<\/p>\n\n\n\n<p>Then there are slightly more fundamental concepts of AI that are good to know, especially in the wake of AI PCs. Whether it\u2019s today or in the near future, conversations on <a href=\"https:\/\/www.digit.in\/features\/laptops\/computex-2024-hands-on-with-snapdragon-x-elite-ai-pc-laptops-from-acer-lenovo-samsung-microsoft.html\" title=\"\">AI PC<\/a> or AI Laptop won\u2019t be complete without the mention of TOPS \u2013 something you must\u2019ve seen across all the recent AI hardware-related articles here in the magazine or on Digit.in. TOPS, or Tera Operations Per Second, is a measure of how many trillion operations a processor can perform every second. In the context of AI hardware, TOPS is a benchmark of processing power, particularly relevant for tasks involving deep learning and neural networks. Higher TOPS indicate a more capable processor, one that can handle complex AI computations with ease. As we start reviewing and writing about AI Laptops, you\u2019ll see TOPS being mentioned as a key parameter going forward.<\/p>\n\n\n\n<iframe width=\"560\" height=\"315\" src=\"https:\/\/www.youtube.com\/embed\/qwtjVD7aAg4\" title=\"YouTube video player\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe>\n\n\n\n<p>Tokens, on the other hand, are a fundamental concept in the software side of AI, particularly when dealing with LLMs or SLMs (Small Language Models). The human brain learns and understands language differently from a computer, which only understands 0s and 1s. Tokens are the numerical pieces of text \u2013 words, characters, or subwords \u2013 that an AI model processes to understand and generate language. They are the individual units that make up the sentences and paragraphs produced by models like ChatGPT. The more tokens a model can process efficiently, the better it performs in generating contextually accurate text for any given scenario, based on its prompt relevance.&nbsp;<\/p>\n\n\n\n<p>A prompt, of course, is another new AI-related term \u2013 quite similar to the idea of a keyword in the traditional search engine setting, but allowing for far more complexity and nuance than a keyword ever can. Hallucination is another term you will encounter a lot with respect to one of the fundamental faults in AI (LLM &amp; GenAI) that researchers and AI companies are trying to fix \u2013 hallucination occurs when an AI model generates incorrect or misleading information presented as fact. One of the ways to reduce hallucination in LLMs is called RAG (Retrieval Augmented Generation), which is a technique to improve the factual accuracy of generative AI content by referencing outside information that wasn\u2019t previously part of the original AI training database.<\/p>\n\n\n\n<figure class=\"wp-block-pullquote\"><blockquote><p>Terms like &#8220;streaming,&#8221; &#8220;Wi-Fi,&#8221; and &#8220;Bluetooth&#8221; were once new and confusing, but today they are part of our everyday vocabulary.<\/p><\/blockquote><\/figure>\n\n\n\n<p>Of course, this is nowhere close to even scratching the surface of AI related jargon. Learning the AI jargon may seem daunting at first, filled with unfamiliar acronyms and complex concepts. However, remember that the same was true for all the early consumer tech jargon 20-25 years ago till now. Terms like &#8220;streaming,&#8221; &#8220;Wi-Fi,&#8221; and &#8220;Bluetooth&#8221; were once new and confusing, but today they are part of our everyday vocabulary. As technology progresses and evolves, our capacity to absorb and speak tech improves, making these initially bewildering terms second nature over time.<\/p>\n\n\n\n<figure class=\"wp-block-image size-large\"><a href=\"https:\/\/static.digit.in\/AI_pcb.png\"><img decoding=\"async\" width=\"1024\" height=\"576\" src=\"https:\/\/static.digit.in\/AI_pcb-1024x576.png\" alt=\"\" class=\"wp-image-598541\" srcset=\"https:\/\/static.digit.in\/AI_pcb-1024x576.png 1024w, https:\/\/static.digit.in\/AI_pcb-300x169.png 300w, https:\/\/static.digit.in\/AI_pcb-768x432.png 768w, https:\/\/static.digit.in\/AI_pcb-1536x864.png 1536w, https:\/\/static.digit.in\/AI_pcb-2048x1152.png 2048w, https:\/\/static.digit.in\/AI_pcb-267x150.png 267w, https:\/\/static.digit.in\/AI_pcb-100x56.png 100w, https:\/\/static.digit.in\/AI_pcb-622x350.png 622w, https:\/\/static.digit.in\/AI_pcb-788x443.png 788w, https:\/\/static.digit.in\/AI_pcb-599x337.png 599w, https:\/\/static.digit.in\/AI_pcb-150x84.png 150w, https:\/\/static.digit.in\/AI_pcb.png 1280w\" sizes=\"(max-width: 1024px) 100vw, 1024px\" \/><\/a><\/figure>\n\n\n\n<p>When it comes to AI, keeping up with the jargon is more than just a necessity \u2013 it\u2019s an invitation to be part of the conversation on a more geeky level. As we continue to navigate the fast evolving AI landscape, familiarising ourselves with terms like TOPS, tokens, and prompts is just the beginning. Digging deeper and further understanding concepts such as model training, inference, neural networks, and the difference between generative and discriminative models will be crucial for anyone looking to stay informed and engaged, and fundamentally have a deeper appreciation of tech and its application in our lives.<\/p>\n\n\n\n<p>Also read: <a href=\"https:\/\/www.digit.in\/features\/general\/ai-hallucination-in-llm-and-beyond-can-it-be-fixed-completely.html\">AI hallucination in LLM and beyond: Will it ever be fixed?<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Whenever a new facet of technology debuts and gathers momentum, it introduces lots of new terminology into the lexicon. From tuning into FM radio to streaming music on-the-go, STD phone calls in the past to emoji and GIF-enabled chats in your favourite app, for example, haven\u2019t we all learned, unlearned, and relearned quite a few [&hellip;]<\/p>\n","protected":false},"author":1934,"featured_media":598516,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_gspb_post_css":"","footnotes":""},"categories":[186989],"tags":[222570,221671,213098],"contenttype":[205],"digitlang":[165350],"dealstore":[],"offerexpiration":[],"acf":[],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/posts\/598515"}],"collection":[{"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/users\/1934"}],"replies":[{"embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/comments?post=598515"}],"version-history":[{"count":6,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/posts\/598515\/revisions"}],"predecessor-version":[{"id":598579,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/posts\/598515\/revisions\/598579"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/media\/598516"}],"wp:attachment":[{"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/media?parent=598515"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/categories?post=598515"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/tags?post=598515"},{"taxonomy":"contenttype","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/contenttype?post=598515"},{"taxonomy":"digitlang","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/digitlang?post=598515"},{"taxonomy":"dealstore","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/dealstore?post=598515"},{"taxonomy":"offerexpiration","embeddable":true,"href":"https:\/\/www.digit.in\/wp-json\/wp\/v2\/offerexpiration?post=598515"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}