<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Generative AI Archives - Creospan</title>
	<atom:link href="https://creospan.com/tag/generative-ai/feed/" rel="self" type="application/rss+xml" />
	<link>https://creospan.com/tag/generative-ai/</link>
	<description>Digital Transformation Consultancy</description>
	<lastBuildDate>Mon, 16 Feb 2026 21:55:31 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>
	<item>
		<title>Why Model Context Protocol Matters: Building Real-World Workflows</title>
		<link>https://creospan.com/why-model-context-protocol-matters-building-real-world-workflows/</link>
		
		<dc:creator><![CDATA[Donna Mathew]]></dc:creator>
		<pubDate>Thu, 22 Jan 2026 17:59:42 +0000</pubDate>
				<category><![CDATA[Insights]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI Transformation]]></category>
		<category><![CDATA[AI Workflows]]></category>
		<category><![CDATA[API]]></category>
		<category><![CDATA[Artificial intelligence]]></category>
		<category><![CDATA[ChatGPT]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[GitHub Copilot]]></category>
		<category><![CDATA[IDE]]></category>
		<category><![CDATA[Large Language Models (LLMs)]]></category>
		<category><![CDATA[Linear]]></category>
		<category><![CDATA[LLM]]></category>
		<category><![CDATA[MCP]]></category>
		<category><![CDATA[Model Context Protocol]]></category>
		<category><![CDATA[Notion]]></category>
		<category><![CDATA[Prompt Engineering]]></category>
		<guid isPermaLink="false">https://creospan.com/?p=1452</guid>

					<description><![CDATA[<p>When large language models (LLMs) first became accessible, most of our interactions with them were bound within a single prompt-response cycle. You asked, they answered. But as developers began embedding AI into real systems (IDE copilot etc.), it became clear that prompts alone couldn’t sustain meaningful workflows. AI needed context, memory, and the ability to act, not just chat. That’s where the Model Context Protocol (MCP) enters the picture (to solve the context and ability needs).  </p>
<p>The post <a href="https://creospan.com/why-model-context-protocol-matters-building-real-world-workflows/">Why Model Context Protocol Matters: Building Real-World Workflows</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>When large language models (LLMs) first became accessible, most of our interactions with them were bound within a single prompt-response cycle. You asked, they answered. But as developers began embedding AI into real systems (IDE copilot etc.), it became clear that prompts alone couldn’t sustain meaningful workflows. AI needed context, memory, and the ability to act, not just chat. That’s where the Model Context Protocol (MCP) enters the picture (to solve the context and ability needs).  </p>



<p>At its core, MCP is an open standard that lets AI models connect to external systems in a structured, context-aware way. Think of it as the connective tissue between an AI and the tools it depends on; databases, project trackers, and code environments. Rather than reinventing integrations for each tool, MCP solves the integration bottleneck for agentic systems and enables real-time, context-aware automation. <em>  </em> </p>



<p><br><strong>Why Not Just Call APIs Directly?</strong> </p>



<p>Why not let the model talk directly to the tool’s API?&nbsp;</p>



<p>The short answer is control and security.&nbsp;<br>&nbsp;<br>MCP defines a client-server pattern that allows AI systems to interact with real-world applications through a common interface. This allows models to securely call external tools, fetch structured data, and perform actions without the LLM needing to know every detail about the API behind it. It standardizes how models “see” tools, what they can access, and how they act to keep everything modular, secure, and interoperable.&nbsp;<br>&nbsp;<br><strong>How it&nbsp;Works</strong>&nbsp;<br>&nbsp;<br>In a typical MCP architecture, an LLM communicates through an MCP client, which routes requests to one or more MCP servers. The client handles translation between the model’s natural language intent and the technical request schema, while the server executes the actual&nbsp;tool&nbsp;actions like storing data, fetching content, or performing updates. Some IDE environments, such as Cursor, already act as an MCP client under the hood, enabling seamless communication with compatible servers. This design separates the language model from the tool’s raw APIs.&nbsp;</p>



<p><strong>Our Workflow: IDE&nbsp;Centered Intelligence with MCP</strong>&nbsp;<br>&nbsp;<br>At&nbsp;Creospan, we deliberately designed our MCP&nbsp;based workflow around a simple but important belief: meaningful engineering decisions require code-level context. While large language models can reason over user stories and tickets in isolation, real prioritization, dependency analysis, and implementation planning only become reliable when the model understands the actual code,&nbsp;it is going to change. This is precisely the gap MCP helps us bridge.&nbsp;</p>



<p>This is why our workflow places the IDE, not the task tracker or project planning tool, at the center.&nbsp;</p>



<p>At&nbsp;Creopsan,&nbsp;Linear&nbsp;serves&nbsp;as our&nbsp;project management tool.&nbsp;Linear is a high-performance project management tool designed to streamline software development workflows through a minimalist interface.&nbsp;It holds user stories, priorities, and labels. However, instead of treating Linear as the place where decisions are made, we treat it as a&nbsp;structured input&nbsp;source. Through an MCP connection, stories flow from Linear directly into the coding environment, where they can be evaluated with full visibility into the codebase using&nbsp;AI&nbsp;Assisted&nbsp;IDE’s&nbsp;context engine.&nbsp;</p>



<p>Once inside the&nbsp;AI Assisted&nbsp;IDE (Cursor, GitHub Copilot, Augment Code,&nbsp;etc.), the LLM operates with two critical forms of context. The first is project management context, fetched from Linear via MCP. The second is implementation context, derived from the&nbsp;code&nbsp;repository itself&nbsp;using&nbsp;the&nbsp;IDE’S&nbsp;context engine&nbsp;which&nbsp;maintains&nbsp;a live understanding of the stack across repositories, services, and code history.&nbsp;</p>



<p>This combination enables a class of reasoning that is difficult to achieve elsewhere. As stories are loaded into the IDE, the LLM can reason across them to surface overlaps, shared implementation paths, and implicit relationships.&nbsp;Similar stories&nbsp;can be grouped not just based on description&nbsp;but based on the parts of the codebase they affect.&nbsp;Common work&nbsp;emerges&nbsp;naturally when multiple tickets map to the same components or abstractions. Ordering concerns surface by inspecting dependencies in code rather than relying solely on ticket-level links.&nbsp;</p>



<p>Importantly, this reasoning is not fully automated or opaque. The LLM proposes insights and prioritization suggestions, but developers&nbsp;remain&nbsp;in the loop. Engineers&nbsp;validate, adjust, or override decisions with a clear understanding of why a particular ordering or grouping was suggested. MCP makes this possible by ensuring that product intent from Linear and technical reality from the codebase&nbsp;using context engine&nbsp;are available together inside the IDE.&nbsp;</p>



<p>Once decisions are&nbsp;validated, the workflow completes its loop. Updates, refinements, and execution outcomes are pushed back into Linear via MCP, keeping the product view synchronized without forcing developers to leave their editor. Developers can then pick up a story, begin implementation, and update its status directly from the IDE. Every change, discussion, and update&nbsp;stay&nbsp;synchronized, giving stakeholders a live view of progress while preserving developer flow.&nbsp;<br>&nbsp;<br><strong>Notion as the Learning Layer</strong>&nbsp;</p>



<p>If Linear captures what we plan to build, Notion captures how we build it. Notion is an all-in-one workspace that blends note-taking, document collaboration, and database management into a single, highly customizable platform.&nbsp;&nbsp;Through a separate MCP server, we log meaningful AI interactions from the IDE into Notion. This includes prompts that led to better architectural decisions, reasoning traces behind prioritization choices, and patterns that repeat across projects. Over time, these logs have evolved into a knowledge dataset, a reflection of how our team collaborates with AI. By analyzing them, we uncover which prompts&nbsp;drive&nbsp;faster development or cleaner code, and which patterns repeat across projects. The most effective ones become shared templates, enabling the entire team to improve collectively rather than individually.&nbsp;</p>



<p>The result is a connected system where planning, implementation, and learning reinforce each other through shared context. MCP’s value lies not in tool integration itself, but in enabling intelligence to&nbsp;operate&nbsp;within the IDE, where code and product intent converges.&nbsp;</p>



<p>At&nbsp;Creospan, we see this as a key step forward for SDLC productivity, where small efficiencies compound across teams and projects. In the end, our implementation shows how AI systems can evolve from reactive to proactive. Tools like Notion and Linear are not just endpoints; they are contexts. With MCP, we give AI the means to understand, navigate, and contribute to those contexts intelligently.&nbsp;</p>



<figure class="wp-block-image size-full"><img fetchpriority="high" decoding="async" width="880" height="451" src="https://creospan.com/wp-content/uploads/2026/01/image.png" alt="" class="wp-image-1453" srcset="https://creospan.com/wp-content/uploads/2026/01/image.png 880w, https://creospan.com/wp-content/uploads/2026/01/image-300x154.png 300w, https://creospan.com/wp-content/uploads/2026/01/image-768x394.png 768w" sizes="(max-width: 880px) 100vw, 880px" /></figure>



<p><strong>Conclusion&nbsp;</strong>&nbsp;</p>



<p>As AI continues to reshape the landscape of software development, MCP stands out as a transformative standard for building agentic, context-aware workflows. By bridging product intent and technical reality within the IDE, MCP empowers both AI and human collaborators to make informed, reliable decisions&nbsp;driving productivity and innovation across teams. The recent evolution of MCP, with enhanced security, structured tool output, and seamless IDE integrations, positions it not just as a technical solution but as a foundation for the next generation of intelligent engineering systems.&nbsp;&nbsp;</p>



<p>Article Written By Dhairya Bhuta </p>



<p></p>
<p>The post <a href="https://creospan.com/why-model-context-protocol-matters-building-real-world-workflows/">Why Model Context Protocol Matters: Building Real-World Workflows</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>What’s Holding You Back from Unlocking AI-Powered Workforce Productivity?</title>
		<link>https://creospan.com/whats-holding-you-back-from-unlocking-ai-powered-workforce-productivity/</link>
		
		<dc:creator><![CDATA[Donna Mathew]]></dc:creator>
		<pubDate>Sat, 24 May 2025 22:40:34 +0000</pubDate>
				<category><![CDATA[Insights]]></category>
		<category><![CDATA[AI]]></category>
		<category><![CDATA[AI Adoption]]></category>
		<category><![CDATA[AI Compliance]]></category>
		<category><![CDATA[AI in the Workplace]]></category>
		<category><![CDATA[AI Productivity]]></category>
		<category><![CDATA[AI Workflows]]></category>
		<category><![CDATA[Data Security]]></category>
		<category><![CDATA[Digital Transformation]]></category>
		<category><![CDATA[Enterprise AI]]></category>
		<category><![CDATA[Future of work]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[Microsoft 365 Copilot]]></category>
		<category><![CDATA[Workplace AI]]></category>
		<guid isPermaLink="false">https://creospan.com/?p=1225</guid>

					<description><![CDATA[<p>Across industries, individual users are embracing AI as their “digital coworker” - one who’s fast, tireless, and surprisingly helpful. Whether they’re drafting blog posts, crunching data, or writing code. AI can do it all. Yet, many organizations hesitate to fully integrate AI into their workflows.</p>
<p>The post <a href="https://creospan.com/whats-holding-you-back-from-unlocking-ai-powered-workforce-productivity/">What’s Holding You Back from Unlocking AI-Powered Workforce Productivity?</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="wpb-content-wrapper"><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>Across industries, individual users are embracing AI as their “digital coworker” &#8211; one who’s fast, tireless, and surprisingly helpful. Whether they’re drafting blog posts, crunching data, or writing code. AI can do it all. Yet, many organizations hesitate to fully integrate AI into their workflows.</p>
<p>Why the disconnect?</p>
<p>Their concerns are valid. Worries about data privacy, fears surrounding misinformation, and uncertainty about how to scale initiatives responsibly do not instill trust in organizations wanting to extend their workflows. However, a well-structured AI adoption strategy can address and overcome these challenges.</p>
<p>In this article, we walk through a 7-stage roadmap for introducing Microsoft 365 Copilot across your organization, helping you accelerate productivity while staying secure and compliant.</p>
<h2>Stage 1: Adopting Microsoft 365 – Laying the Foundation</h2>
<p>The journey begins with Microsoft 365, a comprehensive platform designed to power productivity and collaboration. Many organizations stop at basic familiar and functional tools such as Teams, Excel, Word, Outlook, while missing out on the AI capabilities embedded into the ecosystem such as predictive text suggestions, summarizing, content creation smart templates, real-time collaboration enhancements and automating processes.</p>
<p><strong>Pro Tip:</strong> If you’ve already deployed Microsoft 365, you’re halfway there. The next step is unlocking its AI-enhanced features.</p>
<h2>Stage 2: Introducing Microsoft Copilot – The Productivity Multiplier</h2>
<p>As familiarity with Microsoft 365 grows, so does awareness of Microsoft Copilot, an AI add-on that can automate repetitive tasks, summarize content, generate insights, and more. However, uncertainty around how Copilot fits into daily workflows can slow its adoption.</p>
<p><strong>Pro Tip:</strong> Host internal demos or lunch-and-learns sessions showcasing real-world use cases tailored to finance, HR, or sales roles.</p>
<h2>Stage 3: Addressing Security, Privacy &amp; Compliance</h2>
<p>AI adoption must be built on trust. At this stage, organizations are asking:</p>
<ul>
<li>What data does Copilot access?</li>
<li>Can access be role-based?</li>
<li>How is sensitive information protected?</li>
<li>Is the solution compliant with our regulatory standards?</li>
<li>What safeguards are in place to prevent misuse?</li>
</ul>
<p><strong>Pro Tip:</strong> Partner with IT and compliance teams early in the adoption and integration process. Establish clear documentation on data access, protection protocols, and AI risk mitigation.</p>
<h2>Stage 4: Establishing AI Policies &amp; Governance</h2>
<p>Without a strong governance framework, organizations risk inconsistent adoption and exposure to compliance risks. Key policy areas include:</p>
<ul>
<li>Responsible use guidelines</li>
<li>Data retention and sharing protocols</li>
<li>Alignment with internal and external regulatory standards</li>
<li>Ethical use policies, including bias mitigation</li>
</ul>
<p><strong>Pro Tip:</strong> Create a cross-functional AI Governance Council to steer strategy, policy, and education.</p>
<h2>Stage 5: Prototyping &amp; Piloting for Proof of Value</h2>
<p>Rather than jumping straight to full deployment, many successful organizations begin with targeted pilots. A focused rollout enables teams to:</p>
<ul>
<li>Experiment with real use cases</li>
<li>Identify integration or cultural challenges</li>
<li>Measure productivity uplift</li>
<li>Build internal champions</li>
</ul>
<p><strong>Pro Tip:</strong> Choose a pilot team with measurable KPIs and a high volume of knowledge-work for maximum impact.</p>
<h2>Stage 6: Scaling Across the Enterprise</h2>
<p>Once early wins are documented, scaling can begin. This phase is about:</p>
<ul>
<li>Delivering role-specific training</li>
<li>Embedding Copilot into standard workflows</li>
<li>Ensuring executive sponsorship</li>
<li>Managing resistance and change with empathy</li>
</ul>
<p><strong>Pro Tip:</strong> Track usage analytics and feedback to tailor your training and adoption campaigns.</p>
<h2>Stage 7: Measuring ROI and Driving Continuous Improvement</h2>
<p>Implementation is just the beginning. Leading organizations continuously monitor:</p>
<ul>
<li>Time saved per task or team</li>
<li>Increase in throughput or decision quality</li>
<li>Employee satisfaction and Copilot adoption</li>
<li>Opportunities for new use cases or advanced integration</li>
</ul>
<p><strong>Pro Tip:</strong> Treat this as a feedback loop &#8211; measure, learn, adapt. The path to AI-powered productivity isn’t linear, but with the right plan, you can turn uncertainty into action. When deployed thoughtfully, Microsoft Copilot doesn’t just improve workflows; it transforms them.</p>
<h2>How We Can Help</h2>
<p>Choosing the right partner for your AI adoption journey is critical. Here’s why organizations trust Creospan to help them unlock the full potential of Microsoft Copilot:</p>
<ul>
<li><strong>Expertise in AI Productivity Tools:</strong> Our team has deep experience with Microsoft Copilot and other generative AI solutions, ensuring a smooth and effective implementation.</li>
<li><strong>Tailored Solutions:</strong> We understand that every organization is unique. Our strategies are customized to align with your specific needs, workflows, and goals.</li>
<li><strong>End-to-End Support:</strong> From initial education to enterprise-wide rollout and ongoing optimization, we’re with you at every step of your AI journey.</li>
<li><strong>Focus on Security and Compliance:</strong> We prioritize data security, privacy, and adherence to industry standards, giving you peace of mind as you adopt AI tools</li>
</ul>
<p>Ready to transform your workforce with Microsoft Copilot? Contact us today to start your AI adoption journey.</p>
<p><em>Article written by Davinder Kohli and Shirali Shah.</em></p>

		</div>
	</div>
</div></div></div></div>

<p>&nbsp;</p>
</div><p>The post <a href="https://creospan.com/whats-holding-you-back-from-unlocking-ai-powered-workforce-productivity/">What’s Holding You Back from Unlocking AI-Powered Workforce Productivity?</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Caching Patterns in Retrieval Augmented Generation</title>
		<link>https://creospan.com/caching-patterns-in-retrieval-augmented-generation/</link>
		
		<dc:creator><![CDATA[Donna Mathew]]></dc:creator>
		<pubDate>Sat, 21 Dec 2024 22:36:14 +0000</pubDate>
				<category><![CDATA[Insights]]></category>
		<category><![CDATA[Caching in AI]]></category>
		<category><![CDATA[Chunk-based caching]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[Knowledge tree caching]]></category>
		<category><![CDATA[Multilevel dynamic caching]]></category>
		<category><![CDATA[RAG caching patterns]]></category>
		<category><![CDATA[RAG performance optimization]]></category>
		<category><![CDATA[RAG system efficiency]]></category>
		<category><![CDATA[RAG systems]]></category>
		<category><![CDATA[Retrieval-Augmented Generation]]></category>
		<category><![CDATA[Semantic caching]]></category>
		<guid isPermaLink="false">https://creospan.com/?p=1187</guid>

					<description><![CDATA[<p>The post <a href="https://creospan.com/caching-patterns-in-retrieval-augmented-generation/">Caching Patterns in Retrieval Augmented Generation</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="wpb-content-wrapper"><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>Retrieval-Augmented Generation (RAG) systems are transforming the way we interact with large-scale language models by integrating external knowledge retrieval into the generation process. But as powerful as RAG is, it comes with its own performance challenges, especially when working with massive datasets and high query volumes.</p>
<p>One way to make RAG faster and more efficient? Caching.</p>
<p>By strategically caching data, RAG systems can reduce redundancy, speed up response times, and lower operational costs. Let&#8217;s break down the most effective caching patterns for RAG and the trade-offs you need to be aware of.</p>
<h2>Key RAG Caching Patterns:</h2>
<h3>1. Knowledge Tree Caching:</h3>
<p>Organizes intermediate states of retrieved knowledge in a hierarchical structure, caching them in both GPU and host memory.</p>
<p><strong>Benefits:</strong> Efficiently shares cached knowledge across multiple requests, reducing redundant computations and speeding up response times.</p>
<h3>2. Semantic Caching:</h3>
<p>Identifies and caches similar or identical user requests. When a matching request is found, the system retrieves the corresponding information from the cache. This is the most popular one that is readily available with fully managed cloud service providers.</p>
<p><strong>Benefits:</strong> Reduces the need to fetch information from the original source, improving response times.</p>
<h3>3. Chunk-Based Caching:</h3>
<p>Breaks down large documents into smaller chunks and caches these chunks individually.</p>
<p><strong>Benefits:</strong> Improves retrieval speed and accuracy by focusing on smaller, relevant sections of the document.</p>
<h3>4. Multilevel Dynamic Caching:</h3>
<p>Implements a multilevel caching system that dynamically adjusts based on the characteristics of the RAG system and the underlying hardware.</p>
<p><strong>Benefits:</strong> Optimizes the use of memory and computational resources, enhancing overall system performance.</p>
<h3>5. Replacement Policies</h3>
<p>Uses intelligent replacement policies to manage the cache, ensuring that the most relevant and frequently accessed data is retained.</p>
<p><strong>Benefits:</strong> Maintains cache efficiency and relevance, reducing the likelihood of cache misses.</p>
<p>These caching patterns help RAG systems manage and retrieve large volumes of data more efficiently, leading to faster and more accurate responses.</p>
<p>For any RAG implementation we have to prep and plan the pitfalls:</p>
<h2>Retrieval-Augmented Generation (RAG) Caching Pattern Pitfalls:</h2>
<p><strong>Consistency Issues:</strong> Ensuring consistency between the cached data and the source data can be challenging, especially in distributed systems.</p>
<p><strong>Complexity:</strong> Implementing RAG caching patterns can be complex due to the need to manage both retrieval and generation components effectively. This complexity can lead to higher development and maintenance costs.</p>
<p><strong>Latency:</strong> While caching can reduce retrieval times, it may introduce latency in scenarios where the cache needs to be updated frequently. This can affect the overall performance of the system.</p>
<p><strong>Storage Overhead:</strong> You cannot cache without an additional storage, which can be significant depending on the size and frequency of the data being cached.</p>
<p><strong>Staleness:</strong> Cached data can become outdated, leading to the generation of responses based on obsolete information. This is particularly problematic in dynamic environments where information changes rapidly.</p>
<h2>Conclusion</h2>
<p>Even though these patterns are effective in reducing costs and improving response times, they have to be thoroughly validated to ensure the objectives of the RAG implementation are met with effective invalidation techniques like Staleness. Implement a Semantic pattern first and test the model&#8217;s ability, then try out other options.</p>
<p><em>Article written by Krishnam Raju Bhupathiraju.</em></p>
<p>&nbsp;</p>

		</div>
	</div>
</div></div></div></div>
</div><p>The post <a href="https://creospan.com/caching-patterns-in-retrieval-augmented-generation/">Caching Patterns in Retrieval Augmented Generation</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Power of Generative AI and RAG?</title>
		<link>https://creospan.com/why-generative-ai-and-rag/</link>
		
		<dc:creator><![CDATA[Donna Mathew]]></dc:creator>
		<pubDate>Sat, 12 Oct 2024 22:24:38 +0000</pubDate>
				<category><![CDATA[Insights]]></category>
		<category><![CDATA[AI content generation]]></category>
		<category><![CDATA[AI hallucination]]></category>
		<category><![CDATA[Amazon Bedrock RAG]]></category>
		<category><![CDATA[Amazon Kendra]]></category>
		<category><![CDATA[Amazon SageMaker JumpStart]]></category>
		<category><![CDATA[AWS generative AI services]]></category>
		<category><![CDATA[Fine-tuning AI models]]></category>
		<category><![CDATA[Foundation Models (FMs)]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[Large Language Models (LLMs)]]></category>
		<category><![CDATA[RAG pattern]]></category>
		<category><![CDATA[Retrieval-Augmented Generation (RAG)]]></category>
		<guid isPermaLink="false">https://creospan.com/?p=1182</guid>

					<description><![CDATA[<p>The post <a href="https://creospan.com/why-generative-ai-and-rag/">The Power of Generative AI and RAG?</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="wpb-content-wrapper"><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>This article explores three key areas: Generative AI and its patterns, the Retrieval-Augmented Generation (RAG) framework, and AWS’s role in supporting this journey.</p>
<h2>What is Generative AI?</h2>
<p>Generative AI is a type of artificial intelligence focused on the ability of computers to use models to create content like images, text, code, and synthetic data.</p>
<p>The foundation of Generative AI applications are large language models (LLMs) and foundation models (FMs).</p>
<p>Large Language Models (LLMs) are trained effectively on vast volumes of data and use billions of parameters, Then LLM&#8217;s get the ability to generate original output for tasks like completing sentences, translating languages and answering questions.</p>
<p>Foundation models (FMs) are large ML models are pre-trained with the intention that they are to be fine-tuned for more specific language understanding and generation tasks.</p>
<p>Once these models have completed their learning processes, together they generate statistically probable outputs. On prompted (Queried) they can be employed to accomplish various tasks like:</p>
<ul>
<li>Image generation based on existing ones or utilizing the style of one image to modify or create a new one.</li>
<li>Speech oriented tasks such as translation, question/answer generation, and interpretation of the intent or meaning of text.</li>
</ul>
<h2>Generative AI has the following list of design patterns:</h2>
<ul>
<li><strong>Prompt Engineering:</strong> Crafting specialized prompts to guide LLM behavior</li>
<li><strong>Retrieval Augmented Generation (RAG):</strong> Combining an LLM with external knowledge retrieval. Combining best of two capabilities (most recommended).</li>
<li><strong>Fine-tuning:</strong> Adapting a pre-trained LLM to specific data sets of domains. Eg: Specific for Customer service or in Health Care etc.</li>
<li><strong>Pre-training:</strong> Training an LLM from scratch. Needs lot of computing power/time.</li>
</ul>
<h2>Retrieval Augmented Generation (RAG):</h2>

		</div>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div  class="wpb_single_image wpb_content_element vc_align_center">
		
		<figure class="wpb_wrapper vc_figure">
			<div class="vc_single_image-wrapper   vc_box_border_grey"><img decoding="async" width="736" height="566" src="https://creospan.com/wp-content/uploads/2025/05/1721195442713.png" class="vc_single_image-img attachment-large" alt="" title="1721195442713" srcset="https://creospan.com/wp-content/uploads/2025/05/1721195442713.png 736w, https://creospan.com/wp-content/uploads/2025/05/1721195442713-300x231.png 300w" sizes="(max-width: 736px) 100vw, 736px"  data-dt-location="https://creospan.com/why-generative-ai-and-rag/attachment/1721195442713/" /></div>
		</figure>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>&nbsp;</p>
<p>RAG (Retrieval Augmented Generation) is a method to improve LLM response accuracy by giving your LLM access to external data sources.</p>
<p>LLMs are trained on enormous data sets, but they don’t have specific context for your business, industry, or customer specific needs. RAG adds that crucial layer of information for LLMs to make effective closures.</p>
<h2>To understand RAG, we need to explore the limitations of LLMs.</h2>
<h4>Limitations of LLM&#8217;s:</h4>
<ul>
<li><strong>Hallucination:</strong> LLM&#8217;s try to present false information when it does not have the answer or even there is no answer.</li>
<li><strong>Outdated Info:</strong> Presenting out-of-date or generic information when the user wants a specific, accurate response.</li>
<li><strong>Tech Confusion:</strong> Generating inaccurate responses due to terminology confusion, wherein different training sources use the similar terminology about different things.</li>
<li><strong>Unauthorized:</strong> Creating a response from non-authoritative sources.</li>
</ul>
<h2>RAG works in three stages:</h2>
<ul>
<li><strong>Retrieval:</strong> When a request reaches LLM and the system looks for relevant information that informs the final response.  It searches through an external dataset or document collection to find most relevant pieces of information. This dataset could be a curated knowledge base, or any extensive collection of text, images, videos, and audio or even your local database.</li>
<li><strong>Augmentation:</strong> In this step the query is enhanced with the information retrieved in the previous step.</li>
<li><strong>Generation:</strong> The final augmented response or output is generated. Your LLM uses the additional context provided by the augmented input to produce an answer that is not only relevant to the original query but enriched with information from external sources.</li>
</ul>
<h3>Customer service RAG use cases:</h3>
<p><strong>Personalized recommendations:</strong> Generate personalized product recommendations based on customer&#8217;s browsing patterns or past interactions and preferences</p>
<p><strong>Advanced chatbots:</strong> RAG empowers chatbots to answer complex questions and provide personalized support to customers – improving customer satisfaction and reducing support costs.</p>
<p><strong>Knowledge base search:</strong> Quickly retrieve relevant information from internal knowledge bases to answer customer inquiries faster and more accurately.</p>
<h2>AWS had the following ways of support for RAG:</h2>
<p><strong>Amazon Bedrock:</strong> Is a fully-managed service that offers a choice of high-performing foundation models—along with a broad set of capabilities—to build generative AI applications while simplifying development and maintaining privacy and security. With knowledge bases for Amazon Bedrock, you can connect FMs to your data sources for RAG in just a few clicks. Vector conversions, retrievals, and improved output generation are all handled automatically.</p>
<p><strong>Amazon Kendra:</strong> Is for organizations managing their own RAG.A highly-accurate enterprise search service powered by machine learning. It provides an optimized Kendra Retrieve API that you can use with Amazon Kendra’s high-accuracy semantic ranker as an enterprise retriever for your RAG workflows.</p>
<p><strong>Amazon SageMaker:</strong> Amazon SageMaker &#8211; JumpStart is a ML hub with FMs, built-in algorithms, and prebuilt ML solutions that you can deploy with just a few clicks. You can speed up RAG implementation by referring to existing SageMaker notebooks and code examples.</p>
<p><em>Article written by Krishnam Raju Bhupathiraju.</em></p>
<p>&nbsp;</p>

		</div>
	</div>
</div></div></div></div>


<p></p>
</div><p>The post <a href="https://creospan.com/why-generative-ai-and-rag/">The Power of Generative AI and RAG?</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Small Language Models are the New Big Thing in AI</title>
		<link>https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/</link>
		
		<dc:creator><![CDATA[Donna Mathew]]></dc:creator>
		<pubDate>Wed, 24 Jul 2024 22:04:17 +0000</pubDate>
				<category><![CDATA[Insights]]></category>
		<category><![CDATA[AI for mobile devices]]></category>
		<category><![CDATA[Artificial intelligence]]></category>
		<category><![CDATA[Data Privacy]]></category>
		<category><![CDATA[Edge AI]]></category>
		<category><![CDATA[Edge computing AI]]></category>
		<category><![CDATA[Generative AI]]></category>
		<category><![CDATA[IoT]]></category>
		<category><![CDATA[Large Language Models (LLMs)]]></category>
		<category><![CDATA[Lightweight AI]]></category>
		<category><![CDATA[On-device AI]]></category>
		<category><![CDATA[Small Language Models (SLMs)]]></category>
		<guid isPermaLink="false">https://creospan.com/?p=1213</guid>

					<description><![CDATA[<p>The post <a href="https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/">Small Language Models are the New Big Thing in AI</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="wpb-content-wrapper"><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>Throughout the history of technology, we’ve witnessed the evolution of software applications—from massive monolithic servers to sleek microservices and miniaturized platforms. History, indeed, has a way of repeating itself, and Generative AI is no exception to this cyclical progression.</p>
<p>Today when you consume ChatGPT, Gemini, CoPilot the intelligence comes from the centralized computing:</p>
<h5>Consuming Large Language Model</h5>

		</div>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div  class="wpb_single_image wpb_content_element vc_align_center">
		
		<figure class="wpb_wrapper vc_figure">
			<div class="vc_single_image-wrapper   vc_box_border_grey"><img decoding="async" width="1024" height="620" src="https://creospan.com/wp-content/uploads/2025/05/1-1024x620.png" class="vc_single_image-img attachment-large" alt="" title="1" srcset="https://creospan.com/wp-content/uploads/2025/05/1-1024x620.png 1024w, https://creospan.com/wp-content/uploads/2025/05/1-300x182.png 300w, https://creospan.com/wp-content/uploads/2025/05/1-768x465.png 768w, https://creospan.com/wp-content/uploads/2025/05/1.png 1488w" sizes="(max-width: 1024px) 100vw, 1024px"  data-dt-location="https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/attachment/1/" /></div>
		</figure>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>Now that you&#8217;ve delved into the realm of Large Language Models (LLMs), mastering techniques like Prompt Engineering and Retrieval-Augmented centralized intelligence Generation (RAG) patterns, it&#8217;s time to shift gears. The focus is no longer on centralized intelligence, but rather on bringing AI-driven capabilities closer to customers or end user devices.</p>
<p>With the devices limitation we could not deploying LLM intelligence directly on the devices. Now comes the rise of SLM (Small Language Models), which are built for everyday use.</p>
<h2>The Rise of SLMs in Everyday Use</h2>
<p>From powering smarter mobile applications to revolutionizing real-time translations and beyond, small language models are quietly shaping our digital lives. Developers and researchers are increasingly favoring SLMs for their ability to deliver lightweight yet impactful solutions.</p>
<h2>What are Small Language Models?</h2>
<p>Inherently, Small Language Models (SLMs) are smaller counterparts of Large Language Models. They have fewer parameters and are more lightweight and faster in inference time. We can consider models with billions and trillion of parameters as LLMs (the largest Chat GPT-4o: has 1.8 trillion parameters), demanding resource-heavy training and inferences. The definition of a Small Language Model can vary among different authors.</p>
<h2>How are they different from LLMs?</h2>
<p>Unlike large language models (LLMs), with primary purpose of general-purpose capabilities across a variety of applications, SLMs are optimized for efficiency, making them ideal for deployment in resource-constrained environments such as mobile devices, point of sale, IOT and edge computing systems.</p>
<p>SLMs are compact versions of Language Models, and they excel in two main areas:</p>
<ol>
<li>SLMs are suitable for Edge Devices, offering businesses benefits such as cost reduction, offline usage, or enhanced data privacy.</li>
<li>SLMs facilitate speeding up R&amp;D progress, swiftly testing new ideas, benchmarking at scale, and iterating relatively fast. Even retraining SLMs (even from scratch) is feasible for small groups with access to home-grade GPUs.</li>
</ol>
<h2>SLM (Small Language Model) vs. LLM (Large Language Model) Comparison:</h2>

		</div>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div  class="wpb_single_image wpb_content_element vc_align_center">
		
		<figure class="wpb_wrapper vc_figure">
			<div class="vc_single_image-wrapper   vc_box_border_grey"><img loading="lazy" decoding="async" width="905" height="240" src="https://creospan.com/wp-content/uploads/2025/05/2.png" class="vc_single_image-img attachment-large" alt="" title="2" srcset="https://creospan.com/wp-content/uploads/2025/05/2.png 905w, https://creospan.com/wp-content/uploads/2025/05/2-300x80.png 300w, https://creospan.com/wp-content/uploads/2025/05/2-768x204.png 768w" sizes="(max-width: 905px) 100vw, 905px"  data-dt-location="https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/attachment/2/" /></div>
		</figure>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<p>Small Language Models (SLMs) are designed for efficiency and specialization, making them ideal for a variety of use cases across industries. Here are some notable applications:</p>
<ol>
<li>Real-Time Mobile Apps
<ul>
<li>Customer Support: SLMs can power chatbots and virtual assistants on websites or apps, providing instant responses to customer queries.</li>
<li>Sentiment Analysis: Analyze customer feedback from social media and integrate insights into customer data platforms.</li>
<li>Personalized Offers: Generate tailored promotions and recommendations based on user profiles and behavior.</li>
<li>Self-Healing Systems: SLMs can enable networks to automatically detect and resolve issues without human intervention.</li>
</ul>
</li>
<li>Edge Computing
<ul>
<li>IoT Devices: SLMs enable smart home devices, like thermostats or speakers, to process commands locally without relying on cloud servers or choking the internet.</li>
<li>Connected Cars: They can assist with navigation, voice commands, and diagnostics directly within the vehicle.</li>
</ul>
</li>
<li>Domain-Specific Applications
<ul>
<li>Retail: SLMs can enhance Point-of-Sale (POS) systems by offering personalized recommendations or promotions.</li>
<li>Finance: Used for fraud detection, transaction analysis, and customer service in banking apps.</li>
</ul>
</li>
<li>Privacy-Sensitive Environments
<ul>
<li>Data Masking: SLMs can anonymize sensitive data, such as personally identifiable information (PII), ensuring compliance with privacy regulations.</li>
<li>On-Device Processing: By running locally, SLMs reduce the need to send data to external servers, enhancing security.</li>
<li>5. Specialized Content Creation</li>
<li>Marketing: SLMs can generate targeted ad copy or social media posts for specific audiences.</li>
<li>Technical Writing: Used to create concise and accurate documentation for niche industries.</li>
</ul>
</li>
</ol>
<p>Now let&#8217;s redraw the same image with SLM, the compute can be deployed on every entry with specific customizations:</p>
<h5>Customized Small Language Model&#8217;s for Specific Use Cases</h5>

		</div>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div  class="wpb_single_image wpb_content_element vc_align_center">
		
		<figure class="wpb_wrapper vc_figure">
			<div class="vc_single_image-wrapper   vc_box_border_grey"><img loading="lazy" decoding="async" width="1024" height="530" src="https://creospan.com/wp-content/uploads/2025/05/3-1024x530.png" class="vc_single_image-img attachment-large" alt="" title="3" srcset="https://creospan.com/wp-content/uploads/2025/05/3-1024x530.png 1024w, https://creospan.com/wp-content/uploads/2025/05/3-300x155.png 300w, https://creospan.com/wp-content/uploads/2025/05/3-768x398.png 768w, https://creospan.com/wp-content/uploads/2025/05/3-1536x795.png 1536w, https://creospan.com/wp-content/uploads/2025/05/3.png 1920w" sizes="(max-width: 1024px) 100vw, 1024px"  data-dt-location="https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/attachment/3/" /></div>
		</figure>
	</div>
</div></div></div></div><div class="vc_row wpb_row vc_row-fluid"><div class="wpb_column vc_column_container vc_col-sm-12"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column wpb_content_element " >
		<div class="wpb_wrapper">
			<h2>Conclusion</h2>
<p>Small Language Models (SLMs) are revolutionizing the way we think about AI—bringing the power of intelligent computation closer to end users. They are compact, efficient, and purpose-built to address specific use cases across industries, from retail and IoT devices to connected vehicles and telecom.</p>
<p>By processing data closer to the edge, SLMs not only reduce latency but greatly improve privacy and accessibility, making them the future of responsive, on-device intelligence. As we continue to innovate and adapt these models, the possibilities for seamless integration, improved customer experiences, and optimized operational efficiencies are boundless.</p>
<p>The future of AI isn’t just large-scale intelligence—it’s small, smart, and specialized. Let’s embrace this next frontier.</p>
<p><em>Article Written by Krishnam Raju Bhupathiraju.</em></p>
<p>&nbsp;</p>

		</div>
	</div>
</div></div></div></div>
</div><p>The post <a href="https://creospan.com/small-language-models-are-the-new-big-thing-in-ai/">Small Language Models are the New Big Thing in AI</a> appeared first on <a href="https://creospan.com">Creospan</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
