<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Megaputer Intelligence</title>
	<atom:link href="https://www.megaputer.com/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.megaputer.com</link>
	<description>Your Knowledge Partner</description>
	<lastBuildDate>Tue, 24 Mar 2026 00:02:52 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>hourly</sy:updatePeriod>
	<sy:updateFrequency>1</sy:updateFrequency>
	<generator>https://wordpress.org/?v=5.0.22</generator>

<image>
	<url>https://www.megaputer.com/wp-content/uploads/favicon.png</url>
	<title>Megaputer Intelligence</title>
	<link>https://www.megaputer.com</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Can AI Understand Your Customers&#8217; Needs?</title>
		<link>https://www.megaputer.com/deep-listen-post/</link>
		<pubDate>Tue, 24 Feb 2026 20:54:16 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35564</guid>
		<description><![CDATA[<p>The post <a rel="nofollow" href="https://www.megaputer.com/deep-listen-post/">Can AI Understand Your Customers&#8217; Needs?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Right now, your customer feedback is telling you exactly what&#8217;s wrong – and you&#8217;ll never read it.</p>
<p>Not because you don&#8217;t care – but because the feedback is so diverse and volume is overwhelming. Thousands of reviews, survey responses, support tickets, employee comments, interview transcripts, and open-ended form fields pour in constantly. There&#8217;s simply no practical way for a human team to read through all of it, find the key patterns, and turn those patterns into action. So most of that data sits untouched, a goldmine that never gets mined.</p>
<p>It seems like AI should be able to solve this problem. After all, AI can read all of your data. But if you&#8217;ve tried feeding your data into an AI system and asking for answers, you probably didn&#8217;t get what you wanted. An ordinary AI will just hallucinate, and a RAG system can’t find patterns that only emerge over thousands of records.</p>
<p>What you need is not a smarter prompt, a bigger model, or a more sophisticated RAG system. You need a system that was designed from the ground up to listen for patterns over vast swathes of data. You need a system that reads all your documents to identify your problems, determines which problems are driving the selected KPIs, and writes a report on each one &#8211; with citations from the data to verify that it isn&#8217;t hallucinating.</p>
<p>This is exactly what we have.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><strong>What Is <em>Deep.Listen</em>?</strong></h2>
<p><em><strong>Deep.Listen</strong></em>™ is a specialized system for feedback analysis powered by AI agents that work together to do what no single AI model can do – read all of your data, find the top issues driving your selected KPIs, and write a clear report on each of them. These AI insight reports provide actionable recommendations for fixing the most important problems.</p>

		</div>
	</div>
</div></div></div><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="1024" height="683" src="https://www.megaputer.com/wp-content/uploads/deep-listen-logo-1024x683.png" class="attachment-large size-large" alt="" srcset="https://www.megaputer.com/wp-content/uploads/deep-listen-logo-1024x683.png 1024w, https://www.megaputer.com/wp-content/uploads/deep-listen-logo-300x200.png 300w, https://www.megaputer.com/wp-content/uploads/deep-listen-logo-600x400.png 600w, https://www.megaputer.com/wp-content/uploads/deep-listen-logo-768x512.png 768w, https://www.megaputer.com/wp-content/uploads/deep-listen-logo.png 1536w" sizes="(max-width: 1024px) 100vw, 1024px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>From your perspective, the process is straightforward:</p>
<ol>
<li><strong> You provide the data:</strong> Send us your customer reviews, stakeholder feedback, or other documents in any standard format.</li>
<li><strong> We run the analysis:</strong> Our system analyzes your data, typically completing within the same day, regardless of volume. The system automatically identifies the key issues that are driving your chosen KPIs and then generates a written report on each of them.</li>
<li><strong> Human in the loop:</strong> After our human analysts have reviewed the results, we deliver an AI-generated and analyst-audited report on each of the most pressing issues that are driving your KPIs. This process may take a few days.</li>
</ol>
<p>You give us data. We give you answers. It’s as easy as that.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><strong>A Real-World Example: Finding Hidden Liabilities</strong></h2>
<p>In a recent engagement, a customer had tens of thousands of reviews that were sitting unread on their website. We ran these reviews through Deep.Listen, specifying the star rating as the target KPI. Within hours, the system surfaced something the team hadn&#8217;t flagged: multiple systematic labeling issues were causing customers to feel deceived about their purchases of certain juice products.</p>
<p>The business impact was significant. Customers affected by the confusion rated their purchases an average of 2.2 stars, compared to a 4.6-star average across the full dataset. Reviews were littered with words like &#8220;deceptive&#8221; and &#8220;misleading,&#8221; and some customers complained of health issues.</p>
<p>The problem was complex.</p>
<p><em>Deep.Listen</em> identified dozens of product pages in which a “New Look” image overlaid the product’s hero image in a way that obstructed essential regulatory disclosures. In some cases, customers believed that they were buying 100% juice because the “with added ingredients” disclosure was obstructed by the “New Look” image overlay. Beyond lost trust and repeat purchases, this labeling issue created significant legal liability.</p>
<p><em>Deep.Listen</em> also discovered a labeling issue that was causing some customers to believe they were purchasing pure cranberry juice when, in fact, they were purchasing a blend. Some customers reported allergies to other fruits in the blend, while others reported that they had intended to buy pure cranberry juice to treat a UTI infection, which was not effectively treated by this cranberry-flavored blend. Both of these issues put the customers’ health at risk, degraded customer satisfaction, and generated legal liability for the company.</p>
<p>We did not direct <em>Deep.Listen</em> to look for liability risks or to examine product labeling – it discovered these issues on its own. In its’ drive to maximize the target KPI, it automatically found the problems that mattered most.</p>
<p>These hidden liability risks were just one of several issues that <em>Deep.Listen</em> discovered for the company. These issues were hiding in plain sight across thousands of reviews. It took the AI hours to find them. It could have taken a human team months – if they&#8217;d ever found them at all.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><strong>How It Works</strong></h2>
<p>Under the hood, <em>Deep.Listen</em> is a four-agent system.</p>
<p>First, a pair of cooperating agents reads your documents and organizes them into topics and subtopics. These agents are searching for any issues that could potentially drive your KPIs, not just what you told them to look for. Then a third agent calculates each topic’s significance using machine learning to determine which issues are actually driving your target KPIs.</p>
<p>Finally, a fourth agent investigates each significant topic – reading every single piece of feedback associated with it – and writes a plain-language report that explains the issue and makes specific recommendations for improvement. Every report is reviewed by a human analyst for errors, and is grounded in citations from your data, so you can verify that what you&#8217;re reading is real – not a hallucination. Sometimes these reports are produced through multiple rounds of clarifying conversation between the AI and the human in the loop.</p>
<p>The result is a report with actionable recommendations to unleash your KPIs.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><strong>The Applications Are Broad</strong></h2>
<p>Any organization that collects unstructured text data can benefit from <em>Deep.Listen</em>. A few examples include:</p>
<ul>
<li><strong>Retailers and product companies:</strong> analyzing customer reviews</li>
<li><strong>Healthcare organizations:</strong> mining patient feedback for care gaps and satisfaction drivers</li>
<li><strong>HR:</strong> reviewing employee engagement surveys or exit interviews</li>
<li><strong>Government and NGOs:</strong> analyzing public consultation responses or stakeholder feedback</li>
<li><strong>Market researchers:</strong> reading open-ended survey responses at scale</li>
<li><strong>Financial services firms:</strong> reviewing client feedback, complaints, or advisor notes</li>
</ul>
<p>If you are receiving customer feedback, <em>Deep.Listen</em> can help you organize it and prioritize what really matters.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><strong>Ready to See What&#8217;s Hiding in Your Data?</strong></h2>
<p><em>Deep.Listen</em> can be applied to any data source, any format, and any volume of text. You don&#8217;t need to know what you&#8217;re looking for in advance – you just need data and a target KPI. We will do the rest and provide you with actionable insights and recommendations.</p>
<p>Contact <strong>Megaputer Intelligence</strong> today to discuss your data and needs and receive a custom proposal.</p>
<p>info@megaputer.com</p>
<p><em>Quotes available upon request.</em></p>
<p>&nbsp;</p>
<p style="text-align: center;">View the webinar recording: <strong><em>Deep.Listen &#8211; Actionable Insights from Customer Feedback</em></strong></p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-video ratio_16x9 align_center" style="max-width:768px;"><div class="w-video-h"><iframe src="//www.youtube.com/embed/X8J_YJ5eq18?rel=0" allowfullscreen="1"></iframe></div></div></div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/deep-listen-post/">Can AI Understand Your Customers&#8217; Needs?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Webinar: Accelerating Drug Discovery with AI Agents</title>
		<link>https://www.megaputer.com/webinar-accelerating-drug-discovery-with-ai/</link>
		<pubDate>Wed, 21 Jan 2026 18:08:17 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35511</guid>
		<description><![CDATA[<p>The post <a rel="nofollow" href="https://www.megaputer.com/webinar-accelerating-drug-discovery-with-ai/">Webinar: Accelerating Drug Discovery with AI Agents</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-video ratio_16x9 align_center" style="max-width:768px;"><div class="w-video-h"><iframe src="//www.youtube.com/embed/ALsIlZMtetk?rel=0" allowfullscreen="1"></iframe></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>AI is transforming early-stage drug discovery by accelerating key tasks such as target identification, molecular design, virtual screening, ADMET prediction, and drug repurposing. In this webinar, we’ll focus on one of the most critical steps: <strong>using AI agents to speed up target identification</strong>.</p>
<p>By combining scientific literature with experimental and omics data, AI can help prioritize disease-relevant genes, proteins, and pathways &#8211; enabling faster generation and evaluation of therapeutic hypotheses. This supports both target-based discovery and indication expansion for novel or existing molecules.</p>
<p>Building reliable, pathway-centric knowledge systems isn’t easy. Data volume, quality, bias, and integration remain major challenges.</p>
<p>Topics covered:</p>
<ul>
<li>Prioritizing disease-relevant genes, proteins, and pathways with AI</li>
<li>Creating AI agents without coding</li>
<li>Typical challenges and how to overcome them</li>
<li>Live demo of a practical solution and results</li>
</ul>

		</div>
	</div>
</div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/webinar-accelerating-drug-discovery-with-ai/">Webinar: Accelerating Drug Discovery with AI Agents</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>RxLaunch: AI Agents for Pharma Launch Monitoring</title>
		<link>https://www.megaputer.com/rxlaunch-ai-agents-for-pharma-launch-monitoring/</link>
		<pubDate>Tue, 21 Oct 2025 21:44:12 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35482</guid>
		<description><![CDATA[<p>The post <a rel="nofollow" href="https://www.megaputer.com/rxlaunch-ai-agents-for-pharma-launch-monitoring/">RxLaunch: AI Agents for Pharma Launch Monitoring</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="300" height="144" src="https://www.megaputer.com/wp-content/uploads/rxlauch_logo-300x144.png" class="attachment-medium size-medium" alt="RxLaunch logo" srcset="https://www.megaputer.com/wp-content/uploads/rxlauch_logo-300x144.png 300w, https://www.megaputer.com/wp-content/uploads/rxlauch_logo-600x289.png 600w, https://www.megaputer.com/wp-content/uploads/rxlauch_logo.png 659w" sizes="(max-width: 300px) 100vw, 300px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h5 style="text-align: center;"><strong>Catch issues before they derail your product launch.</strong></h5>
<h4><strong>The Problem: When Insights Arrive Too Late</strong></h4>
<p>It&#8217;s week three after the new product launch. Early monitoring shows a growing spike of negative sentiment toward your product compared to competing products – though sentiment was well above parity on launch day. Pharmacies are reporting patient concerns. Medical information calls are higher than expected. You have thousands of pieces of feedback but no clear picture of what&#8217;s driving the problem.</p>
<p>Using ordinary tools, your team might take weeks to comb through the textual feedback to reveal those drivers of negative sentiment that weren’t flagged for monitoring in advance of the launch day. Three weeks from now, you&#8217;ll be at week six. Thousands of patients will have discontinued. Prescribers will have formed opinions. The launch trajectory will be set.</p>
<p>The speed of identifying and resolving key issues determines launch success – or failure. By the time your analysis delivers answers several weeks later, the critical early period has already passed: you&#8217;re just getting an explanation of what went wrong instead of arming yourself with the insights you need to fix the situation.</p>
<h4><strong>RxLaunch™ Provides Timely Insights</strong></h4>
<p>RxLaunch™ by Megaputer Intelligence is a pharmaceutical launch intelligence platform that uses AI agents to transform launch feedback into actionable insights in just a matter of hours, allowing your team to recognize and respond to key issues as they emerge in real-time.</p>
<p>You load in all your feedback: patient comments, prescriber notes, medical information calls, pharmacy reports, support tickets, and other data. RxLaunch&#8217;s AI agents perform semantic analysis of the texts and generate a hierarchy of informative tags that organize the feedback. Next, the system determines which issues are driving the key performance indicators (KPIs) such as stakeholder sentiment, adherence, or patient discontinuation. Notably, RxLaunch does not require you to specify the patterns that you are looking for in advance.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="600" height="308" src="https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-600x308.png" class="attachment-shop_single size-shop_single" alt="" srcset="https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-600x308.png 600w, https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-300x154.png 300w, https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-1024x525.png 1024w, https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-768x394.png 768w, https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1-780x400.png 780w, https://www.megaputer.com/wp-content/uploads/rxlaunch_explained-1.png 1520w" sizes="(max-width: 600px) 100vw, 600px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;"><em>RxLaunch™ ingests launch feedback, discovers emerging issues, and ranks the issues by impact on selected KPIs</em></p>
<p>&nbsp;</p>
<p>You receive a prioritized action list showing exactly which issues to address first. This list shows which issues are actually driving your KPIs, and which issues can be safely ignored. No need to train a neural network. No category definitions. No technical expertise required. The RxLaunch agents do all the work and deliver insights in an easy to comprehend form.</p>
<h4><strong>How One Pharma Company Saved Its Launch</strong></h4>
<p><em>Note: The following scenario is a hypothetical example designed to illustrate potential applications and outcomes. Specific metrics and results are for illustrative purposes only.</em></p>
<p>A pharmaceutical company launched a chronic disease medication. Launch day went smoothly. Week one looked promising. But by week three, early data showed a problem: sentiments were slipping compared to competing products. The launch was losing momentum at the worst possible time &#8211; during the critical early weeks when market position gets established.</p>
<p>The company had the feedback: pharmacist notes, medical information calls, patient comments, online discussions &#8211; tens of thousands of text records in all. What they didn&#8217;t have was clarity. But the use of RxLaunch enabled them to analyze the incoming data daily, automatically organizing and prioritizing issues as they emerged.</p>
<p>RxLaunch revealed that the drop in sentiment from patients and providers was being driven by a surprising issue &#8211; “refrigeration”. This was unexpected, since the product did not require refrigeration. Drilling down into the subtopics of the refrigeration category showed “can&#8217;t travel with medication,” “does not fit with lifestyle,” and surprisingly, mentions of a specific injection pen for an unrelated product in a different therapeutic area.</p>
<p>A closer inspection of the records showed that the product’s injection pen bore a superficial resemblance to another widely-used injection pen for a medication that required refrigeration. Even though the injection pen was for a different therapeutic category, and made by a different manufacturer, the superficial similarities in mechanism, appearance, and size caused some patients to draw a parallel. Those patients who&#8217;d used the other injector pen were instinctively storing the new medication the same way &#8211; and convincing others to do the same.</p>
<p>The refrigeration myth was spreading virally. During week one, just 2-3% of patients were refrigerating due to individual confusion. But by week three, that number had risen to 14%. Projections showed that the myth would have spread to a majority of patients by week eight if the trend had continued.</p>
<p>Refrigeration didn&#8217;t harm the drug, but it destroyed adherence. Patients could forget doses when traveling, skip injections when away from home overnight, or just find the cold-storage routine burdensome. Many who believed the myth said that the product was &#8220;too inconvenient,&#8221; and often stopped refilling.</p>
<h4><strong>Rapid Response and Recovery</strong></h4>
<p>Acting on this insight, the company moved immediately. Within a day, the patient support and communications teams began posting clarifications in online communities where misinformation was spreading. In parallel, pharmacy hotlines and patient support staff were briefed to proactively correct the misconception during inbound calls and chat interactions. Within a week, pharmacists received an updated counseling cue card: “Storage: room temperature; do not refrigerate.” Automated text messages on day two and day seven reminded patients to store the product at room temperature.</p>
<p>Within eight weeks, belief in the refrigeration myth had dropped to almost zero. Patient and provider sentiment recovered, regaining full parity with competing products. Six-month refill rates came in at 72%, two points above projection.</p>
<p>Traditional analysis might have taken weeks to find the issue. In those extra weeks, the product’s image would have been irreparably tarnished. While the myth would have eventually been discovered and put down, first impressions among patients and healthcare providers cannot easily be corrected. The product would have been permanently positioned as &#8220;the inconvenient one,&#8221; with low adherence and high discontinuation.</p>
<p>RxLaunch found what no one thought to look for, quantified what might have been dismissed, and delivered it before viral spread made the problem unfixable. In this case, the speed of discovery and response to new insights drew the line between success and failure.</p>
<h4><strong>What make RxLaunch Different</strong></h4>
<p>The refrigeration myth was just one of many insights brought to light by RxLaunch. While competing product launch monitoring solutions rely on crude algorithms that struggle to find patterns beyond what they are told to look for, RxLaunch takes a fundamentally different approach. Our AI agents work like expert human analysts &#8211; discovering and organizing issues in real-time, making intelligent decisions at every step.</p>
<p>These agents don&#8217;t just search for predefined problems &#8211; they autonomously identify what&#8217;s actually driving your KPIs, even unexpected issues you never thought to monitor. As new patterns emerge in your feedback, the agents adapt and refine their analysis automatically.</p>
<p>The result: comprehensive, intelligent, actionable insights from day one. No blind spots from forgotten categories. No critical patterns buried in the noise of massive unstructured datasets. Just intelligent agents working around the clock to surface the issues that matter most for launch success.</p>
<h4><strong>Who This Is For</strong></h4>
<p>RxLaunch is designed for pharmaceutical teams responsible for product launch success:</p>
<ul>
<li><strong>Launch Directors</strong> who need real-time visibility into emerging issues that could derail commercial performance during the critical early weeks</li>
<li><strong>Brand Teams</strong> who must protect market positioning and respond quickly to unexpected competitive dynamics or patient perception shifts</li>
<li><strong>Medical Affairs</strong> who handle complex medical information inquiries and need to identify systemic clinical or product questions emerging across touchpoints</li>
<li><strong>Patient Support Teams</strong> who manage frontline feedback but lack tools to quickly identify patterns that signal broader adherence or satisfaction issues</li>
<li><strong>Commercial Operations Leaders</strong> who must make fast, data-driven decisions to protect launch investments and adjust tactics in real-time</li>
<li><strong>Market Access Teams</strong> who need to understand and address payer, pharmacy, and access barriers before they become widespread adoption blockers</li>
<li><strong>Field Teams</strong> who need actionable intelligence about prescriber concerns and patient feedback to guide their conversations and remove obstacles to uptake</li>
</ul>
<p>If your team is responsible for a product launch and you&#8217;re swimming in feedback data but struggling to identify which issues actually matter, consider RxLaunch. It transforms unstructured data into clear, prioritized action items &#8211; before problems become unfixable.</p>
<p>&nbsp;</p>
<p>Your next launch is coming. The feedback will accumulate. The critical issues will emerge in early data. The only question: will you have insights when the launch opportunity window is still open, or will you only receive a report explaining what went wrong when it&#8217;s too late?</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;">See <a href="https://www.megaputer.com/contact/"><strong>how RxLaunch works</strong></a></p>
<p style="text-align: center;">Read more about the <strong><a href="https://www.megaputer.com/what-is-just-in-time-taxonomy/">AI agents powering RxLaunch</a></strong></p>

		</div>
	</div>
</div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/rxlaunch-ai-agents-for-pharma-launch-monitoring/">RxLaunch: AI Agents for Pharma Launch Monitoring</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>What is Just-in-Time Taxonomy?</title>
		<link>https://www.megaputer.com/what-is-just-in-time-taxonomy/</link>
		<pubDate>Tue, 23 Sep 2025 21:35:00 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35454</guid>
		<description><![CDATA[<p>The post <a rel="nofollow" href="https://www.megaputer.com/what-is-just-in-time-taxonomy/">What is Just-in-Time Taxonomy?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Today is the big day. Maybe your product just launched, your new feature went live, or the results of the survey are in. Thousands of comments, reviews, support tickets (or whatever other data you can imagine) are flooding your system, and your team desperately needs to make sense of it all. But first, someone has to organize all this feedback into a meaningful form.</p>
<p>Using traditional methods, a team might need 3-4 weeks to taxonomize the data, assuming they can work on the problem full-time. By then, the crucial issues will have festered, opportunities will have passed, and the critical window for rapid response will have already slammed shut. You can also use a pre-built taxonomy, but this will capture only the categories that you were able to anticipate in advance.</p>
<p>An AI-powered Just-in-Time Taxonomy system solves this problem. Instead of spending weeks manually creating categories or trying to force your data into prechosen boxes where it does not really fit, the Just-in-Time Taxonomy system employs a pair of collaborating AI agents to analyze your data and build a custom taxonomy in under an hour. This gives you a taxonomy that was built<em> from your own data</em> on day one.</p>
<p>But how does this work? And when does it make sense to use this approach instead of traditional methods?</p>
<h2>How it works: A two-agent system</h2>
<p>The Just-in-Time taxonomy system uses two specialized AI agents that work in tandem to analyze your data.</p>
<p>The <strong>Category Building Agent</strong> (aka Category Builder) is responsible for discovering what categories actually exist in your data. It works through a three-stage process: first, it reads through your content and generates candidate labels for the themes it finds. For example, if your data was customer feedback about a mobile app, the agent might identify candidate labels like &#8220;battery drain,&#8221; &#8220;login problems,&#8221; &#8220;slow loading,&#8221; &#8220;confusing navigation,&#8221; &#8220;crashes on startup,&#8221; &#8220;great design,&#8221; &#8220;helpful support,&#8221; and dozens of others as it processes each piece of feedback.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="1024" height="683" src="https://www.megaputer.com/wp-content/uploads/category-building-agent-1-1024x683.png" class="attachment-large size-large" alt="Category Building Agent" srcset="https://www.megaputer.com/wp-content/uploads/category-building-agent-1-1024x683.png 1024w, https://www.megaputer.com/wp-content/uploads/category-building-agent-1-300x200.png 300w, https://www.megaputer.com/wp-content/uploads/category-building-agent-1-600x400.png 600w, https://www.megaputer.com/wp-content/uploads/category-building-agent-1-768x512.png 768w, https://www.megaputer.com/wp-content/uploads/category-building-agent-1.png 1536w" sizes="(max-width: 1024px) 100vw, 1024px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;"><em>The Category Builder is a three-stage pipeline</em></p>
<p>&nbsp;</p>
<p>In the second stage, the agent groups these candidate labels through semantic clustering. It recognizes when different labels are referring to the same concept. For example, if customers mentioned &#8220;battery life,&#8221; &#8220;battery drain,&#8221; and &#8220;poor battery performance,&#8221; these would be clustered together because they&#8217;re all talking about the same issue using different words. Similarly, &#8220;log in issues,&#8221; &#8220;login problems,&#8221; and &#8220;can&#8217;t sign in&#8221; would group into one cluster, while &#8220;crashes,&#8221; &#8220;app crashes,&#8221; and &#8220;freezing&#8221; might form another cluster.</p>
<p>Finally, in the third stage, the agent selects the best representative label for each cluster. From the battery-related cluster, it might choose &#8220;Battery Issues&#8221; as the clearest label. From the login cluster, it might select &#8220;Login Problems&#8221; as the most straightforward term.</p>
<p>This process eliminates redundancy and creates a clean taxonomy where each category represents a distinct concept, even when customers express the same issue in many different ways.</p>
<p>The <strong>Document Tagging Agent</strong> (aka Document Tagger) takes any new piece of content and maps it to the appropriate categories from the taxonomy the Category Builder created. Think of it as the agent that actually does the organizing work once the categories exist.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="1024" height="683" src="https://www.megaputer.com/wp-content/uploads/document-tagging-agent-1024x683.png" class="attachment-large size-large" alt="Document Tagging Agent" srcset="https://www.megaputer.com/wp-content/uploads/document-tagging-agent-1024x683.png 1024w, https://www.megaputer.com/wp-content/uploads/document-tagging-agent-300x200.png 300w, https://www.megaputer.com/wp-content/uploads/document-tagging-agent-600x400.png 600w, https://www.megaputer.com/wp-content/uploads/document-tagging-agent-768x512.png 768w, https://www.megaputer.com/wp-content/uploads/document-tagging-agent.png 1536w" sizes="(max-width: 1024px) 100vw, 1024px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;"><em>The Document Tagger maps records to the categories discovered by the Category Builder</em></p>
<p>&nbsp;</p>
<p>When a new piece of feedback comes in—let&#8217;s say a customer writes &#8220;This app is killing my phone&#8217;s battery and takes forever to load anything&#8221;—the Document Tagger reads this content and identifies which categories apply. In this case, it would likely tag this feedback with both &#8220;Battery Issues&#8221; and whatever category the system created for loading/performance problems.</p>
<p>The Document Tagger doesn&#8217;t just look for exact keyword matches. It understands the meaning and context of the content. So if someone writes &#8220;My phone dies so fast when I use this app,&#8221; the agent recognizes this as a battery issue even though they didn&#8217;t use the word &#8220;battery.&#8221; Similarly, &#8220;The app is super sluggish&#8221; would get tagged appropriately even though &#8220;sluggish&#8221; wasn&#8217;t in the original candidate labels.</p>
<p>This agent can process content as it arrives, instantly categorizing new feedback, support tickets, or survey responses using the taxonomy that emerged from your initial dataset. It can also handle cases where a single piece of content belongs in multiple categories—like our example above that touched on both battery and performance issues.</p>
<h2>Building a Synergy: Agentic Collaboration</h2>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="600" height="400" src="https://www.megaputer.com/wp-content/uploads/taxonomy-600x400.png" class="attachment-shop_single size-shop_single" alt="Three-level taxonomy fragment" srcset="https://www.megaputer.com/wp-content/uploads/taxonomy-600x400.png 600w, https://www.megaputer.com/wp-content/uploads/taxonomy-300x200.png 300w, https://www.megaputer.com/wp-content/uploads/taxonomy-1024x683.png 1024w, https://www.megaputer.com/wp-content/uploads/taxonomy-768x512.png 768w, https://www.megaputer.com/wp-content/uploads/taxonomy.png 1536w" sizes="(max-width: 600px) 100vw, 600px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;"><em>A snapshot of a three-level taxonomy</em></p>
<p>&nbsp;</p>
<p>The Category Builder and the Document Tagger work together in a closed loop to build multi-level taxonomies that capture both broad themes and specific details.</p>
<p>Let’s continue with our mobile app example. During the initial three-stage process, the Category Builder might create a top-level category like “Performance Issues.” But as the Document Tagger processes more content, it notices that “Performance Issues” is getting tagged very frequently and contains distinct sub-themes.</p>
<p>When the Category Builder runs another round of analysis, it discovers that Performance Issues actually break down into: “Battery Drain,” “Slow Loading,” and “App Crashes.” The Category Builder creates this second level of hierarchy, with “Performance Issues” as the parent category and these three as subcategories.</p>
<p>Now the Document Tagger can work with this richer structure. When new feedback comes in saying “The app froze three times today and completely drained my battery,” the agent can tag it at multiple levels: the broad “Performance Issues” category, and the specific “App Crashes” and “Battery Drain” subcategories.</p>
<p>This collaborative process can continue recursively. If “App Crashes” becomes heavily populated, the agents might discover it breaks down further into “Crashes on Startup,” “Crashes During Use,” and “Crashes When Closing.” The taxonomy grows more sophisticated as it encounters more data, always staying grounded in what’s actually present in your content.</p>
<h2>Use Cases: Beyond Product Launch Monitoring</h2>
<p>While product launch feedback is a compelling example, Just-in-Time taxonomies (for short, JIT taxonomies) solve organizational challenges across many different types of unstructured text data.</p>
<p><strong>Survey Analysis</strong> represents one of the most powerful applications. Open-ended survey responses often contain the richest insights, but they&#8217;re also the hardest to analyze at scale. Traditional approaches require manually reading through hundreds or thousands of responses to identify themes, then coding each response by hand. With Just-in-Time taxonomies, the agents automatically discover the themes present in your responses and categorize them accordingly. More importantly, you get to see the entire response distribution—not just summary statistics, but how responses cluster around different themes and what the full spectrum of sentiment looks like within each category.</p>
<p><strong>Social Listening at Scale</strong> becomes feasible in ways that weren&#8217;t possible before. Because the taxonomies are auto-generated, you can monitor thousands of influencers, social media accounts, and online conversations simultaneously. Traditional social listening requires pre-defining what topics and keywords to track, which means you miss unexpected conversations and emerging themes. With Just-in-Time taxonomies, you can cast a wide net across your entire industry ecosystem and let the system discover what people are actually talking about, including topics you hadn&#8217;t thought to monitor.</p>
<p><strong>Voice of Customer (VOC) programs</strong> generate massive amounts of unstructured feedback from multiple touchpoints: social media mentions, review sites, email feedback, focus groups, and customer interviews. Each of these may get their own taxonomy, or alternatively, a single monolithic taxonomy may be built that unifies all of these different data sources under a single organizational structure.</p>
<p><strong>Customer support conversation analysis</strong> is another natural fit. Support tickets, chat logs, and call transcripts contain valuable information about recurring issues, but this knowledge often stays trapped in individual conversations. By automatically categorizing support interactions, teams can quickly identify systemic problems, track resolution patterns, and understand which issues are driving the most contact volume.</p>
<p>The common thread across all these applications is the same fundamental challenge: you have large volumes of text that need to be tagged, structured, and made searchable so that patterns become visible and actionable insights can emerge. Whether it&#8217;s understanding why customers are calling support, discovering unexpected themes in survey data, or tracking sentiment across multiple feedback channels, Just-in-Time taxonomies transforms unstructured text into organized, analyzable data.</p>
<h2>Enriching the Taxonomy: Finding the KPI drivers</h2>
<p>Creating the taxonomy can be just the first step in a larger business solution. The next step is enriching the taxonomy by calculating &#8220;importance scores&#8221; that reveal which categories actually matter for your business outcomes. We do this by using the taxonomic categories as predictors of a key performance metric—like customer satisfaction scores in survey data, or support quality ratings in customer service conversations. The strongest predictors are the most important drivers of your KPI. If these performance indicators are not initially present in the dataset, we may derive them using AI.</p>
<p>When the categories are sorted by importance scores, the most important categories are visually obvious. It is often a good strategy to go down the list in order, addressing the issues identified by the AI in the order that they occur. This simple strategy can have profound effects on KPIs.</p>
<p>In one case, a software company wanted to improve its customer support, but had thousands of support conversations with no quality ratings. We used AI to evaluate each conversation and generate a &#8220;support quality&#8221; score.</p>
<p>Once we had the AI-derived quality score, we used the taxonomic categories as predictors to see which conversation topics were associated with poor support outcomes. The results were surprising. The company expected complex technical issues to drive poor support quality, but &#8220;password resets&#8221; appeared as the strongest predictor of low-quality interactions.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-image align_center"><div class="w-image-h"><img width="600" height="324" src="https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-600x324.jpg" class="attachment-shop_single size-shop_single" alt="Key KPI Drivers" srcset="https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-600x324.jpg 600w, https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-300x162.jpg 300w, https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-1024x553.jpg 1024w, https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-768x415.jpg 768w, https://www.megaputer.com/wp-content/uploads/key-kpi-drivers-741x400.jpg 741w, https://www.megaputer.com/wp-content/uploads/key-kpi-drivers.jpg 1278w" sizes="(max-width: 600px) 100vw, 600px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p style="text-align: center;"><em>Top customer satisfaction score (CSAT) drivers ranked by importance</em></p>
<p>&nbsp;</p>
<p>Drilling down revealed that password reset conversations consistently involved long hold times, multiple transfers between agents, and frustrated customers. The company discovered that its password reset system was broken in a way that made simple requests unnecessarily complicated, but because these seemed like &#8220;easy&#8221; tickets, management had never paid attention to them.</p>
<p>They fixed the underlying system issues and streamlined the password reset process. As a result of fixing this <em>single issue</em>, overall support quality scores improved by 20%, and the volume of password-related tickets dropped by 60%. After all issues had been addressed, support quality improved by 40%.</p>
<p>Without the AI-derived quality metric and taxonomic analysis, they would have continued focusing on complex technical issues while a simple broken system was quietly frustrating thousands of customers with what should have been routine requests.</p>
<h2>When do you need a JIT taxonomy?</h2>
<p>Just-in-Time (JIT) taxonomies fill a gap that human taxonomists cannot fill: If you need data-driven insights faster than a human taxonomist can provide them, a JIT taxonomy is the only option. That makes them perfect for time-sensitive analysis, where waiting weeks for manual categorization means missing the opportunity entirely. But the applications extend far beyond just speed.</p>
<p>JIT taxonomies really shine when you&#8217;re exploring uncharted territory—entering new markets, launching innovative products, or researching unfamiliar customer segments where you genuinely don&#8217;t know what themes will emerge. Traditional taxonomies require you to define categories upfront, but sometimes the most valuable insights come from patterns you never would have anticipated. This makes them invaluable for competitive intelligence, emerging trend detection, or any situation where you&#8217;re trying to understand &#8220;what&#8217;s really going on&#8221; rather than confirming what you already suspect.</p>
<p>They&#8217;re also ideal for evolving content streams where the conversation itself is constantly changing. Social media discussions, ongoing customer feedback, live chat transcripts, and user-generated content don&#8217;t stay static—new themes emerge, language evolves, and yesterday&#8217;s taxonomy might miss today&#8217;s most important topics. A JIT taxonomy adapts as these changes happen, making it perfect for brand monitoring, community management, and any application where staying current matters more than maintaining rigid consistency.</p>
<p>The approach works particularly well with moderate-scale datasets—typically thousands to tens of thousands of documents. This covers most business scenarios perfectly: quarterly feedback analysis, event-driven research, pilot program evaluation, or departmental data organization. It&#8217;s large enough that manual analysis becomes impractical, but focused enough that AI can identify genuinely meaningful patterns rather than getting lost in noise.</p>
<p>JIT taxonomies are also excellent for proof-of-concept work. Before committing to a major taxonomy project, they can quickly demonstrate what&#8217;s possible with your specific data. This makes them invaluable for pilot programs, budget justification, or simply exploring whether text analysis will provide value for your particular use case. And when you&#8217;re resource-constrained—whether that means lacking dedicated research staff, facing tight deadlines, or working within a limited budget—they can deliver insights that would otherwise require weeks of manual work or expensive consulting engagements.</p>
<p>JIT taxonomies are less suitable for massive enterprise taxonomies requiring thousands of predefined categories, situations demanding strict compliance with industry standards or regulatory frameworks, or long-term strategic classification systems where consistency across years matters more than speed. If you&#8217;re integrating with established enterprise systems or building something that needs to conform to specific legal or industry requirements, traditional approaches may be more appropriate.</p>
<p>But rather than viewing this as an either-or choice, many organizations find the most value in hybrid approaches where JIT taxonomies extend and enhance traditional human-built structures. Consider a large enterprise that has spent years developing a comprehensive taxonomy for customer feedback—they have well-established top-level categories like &#8220;Product Quality,&#8221; &#8220;Service Experience,&#8221; and &#8220;Billing Issues.&#8221; These broad categories work well and align with their reporting structures and business processes.</p>
<p>However, at the detailed level, their traditional taxonomy might categorize everything under &#8220;Product Quality&#8221; without capturing the specific nuances that emerge in real customer feedback. This is where JIT taxonomies excel. By running the system specifically on feedback already categorized as &#8220;Product Quality,&#8221; it can automatically discover granular subcategories that reflect what customers are actually saying: &#8220;Packaging Damage,&#8221; &#8220;Assembly Instructions,&#8221; &#8220;Color Accuracy,&#8221; &#8220;Durability After 6 Months,&#8221; and dozens of other specific issues that human taxonomists might not have anticipated or prioritized.</p>
<p>These AI-generated extensions can serve end users directly, giving them immediate access to detailed categorization while the broader enterprise taxonomy remains stable and compliant. Alternatively, the JIT results can guide human taxonomists, showing them exactly where their existing categories are insufficient and what new subcategories would capture the most meaningful distinctions in the data. Instead of guessing which detailed categories to add, they can see empirically what themes are emerging and how frequently they appear.</p>
<p>This approach gives organizations the best of both worlds: the governance and consistency of traditional taxonomies at the strategic level, combined with the responsiveness and data-driven precision of JIT taxonomies where detailed categorization matters most.</p>
<h2>Conclusion: The Future of Data Organization</h2>
<p>Just-in-Time taxonomies represent a fundamental shift in how we approach unstructured data. Instead of forcing your data into predetermined boxes or waiting weeks for manual analysis, you can let the data reveal its own structure and start extracting insights immediately.</p>
<p>The technology works because it mirrors how human experts actually think about categorization—identifying patterns, grouping similar concepts, and refining understanding as more information becomes available. But it does this at machine speed and scale, processing thousands of documents in the time it would take a human to read dozens.</p>
<p>Whether you&#8217;re launching a product, analyzing survey responses, monitoring social conversations, or trying to understand customer support patterns, the fundamental challenge remains the same: transforming unstructured text into actionable intelligence. Just-in-Time taxonomies make this transformation immediate rather than eventual, comprehensive rather than sampled, and adaptive rather than static.</p>
<p>The question isn&#8217;t whether AI will change how we organize and analyze text data—it already has. The question is whether you&#8217;ll take advantage of these capabilities to understand your data while the insights still matter, or continue waiting for perfect taxonomies that arrive too late to drive decisions.</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><em>Would you like to see how JIT taxonomy works? Interested in trying it on a sample of your data?</em> <a href="https://www.megaputer.com/contact/">Let&#8217;s talk!</a></p>

		</div>
	</div>
</div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/what-is-just-in-time-taxonomy/">What is Just-in-Time Taxonomy?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Data Curation Made Easy: Hybrid AI Approach</title>
		<link>https://www.megaputer.com/data-curation-made-easy-hybrid-ai-approach/</link>
		<pubDate>Tue, 03 Jun 2025 20:19:47 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35317</guid>
		<description><![CDATA[<p>Across industries, businesses constantly struggle to extract valuable insights from complex documents such as financial reports or research papers. These documents often contain a mix of dense text, tables packed with data, and intricate charts. Manually pulling insights from such material is not only tedious but also slows down decision-making, introduces risk, and can limit...</p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/data-curation-made-easy-hybrid-ai-approach/">Data Curation Made Easy: Hybrid AI Approach</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p>Across industries, businesses constantly struggle to extract valuable insights from complex documents such as financial reports or research papers. These documents often contain a mix of dense text, tables packed with data, and intricate charts. Manually pulling insights from such material is not only tedious but also slows down decision-making, introduces risk, and can limit a company&#8217;s ability to stay ahead in fast-paced markets.</p>
<p>Now, imagine a solution that automates this process. Unstructured and semi-structured documents can be instantly transformed into actionable data, and your team can easily change the focus of analysis. This is the power of next-generation data curation, which is fundamentally changing how businesses approach information analysis and reporting.</p>
<p>Consider the pharmaceutical industry as an example. Picture a data analyst at a leading pharma company responsible for reviewing each month dozens of research papers describing results of clinical trials. These reports are filled with highly technical text, detailed tables, and a variety of charts. In the past, the analyst might have spent weeks extracting, organizing, and interpreting this information before presenting it to decision-makers. This manual approach is slow and increases the risk of overlooking critical information.</p>
<p>We developed a Hybrid AI solution that addresses these challenges. By integrating the robust analytical platform PolyAnalyst with advanced Generative AI models, such as ChatGPT and Claude, we have created a solution that automates and accelerates the entire data curation process. Organizations in finance, pharma, life sciences, academia, and other sectors can now analyze complex documents with ease. Vast volumes of unstructured information get converted into actionable intelligence in a fraction of the usual time, requiring little human intervention.</p>
<p><img class="alignnone wp-image-35318 aligncenter" src="https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-240x300.png" alt="" width="429" height="536" srcset="https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-240x300.png 240w, https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-820x1024.png 820w, https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-768x959.png 768w, https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-320x400.png 320w, https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics-600x749.png 600w, https://www.megaputer.com/wp-content/uploads/fig1-data-curation-infographics.png 1230w" sizes="(max-width: 429px) 100vw, 429px" /></p>
<p style="text-align: center;">Figure 1. AI Data Curation solution overview</p>
<h2>How It Works</h2>
<p>AI Data Curation solution uses a streamlined workflow organized into three main stages:</p>
<h4><strong>Stage 1: Converting documents to machine readable form</strong></h4>
<ol>
<li><strong>Object separation:<br />
</strong>When the user uploads documents to the platform, PolyAnalyst initiates the analysis workflow. Using a Vision-Language Model (VLM), the system detects and separates plain text, tables, and charts on each page.<img class=" wp-image-35319 aligncenter" src="https://www.megaputer.com/wp-content/uploads/fig2-object-detection-300x244.png" alt="" width="430" height="350" srcset="https://www.megaputer.com/wp-content/uploads/fig2-object-detection-300x244.png 300w, https://www.megaputer.com/wp-content/uploads/fig2-object-detection.png 324w" sizes="(max-width: 430px) 100vw, 430px" /></p>
<p style="text-align: center;"><span style="font-size: 16px;">Figure 2. Detecting objects on a page</span></p>
<p>&nbsp;</li>
<li><strong>OCR:<br />
</strong>Neural network-based OCR engine converts textual and tabular content into machine-readable form.</li>
<li><strong>Figure Interpretation:<br />
</strong>A VLM interprets graphical figures, such as line charts or bar charts and converts them into structured data tables, retaining their positions in the original document.</li>
</ol>
<p><img class="alignnone wp-image-35320 aligncenter" src="https://www.megaputer.com/wp-content/uploads/fig3-chart-to-table-300x113.png" alt="" width="640" height="241" srcset="https://www.megaputer.com/wp-content/uploads/fig3-chart-to-table-300x113.png 300w, https://www.megaputer.com/wp-content/uploads/fig3-chart-to-table.png 553w" sizes="(max-width: 640px) 100vw, 640px" /></p>
<p style="text-align: center;">Figure 3. Converting chart to table</p>
<h4><strong>Stage 2: AI-driven Analysis</strong></h4>
<ol start="4">
<li><strong>Stating Facts of Interest:</strong><br />
The user describes target facts of interest through a simple interface.</li>
<li><strong>Extracting facts of interest:</strong><br />
An LLM finds and extracts facts specified by the user.</li>
</ol>
<h4><strong>Stage 3: Results Integration and Presentation</strong></h4>
<ol start="6">
<li><strong>Data Integration and Post-processing:<br />
</strong>Facts extracted by an LLM are cleansed and organized into unified datasets using sophisticated post-processing algorithms available in PolyAnalyst. This provides coherent insights for the entire document.</li>
<li><strong>Indexing and highlighting:<br />
</strong>The solution automatically creates linguistic rules that help navigate the original documents and highlight patterns supporting the performed fact extraction.<br />
<img class=" wp-image-35321 aligncenter" src="https://www.megaputer.com/wp-content/uploads/fig4-create-linguistic-rules-300x130.png" alt="" width="505" height="219" srcset="https://www.megaputer.com/wp-content/uploads/fig4-create-linguistic-rules-300x130.png 300w, https://www.megaputer.com/wp-content/uploads/fig4-create-linguistic-rules.png 493w" sizes="(max-width: 505px) 100vw, 505px" />                                                             Figure 4. Linguistic rules for pattern highlighting</li>
<li><strong>Interactive Web Reports:<br />
</strong>Users can access graphical reports, view the extraction results alongside the original documents with highlights, and validate results of extraction using an interactive interface. This enhances transparency and trust.</li>
</ol>
<p><img class="alignnone wp-image-35322 aligncenter" src="https://www.megaputer.com/wp-content/uploads/fig5-report-300x118.png" alt="" width="646" height="254" srcset="https://www.megaputer.com/wp-content/uploads/fig5-report-300x118.png 300w, https://www.megaputer.com/wp-content/uploads/fig5-report.png 519w" sizes="(max-width: 646px) 100vw, 646px" /></p>
<p style="text-align: center;">Figure 5. Interface for navigating and validating the results of AI-based extraction</p>
<h2>Key Advantages of AI Data Curation</h2>
<ul>
<li><strong>Efficiency:</strong> The automated extraction process dramatically reduces time and manual effort.</li>
<li><strong>Precision:</strong> Robust data transformation and analysis software works alongside state-of-the-art Generative AI to provide highly accurate extraction and analysis.</li>
<li><strong>Scalability:</strong> The system can process numerous complex documents at once, making it suitable for both small projects and enterprise-level data initiatives.</li>
<li><strong>Simple Manual Validation:</strong> An interactive interface enables easy validation, which minimizes errors and ensures high-quality results.</li>
<li><strong>Minimal Human Intervention: </strong>Our multi-model AI approach allows different AI engines to collaborate seamlessly and share data and insights automatically. The user only specifies a collection of facts of interest.</li>
<li><strong>Broad Applicability: </strong>The user can easily specify a different set of facts of interest to make the solution solve a different task. This makes the solution easily adaptable to a wide range of industries and document types. Moreover, in the “free hunting” mode (to be described in a separate blog article soon) the system requires little to no domain-specific input from users. Instead, it can use the contents of documents themselves to extract and organize facts of potential interest and then present them to the user.</li>
</ul>
<p>By blending analytical capabilities of PolyAnalyst with advanced Generative AI, our AI Data Curation solution automates and simplifies the process of extracting facts of interest from complex documents. Actionable insights are delivered faster and with higher accuracy.</p>
<p>To see AI Data Curation solution in action, <a href="https://www.megaputer.com/contact/">request a demo</a>.</p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/data-curation-made-easy-hybrid-ai-approach/">Data Curation Made Easy: Hybrid AI Approach</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Retrieval Augmented Generation (RAG) in simple business terms</title>
		<link>https://www.megaputer.com/rag-in-simple-business-terms/</link>
		<pubDate>Tue, 03 Jun 2025 15:00:34 +0000</pubDate>
		<dc:creator><![CDATA[Sergei Ananyan]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35314</guid>
		<description><![CDATA[<p>1. What Is RAG? Imagine pairing a super‑fast semantic search engine with a large language model (LLM). The search engine grabs the most relevant facts found in the loaded corpus of documents; the LLM turns those facts into clear answers. That combo is Retrieval‑Augmented Generation (RAG). 2. How Does It Work? Retrieve – The system...</p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/rag-in-simple-business-terms/">Retrieval Augmented Generation (RAG) in simple business terms</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p id="ember527" class="ember-view reader-text-block__paragraph"><strong>1. What Is RAG?</strong></p>
<p id="ember528" class="ember-view reader-text-block__paragraph">Imagine pairing a super‑fast semantic search engine with a large language model (LLM). The search engine grabs the most relevant facts found in the loaded corpus of documents; the LLM turns those facts into clear answers. That combo is <strong>Retrieval‑Augmented Generation (RAG)</strong>.</p>
<p id="ember529" class="ember-view reader-text-block__paragraph"><strong>2. How Does It Work?</strong></p>
<ol>
<li><strong>Retrieve</strong> – The system quickly searches your chosen knowledge sources (public sites, private docs, etc.) and pulls out the fragments most related to the question.</li>
<li><strong>Augment</strong> – The matching fragments are presented to the LLM in a prompt so the model “sees” them.</li>
<li><strong>Generate</strong> – The LLM writes an answer grounded strictly in these fragments and provides reference to the right places in the original documents.</li>
</ol>
<p id="ember531" class="ember-view reader-text-block__paragraph">Think of it as having a clever librarian (retriever) who hands the expert writer (LLM) exactly the pages they need to produce an informative answer to the user question.</p>
<p id="ember532" class="ember-view reader-text-block__paragraph"><strong>3. Why Your Business Cares</strong></p>
<ul>
<li><strong>Up‑to‑Date Answers</strong> – Since it reads live data, RAG isn’t stuck with an old knowledge cutoff.</li>
<li><strong>Uses Your Proprietary Info</strong> – It can securely tap your internal wikis, policies, or product docs.</li>
<li><strong>Fewer Hallucinations</strong> – Grounding the model in real text sharply reduces the number of made‑up answers.</li>
<li><strong>Transparent</strong> – The system can show where each fact came from.</li>
</ul>
<p id="ember534" class="ember-view reader-text-block__paragraph"><strong>4. Things to Keep in Mind</strong></p>
<ul>
<li id="ember535" class="ember-view reader-text-block__paragraph"><strong>Retrieval Quality. </strong>Accurate answers depend on accurate source documents. Keep your content clean and current.</li>
<li id="ember536" class="ember-view reader-text-block__paragraph"><strong>Speed vs. Depth.</strong> More sources or deeper checks add a bit of response time. We tune this to fit your use case.</li>
<li id="ember537" class="ember-view reader-text-block__paragraph"><strong>Security.</strong> Access controls ensure only authorized users and docs are included.</li>
<li id="ember538" class="ember-view reader-text-block__paragraph"><strong>Bias &#038; Accuracy.</strong> The system reflects the documents it reads. We monitor and filter for fairness and correctness.</li>
</ul>
<p id="ember539" class="ember-view reader-text-block__paragraph"><strong>5. Key Takeaways</strong></p>
<ul>
<li>RAG combines search + generative AI so customers and staff get accurate, cited answers in seconds.</li>
<li>It lets you unlock the value of the information you already have without constant model retraining.</li>
<li>Done right, it is fast, secure, and dramatically reduces AI “hallucinations.”</li>
</ul>
<p id="ember541" class="ember-view reader-text-block__paragraph"><em>Want to see RAG in action on your data? </em><a class="FYFwKXYNhRNjjSvOyrNhgyvQhtApFmdaFNvUxIg " tabindex="0" href="https://www.megaputer.com/contact/" target="_self" data-test-app-aware-link=""><em>Contact our team for a demo</em></a><em>.</em></p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/rag-in-simple-business-terms/">Retrieval Augmented Generation (RAG) in simple business terms</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Query languages—the Swiss army knife of information extraction</title>
		<link>https://www.megaputer.com/query-languages-the-swiss-army-knife-of-information-extraction/</link>
		<pubDate>Tue, 06 Feb 2024 05:19:41 +0000</pubDate>
		<dc:creator><![CDATA[Echo Lu]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Big Data]]></category>
		<category><![CDATA[Business Intelligence]]></category>
		<category><![CDATA[Computational Linguistics]]></category>
		<category><![CDATA[Data Analytics]]></category>
		<category><![CDATA[Data Mining]]></category>
		<category><![CDATA[Entity Extraction]]></category>
		<category><![CDATA[Fuzzy Matching]]></category>
		<category><![CDATA[Machine Learning]]></category>
		<category><![CDATA[Morphology]]></category>
		<category><![CDATA[Pattern Definition Language]]></category>
		<category><![CDATA[Text Analytics]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35231</guid>
		<description><![CDATA[<p>Text mining, the art of extracting information from text, requires the formulation of efficient queries that retrieve information based on user input. To do this, the user requires a language for writing queries. For the most basic use cases, the language operators could be regex or string search. But while regex and string search are...</p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/query-languages-the-swiss-army-knife-of-information-extraction/">Query languages—the Swiss army knife of information extraction</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<p><span style="font-weight: 400;">Text mining, the art of extracting information from text, requires the formulation of efficient queries that retrieve information based on user input. To do this, the user requires a language for writing queries. For the most basic use cases, the language operators could be regex or string search. But while regex and string search are indispensable for text mining, their utility hits a hard ceiling when semantic or meaningful search is required. They cannot, for example, capture complex entities such as human names, corporations, and drugs. For handling tasks like these, we need a more powerful query language that has semantic understanding, such as Megaputer’s PDL.</span></p>
<p><span style="font-weight: 400;">So, what is PDL, and how does it achieve semantic understanding while regex and string search do not? Let’s take a look at an example to find out.</span></p>
<p><span style="font-weight: 400;">First of all, PDL does not just search for the literal form of the word in the query: instead, it automatically extends its search to all morphological forms of the word. For example, when searching for the word “company” in financial news articles, the PDL query will not only find “company”, but also “companies,” the plural form. This feature often comes in handy, especially when the search involves a verb. Suppose that you are interested in extracting </span><i><span style="font-weight: 400;">what the CEOs said. </span></i><span style="font-weight: 400;">With regex or other substring search, you will need to list all possible verb forms such as “say,” “saying,” “says,” and “said.” With PDL, simply entering “say” in the query will automatically fetch all possible verb forms. This behavior can also be turned off by enclosing the word in the </span><i><span style="font-weight: 400;">form </span></i><span style="font-weight: 400;">function</span><span style="font-weight: 400;">,</span><span style="font-weight: 400;"> which will then restrict the search to the literal form of the word, such as in the example below.</span></p>
<p><img class="wp-image-35249 aligncenter" src="https://www.megaputer.com/wp-content/uploads/comparison_pdl-1.png" alt="" width="800" height="497" /><br />
<!-- <img class="wp-image-35232 aligncenter" src="https://www.megaputer.com/wp-content/uploads/pdl-image-1-300x186.jpg" alt="" width="710" height="440" srcset="https://www.megaputer.com/wp-content/uploads/pdl-image-1-300x186.jpg 300w, https://www.megaputer.com/wp-content/uploads/pdl-image-1-1024x636.jpg 1024w, https://www.megaputer.com/wp-content/uploads/pdl-image-1-768x477.jpg 768w, https://www.megaputer.com/wp-content/uploads/pdl-image-1-644x400.jpg 644w, https://www.megaputer.com/wp-content/uploads/pdl-image-1-600x373.jpg 600w" sizes="(max-width: 710px) 100vw, 710px" /> --></p>
<p><span style="font-weight: 400;">Another notable feature of the PDL language is its capability for users to tailor the scope of their searches using a range of built-in functions. Returning to the previous example, you may not wish to confine your search exclusively to the specific verb “say,” but rather include other synonymous verbs like “tell” or “mention.” Achieving this is straightforward with PDL – users can invoke the </span><i><span style="font-weight: 400;">synonym</span></i> <span style="font-weight: 400;">function with the verb &#8220;say,&#8221; as demonstrated in (a) below. As the subsequent results table (b) illustrates, the captured text now includes various speech verbs such as “tell,” “emphasize,” and “claim,” in addition to the word “say,” capturing them in all possible verb forms. For additional flexibility, the user can also create and modify synonym dictionaries.</span></p>
<p><img class="wp-image-35250 aligncenter" src="https://www.megaputer.com/wp-content/uploads/comparison_pdl-2.png" alt="" width="800" height="497" /><br />
<!-- <img class="wp-image-35235 aligncenter" src="https://www.megaputer.com/wp-content/uploads/pdl-image-2-300x269.jpg" alt="" width="737" height="661" srcset="https://www.megaputer.com/wp-content/uploads/pdl-image-2-300x269.jpg 300w, https://www.megaputer.com/wp-content/uploads/pdl-image-2-1024x919.jpg 1024w, https://www.megaputer.com/wp-content/uploads/pdl-image-2-768x689.jpg 768w, https://www.megaputer.com/wp-content/uploads/pdl-image-2-446x400.jpg 446w, https://www.megaputer.com/wp-content/uploads/pdl-image-2-600x538.jpg 600w" sizes="(max-width: 737px) 100vw, 737px" /> --></p>
<p><span style="font-weight: 400;">The PDL language offers various modes of information extraction, including proximity search (e.g., finding words A and B within a sentence, or within a 3-words range), syntactic relation (e.g., finding word A that is the subject or object of B), semantic relation (e.g., finding words that are synonyms/antonyms to word A), access to dictionaries and ontologies, and more. This language is expressive enough to capture complex patterns, and yet relatively easy to use, </span><span style="font-weight: 400;">having a syntax that closely resembles English. Having access to this versatile query language significantly enhances the power and quality of text mining operations.</span></p>
<p><span style="font-weight: 400;">In conclusion, PDL is a powerful and versatile query language that enables users to extract meaningful information from text with greater efficiency and accuracy than competing methods like regex or string search. Its ability to understand and capture morphological forms, synonyms, and other complex patterns makes it an indispensable tool for solving text mining tasks that require semantic understanding. By leveraging the capabilities of PDL, users can enhance their information extraction processes and gain valuable insights from their data, making it a true Swiss army knife of information extraction.</span></p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/query-languages-the-swiss-army-knife-of-information-extraction/">Query languages—the Swiss army knife of information extraction</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>What is Linguistic Corpus?</title>
		<link>https://www.megaputer.com/what-is-linguistic-corpus/</link>
		<pubDate>Tue, 05 Dec 2023 20:55:10 +0000</pubDate>
		<dc:creator><![CDATA[Echo Lu]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Computational Linguistics]]></category>
		<category><![CDATA[NLP]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=35025</guid>
		<description><![CDATA[<p>The post <a rel="nofollow" href="https://www.megaputer.com/what-is-linguistic-corpus/">What is Linguistic Corpus?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">At the core of Natural Language Processing lies extensive language data. The availability of copious amounts of text or speech, faithfully representing natural language, is pivotal for the efficacy of artificial language systems. Diverse forms of text, from Wikipedia articles to Twitter threads, find application in various language models. These substantial language datasets collectively constitute what is known as a corpus in linguistics—an extensive and principled collection of authentic, naturally occurring language.</span></p>

		</div>
	</div>
<div class="w-separator size_small"></div><div class="w-image align_center"><div class="w-image-h"><img width="564" height="374" src="https://www.megaputer.com/wp-content/uploads/photo_2023-11-20_12-03-27.jpg" class="attachment-full size-full" alt="Random collection of different words and word-forms on magnetic tiles" srcset="https://www.megaputer.com/wp-content/uploads/photo_2023-11-20_12-03-27.jpg 564w, https://www.megaputer.com/wp-content/uploads/photo_2023-11-20_12-03-27-300x199.jpg 300w" sizes="(max-width: 564px) 100vw, 564px" /></div></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">Despite gaining increased attention with the advent of LLMs, the concept of a corpus is not new in linguistics, dating back to the early 20th century or earlier. During that time, lexicographers manually compiled substantial language samples for dictionary construction, representing a rudimentary form of a corpus. Beyond lexicographers and grammarians, linguists manually compiled corpora for various purposes, such as language pedagogy or the study of language acquisition in children. Moreover, the so-called &#8220;structural&#8221; linguists, who dominated the field until the 1950s, emphasized the utility of corpora, believing in the analytical approach to the distributional characteristics of language. With this idea, linguist Charles C. Fries, for example, manually gathered telephone conversations from approximately 300 speakers, encompassing 250,000 words. Field linguist Franz Boas dedicated himself to establishing a substantial corpus of texts from Native Americans.</span></p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">In the 1960s, however, the first electronic corpus appeared. W. Nelson Francis and Henry Kučera, based at Brown University, crafted a corpus of 1 million American English words, famously known as the Brown corpus. Thereafter, linguistic corpora rapidly evolved and diversified in terms of size and scope, driven by advancements in computer hardware. For instance, </span><a href="https://www.english-corpora.org/coca/"><span style="font-weight: 400;">the Corpus of Contemporary American English (COCA)</span></a><span style="font-weight: 400;">, currently </span><span style="font-weight: 400;">one of the most widely used American English corpora, encompasses over one billion words sourced from eight different genres.</span></p>

		</div>
	</div>
<div class="w-separator size_medium"></div><div class="w-image align_center"><div class="w-image-h"><img width="600" height="451" src="https://www.megaputer.com/wp-content/uploads/istock-1439146187-600x451.jpg" class="attachment-shop_single size-shop_single" alt="Abstract source code background, Big data database app, Computer code, Lines of HTML code by programmer, active online data transformation and exchange" srcset="https://www.megaputer.com/wp-content/uploads/istock-1439146187-600x451.jpg 600w, https://www.megaputer.com/wp-content/uploads/istock-1439146187-300x225.jpg 300w, https://www.megaputer.com/wp-content/uploads/istock-1439146187-1024x769.jpg 1024w, https://www.megaputer.com/wp-content/uploads/istock-1439146187-768x577.jpg 768w, https://www.megaputer.com/wp-content/uploads/istock-1439146187-533x400.jpg 533w" sizes="(max-width: 600px) 100vw, 600px" /></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><b>Types of Corpora and their use</b></h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">Corpora can be broadly categorized into two types: generalized and specialized. A generalized corpus aims to represent all aspects of a language and typically consists of a large number of words and/or documents. Examples of generalized corpora include the </span><a href="http://www.natcorp.ox.ac.uk/"><span style="font-weight: 400;">British National Corpus</span></a><span style="font-weight: 400;"> and </span><a href="https://www.english-corpora.org/coca/"><span style="font-weight: 400;">COCA</span></a><span style="font-weight: 400;">, mentioned earlier. On the other hand, a specialized corpus is constructed to represent specific aspects of a language, such as language produced by particular user groups (e.g., children, teenagers, second language learners, aphasics) or in specific settings (e.g., university, business). The </span><a href="https://childes.talkbank.org/access/"><span style="font-weight: 400;">Child Language Data Exchange System (CHILDES)</span></a><span style="font-weight: 400;">, a large corpus of language produced by developing children, and the </span><a href="https://scnweb.japanknowledge.com/register/PERC/index.html"><span style="font-weight: 400;">Professional English Research Consortium (PERC)</span></a><span style="font-weight: 400;"> corpus, which contains English academic journal texts in science, engineering, technology, and other fields, are examples of specialized corpora.</span></p>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><img style="float: right;" src="https://www.megaputer.com/wp-content/uploads/frame-2.png" alt="Parts of Speech" width="500" /><span style="font-weight: 400;">Corpora are frequently enriched with annotations containing linguistic information. Typically, texts within corpora undergo Part of Speech (POS) tagging, wherein the grammatical category of each word is identified. Beyond POS tagging, corpora can be further enhanced with annotations covering syntactic structure, semantic roles (denoting the roles played by each noun argument in a sentence, such as agent, patient, instrument, etc.), Named Entity Recognition (NER), sentiment analysis, and more.</span>To illustrate, the <a href="https://catalog.ldc.upenn.edu/LDC99T42"><span style="font-weight: 400;">Penn Treebank</span></a><span style="font-weight: 400;"> project annotates a collection of Wall Street Journal articles with both syntactic dependency information and POS tagging. Another type of syntactic corpus might be the </span><a href="http://commondatastorage.googleapis.com/books/syntactic-ngrams/index.html"><span style="font-weight: 400;">Google Books Syntactic N-grams dataset</span></a><span style="font-weight: 400;">, comprising syntactic tree fragments automatically extracted from the Google English Books collection. Some corpora are annotated with semantic information. The </span><a href="https://propbank.github.io/"><span style="font-weight: 400;">Proposition Bank</span></a><span style="font-weight: 400;">, for example, provides a collection of English sentences that are manually annotated with various semantic roles. Some corpora also feature Named Entity Recognition. For example, </span><a href="https://bmcbioinformatics.biomedcentral.com/articles/10.1186/1471-2105-6-S1-S3"><span style="font-weight: 400;">GENETAG</span></a><span style="font-weight: 400;"> is a corpus of 20,000 sentences from the MEDLINE database annotated with gene/protein NER.</span></p>

		</div>
	</div>
</div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2><b>Use of Corpora in NLP</b></h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">Corpora have traditionally served as tools to investigate specific aspects of language use, including collocation, frequency, preferred constructions, patterns of errors, and more. With the emergence of computer technology and machine learning algorithms, they have evolved into essential resources for model training and validation. Large Language Models are now trained on vast amounts of language data. For instance, OpenAI&#8217;s GPT-3 175B model is trained with 300 billion tokens gathered from sources such as Common Crawl, WebText2, and Wikipedia. Meta&#8217;s LLaMa 65B model, on the other hand, is trained with 1.4 trillion tokens sourced from various platforms, including English CommonCrawl, C4, Github, Wikipedia, ArXiv, and Stack Exchange. In our PolyAnalyst, diverse corpora like </span><a href="https://anc.org/"><span style="font-weight: 400;">Open American National Corpus</span></a><span style="font-weight: 400;">, the </span><a href="https://catalog.ldc.upenn.edu/LDC99T42"><span style="font-weight: 400;">Penn Treebank</span></a><span style="font-weight: 400;">, CoNLL Corpus, </span><a href="https://opus.nlpl.eu/Europarl-v3.php"><span style="font-weight: 400;">Europarl</span></a><span style="font-weight: 400;">, the </span><a href="https://ruscorpora.ru/en/"><span style="font-weight: 400;">Russian National Corpus</span></a><span style="font-weight: 400;">, and Wikipedia are employed for various modules. The utilization of large language datasets is poised to expand further in the future, making the adept handling of such data a crucial skill in the field of Natural Language Processing.</span></p>

		</div>
	</div>
</div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/what-is-linguistic-corpus/">What is Linguistic Corpus?</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Medical Ontology and Its Use in Text Analysis</title>
		<link>https://www.megaputer.com/medical-ontology-and-its-use-in-text-analysis/</link>
		<pubDate>Tue, 24 Oct 2023 14:00:32 +0000</pubDate>
		<dc:creator><![CDATA[Echo Lu]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Text Analytics]]></category>
		<category><![CDATA[PolyAnalyst]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=34772</guid>
		<description><![CDATA[<p>Medical ontologies are a treasure trove of information. They help distill complex knowledge domains into a hierarchical structure of facts.  In this article, we will walk through what exactly an ontology is and how we can use ontologies for analysis tasks in the medical domain.</p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/medical-ontology-and-its-use-in-text-analysis/">Medical Ontology and Its Use in Text Analysis</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="ult-spacer spacer-69d434ecce074" data-id="69d434ecce074" data-height="10" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div><div class="w-image align_none"><div class="w-image-h"><img width="300" height="300" src="https://www.megaputer.com/wp-content/uploads/healthcare-coding2.png" class="attachment-shop_catalog size-shop_catalog" alt="A picture of a filing cabinet where each side of the cabinet has drawers, some partially open" srcset="https://www.megaputer.com/wp-content/uploads/healthcare-coding2.png 500w, https://www.megaputer.com/wp-content/uploads/healthcare-coding2-300x300.png 300w, https://www.megaputer.com/wp-content/uploads/healthcare-coding2-150x150.png 150w, https://www.megaputer.com/wp-content/uploads/healthcare-coding2-350x350.png 350w" sizes="(max-width: 300px) 100vw, 300px" /></div></div></div></div></div><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="ult-spacer spacer-69d434eccf092" data-id="69d434eccf092" data-height="20" data-height-mobile="30" data-height-tab="20" data-height-tab-portrait="20" data-height-mobile-landscape="30" style="clear:both;display:block;"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">The complexity and variability of biomedical language pose a challenge in fields dealing with biomedical data. A vast array of terms, referring to various entities, including living organisms, biochemicals, drugs, biological pathways, diseases, and disease symptoms in patients, is utilized in biomedical research and closely associated healthcare industries. The intricacy inherent in this domain necessitates both a structured representation of knowledge and the use of a standardized language to describe these entities. Medical ontology serves as a valuable tool for fulfilling this crucial purpose.</span></p>

		</div>
	</div>
</div></div></div></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2>What is Ontology?</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">Originally a philosophical study of existence, ontology, in the context of information science, refers to a formal representation of knowledge or information within a specific domain. Ontology provides a structured framework for organizing entities, their properties, and relationships. To illustrate, let&#8217;s consider the ontology of a country. As shown below, it encompasses entities such as states, cities, towns, and villages, along with their hierarchy and relationships. Interestingly, a similar ontology focused on geographical information serves as the foundation for the widely used Geographic Information System (GIS).</span></p>

		</div>
	</div>
<div class="w-image align_center"><div class="w-image-h"><img width="600" height="338" src="https://www.megaputer.com/wp-content/uploads/fig01_slide1-600x338.jpg" class="attachment-shop_single size-shop_single" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig01_slide1-600x338.jpg 600w, https://www.megaputer.com/wp-content/uploads/fig01_slide1-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig01_slide1-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig01_slide1-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig01_slide1.jpg 1280w" sizes="(max-width: 600px) 100vw, 600px" /></div></div><div class="ult-spacer spacer-69d434ecd0d23" data-id="69d434ecd0d23" data-height="85" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">By extension, medical data is also highly complex—perhaps even more so—due to the vast diversity of biological entities and their interdependent relationships and interactions. This inherent complexity has prompted the development of numerous types of ontologies, which are designed to maintain consistency and accuracy in medical data, records, and research. At present, there are hundreds of medical ontologies, and you can find many of them by following </span><a href="http://bioportal.bioontology.org/" target="_blank" rel="noopener"><span style="font-weight: 400;">this</span></a><span style="font-weight: 400;"> link. </span></p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">There are numerous advantages to utilizing ontologies. For biomedical researchers, the use of medical ontologies can assist them in annotating and categorizing data with standardized terminology. This, in turn, leads to clear communication and makes it easier to identify relationships, patterns, and other elements within the data. For medical practitioners, ontologies facilitate the exchange of patient information among healthcare providers and entities, ensuring that data sent and received by different systems are interpreted consistently. This not only enhances the accuracy and consistency of medical information but also promotes interoperability between different healthcare systems. Furthermore, it supports the development of intelligent systems for medical diagnosis and treatment.</span></p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><span style="font-weight: 400;">Apart from the benefits in research and practice, medical ontology also plays a crucial role in text mining. There is a constant influx of biomedical text from various sources every day. However, due to the intricacies of biomedical concepts and the inherent ambiguity in natural language, precise data extraction from large and heterogeneous data sources is not feasible without a system of standardization. Medical ontology helps mitigate this ambiguity and variability in text data. Furthermore, medical ontology is also pivotal in the Semantic Web, where web data are represented through structured ontological representations, allowing interpretation by both humans and machines. In short, the use of medical ontology leads to more intelligent search, enhanced data integration, and improved knowledge discovery across a wide range of medical and healthcare-related resources.</span></p>

		</div>
	</div>
<div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2>An overview of commonly used medical ontologies</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>There are a number of ontologies that are particularly helpful when building text analysis solutions for the medical field. In what follows, we provide an overview of some of the more commonly used ontologies.</p>

		</div>
	</div>
</div></div></div></div><div class="w-separator size_small"></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3>MeSH</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The <a href="https://www.nlm.nih.gov/mesh/meshhome.html" target="_blank" rel="noopener">Medical Subject Headings</a> (MeSH) dictionary is a hierarchical collection of medical terms describing various conditions, diseases, and symptoms. It was created by the US National Library of Medicine (NLM) at the National Institutes of Health (NIH) in the United States.</p>
<p>MeSH is utilized for indexing and searching biomedical information found in national medical databases, including MEDLINE/PubMed. Many biomedical articles are now indexed with MeSH terms, either manually or automatically by computer, enabling users to retrieve records on a specific subject with accuracy and efficiency.</p>

		</div>
	</div>
</div></div></div><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_huge"></div><div class="w-image align_center"><div class="w-image-h"><img width="1280" height="720" src="https://www.megaputer.com/wp-content/uploads/fig03_slide3.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig03_slide3.jpg 1280w, https://www.megaputer.com/wp-content/uploads/fig03_slide3-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig03_slide3-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig03_slide3-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig03_slide3-600x338.jpg 600w" sizes="(max-width: 1280px) 100vw, 1280px" /></div></div></div></div></div></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The MeSH tree structure comprises 16 main branches, such as Anatomy, Organisms, Diseases, and more. Each branch is further divided into hierarchical levels, ranging from more general to more specific categories. These levels represent the taxonomic relationships between terms, as follows:</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<ul>
<li aria-level="1"><b>synonyms</b>: terms with the same, or nearly the same, meaning. (e.g., “dwelling” and “residence” are synonyms of “home”)</li>
</ul>
<ul>
<li aria-level="1"><b>hypernyms</b>: terms with a broad meaning for a group of specific items. (e.g., “animal” is a hypernym for “dog” and “cat”)</li>
</ul>
<ul>
<li aria-level="1"><b>hyponyms</b>: terms with a more specific meaning than a general term. (e.g., “apple”, “banana”, and “orange” are hyponyms of “fruit”)</li>
</ul>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>For example, Angina Pectoris, a medical term for chest pain due to coronary heart disease, would be mapped out as shown below:</p>

		</div>
	</div>
<div class="w-image align_center"><div class="w-image-h"><img width="1920" height="1080" src="https://www.megaputer.com/wp-content/uploads/mesh_ontology_1.gif" class="attachment-full size-full" alt="" /></div></div><div class="w-separator size_large"></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3>ICD-10</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The <a href="https://icd10cmtool.cdc.gov/" target="_blank" rel="noopener">International Statistical Classification of Diseases and Related Health Problems</a> (ICD) is a standardized system for coding medical diagnoses and procedures. It contains a description of all known diseases and injuries across languages and allows quick access to disease information, including symptoms, treatment, and mortality rates.</p>

		</div>
	</div>
</div></div></div><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_small"></div><div class="w-image align_none"><div class="w-image-h"><img width="1280" height="720" src="https://www.megaputer.com/wp-content/uploads/fig05_slide6.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig05_slide6.jpg 1280w, https://www.megaputer.com/wp-content/uploads/fig05_slide6-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig05_slide6-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig05_slide6-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig05_slide6-600x338.jpg 600w" sizes="(max-width: 1280px) 100vw, 1280px" /></div></div></div></div></div></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The ICD-10 system has been adopted by numerous countries worldwide and finds use in a wide array of healthcare settings, including hospitals, clinics, and various healthcare facilities. Fully implemented in 2015, ICD-10 boasts an impressive array of over 150,000 codes covering a broad spectrum of medical conditions and treatments. The ICD ontology continues to evolve, and while the next version, ICD-11, is already available for implementation, it is not expected to be fully launched until 2025 or later.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Codes in ICD-10 can have 3 to 7 characters, where the diagnosis category is the first 3 characters. The next set of characters in the code constitutes the related etiology, anatomic site, severity, or other details. The seventh character (if used) specifies any extension classifier describing the episode of care. For example, if a medical record stated that “the patient developed a chalazion of the left upper eyelid,” this would be coded as H00.14 in the ICD-10 coding system.</p>

		</div>
	</div>
<div class="w-image align_center"><div class="w-image-h"><img width="600" height="338" src="https://www.megaputer.com/wp-content/uploads/fig06_slide7-600x338.jpg" class="attachment-shop_single size-shop_single" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig06_slide7-600x338.jpg 600w, https://www.megaputer.com/wp-content/uploads/fig06_slide7-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig06_slide7-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig06_slide7-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig06_slide7.jpg 1280w" sizes="(max-width: 600px) 100vw, 600px" /></div></div><div class="w-separator size_medium"></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3>MedDRA</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The <a href="https://www.meddra.org/" target="_blank" rel="noopener">Medical Dictionary for Regulatory Activities</a> (MedDRA) is an ontology for classifying and sharing regulatory information. This medical ontology is typically used by pharmaceutical and biotechnology companies for FDA pharmacovigilance reporting on drugs and therapies. MedDRA is used for registration, documentation, and safety monitoring of medical products before and after a product has been authorized for sale. This includes pharmaceuticals vaccines and drug-device combination products.</p>

		</div>
	</div>
</div></div></div><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_small"></div><div class="w-image align_none"><div class="w-image-h"><img width="1280" height="720" src="https://www.megaputer.com/wp-content/uploads/fig07_slide4.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig07_slide4.jpg 1280w, https://www.megaputer.com/wp-content/uploads/fig07_slide4-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig07_slide4-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig07_slide4-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig07_slide4-600x338.jpg 600w" sizes="(max-width: 1280px) 100vw, 1280px" /></div></div></div></div></div></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>MedDRA contains over 70,000 standard terms for various signs, symptoms, diseases, diagnoses, and indications, and is updated twice a year. It consists of a 5-level taxonomy, arranging symptoms and diseases from very specific to very general. For example, consider the condition known as “chest pain” (stenocardia). The 5-level hierarchical taxonomy describing this is summarized by the following animation.</p>

		</div>
	</div>
<div class="w-image align_center"><div class="w-image-h"><img width="1920" height="1080" src="https://www.megaputer.com/wp-content/uploads/medontology_meddra_1.gif" class="attachment-full size-full" alt="" /></div></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>MedDRA is dictated by the FDA in the United States. The FDA requires the use of MedDRA in the submission of safety reports and other regulatory documents related to drug development and approval. It&#8217;s widely recognized as a key standard for the classification and analysis of medical data in the pharmaceutical industry.</p>

		</div>
	</div>
</div></div></div></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-8 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3>RxNorm</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p><a href="https://www.nlm.nih.gov/research/umls/rxnorm/index.html">RxNorm</a> is a comprehensive database containing over 100,000 medications and chemicals. It serves as a standardized nomenclature system that establishes connections between brand names and generic synonyms for clinical drugs. This mapping enables medical care providers to effectively manage the diverse vocabulary associated with drug and pharmaceutical products across various databases.</p>

		</div>
	</div>
</div></div></div><div class="vc_col-sm-4 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_small"></div><div class="w-image align_none"><div class="w-image-h"><img width="1280" height="720" src="https://www.megaputer.com/wp-content/uploads/fig09_slide5.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig09_slide5.jpg 1280w, https://www.megaputer.com/wp-content/uploads/fig09_slide5-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig09_slide5-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig09_slide5-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig09_slide5-600x338.jpg 600w" sizes="(max-width: 1280px) 100vw, 1280px" /></div></div></div></div></div></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Moreover, RxNorm&#8217;s primary aim is to facilitate efficient and accurate communication of drug-related information among computer systems. It is diligently maintained and updated on a weekly basis by the National Library of Medicine (NLM) and encompasses all medications for humans available on the US market.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>RxNorm serves as an exemplary instance of a highly intricate ontology, as demonstrated in the example below that showcases the connections for Zyrtec® (cetirizine hydrochloride), an over-the-counter antihistamine. This not only provides the standard drug names but also highlights the relationships between different drug entities.</p>

		</div>
	</div>
<div class="w-image align_center"><div class="w-image-h"><img width="1280" height="720" src="https://www.megaputer.com/wp-content/uploads/fig10_slide10.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig10_slide10.jpg 1280w, https://www.megaputer.com/wp-content/uploads/fig10_slide10-300x169.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig10_slide10-1024x576.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig10_slide10-768x432.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig10_slide10-600x338.jpg 600w" sizes="(max-width: 1280px) 100vw, 1280px" /></div></div><div class="ult-spacer spacer-69d434ecdd1cf" data-id="69d434ecdd1cf" data-height="80" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3>Other sources for ontologies</h3>
<p>A few additional sources of coding systems and ontologies include:</p>
<ul>
<li><a href="https://www.cms.gov/medicare/fraud-and-abuse/physicianselfreferral/list_of_codes" target="_blank" rel="noopener">Current Procedural Terminology (CPT) codes</a>: an ontology used to report procedures and diagnostic services to healthcare providers, insurance companies, and accredited institutions.</li>
<li><a href="https://www.nlm.nih.gov/healthit/snomedct/index.html" target="_blank" rel="noopener">Systemized Nomenclature of Medicine – Clinical Terms (SNOMED-CT)</a>: an ontology used mainly by healthcare providers to share information in a more standardized manner.</li>
<li><a href="https://www.cms.gov/medicare/medicare-fee-for-service-payment/acuteinpatientpps/ms-drg-classifications-and-software" target="_blank" rel="noopener">Diagnosis-Related Group (DRG) codes</a>: a system used to classify hospital cases into approximately 500 groups for the purpose of payment.</li>
<li><a href="http://geneontology.org/" target="_blank" rel="noopener">Gene Ontology Consortium</a>: a set of frameworks for modeling complex biological systems and concepts.</li>
</ul>

		</div>
	</div>
</div></div></div></div><div class="g-cols wpb_row type_default valign_top vc_inner "><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2>How ontologies are used in PolyAnalyst™</h2>
<p>PolyAnalyst™ utilizes ontologies to search for and extract pertinent information. Within PolyAnalyst™, ontologies are imported as distinct &#8220;semantic dictionaries,&#8221; which are then used by PDL (Pattern Definition Language) to extract text patterns that correspond to various medical entities and their relationships as defined within those semantic dictionaries. In addition to the existing ontologies, users have the ability to create their own ontologies.</p>
<p>PolyAnalyst can import and start working with any kinds of ontologies. Some of these ontologies can be used free of charge, others require the user to obtain a license from the ontology owner. Currently, the following ontologies are provided along with PolyAnalyst™ (out-of-the box):</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<ul>
<li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8728220/" target="_blank" rel="noopener">Human Disease Ontology</a>: provides a framework for organizing and integrating information related to diseases in humans.</li>
<li><a href="https://hpo.jax.org/app/" target="_blank" rel="noopener">Human Phenotype Ontology</a>: provides a standardized vocabulary of phenotypic abnormalities encountered in human disease.</li>
<li><a href="https://www.nlm.nih.gov/mesh/meshhome.html" target="_blank" rel="noopener">MeSH</a>: covers a wide range of topics related to health and medicine.</li>
<li><a href="https://www.nlm.nih.gov/research/umls/rxnorm/index.html" target="_blank" rel="noopener">RXNorm</a>: contains information on medication names and their active ingredients.</li>
<li><a href="https://wordnet.princeton.edu/" target="_blank" rel="noopener">WordNet</a>: a broad semantic ontology that attempts to build a lexical database of English nouns, verbs, adjectives, and the relationships between them.</li>
</ul>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Other ontologies can be imported by the user (possibly, after acquiring the necessary license).</p>
<p>The Dictionary Manager in PolyAnalyst™ is the most direct way to access, edit, and view any imported ontologies in PolyAnalyst™. For example, you can import and view MeSH dictionary in PolyAnalyst™, searching for terms such as “CVA” to view all the associated information.</p>

		</div>
	</div>
<div class="ult-spacer spacer-69d434ecde2b6" data-id="69d434ecde2b6" data-height="80" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div><div class="w-image align_center"><div class="w-image-h"><img width="1920" height="1080" src="https://www.megaputer.com/wp-content/uploads/fig11_gif_medontology_pa_meshdict.gif" class="attachment-full size-full" alt="" /></div></div><div class="ult-spacer spacer-69d434ecdeed8" data-id="69d434ecdeed8" data-height="80" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>These ontologies can also be accessed within the project nodes. Some key nodes that utilize these types of ontologies include Entity Extraction and Taxonomy. Within these nodes, users have the option to select which ontology to import. Subsequently, they can craft a query to identify semantically related terms, such as synonyms, hyponyms, or hypernyms, or any other semantic relations, as defined within that particular ontology.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>The example below illustrates how a query searching for medical conditions within the &#8220;influenza and pneumonia&#8221; category, as per ICD-10, can be simplified by leveraging the PDL functions integrated into PolyAnalyst™ in conjunction with the MeSH ontology. By employing the &#8220;hyponym&#8221; function, as shown below, it becomes possible to capture all the subtypes of influenza and pneumonia without the need to list them individually.</p>

		</div>
	</div>
<div class="ult-spacer spacer-69d434ecdf355" data-id="69d434ecdf355" data-height="80" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div><div class="w-image align_center"><div class="w-image-h"><img width="1374" height="436" src="https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3.jpg" class="attachment-full size-full" alt="" srcset="https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3.jpg 1374w, https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3-300x95.jpg 300w, https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3-1024x325.jpg 1024w, https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3-768x244.jpg 768w, https://www.megaputer.com/wp-content/uploads/fig12_mesh-in-pa3-600x190.jpg 600w" sizes="(max-width: 1374px) 100vw, 1374px" /></div></div><div class="ult-spacer spacer-69d434ece00b3" data-id="69d434ece00b3" data-height="80" data-height-mobile="10" data-height-tab="10" data-height-tab-portrait="10" data-height-mobile-landscape="10" style="clear:both;display:block;"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Lastly, one of the challenges of using ontologies is the issue of ontology cleaning, which involves removing or correcting errors, inconsistencies and irrelevant or outdated information in the ontology. This is particularly important in specialized domains, such as chemistry or toxicology, where the ontology may contain a large number of concepts and entities that are not relevant to the domain or may cause confusion or misinterpretation (i.e., “milk” or “coffee” being present within a chemical substance ontology). Thankfully, the task of modifying ontologies is quite straight-forward within the Dictionaries module of PolyAnalyst™.</p>

		</div>
	</div>
<div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2>Ontologies and beyond</h2>
<p>Ontologies are powerful tools in information science, with potential that far exceeds the medical domain of the topics discussed here. Additional resources on using dictionaries in PolyAnalyst™ and how to import and utilize ontologies and specialized dictionaries can be found in our support and training materials.</p>

		</div>
	</div>
</div></div></div></div></div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/medical-ontology-and-its-use-in-text-analysis/">Medical Ontology and Its Use in Text Analysis</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
		<item>
		<title>Mastering Language Models: A Deep Dive into Input Parameters</title>
		<link>https://www.megaputer.com/mastering-language-models-a-deep-dive-into-input-parameters/</link>
		<pubDate>Thu, 28 Sep 2023 21:57:45 +0000</pubDate>
		<dc:creator><![CDATA[Echo Lu]]></dc:creator>
				<category><![CDATA[Uncategorized]]></category>
		<category><![CDATA[Text Analytics]]></category>
		<category><![CDATA[Artificial Intelligence]]></category>
		<category><![CDATA[Data Science]]></category>
		<category><![CDATA[Large Language Models]]></category>
		<category><![CDATA[NLP]]></category>

		<guid isPermaLink="false">https://www.megaputer.com/?p=34673</guid>
		<description><![CDATA[<p>Helping data analysts create effective dashboards is a key task for User Experience designers.  Even for experienced UX designers, a great amount of effort is spent on distilling large amounts of complex information into a simple, clear storytelling report to display to the client. As PolyAnalyst software UX designers, we work hard to incorporate web reporting features that help users create meaningful and interactive dashboards. Recently, we received some general requests from our users asking us for some tips on how to beautify their reports and make them more effective. So in response, we decided to share a few tips on creating better dashboards and web reports, including some of the DO’s and DON’Ts of dashboard design. </p>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/mastering-language-models-a-deep-dive-into-input-parameters/">Mastering Language Models: A Deep Dive into Input Parameters</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></description>
				<content:encoded><![CDATA[<section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>If you have ever used a language model through a playground or an API, you may have been asked to choose some input parameters. For many of us, the meaning of these parameters (and the right way to use them) is less than totally clear.</p>

		</div>
	</div>
<div class="w-separator size_small"></div><div class="w-image align_center"><div class="w-image-h"><img width="400" height="240" src="https://www.megaputer.com/wp-content/uploads/parameter-selection_in_the_sillytavern_interface.png" class="attachment-large size-large" alt="A screenshot showing parameter selection in the SillyTavern interface. Image by the author." srcset="https://www.megaputer.com/wp-content/uploads/parameter-selection_in_the_sillytavern_interface.png 400w, https://www.megaputer.com/wp-content/uploads/parameter-selection_in_the_sillytavern_interface-300x180.png 300w" sizes="(max-width: 400px) 100vw, 400px" /></div></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>This article will teach you how to use these parameters to control hallucinations, inject creativity into your model’s outputs, and make other fine-grained adjustments to optimize behavior. Much like prompt engineering, input parameter tuning can get your model running at 110%.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>By the end of this article, you’ll be an expert on five essential input parameters — temperature, top-p, top-k, frequency penalty, and presence penalty. You’ll also learn how each of these parameters helps us navigate the quality-diversity tradeoff.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>So, grab a coffee, and let’s get started!</p>

		</div>
	</div>
</div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<div id="table-of-contents">
<h3>Table of Contents</h3>
<ul>
<li><a href="#section1">Background</a></li>
<li><a href="#section2">Quality, Diversity, and Temperature</a></li>
<li><a href="#section3">Top-k and Top-p</a></li>
<li><a href="#section4">Frequency and Presence Penalties</a></li>
<li><a href="#section5">The Parameter-Tuning Cheat Sheet</a></li>
<li><a href="#section6">Wrapping up</a></li>
</ul>
<p><!-- Add more entries as needed --></p>
</div>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section1"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="5876" class="nb nc fo be nd ne nf go ng nh ni gr nj nk nl nm nn no np nq nr ns nt nu nv nw bj">Background</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Before we start, we will need to go over some background information about how these models work. Let’s start our deep dive by reviewing the fundamentals.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">To read a document, a language model breaks it down into a sequence of tokens. A token is just a small chunk of text that the model can easily understand: It could be a word, a syllable, or a character. For example, “Megaputer is great!” could be broken down into five tokens: [“Mega”, “puter ”, “is ”, “ great”, “!”]. This is done by the tokenizer.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Most language models we are familiar with operate by repeatedly generating the next token in a sequence. Each time the model wants to generate another token, it re-reads the entire sequence and then predicts the token that should come next. This strategy is known as <em class="oc">autoregressive </em>generation.</p>

		</div>
	</div>
<div class="w-image align_center meta_simple"><a class="w-image-h" href="https://developer.nvidia.com/blog/how-to-get-better-outputs-from-your-large-language-model/" title="How to Get Better Outputs from Your Large Language Model" target="_blank"><img width="320" height="320" src="https://www.megaputer.com/wp-content/uploads/autoregressive_generation.gif" class="attachment-large size-large" alt="GIF by Echo Lu, containing a modification of an image by Annie Surla from NVIDIA. Modified with permission from owner." /><div class="w-image-meta"><div class="w-image-title">Autoregressive generation of tokens by a language model.</div><div class="w-image-description">Contains a modification of an image by Annie Surla from NVIDIA.</div></div></a></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">This explains why ChatGPT prints the words out one at a time: It is streaming the words to you as it writes them.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">To generate a token, a language model assigns a likelihood score to each token in its vocabulary. A token gets a high likelihood score if it is a good continuation of the text and a low likelihood score if it is a poor continuation of the text, as assessed by the model. The sum of the likelihood scores over all tokens in the model’s vocabulary is always exactly equal to one.</p>

		</div>
	</div>
<div class="w-image align_center meta_simple"><a class="w-image-h" href="https://developer.nvidia.com/blog/how-to-get-better-outputs-from-your-large-language-model/" title="How to Get Better Outputs from Your Large Language Model" target="_blank"><img width="1024" height="260" src="https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process-1024x260.png" class="attachment-large size-large" alt="A language model assigns likelihood scores to predict the next token in the sequence. Original image by Annie Surla from NVIDIA, modified by Echo Lu with permission from owner." srcset="https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process-1024x260.png 1024w, https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process-300x76.png 300w, https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process-768x195.png 768w, https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process-600x152.png 600w, https://www.megaputer.com/wp-content/uploads/llm_token_sampling_process.png 1095w" sizes="(max-width: 1024px) 100vw, 1024px" /><div class="w-image-meta"><div class="w-image-title">A language model assigns likelihood scores to predict the next token in the sequence.</div><div class="w-image-description">Original image by Annie Surla from NVIDIA, modified with permission from owner.</div></div></a></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">After the likelihood scores are assigned, a token sampling scheme is used to pick the next token in the sequence. The token sampling scheme may incorporate some randomness so that the language model does not answer the same question in the same way every time. This randomness can be a nice feature in chatbots, as well as in some other applications.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="59cd" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em class="oc">TLDR: Language models break down text into chunks, predict the next chunk in the sequence, and mix in some randomness. Repeat as needed to generate output.</em></p>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section2"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="1164" class="oc nc ev be nd od oe fv nh of og fy nl oh oi oj ok ol om on oo op oq or os ot bj">Quality, Diversity, and Temperature</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>But why would we ever want to pick the second-best token, the third-best token, or any other token besides the best, for that matter? Wouldn’t we want to pick the best continuation (the one with the highest likelihood score) every time? Often, we do. But if we picked the best answer every time, we would get the same answer every time. If we want a diverse range of answers, we may have to give up some quality to get it. This sacrifice of quality for diversity is called the quality-diversity tradeoff.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>With this in mind, <em class="oc">temperature</em> tells the machine how to navigate the quality-diversity tradeoff. Low temperatures mean more quality, while high temperatures mean more diversity. When the temperature is set to zero, the model always samples the token with the highest likelihood score, resulting in zero diversity between queries, but ensuring that we always pick the highest quality continuation as assessed by the model.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p>Most commonly, we will want to set the temperature to zero for text analytics tasks. This is because, in text analytics, there is often a single “right” answer that we want to get every time. At temperature zero, we have the best chance of arriving at this answer in one shot. I like to set the temperature to zero for entity extraction, fact extraction, sentiment analysis, and most other things that I get up to in my job as an analyst. As a rule, you should always choose temperature zero for any prompt that you will only pass to the model once, as this is most likely to get you a good answer.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a120" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">At higher temperatures, we see more garbage and hallucinations, less coherence, and lower quality of responses on average, but also more creative and diverse responses. We use temperatures higher than zero when we want to pass the same prompt to the model many times and get many creative responses. We recommend that you should <em class="oc">only </em>use non-zero temperatures when you want to ask the same question twice and get two different answers.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>
<div class="w-separator size_small"></div><div class="w-image align_center meta_simple"><div class="w-image-h"><img width="423" height="355" src="https://www.megaputer.com/wp-content/uploads/quality-diversity_tradeoff.png" class="attachment-large size-large" alt="Higher temperatures bring diversity, creativity, and multiplicity of answers but also add garbage, incoherence, and hallucinations. Image by Echo Lu." srcset="https://www.megaputer.com/wp-content/uploads/quality-diversity_tradeoff.png 423w, https://www.megaputer.com/wp-content/uploads/quality-diversity_tradeoff-300x252.png 300w" sizes="(max-width: 423px) 100vw, 423px" /><div class="w-image-meta"><div class="w-image-title">Quality Diversity Tradeoff</div><div class="w-image-description">Temperature adds diversity, creativity, and multiplicity of answers but also garbage, incoherence, and hallucinations.</div></div></div></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a120" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">And why would we ever want two different answers to the same prompt? In some cases, having many answers to one prompt can be very useful. For example, there is a technique in which we generate many answers to a prompt and keep only the best one, which often produces better answers than a single query at temperature zero. Another use case is synthetic data generation: We want many synthetic data points, not just one. We may discuss these use cases (and others) in later articles, but more often than not, we need only one answer. <strong class="lr fp">When in doubt, choose temperature zero!</strong></p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a120" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">It is important to note that while temperature zero should <em class="oc">in theory</em> produce the same answer every time, this may not be true in practice! This is because the GPUs the model is running on can be prone to small miscalculations, such as rounding errors. These errors introduce a low level of randomness into the calculations, even at temperature zero. Since changing one token in a text can significantly alter its meaning, a single error may cause a cascade of different token choices later in the text, resulting in an almost totally different output. But rest assured that this usually has a negligible impact on quality. We only mention it so that you’re not surprised when you get some randomness at temperatures zero.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a120" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">There are more ways to navigate the quality-diversity tradeoff than temperature alone. In the next section, we will discuss some modifications to the temperature sampling technique. But if you are content with using temperature zero, feel free to skip it for now. You may rest soundly knowing that your choice of these parameters at temperature zero will not affect your answer.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a120" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em class="oc">TLDR: Temperature increases diversity but decreases quality by adding randomness to the model’s outputs.</em></p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section3"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="a476" class="nb nc fo be nd ne nf go ng nh ni gr nj nk nl nm nn no np nq nr ns nt nu nv nw bj">Top-k and Top-p sampling</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="6dd4" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">One common way to tweak our token-sampling formula is called top-k sampling. Top-k sampling is a lot like ordinary temperature sampling, except that the lowest likelihood tokens are excluded from being picked: Only the “top k” best choices are considered, which is where we get the name. The advantage of this method is that it stops us from picking truly bad tokens.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Let’s suppose, for example, that we are trying to make a completion for “The sun rises in the…” Then, without top-k sampling, the model considers every token in its vocabulary as a possible continuation of the sequence. Then there is some non-zero chance that it will write something ridiculous like “The sun rises in the refrigerator.” With top-k sampling, the model filters out these truly bad picks and only considers the k best options. By clipping off the long tail, we lose a little diversity, but our quality shoots way up.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>
<div class="w-separator size_small"></div><div class="w-image align_center meta_simple"><div class="w-image-h"><img width="565" height="453" src="https://www.megaputer.com/wp-content/uploads/unreliable-tail.png" class="attachment-large size-large" alt="Top-k sampling improves quality by keeping only the k best candidate tokens and throwing out the rest. Image by Echo Lu." srcset="https://www.megaputer.com/wp-content/uploads/unreliable-tail.png 565w, https://www.megaputer.com/wp-content/uploads/unreliable-tail-300x241.png 300w, https://www.megaputer.com/wp-content/uploads/unreliable-tail-499x400.png 499w" sizes="(max-width: 565px) 100vw, 565px" /><div class="w-image-meta"><div class="w-image-title">Top-k sampling</div><div class="w-image-description">Top-k sampling improves quality by keeping only the k best candidate tokens and throwing out the rest. </div></div></div></div><div class="w-separator size_medium"></div>
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Top-k sampling is a way to have your cake and eat it too: It gets you the diversity you need at a smaller cost to quality than with temperature alone. Since this technique is so wildly effective, it has inspired many variants.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">One common variant of top-k sampling is called top-p sampling, which is also known as nucleus sampling. Top-p sampling is a lot like top-k, except that it uses likelihood scores instead of token ranks to determine where it clips the tail. More specifically, it only considers those top-ranked tokens whose combined likelihood exceeds the threshold p, throwing out the rest.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">The power of top-p sampling compared to top-k sampling becomes evident when there are many poor or mediocre continuations. Suppose, for example, that there are only a handful of good picks for the next token, and there are dozens that just vaguely make sense. If we were using top-k sampling with k=25, we would be considering many poor continuations. In contrast, if we used top-p sampling to filter out the bottom 10% of the probability distribution, we might only consider those good tokens while filtering out the rest.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">In practice, top-p sampling tends to give better results compared to top-k sampling. By focusing on the cumulative likelihood, it adapts to the context of the input and provides a more flexible cut-off. So, in conclusion, top-p and top-k sampling can both be used at non-zero temperatures to capture diversity at a lower quality cost, but top-p sampling usually does it better.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Tip: For both of these settings, lower value = more filtering. At zero, they will filter out all but the top-ranked token, which has the same effect as setting the temperature to zero. So please use these parameters, but be aware that setting them too low will give up all of your diversity.</p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="8e2d" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em class="oc">TLDR: Top-k and top-p increase quality at only a small cost to diversity. They achieve this by removing the worst token choices before random sampling.</em></p>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>
<figure class="mo mp mq mr ms mt ml mm paragraph-image"></figure>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section4"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="e517" class="nb nc fo be nd ne nf go ng nh ni gr nj nk nl nm nn no np nq nr ns nt nu nv nw bj">Frequency and Presence Penalties</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="2cf9" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">We have just two more parameters to discuss before we start to wrap things up: The frequency and presence penalties. These parameters are — big surprise— yet another way to navigate the quality-diversity tradeoff. But while the temperature parameter achieves diversity by adding randomness to the token sampling procedure, the frequency and presence penalties add diversity by penalizing the reuse of tokens that have already occurred in the text. This makes the sampling of old and overused tokens less likely, influencing the model to make more novel token choices.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">The frequency penalty adds a penalty to a token for each time it has occurred in the text. This discourages repeated use of the same tokens/words/phrases and also has the side effect of causing the model to discuss more diverse subject matter and change topics more often. On the other hand, the presence penalty is a flat penalty that is applied if a token has already occurred in the text. This causes the model to introduce more new tokens/words/phrases, which causes it to discuss more diverse subject matter and change topics more often without significantly discouraging the repetition of frequently used words.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Much like temperature, the frequency and presence penalties lead us away from the “best” possible answer and toward a more creative one. But instead of doing this with randomness, they add targeted penalties that are carefully calculated to inject diversity into the answer. On some of those rare tasks requiring a non-zero temperature (when you require many answers to the same prompt), you might also consider adding a small frequency or presence penalty to the mix to boost creativity. But for prompts having just one right answer that you want to find in just one try, your odds of success are highest when you set all of these parameters to zero.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">As a rule, when there is one right answer, and you are asking just one time, you should set the frequency and presence penalties to zero. But what if there are many right answers, such as in text summarization? In this case, you have a little discretion. If you find a model’s outputs boring, uncreative, repetitive, or limited in scope, judicious application of the frequency or presence penalties could be a good way to spice things up. But our final suggestion for these parameters is the same as for temperature: <strong class="lr fp">When in doubt, choose zero!</strong></p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">We should note that while temperature and frequency/presence penalties both add diversity to the model’s responses, the kind of diversity that they add is not the same. The frequency/presence penalties increase the diversity <em class="oc">within a single response. </em>This means that a response will have more distinct words, phrases, topics, and subject matters than it would have without these penalties. But when you pass the same prompt twice, you are not more likely to get two different answers. This is in contrast with temperature, which increases diversity <em class="oc">between responses:</em> At higher temperatures, you will get a more diverse range of answers when passing the same prompt to the model many times.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">I like to refer to this distinction as within-response diversity vs. between-response diversity. The temperature parameter adds both within-response AND between-response diversity, while the frequency/presence penalties add only within-response diversity. So, when we need diversity, our choice of parameters should depend on the kind of diversity we need.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="d640" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em class="oc">TLDR: The frequency and presence penalties increase the diversity of subject matters discussed by a model and make it change topics more often. The frequency penalty also increases the diversity of word choice by reducing the repetition of words and phrases.</em></p>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section5"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="c9f4" class="nb nc fo be nd ne nf go ng nh ni gr nj nk nl nm nn no np nq nr ns nt nu nv nw bj">The Parameter-Tuning Cheat Sheet</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="328a" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">This section is intended as a practical guide for choosing your model’s input parameters. We first provide some hard-and-fast rules for deciding which values to set to zero. Then, we give some tips to help you find the right values for your non-zero parameters.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="328a" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">I strongly encourage you to use this cheat sheet when choosing your input parameters. Go ahead and bookmark this page now so you don’t lose it!</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3 id="9e0d" class="om nc fo be nd on oo op ng oq or os nj ly ot ou ov mc ow ox oy mg oz pa pb pc bj">Rules for setting parameters to zero:</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="88ed" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">Temperature:</p>
<ul>
<li>For a <strong class="lr fp">single answer</strong> per prompt: Zero.</li>
<li>For <strong class="lr fp">many answers</strong> per prompt: Non-zero.</li>
</ul>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="60cf" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Frequency and Presence Penalties:</p>
<ul>
<li>When there is<strong class="lr fp"> one correct answer:</strong> Zero.</li>
<li>When there are <strong class="lr fp">many correct answers:</strong> Optional.</li>
</ul>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="1d77" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Top-p/Top-k:</p>
<ul class="">
<li id="be8e" class="lp lq fo lr b gm ls lt lu gp lv lw lx ly pd ma mb mc pe me mf mg pf mi mj mk pg ph pi bj" data-selectable-paragraph="">With <strong class="lr fp">zero temperature</strong>: The output is not affected.</li>
<li id="9713" class="lp lq fo lr b gm pj lt lu gp pk lw lx ly pl ma mb mc pm me mf mg pn mi mj mk pg ph pi bj" data-selectable-paragraph="">With <strong class="lr fp">non-zero temperature</strong>: Non-zero.</li>
</ul>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="1d77" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">If your language model has additional parameters not listed here, leave them at their default values.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h3 id="32f8" class="om nc fo be nd on oo op ng oq or os nj ly ot ou ov mc ow ox oy mg oz pa pb pc bj">Tips for tuning the non-zero parameters:</h3>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="1d77" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Make a list of those parameters that should have non-zero values, and then go to a playground and fiddle around with some test prompts to see what works. <em>But if the rules above say to leave a parameter at zero, leave it at zero!</em></p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="1d77" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Tuning temperature/top-p/top-k:</p>
<ol class="">
<li id="1ea3" class="lp lq ev lr b ft ls lt lu fw lv lw lx ly pd ma mb mc pe me mf mg pf mi mj mk po ph pi bj" data-selectable-paragraph="">For more diversity/randomness, increase the temperature.</li>
<li id="641b" class="lp lq ev lr b ft pj lt lu fw pk lw lx ly pl ma mb mc pm me mf mg pn mi mj mk po ph pi bj" data-selectable-paragraph="">With non-zero temperatures, start with a top-p around 0.95 (or top-k around 250) and lower it as needed.</li>
</ol>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="950f" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Troubleshooting:</p>
<ol class="">
<li id="f1ed" class="lp lq ev lr b ft ls lt lu fw lv lw lx ly pd ma mb mc pe me mf mg pf mi mj mk po ph pi bj" data-selectable-paragraph="">If there is too much nonsense, garbage, or hallucination, decrease temperature and/or decrease top-p/top-k.</li>
<li id="77ec" class="lp lq ev lr b ft pj lt lu fw pk lw lx ly pl ma mb mc pm me mf mg pn mi mj mk po ph pi bj" data-selectable-paragraph="">If the temperature is high and diversity is low, increase top-p/top-k.</li>
</ol>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="950f" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Tip: While some interfaces allow you to use top-p and top-k at the same time, we prefer to keep things simple by choosing one or the other. Top-k is easier to use and understand, but top-p is often more effective.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="caaa" class="pw-post-body-paragraph lp lq ev lr b ft ls lt lu fw lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk eo bj" data-selectable-paragraph="">Tuning frequency penalty and presence penalty:</p>
<ol class="">
<li id="f568" class="lp lq ev lr b ft ls lt lu fw lv lw lx ly pd ma mb mc pe me mf mg pf mi mj mk po ph pi bj" data-selectable-paragraph="">For more diverse topics and subject matters, increase the presence penalty.</li>
<li id="db7a" class="lp lq ev lr b ft pj lt lu fw pk lw lx ly pl ma mb mc pm me mf mg pn mi mj mk po ph pi bj" data-selectable-paragraph="">For more diverse and less repetitive language, increase the frequency penalty.</li>
</ol>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="950f" class="pw-post-body-paragraph lp lq ev lr b ft ls lt lu fw lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk eo bj" data-selectable-paragraph="">Troubleshooting:</p>
<ol class="">
<li id="562f" class="lp lq ev lr b ft ls lt lu fw lv lw lx ly pd ma mb mc pe me mf mg pf mi mj mk po ph pi bj" data-selectable-paragraph="">If the outputs seem scattered and change topics too quickly, decrease the presence penalty.</li>
<li id="108e" class="lp lq ev lr b ft pj lt lu fw pk lw lx ly pl ma mb mc pm me mf mg pn mi mj mk po ph pi bj" data-selectable-paragraph="">If there are too many new and unusual words, or if the presence penalty is set to zero and you still get too many topic changes, decrease the frequency penalty.</li>
</ol>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="950f" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em>TLDR: You can use this section as a cheat sheet for tuning language models. You are <strong>definitely</strong> going to forget these rules, so bookmark this page and use it later as a reference.</em></p>

		</div>
	</div>
<div class="g-cols wpb_row type_default valign_top vc_inner " id="section6"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper"></div></div></div></div></div></div></div></div></div></section><section class="l-section wpb_row height_small"><div class="l-section-h i-cf"><div class="g-cols vc_row type_default valign_top"><div class="vc_col-sm-12 wpb_column vc_column_container"><div class="vc_column-inner"><div class="wpb_wrapper">
	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<h2 id="7610" class="nb nc fo be nd ne nf go ng nh ni gr nj nk nl nm nn no np nq nr ns nt nu nv nw bj">Wrapping up</h2>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="a507" class="pw-post-body-paragraph lp lq fo lr b gm nx lt lu gp ny lw lx ly nz ma mb mc oa me mf mg ob mi mj mk fh bj" data-selectable-paragraph="">There is no limit to the possible token-sampling strategies out there. Other notable strategies include beam search and adaptive sampling. However, the ones we’ve discussed here — temperature, top-k, top-p, frequency penalty, and presence penalty — are the most commonly used parameters. These are the parameters that you can expect to find in models like Claude, Llama, and the GPT series. In this article, we have shown that all of these parameters are really just here to help us navigate the quality-diversity tradeoff.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="e36b" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">Before we go, there is one last input parameter to mention: maximum token length. The maximum token length is just the cutoff where the model stops printing its answer, even if it is not finished responding. After this complex discussion, we hope this one is self-explanatory. ?</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="e36b" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph="">As we move further in this series, we’ll do more deep dives into advanced topics. Future articles will discuss prompt engineering, choosing the right language model for your use case, and more! Stay tuned, and keep exploring the vast horizons of language models with Megaputer Intelligence.</p>

		</div>
	</div>

	<div class="wpb_text_column ">
		<div class="wpb_wrapper">
			<p id="e36b" class="pw-post-body-paragraph lp lq fo lr b gm ls lt lu gp lv lw lx ly lz ma mb mc md me mf mg mh mi mj mk fh bj" data-selectable-paragraph=""><em class="oc">TLDR: When in doubt, set the temperature, frequency penalty, and presence penalty to zero. If that doesn’t work for you, reference the cheat sheet above.</em></p>

		</div>
	</div>
</div></div></div></div></div></section>
<p>The post <a rel="nofollow" href="https://www.megaputer.com/mastering-language-models-a-deep-dive-into-input-parameters/">Mastering Language Models: A Deep Dive into Input Parameters</a> appeared first on <a rel="nofollow" href="https://www.megaputer.com">Megaputer Intelligence</a>.</p>
]]></content:encoded>
			</item>
	</channel>
</rss>

<!--
Performance optimized by W3 Total Cache. Learn more: https://www.w3-edge.com/products/

Page Caching using disk: enhanced 
Minified using disk

Served from: www.megaputer.com @ 2026-04-06 17:34:21 by W3 Total Cache
-->