<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Paul Golding, Chief Technologist (AI)</title>
	<atom:link href="http://paulgolding.com/feed/" rel="self" type="application/rss+xml" />
	<link>http://paulgolding.com</link>
	<description>I use AI to generate business growth</description>
	<lastBuildDate>Sat, 22 Apr 2023 21:12:11 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.2.6</generator>
	<item>
		<title>Companies not Using AI Will Lose Sales</title>
		<link>http://paulgolding.com/2023/03/25/companies-not-using-ai-will-lose-sales/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Sat, 25 Mar 2023 19:50:06 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1092</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_0 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_0">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_image et_pb_image_0">
				
				
				
				
				<span class="et_pb_image_wrap "><img decoding="async" width="768" height="507" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/03/mind_meld.png" alt="" title="mind_meld" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/03/mind_meld.png 768w, http://paulgolding.com/wp-content/uploads/sites/2/2023/03/mind_meld-480x317.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 768px, 100vw" class="wp-image-1094" /></span>
			</div>
			</div>
				
				
				
				
			</div><div class="et_pb_row et_pb_row_1">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_1  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><h2 class="reader-text-block__heading1">Mind Meld with the Customer</h2>
<p class="reader-text-block__paragraph">How would you like to mind-meld with the customer, as if you had perfect knowledge of all information pertinent to the sale?</p>
<p class="reader-text-block__paragraph">With AI, this might just become a reality.</p>
<p class="reader-text-block__paragraph">Thanks to large-scale computation, AI has now exceeded beyond-human performance in many economically valuable cognitive tasks, especially related to language through the use of so-called Large Language Models (LLMs).</p>
<p class="reader-text-block__paragraph">Naturally, many business processes will be amenable to beyond-human automation, including sales.</p>
<p class="reader-text-block__paragraph">In full disclosure, I credit inspiration to Irzana Golding for her article<span> </span><a href="https://igolding.medium.com/operationalizing-competitive-intelligence-including-chatgp-ff5be9293f77">&#8220;Operationalizing Competitive Intelligence, including ChatGPT&#8221;</a><span> </span>in which she introduces the notion of a &#8220;SalesGPT&#8221; for competitive-intelligence questions.</p>
<p class="reader-text-block__paragraph">She posits how a locally-informed LLM could bring critical competitive intelligence (CI) to the salesperson&#8217;s attention in a way that might clinch the deal.</p>
<p class="reader-text-block__paragraph">But how?</p>
<h2 class="reader-text-block__heading1">LLMs are Not Just Language Models</h2>
<p class="reader-text-block__paragraph">It is often revealed via<span> </span><a href="https://www.copper.com/resources/win-loss-analysis">win-loss analysis</a><span> </span>that lost deals (some say up to &#8220;50%&#8221;) were winnable, but for some missing information.</p>
<p class="reader-text-block__paragraph">It&#8217;s likely that the information was available, but unobserved. As Sherlock Holmes said: &#8220;Watson, you see, but you do not observe.&#8221;</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQGaLxXZkXDyHQ/article-inline_image-shrink_1000_1488/0/1679704742471?e=1684972800&amp;v=beta&amp;t=DRxUPcwLerYM6sCDKzda06h50-oIPm8Li9WClMg6h54" loading="lazy" alt="No alt text provided for this image" id="ember3473" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div>
</figure>
</div>
<p class="reader-text-block__paragraph">In the realm of deductive reasoning, observation is akin to connecting the dots. For humans, the process is far from simple and fraught with mishaps and biases. Enter the LLM, a powerful tool that recognizes semantic patterns and can apply them to knowledge across disparate datasets.</p>
<p class="reader-text-block__paragraph">Beyond being a mere sentence generator, an LLM possesses a prodigious memory of all the data it has encountered. With this wealth of information, an LLM can expertly connect disparate pieces of data and generate insights that might have been impossible for humans to discern.</p>
<p class="reader-text-block__paragraph">This ability to identify patterns in data is especially useful for businesses seeking to gain a competitive edge. An LLM could identify connections between seemingly unrelated data sets and illuminate the potential impact of a competitor&#8217;s latest product feature on a company&#8217;s sales play.</p>
<p class="reader-text-block__paragraph">Consider the following prompt provided to ChatGPT about going up against a fictitious company called Volanti:</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQF0gGDdBjgoFg/article-inline_image-shrink_1500_2232/0/1679693472878?e=1684972800&amp;v=beta&amp;t=LTe5057A_WzmzdSTCUrrgh8KXnvpcsy-86QMb1Hryho" loading="lazy" alt="No alt text provided for this image" id="ember3474" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Competitive Intelligence Provided to ChatGPT</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">Note the blue highlights are the contextual prompts that I entered manually for the demo. For an integrated solution (&#8220;SalesGPT&#8221;), this data would be retrieved dynamically (from a vector database like<span> </span><a href="https://www.pinecone.io/">Pinecone</a>) in response to the question. The aggregated data would be sourced from various CI data sources.</p>
<p class="reader-text-block__paragraph">Now let&#8217;s pose the question:<span> </span><strong><em>&#8220;How might we win the deal selling to a client interested in Volanti because of their Zero Trust offering?&#8221;</em></strong></p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQHYcj5bZadqHA/article-inline_image-shrink_1500_2232/0/1679693717054?e=1684972800&amp;v=beta&amp;t=zTYfRznEuG4hCIx2YCV1EamWQYN6p_LYB6gqwX_ELHg" loading="lazy" alt="No alt text provided for this image" id="ember3475" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Response from ChatGPT about sales play informed by CI data</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">Note how the response is a synthesis–&#8221;connecting the dots&#8221;. Consider that the salesperson might not have noticed the CRM note by a colleague, but that&#8217;s OK because &#8220;SalesGPT&#8221; did.</p>
<p class="reader-text-block__paragraph">The response is synthesized from a range of disparate CI sources, turning that data into something potentially cogent and salient, given the circumstances–i.e. how to deal with the competitor&#8217;s feature. Notice the statement:</p>
<blockquote class="reader-text-block__quote"><p>&#8220;Highlight our roadmap: While we may not currently offer Zero Trust, we have a roadmap to introduce it in the future. It is important to communicate this to the client and emphasize that our implementation will be built from the ground up, like our flagship product, ensuring optimal integration and cost efficiency.</p></blockquote>
<p class="reader-text-block__paragraph">This is a classic sales play to recognize an objection and then turn it around to become an advantage–i.e. yes, we&#8217;ll be later, but better and lower cost &#8212; here&#8217;s why!</p>
<p class="reader-text-block__paragraph">Maybe this contrived example is too crude and obvious, but you get the idea. And don&#8217;t overestimate the ability of a human sales person to do all this synthesis for themselves. There are plenty of places to go wrong: information blindness, distraction, cognitive biases, information overload etc.</p>
<h2 class="reader-text-block__heading1">Zero Sum Game or Secret Sauce?</h2>
<p class="reader-text-block__paragraph">From my recent experiments with a number of business use cases, the exciting feature of LLMs is how much latent power there is waiting to be utilized via prompt-engineering, data augmentation, fine-tuning and various architectural tricks. LLMs offer a new toolkit for conducting information retrieval experiments and then quickly putting them into production.</p>
<p class="reader-text-block__paragraph">But what if everyone is using the same LLM and similar competitive data? Isn&#8217;t this a zero-sum game? And, what if the client is using the same technology, second-guessing the sales pitch? (This seems likely.)</p>
<p class="reader-text-block__paragraph">The winner will be whoever uses LLMs creatively enough. The right approach could lead to unique in-house advantages, making LLM-hacking &#8220;recipes&#8221; the secret sauce for sales.</p>
<p class="reader-text-block__paragraph">Moreover, with recent announcements, like<span> </span><a href="https://www.databricks.com/blog/2023/03/24/hello-dolly-democratizing-magic-chatgpt-open-models.html">Databrick&#8217;s work with Dolly</a>, or<span> </span><a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">Stanford&#8217;s work with Alpaca</a>, there is every reason to believe that enterprises could build their own LLMs without requiring the gargantuan budgets used to build GPT-4.</p>
<p class="reader-text-block__paragraph">There are so many possibilities, such as the use of<span> </span><a href="https://arxiv.org/pdf/2210.06280.pdf">LLMs to create tabular datasets</a><span> </span>from unstructured, semi-structured and structured data (e.g. the CRM, sales reports and various metrics tables) and feed those into powerful prediction frameworks like<span> </span><a href="https://xgboost.readthedocs.io/en/stable/index.html">XGBoost</a><span> </span>or<span> </span><a href="https://www.microsoft.com/en-us/research/project/econml/">EconML</a><span> </span>to generate predictive and causal outlooks. Better still, these outlooks are blended back into the LLM generative toolchain to enable sales teams to ask speculative or causal questions:</p>
<blockquote class="reader-text-block__quote"><p>Which of the three contacts in the client is most likely to have objections and what might they be?</p></blockquote>
<blockquote class="reader-text-block__quote"><p>Why did the customer to ask about our Composable API roadmap?</p></blockquote>
<p class="reader-text-block__paragraph">Well, this is just the<span> </span><a href="https://www.linkedin.com/pulse/chatgpt-tip-iceberg-paul-golding/">tip of iceberg for enterprise applications of LLMs</a>.</p>
<p class="reader-text-block__paragraph">If you want to know more about how to do your own LLM-hacking for your use case, feel free to<span> </span><a href="https://paulgolding.com/">get in touch</a><span> </span>(or connect on LinkedIn).</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The AI Paradigm: Scaling</title>
		<link>http://paulgolding.com/2023/03/14/the-ai-paradigm-scaling/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Tue, 14 Mar 2023 19:47:43 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1088</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_1 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_2">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_2  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_image et_pb_image_1">
				
				
				
				
				<span class="et_pb_image_wrap "><img decoding="async" width="767" height="513" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/03/complexity.png" alt="" title="complexity" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/03/complexity.png 767w, http://paulgolding.com/wp-content/uploads/sites/2/2023/03/complexity-480x321.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 767px, 100vw" class="wp-image-1090" /></span>
			</div>
			</div>
				
				
				
				
			</div><div class="et_pb_row et_pb_row_3">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_3  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_1  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p class="reader-text-block__paragraph">The emergence of Large Language Models with their impressive beyond-human performance (in many<span> </span><a href="https://gluebenchmark.com/">benchmarks</a>) gives us reason to ask:<span> </span><em>what is the strategic value of the underlying principles to the enterprise?</em></p>
<p class="reader-text-block__paragraph">As an antidote to many AI books with faux-paradigms, I explore the at least one core paradigm of Large AI, which is the ability to address Complexity through<span> </span><a href="https://medium.com/sfi-30-foundations-frontiers/scaling-the-surprising-mathematics-of-life-and-civilization-49ee18640a8">scaling</a><span> </span>(Large AI models are Complex Systems).</p>
<p class="reader-text-block__paragraph">Because enterprises are Complex systems, we should expect that they too are amenable to computation using Large AI in the same way human language (also a Complex system) has proven amenable.</p>
<p class="reader-text-block__paragraph">I call these new models<span> </span><em>Large Enterprise Models (LEMs)</em>. Their emergence could have profound affect upon corporate strategy generally, but certainly IT strategy.</p>
<h2 class="reader-text-block__heading1">Introduction</h2>
<p class="reader-text-block__paragraph">Many AI solutions in the enterprise exist within a standard digital transformation paradigm: AI as an embedded technology choice addressing a localized use case (e.g. supply chain management or product recommendation).</p>
<p class="reader-text-block__paragraph">Some biz leaders are trying to parse what the apparent inflection-point around Large AI (e.g. ChatGPT) means in a more strategic sense, if anything.</p>
<p class="reader-text-block__paragraph">The motivation of this article is to provoke thought about an &#8220;AI-first&#8221; mindset that asks:<span> </span><em>What happens if we build our enterprise around Large AI?</em></p>
<p class="reader-text-block__paragraph">My post will be short and non-technical, leaving technical justifications for a long-form elsewhere, most likely<span> </span><a href="https://paulgolding.com/">on my site</a>.</p>
<h2 class="reader-text-block__heading1">First, Beware of &#8220;Pop&#8221; Paradigms</h2>
<p class="reader-text-block__paragraph">Inevitably, some biz leaders are seeking answers in AI books or magazines.</p>
<p class="reader-text-block__paragraph">However, many follow the same glib pattern of introducing quirky metaphors (&#8220;Tipping Points&#8221;) sprinkled with uncritical marketing case studies. They sound convincing, only to yield little actionable substance.</p>
<p class="reader-text-block__paragraph">Many case studies lacks critical details, such as why, exactly, AI was the best choice and what percentage of improvement was uniquely attributable to the AI. (You should always ask this question.)</p>
<p class="reader-text-block__paragraph">Some books invent concepts – e.g. &#8220;Cognitive Enterprise&#8221;. It sounds convincing, but is marketing blurb that wouldn&#8217;t translate into a CIO, CTO or CMO world. (Read<span> </span><a href="https://bigthink.com/personal-growth/how-to-use-the-feynman-technique-to-identify-pseudoscience/">Richard Feynman&#8217;s advice</a><span> </span>on meaning versus marketing jargon.)</p>
<p class="reader-text-block__paragraph">Analyst outputs are often no better, with uncritical claims like &#8220;AI high flyers&#8221; attributing 30% of revenues to AI.</p>
<p class="reader-text-block__paragraph">Really? How, exactly?</p>
<p class="reader-text-block__paragraph">Lacking counterfactuals and critical contextual details, we don&#8217;t know.</p>
<p class="reader-text-block__paragraph">Many transformation improvements are not driven by a single technology, but rather come about via refactoring and modernization. Solutions might contain AI, hence successes will be attributed to AI versus refactoring.</p>
<p class="reader-text-block__paragraph">Given the poor success rate of many transformations, the overlap of this with the oft-reported &#8220;80% of AI projects fail&#8221; might not be a coincidence. As<span> </span><a href="https://www.linkedin.com/in/ACoAAB6FDOABije2nlT1yX0kWln606bCPQU3Iwk">Tony Saldanha</a><span> </span>argues well in &#8220;Why Digital Transformations Fail&#8221;, a common failure mode is focus on technology over value. Was AI-for-AI&#8217;s sake the culprit?</p>
<p class="reader-text-block__paragraph">More often, the reason for failure is simple: the organization wasn&#8217;t ready to switch to an automated process. This happens when AI programs are developed in isolation of change management or suitable operating models, like<span> </span><a href="https://www.cio.com/article/193401/crossing-the-project-to-product-chasm.html">product-centric IT.</a></p>
<h2 class="reader-text-block__heading1">AI-assisted Strategy?</h2>
<p class="reader-text-block__paragraph">Given the connection between strategy and innovation, one framing for exploring the application of AI is via the different types of innovation, per Geoffrey Moore, This is the common method, albeit perhaps via value-streams, not necessarily aligned with these innovation types:</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQG92teaSJvaKA/article-inline_image-shrink_1500_2232/0/1678734583254?e=1684972800&amp;v=beta&amp;t=CEIdYcC-HczUj1sppjp83-kNBz7HeMjKWsZfqgUVxFI" loading="lazy" alt="No alt text provided for this image" id="ember3254" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Source: click on image</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">Each types will have a related strategies such as Marketing Strategy, Information Strategy, and so on, hence we can ask how AI might impact these strategies, however they are formulated according to the current operating model:</p>
<blockquote class="reader-text-block__quote"><p>How might AI assist my &lt;innovation type&gt; strategy?</p></blockquote>
<p class="reader-text-block__paragraph">In reality, no such question is formally posed. In many of these cases, the choice of a solution will be based upon performance against an existing use case.</p>
<p class="reader-text-block__paragraph">You pick the &#8220;AI solution&#8221; if it offers significant performance in relation to alternatives or if it unlocks an otherwise stubborn use case (e.g. something requiring NLP). Increasingly, the AI is embedded into a vendor solution, but sometimes a data science team might devise custom AI solutions.</p>
<p class="reader-text-block__paragraph">These selections become strategic when part of a long-term rollout to achieve whatever the overriding strategy is.</p>
<p class="reader-text-block__paragraph">Within this rubric, AI can become a core strategic enabler if it&#8217;s a common undergirding theme. Typically, this translates into data strategy e.g. to allow better access to joined-up datasets, data augmentation, secure data federation, and so on.</p>
<p class="reader-text-block__paragraph">More general strategic imperatives might fall out of such considerations, such as the need to scale data, which cannot be done overnight (can take years) and so definitely requires strategy and might even affect partnerships or ecosystem strategies etc.</p>
<h2 class="reader-text-block__heading1">AI Paradigm?</h2>
<p class="reader-text-block__paragraph">More fundamentally, we need to ask whether or not AI presents a paradigm shift, like with the invention of the web browser did.</p>
<p class="reader-text-block__paragraph">If so, what is the paradigm?</p>
<p class="reader-text-block__paragraph"><em>There is one problem that AI uniquely solves</em>, with strong evidence from breakthroughs in Large Language Models (LLMs), like ChatGPT.</p>
<p class="reader-text-block__paragraph"><strong>The unique paradigm is solving<span> </span></strong><a href="https://medium.com/@junp01/an-introduction-to-complexity-theory-3c20695725f8"><strong>Complexity</strong></a><span> </span>(with a &#8220;C&#8221;)<strong>.</strong></p>
<p class="reader-text-block__paragraph">AI magic has mostly appeared thanks to one thing: scaling. The &#8220;magic models&#8221; are huge: 175B parameters for GPT-3, and climbing:</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQH7vSsY6o0Kgw/article-inline_image-shrink_1000_1488/0/1678756961913?e=1684972800&amp;v=beta&amp;t=TU0AHS9kYm5BYTDXrNNZBd8oIhn2OCyOjzIIjcytbvk" loading="lazy" alt="No alt text provided for this image" id="ember3255" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Scaling of parameters in large language models: click for source (Huggingface)</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">
<h3 class="reader-text-block__heading2">What does scaling do?</h3>
<p class="reader-text-block__paragraph">It allows a model to encompass complexity with its impenetrable web of interactions between many smaller components (e.g. words).<span> </span><strong>AI makes complexity amenable to computation!</strong></p>
<p class="reader-text-block__paragraph">Note that<span> </span><a href="https://medium.com/sfi-30-foundations-frontiers/scaling-the-surprising-mathematics-of-life-and-civilization-49ee18640a8">scaling is a concept already embedded within complexity theory</a>.</p>
<p class="reader-text-block__paragraph">There is a maxim in complexity theory, that complexity can only be &#8220;met&#8221; with complexity.<span> </span><a href="https://www.santafe.edu/research/results/working-papers/language-is-a-complex-adaptive-system">Human language is a complex dynamical system</a>, which is why Large AI works.</p>
<p class="reader-text-block__paragraph">LLMs are perhaps the first real evidence that a<span> </span><a href="https://onlinelibrary.wiley.com/doi/10.1111/cogs.13256">complex system (175B parameters) can indeed encompass a complex mechanism</a><span> </span>(language).</p>
<p class="reader-text-block__paragraph">The magic levels of performance only became apparent<span> </span><em>once the models were scaled significantly</em><span> </span>(with no other changes). Smaller versions were relatively unimpressive.</p>
<p class="reader-text-block__paragraph">Moreover, unexpected capabilities of LLMs continue to<span> </span><em>emerge</em><span> </span>merely by tinkering with the inputs (&#8220;prompt engineering&#8221; – apparently<span> </span><a href="https://www.datanami.com/2023/02/14/prompt-engineer-the-next-hot-job-in-ai/">the next hot job</a>).</p>
<p class="reader-text-block__paragraph">This has led some thinkers to use the term<span> </span><a href="https://arxiv.org/abs/2206.07682">emergence in relation to LLMs</a>. Well, we should not be surprised because<span> </span><a href="https://necsi.edu/emergence">emergence is a key characteristic of a complex system.</a></p>
<h2 class="reader-text-block__heading1">Large Enterprise Models</h2>
<p class="reader-text-block__paragraph">What then, does scaling mean for the enterprise?</p>
<p class="reader-text-block__paragraph">Two strategic opportunities arise:</p>
<ol>
<li>Systems like LLMs can play a big role in enterprise management, given the undeniable claim that<span> </span><a href="https://clear-impact.com/wp-content/uploads/2018/04/Intro-to-Cynefin-Model.pdf">enterprises are also Complex systems</a><span> </span>(which is what<span> </span><a href="https://www.santafe.edu/research/themes/invention-innovation">drives innovation</a>) and much of that complexity is presented in language-centric processes. LLMs make enterprise knowledge computable.</li>
<li>Scaling should also play a role directly in enterprise management – i.e. via novel models applied to the entire information state-space of the enterprise:<span> </span><em>Large Enterprise Models, or LEMs</em>.</li>
</ol>
<p class="reader-text-block__paragraph">Both of these present significant strategic implications that few enterprises are prepared for, such as:</p>
<ol>
<li>Making<span> </span><em>all data</em><span> </span>in the enterprise available to AI models. We are still in the early days of concepts like<span> </span><a href="https://martinfowler.com/articles/data-mesh-principles.html">Data Mesh</a> and <a href="https://www.gartner.com/smarterwithgartner/data-fabric-architecture-is-key-to-modernizing-data-management-and-integration">Data Fabric</a>, but these are a step in the right direction and will involve AI (e.g. via LLMs) for sense-making<span> </span><em>prior to use in larger models</em>. How do we architect and manage this?</li>
<li>Do we have a single multi-modal Large Enterprise Model or a highly distributed ensemble of many weaker LEMs (plus other models)? What does that look like?</li>
<li>How do we operate with increasingly opaque models?</li>
<li>What is the impact upon<span> </span><a href="https://neilfaganblog.wordpress.com/2018/04/24/systems-thinking-vs-complexity-theory-or-why-building-a-strategy-could-be-the-wrong-thing-to-do/">organizational models</a>?</li>
<li>What kind of skill sets are required to work in these hyper organizations?</li>
<li>Do we need new ways of thinking about partnerships–e.g. at the federated &#8220;model&#8221; level?</li>
<li>Do we need a pivot towards<span> </span><a href="https://www.researchgate.net/publication/292653679_Complexity_and_systems_thinking">Systems Thinking</a>? Where are the frontiers of &#8220;The System&#8221; in a highly connected marketplace and ecosystem?</li>
</ol>
<p class="reader-text-block__paragraph">All of the above is largely uncharted territory and nobody quite yet knows what kind of new strategy emerges from the paradigm. But the question &#8220;What is AI strategy&#8221; must recognize the above and figure out how to deal with the new paradigm as applied to all facets of business: customers, processes, markets, people, regulations, sustainability etc.</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Mindless Data</title>
		<link>http://paulgolding.com/2023/03/08/mindless-data/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Wed, 08 Mar 2023 19:44:55 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1083</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_2 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_4">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_4  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_image et_pb_image_2">
				
				
				
				
				<span class="et_pb_image_wrap "><img decoding="async" width="747" height="501" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/03/mindless_data.png" alt="" title="mindless_data" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/03/mindless_data.png 747w, http://paulgolding.com/wp-content/uploads/sites/2/2023/03/mindless_data-480x322.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 747px, 100vw" class="wp-image-1085" /></span>
			</div>
			</div>
				
				
				
				
			</div><div class="et_pb_row et_pb_row_5">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_5  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_2  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p class="reader-text-block__paragraph">I my last post, I wrote about being<span> </span><a href="https://www.linkedin.com/posts/paulgolding_i-just-read-the-blurb-to-an-expensive-book-activity-7038690501776068608-_7Yf?utm_source=share&amp;utm_medium=member_desktop">&#8220;data driven&#8221; operationally</a>, and in the post before that I wrote about the<span> </span><a href="https://www.linkedin.com/posts/paulgolding_modeling-mindsets-activity-7037519186067165184-9iNe?utm_source=share&amp;utm_medium=member_desktop">mindsets that surround using machine learning models</a>.</p>
<p class="reader-text-block__paragraph">Let me try to marry the two ideas.</p>
<p class="reader-text-block__paragraph">There is a widely shared mindset that models are somehow free of bias. And I don&#8217;t mean statistical bias. I mean prejudicial thinking. We may not recognize this at first because the name &#8220;data science&#8221; seems to suggest that modeling is a scientific endeavor.</p>
<p class="reader-text-block__paragraph">Let&#8217;s say it is scientific, then what kind of science?</p>
<p class="reader-text-block__paragraph">These methods, as used in enterprises, are often closer to social science.</p>
<p class="reader-text-block__paragraph">I say that in the sense that we are often trying to account for behavioral mechanisms (in the enterprise) using statistics.</p>
<p class="reader-text-block__paragraph">The hypotheses formed about the enterprise are theories about human activities, not, say, balls rolling down a slope (laws of physics).</p>
<p class="reader-text-block__paragraph">As such, business leaders need to be aware that this entire gamut of activities can easily be gamed to make the statistics say whatever we want.</p>
<p class="reader-text-block__paragraph">A damning analysis of how this happens is well documented by Gerd Gigerenzer in his paper:<span> </span><a href="https://lnkd.in/gYjF-Nsm">&#8220;Mindless statistics&#8221;</a><span> </span>&#8212; the inspiration for my article&#8217;s title.</p>
<p class="reader-text-block__paragraph">In it, he gives an account of how a statistics textbook written by a university professor was revised to remove a broader discussion of statistical techniques in order to focus on only one. When asked why, the professor said:</p>
<blockquote class="reader-text-block__quote"><p><strong>Most researchers [&#8230;] are not really interested in statistical thinking, but only in how to get their papers published.</strong></p></blockquote>
<p class="reader-text-block__paragraph">We could easily substitute this for:<span> </span><em>&#8220;Most managers are not really interested in statistical thinking, but only how to get their ideas accepted.&#8221;</em></p>
<p class="reader-text-block__paragraph"><strong>This reality is what data-driven operating models need to address.</strong></p>
<p class="reader-text-block__paragraph">Yet, the discussion is often about &#8220;Data Governance&#8221;, or the like, as if assuring quality and control of data will translate into statistically-valid decisions.</p>
<p class="reader-text-block__paragraph">Perhaps in the oft-quoted<span> </span><a href="https://www.teradata.com/Glossary/What-are-the-5-V-s-of-Big-Data#:~:text=Big%20data%20is%20a%20collection,variety%2C%20velocity%2C%20and%20veracity.">5 Vs of Big Data</a>, there needs to be a sixth V: Veridicality, as in<span> </span><strong>coinciding with reality</strong>.</p>
<p class="reader-text-block__paragraph">There are no easy answers here.</p>
<p class="reader-text-block__paragraph">We can wave our hands about solutions, yet the<span> </span><a href="https://deepblue.lib.umich.edu/bitstream/handle/2027.42/92168/TheUseOfStatisticalHeuristics.pdf%3Bjsessionid%3D99D76EB98D7D1C97F9B387FE4A1F8868?sequence%3D1">statistical-heuristical dichotomy</a><span> </span>pervades how we operate our businesses.</p>
<p class="reader-text-block__paragraph">I can say two things as food for thought:</p>
<h3 class="reader-text-block__heading2">Nurturing a Numerical Culture &amp; Systems Thinking</h3>
<p class="reader-text-block__paragraph">First, as mentioned in the original mindsets-post, all processes are socio-technical in nature. Social pertains to culture. Before we talk about AI-first, we need to talk about building a data-first culture, possible even with the humble spreadsheet or, heavens-forbid, a whiteboard or napkin.</p>
<p class="reader-text-block__paragraph">We should all be hands-on with numbers and strive to explain our arguments using numerical reasoning versus our tendency to resort to qualitative narratives. Business leaders should, well, take the lead, and be sure to encourage numerically-reasoned arguments, including causality assumptions.</p>
<p class="reader-text-block__paragraph">This is hard to do, but that&#8217;s good because it means that it requires<span> </span><a href="https://thedecisionlab.com/reference-guide/philosophy/system-1-and-system-2-thinking">System 2 thinking</a><span> </span>that is supposedly less vulnerable to bias. Its deliberate application to various actives has been fairly well<span> </span><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5344059/#:~:text=System%202%20is%20the%20more,through%20past%20learning%20and%20experience.">studied</a><span> </span>(beyond the problematic Thinking Fast &amp; Slow book).</p>
<p class="reader-text-block__paragraph">Much has been said about the value of<span> </span><a href="https://www.zdnet.com/article/7-ways-design-thinking-is-percolating-into-the-enterprise-application-space/">design thinking</a>, with many companies, like<span> </span><a href="https://www.ibm.com/design/thinking/static/Enterprise-Design-Thinking-Report-8ab1e9e1622899654844a5fe1d760ed5.pdf">IBM, adopting it systematically</a>. But little is said about numerical thinking or its grander counterpart systems-thinking, which is often still treated as a kind of esoteric topic favored by management theorists and Complexity fans.</p>
<p class="reader-text-block__paragraph">Just as we should all become design thinkers, we equally need to become systems thinkers. (See this<span> </span><a href="https://www.eckerson.com/articles/a-systems-thinking-view-of-analytics-part-i-introduction-to-systems-thinking">great post</a><span> </span>by<span> </span><a href="https://www.linkedin.com/in/dlwells/">Dave Wells</a><span> </span>on<span> </span><a href="https://www.eckerson.com/articles/looking-at-the-future-through-analytics-predictive-vs-prognostic">prognostic analytics</a><span> </span>and systems thinking.)</p>
<h3 class="reader-text-block__heading2">Discipline: Start as You Mean to Finish</h3>
<p class="reader-text-block__paragraph">Still related to culture, leaders have to set the tone about what kinds and levels of discipline they want to operate with. We all know that our<span> </span><a href="https://fs.blog/bias-incentives-reinforcement/">behaviors align to incentives</a>. If good data discipline isn&#8217;t incentivized, then who&#8217;s going to bother?</p>
<p class="reader-text-block__paragraph">It is well documented that the manner in which a start-up sets out in terms of adopting good practices is how it ends up. Indeed, they are related because good practices are what makes a business repeatable and scalable. We have known this since Six Sigma, which saved many US companies from the onslaught of process-obsessed Japanese industries.</p>
<p class="reader-text-block__paragraph">Yet, just as we learned that you cannot inspect robustness into a product or process (hence Design for Manufacturability), we have seemingly not learned that you cannot inspect robustness into your data processes or culture.</p>
<p class="reader-text-block__paragraph">This is still the dominant mindset of so many enterprises and is a reflection of the same lack of strategic mindset (i.e. dominance of short-term thinking &#8212; action first, data second) that dogged US industry prior to Six Sigma.</p>
<p class="reader-text-block__paragraph">Cultivating a mindset of doing things &#8220;the right way&#8221; is essential. In some areas, like software engineering, the principles of &#8220;the right way&#8221; are now well established that they are infused into the wider culture of software participants almost as a code of conduct.</p>
<p class="reader-text-block__paragraph">These methods are beginning to percolate into data via the DataOps gamut of ideas, but still slow to penetrate wider organizational behaviors.</p>
<p class="reader-text-block__paragraph">I would argue that Design for Measurement ought to become part of Design Thinking. Indeed, like all new things, it is actually an<span> </span><a href="https://www.creative-wisdom.com/teaching/WBI/dma.shtml">old idea</a>, but still broadly missing from our codes of conduct.</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>ChatGPT is the Tip of the Iceberg</title>
		<link>http://paulgolding.com/2023/02/27/chatgpt-is-the-tip-of-the-iceberg/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Mon, 27 Feb 2023 19:42:38 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1079</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_3 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_6">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_6  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_image et_pb_image_3">
				
				
				
				
				<span class="et_pb_image_wrap "><img decoding="async" width="900" height="600" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/03/tip_of_iceberg.png" alt="" title="tip_of_iceberg" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/03/tip_of_iceberg.png 900w, http://paulgolding.com/wp-content/uploads/sites/2/2023/03/tip_of_iceberg-480x320.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 900px, 100vw" class="wp-image-1081" /></span>
			</div>
			</div>
				
				
				
				
			</div><div class="et_pb_row et_pb_row_7">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_7  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_3  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p class="reader-text-block__paragraph">ChatGPT is just the tip of the iceberg that enterprises find themselves crashing into even though the underlying mass – Large Language Models (LLMs) – was introduced back in 2017. In AI years,<span> </span><strong>that’s a lifetime ago</strong>.</p>
<p class="reader-text-block__paragraph">LLMs ought to have already featured in digital transformation programs, but many CIOs and technology leaders are still figuring this out.</p>
<p class="reader-text-block__paragraph">Whilst there has been a rush of explanations of ChatGPT phenomena and hundreds of proposed applications, from copy-writing to ideation to citizen coding to<span> </span><a href="https://hai.stanford.edu/news/how-large-language-models-will-transform-science-society-and-ai">science</a><span> </span>to<span> </span><a href="https://www.zebrium.com/blog/using-gpt-3-with-zebrium-for-plain-language-incident-root-cause-from-logs">log-processing</a>, very few organizations have considered the strategic use of LLMs.</p>
<p class="reader-text-block__paragraph">Stanford declared LLMs as Foundation Models via a<span> </span><a href="https://arxiv.org/abs/2108.07258">rather long, yet instructive, paper</a>. Due to their incredible, almost incredulous, performance, LLMs have proven to be foundational to many applications using Natural Language Processing (NLP).</p>
<p class="reader-text-block__paragraph">This power, mostly thanks to the size of the models, has taken many by surprise. It is indeed curious that the underlying training scheme of guessing missing words from large corpuses of text should turn out to be so useful in tackling almost every NLP task imaginable.</p>
<p class="reader-text-block__paragraph">What is even more staggering is how this supercharged word-power is easily accessible via just a few lines of Python code thanks to the rise of services like<span> </span><a href="https://huggingface.co/">Huggingface</a><span> </span>that have made these models available beyond research labs.</p>
<p class="reader-text-block__paragraph">Such is the power at the fingertips of many a coder, plenty of pre-LLM AI tools have been made obsolete.<span> </span><em>Here’s a quick tip: for every AI solution touted, check to see if LLMs could replace or improve it. The answer is increasingly yes.</em></p>
<p class="reader-text-block__paragraph">Suppose you want to add keyword labels to call transcripts or contract clauses, in order to<span> </span><a href="https://huggingface.co/docs/transformers/tasks/sequence_classification">classify them</a>, this is literally a few lines of code. As of the time of this writing, Huggingface has<span> </span><a href="https://huggingface.co/models?pipeline_tag=text-classification&amp;sort=downloads">16,757 text classification models in their library.</a><span> </span>Of these, perhaps a hundred are foundational whilst the rest are fine-tuned riffs, like<span> </span><a href="https://huggingface.co/LiYuan/amazon-review-sentiment-analysis">this one, tuned on US Amazon product reviews.</a></p>
<p class="reader-text-block__paragraph">
<h2 class="reader-text-block__heading1">LLMs for Digital Transformation</h2>
<p class="reader-text-block__paragraph">There are many frameworks of digital transformation, but let’s consider their context, namely the range of innovation types per<span> </span><a href="https://en.wikipedia.org/wiki/Geoffrey_Moore">Geoffrey Moore</a>:</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D4D12AQFka1OjZGp65w/article-inline_image-shrink_1500_2232/0/1677541802203?e=1684972800&amp;v=beta&amp;t=H8_vEGmPlp2aLT_Z5qm3YcvrUoxtYPxZrxXWkrX7LnE" loading="lazy" alt="No alt text provided for this image" id="ember2660" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Source: kikku website</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">
<p class="reader-text-block__paragraph"><a href="https://www.scaledagileframework.com/value-streams/">Agile value streams</a><span> </span>will span the entire gamut of innovation types and<span> </span><strong>there isn’t one that won&#8217;t be impacted by LLMs</strong>, or even the core idea: Attention. </p>
<p class="reader-text-block__paragraph">Let’s take a quick detour of Attention (if you want a longer code-level explanation, see<span> </span><a href="https://github.com/pgolding/transformer">my repo on Github</a>). This simple invention (a differentiable memory) gives rise to two enterprise opportunities:</p>
<ol>
<li>Models that can understand the language used across your enterprise</li>
<li>Scalable pattern-finding in many kinds of enterprise data besides language</li>
</ol>
<p class="reader-text-block__paragraph">Attention is powerful at finding patterns in<span> </span><a href="https://en.wikipedia.org/wiki/Set_(mathematics)">set-based data</a>. Indeed, the invention was motivated by mapping one kind of set to another (<a href="https://huggingface.co/tasks/translation">language translation</a>).</p>
<p class="reader-text-block__paragraph">It’s called Attention because it figures out how members of a set “pay attention” to each other. For example, the word<span> </span><strong>He</strong><span> </span>in the sentence “<strong>He ran</strong><span> </span>across the road” pays close attention to the word<span> </span><strong>ran</strong>, but not so much to the word road.</p>
<p class="reader-text-block__paragraph">
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D4D12AQFHL7YhkpTTog/article-inline_image-shrink_1500_2232/0/1677541927976?e=1684972800&amp;v=beta&amp;t=USr4T_Ga-YAXjBxoYYrt8uxIzQsZG6ByBGDSQKhpwDI" loading="lazy" alt="No alt text provided for this image" id="ember2661" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Attention Mapping</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">
<p class="reader-text-block__paragraph">Attention doesn’t just figure out that<span> </span><strong>he</strong><span> </span>attends to<span> </span><strong>ran<span> </span></strong><em>in particular</em>. Nor does it figure out that the first and second word are related, as they will often be. It figures out that “He ran” is a<span> </span><strong>common kind of pattern</strong>, one that us humans would recognize as subject-verb correspondence. </p>
<p class="reader-text-block__paragraph">Given enough examples and enough layers of Attention, the Transformer architecture can find many patterns, even over longer spans, like “<strong>He</strong>, the tall man,<span> </span><strong>ran</strong><span> </span>surprisingly swiftly.” Here it will also associate pronoun (He) with subject (man) or, more generally, subjects with predicates.</p>
<p class="reader-text-block__paragraph">Attention doesn’t know what these patterns are, which is why an LLM doesn’t “know” the rules of grammar per se (and, arguably<span> </span><a href="https://arxiv.org/pdf/2008.05580.pdf">can never know</a>). But grammatical structures, formal and informal, get<span> </span><em>approximately</em><span> </span><em>embedded</em><span> </span>into the model via this pattern-finding process, as does knowledge encoded in the source text.</p>
<p class="reader-text-block__paragraph">Even when approximately embedded, these grammatical rules prove useful as the building blocks for so many NLP tasks built using LLMs.</p>
<p class="reader-text-block__paragraph">But what if your enterprise is full of arcane texts specific to business, like bills of materials or technical product descriptions or specific contract legalese, etc?</p>
<p class="reader-text-block__paragraph">Using a mechanism called<span> </span><a href="https://www.seldon.io/transfer-learning">Transfer Learning</a>, the power of a foundation LLM can be transferred to a custom one fine-tuned for your enterprise. For example, the Fairmont hotel group could fine-tune a model to become the “Fairmont-LLM” that understands “hotel speak”. Silicon Valley Bank could build a “venture debt speak” model, and so on.</p>
<p class="reader-text-block__paragraph">Let’s explore what this looks like in the context of digital transformation.</p>
<h2 class="reader-text-block__heading1">Fairmont LLM: &#8220;Hotel Speak&#8221;</h2>
<p class="reader-text-block__paragraph">Imagine a value-stream in the Fairmont hotel&#8217;s transformation programs that is linked to deeply personalized experiences across the customer journey. Here’s a journey map from GCH Hotel Group as a reference.</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D4D12AQGqoklpC8ScSA/article-inline_image-shrink_1500_2232/0/1677542082966?e=1684972800&amp;v=beta&amp;t=RtzpHeYMHB8PlCdHOJNODbSaOr6o0US6sPpEuQS27kA" loading="lazy" alt="No alt text provided for this image" id="ember2662" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">Source: GCH Hotel Group</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">Source:<span> </span><a href="https://www.gchhotelgroup.com/en/newsroom/blog/customer_journey/3258?rel=/en/newsroom/blog">GCH Hotel Group</a></p>
<p class="reader-text-block__paragraph">What do all of these interfaces have in common? They all use language.</p>
<p class="reader-text-block__paragraph">Given the foundational power of LLMs, it is entirely possible that the “Fairmont-LLM” could power many NLP services within myriad touch-point components:</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex
        
        "><img decoding="async" src="https://media.licdn.com/dms/image/D4D12AQEKHfCvyPlQug/article-inline_image-shrink_1500_2232/0/1677542187033?e=1684972800&amp;v=beta&amp;t=PRDRMJbD8duDH13z_rUJ48VbLL04rLJd5UM8By4l9OM" loading="lazy" alt="No alt text provided for this image" id="ember2663" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div><figcaption class="display-block mt2 full-width text-body-small-open t-sans text-align-center t-black--light">LLM Integrated into Touch-Point Services</figcaption></figure>
</div>
<p class="reader-text-block__paragraph">
<p class="reader-text-block__paragraph">This has obvious implications for digital transformation architecture because a unified custom-LLM potentially impacts many innovation types and related value streams. Failure to notice this will stymy efforts to produce consistent customer-facing experiences. They will cost more and deliver less.</p>
<p class="reader-text-block__paragraph"><a href="http://bit.ly/2023trendsreport">Google’s 2023 AI and Data report</a><span> </span>confirms a trend set by software experts (like<span> </span><a href="https://martinfowler.com/articles/data-mesh-principles.html">Martin Fowler</a>) who have argued for unified<span> </span><a href="https://martinfowler.com/articles/data-mesh-principles.html">Data Mesh</a><span> </span>and<span> </span><a href="https://www.gartner.com/smarterwithgartner/data-fabric-architecture-is-key-to-modernizing-data-management-and-integration">Data Fabric</a><span> </span>architectures that<span> </span><strong>put an end to data silos</strong>. This thinking applies equally to AI, in particular LLMs. </p>
<p class="reader-text-block__paragraph">For example, fine-tuning will require constant updates and related services, such as data-quality to ensure human alignment–i.e. making sure that any generated text is compliant with policies, brand guidelines and ethical standards.</p>
<p class="reader-text-block__paragraph">This will require a<span> </span><strong>composable</strong><span> </span>data architecture and a distributed MLOps operating model, more akin to a mesh. This is a far cry from many of today’s embedded enterprise AI solutions with opaque data models–an anti-pattern for scalable value from LLM models and services.</p>
<h2 class="reader-text-block__heading1">Transformers: LLMs in Disguise</h2>
<p class="reader-text-block__paragraph">There are many ways custom LLMs could/should be deployed in the enterprise as a strategic IT resource versus a point solution. Various architectural and ops patterns are still emerging.</p>
<p class="reader-text-block__paragraph">With different business cases and transformation mindsets to consider, it pays to have awareness of the potentially strategic role of custom LLMs before committing to yet another embedded AI point-solution that lacks flexibility or is heavily tied to a vendor.</p>
<p class="reader-text-block__paragraph">Going a step further, let’s return to our earlier claim that Transformers can pay attention to patterns in all kinds of set-based data. It turns out that many enterprise data sources are sets that contain patterns amenable to discovery via Attention:<span> </span><a href="https://www.sciencedirect.com/science/article/pii/S0957417422004146">sales forecasting</a>,<span> </span><a href="https://dl.acm.org/doi/abs/10.1145/3556702.3556844">actions on a website</a>,<span> </span><a href="https://catalog.ngc.nvidia.com/orgs/nvidia/teams/clara/models/megamolbart">molecules in a drug</a>,<span> </span><a href="https://www.sciencedirect.com/science/article/pii/S2352484722004310">power trends in a wind farm</a><span> </span>etc. (I did research to apply Transformers to no-code website generation because website layouts are inherently set-based.)</p>
<p class="reader-text-block__paragraph">Where it gets really interesting is when different sets get combined into novel multi-modal models capable of solving even larger classes of problems such as how language used in sales might predict revenue.</p>
<p class="reader-text-block__paragraph">Many datasets within the enterprise are ripe for multi-modal modeling, which tends to suggest that Transformer-based modeling might become a horizontal data-science capability, maybe via new low-code tools, not just a feature of an embedded AI product. Certainly, the availability of Huggingface models makes this a realistic proposition.</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Killer Use Case for Generative AI is Empowering Enterprise Citizens</title>
		<link>http://paulgolding.com/2023/02/19/the-killer-use-case-for-generative-ai-is-empowering-enterprise-citizens/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Sun, 19 Feb 2023 19:36:45 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1073</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_4 et_pb_fullwidth_section et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_module et_pb_fullwidth_image et_pb_fullwidth_image_0">
				
				
				
				
				<img decoding="async" loading="lazy" width="800" height="533" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy.png" alt="" title="1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c copy" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy.png 800w, http://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy-480x320.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 800px, 100vw" class="wp-image-923" />
			
			</div>
				
				
			</div><div class="et_pb_section et_pb_section_5 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_8">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_8  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_4  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div>
<div dir="ltr" class="reader-article-content reader-article-content--content-blocks">
<h2 class="reader-text-block__heading1">Digital Democratization: Tech Osmosis</h2>
<p class="reader-text-block__paragraph">Following on from a recent post about<span> </span><a href="https://hbr.org/2022/05/democratizing-transformation">digital democratization</a>, let&#8217;s explore where Generative AI fits into the picture in order to arrive at a more interesting use case for Large Language Models (LLMs) like GPT-3, ChatGPT and<span> </span><a href="https://github.com/features/copilot">Github Co-pilot</a>.</p>
<p class="reader-text-block__paragraph"><a href="https://www.gartner.com/en/information-technology/glossary/citizen-developer"><strong>Citizen Coders</strong></a><strong><span> </span>and Citizen Technologists (henceforth both called Citizens) are vital for the future of digitally complex enterprises.</strong><span> </span>I have argued that the only way to accelerate digital dividends is to amplify the efforts of workers closer to the customer and/or closer to the in-flight business problem.</p>
<p class="reader-text-block__paragraph">Indeed,<span> </span><a href="https://www.gartner.com/en/doc/fusion-teams-democratized-and-distributed-technology-delivery-for-digital">per Gartner</a>, CEOs increasingly hold the belief that the answer to unlocking greater digital dividends is placing more technology into the hands of business operators. They expect their CIOs to deliver.</p>
<p class="reader-text-block__paragraph">The reality of modern software tooling is that it is often easier for a business-domain worker to gain access to automation via new tools (like embedded data science &#8212; e.g. in Tableau &#8212; or no-code super-apps or reverse-ETL) than it is for a highly technical engineer or data scientist to grasp the in-moment need and prioritization of an in-flight business problem. This is thanks to the ongoing march of automation.</p>
<p class="reader-text-block__paragraph">There is a kind of<span> </span><em>digital osmosis</em><span> </span>whereby power flows more easily from technology to the business domain than business knowledge can flow to the technical domain, if it isn&#8217;t lost in translation.</p>
<p class="reader-text-block__paragraph"><em>This is especially true under conditions of uncertainty and complexity</em><span> </span>that are here to stay in the modern enterprise.</p>
<div class="reader-image-block reader-image-block--full-width">
<figure class="reader-image-block__figure">
<div class="ivm-image-view-model   ">
<div class="ivm-view-attr__img-wrapper ivm-view-attr__img-wrapper--use-img-tag display-flex                  "><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQE1RBt3OvMMAw/article-inline_image-shrink_1500_2232/0/1676834950849?e=1684972800&amp;v=beta&amp;t=_0ktlJ69pefv9WN8FvqmRNKmVe6Kij_O0kI1l4ryUww" loading="lazy" alt="No alt text provided for this image" id="ember2325" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></div>
</div>
</figure>
</div>
<p class="reader-text-block__paragraph">
<p class="reader-text-block__paragraph">Let&#8217;s put it another way: if a business worker has to spend more effort explaining and justifying a biz problem to a data scientist, data engineer or salesforce operator than they could theoretically spin up their own solution using self-serve techniques,<span> </span><strong>then you&#8217;re already too slow and likely to suffer the dreaded<span> </span></strong><a href="https://www.jstor.org/stable/20159494"><strong>Red Queen effect</strong></a><span> </span>that business strategists like to remind us of.</p>
<h2 class="reader-text-block__heading1">Generative AI and Citizen Productivity</h2>
<p class="reader-text-block__paragraph">A key automation force will be Generative AI in the form of code generation (e.g.<span> </span><a href="https://github.com/features/copilot">Github Co-pilot</a>) and bot-like interfaces like variants of ChatGPT.</p>
<p class="reader-text-block__paragraph">Indeed, I predict that the sweet spot for these technological possibilities will come when Citizens are able to create their own hybrid-chat interfaces &#8212; think thousands of helper bots, or chat-powered UIs (with embedded<span> </span><a href="https://www.ibm.com/topics/low-code#:~:text=Low%2Dcode%20is%20a%20visual,applications%20through%20minimal%20hand%2Dcoding.">low-code widgets</a>), not one giant &#8220;Siri for the enterprise&#8221;.</p>
<p class="reader-text-block__paragraph">Over time, Generative capabilities will widen to include more tools. For example, in a recent assignment for a major digital experience vendor, I proved that it is possible to generate website pages using the same Transformer technologies that underly LLMs. We know that start-ups like<span> </span><a href="https://www.adept.ai/">Adept</a><span> </span>are working on generalized UI-learning machines in the same manner.</p>
<p class="reader-text-block__paragraph">There will be several outcomes that accelerate digital dividends:</p>
<ol>
<li>Citizens will become 10x more productive at solving in-flight business problems.</li>
<li>$-for-$, Citizen productivity curves will outstrip the productivity curves of in-house technical specialists.</li>
<li>Specialists will migrate to core technical innovation tasks, such as customized AI model production which, in turn, will empower Citizens.</li>
<li>The delivery model of specialists will increasingly be via generalizable Citizen-accessible interfaces.</li>
</ol>
<p class="reader-text-block__paragraph">
<h2 class="reader-text-block__heading1">Accelerated Context Switching</h2>
<p class="reader-text-block__paragraph">A key part of the puzzle is knowledge management. Over time, the ability for Citizens to become successful will depend upon their ability to understand various moving parts (like datasets) quickly. Generative techniques will somewhat lessen the burden of having to understand details, leaving that to models that sit in constant fine-tuning loops.</p>
<p class="reader-text-block__paragraph"><strong>Citizens will recognize that their ability to be successful will depend upon context-switching efficiencies &#8212; how easy it is to switch from one task to another</strong>.</p>
<p class="reader-text-block__paragraph">As I argued elsewhere,<span> </span><a href="https://paulgolding.com/2023/01/09/chatgpt-great-for-unblocking/">the great potential for apps like ChatGPT is a &#8220;mental un-blocker&#8221; apps</a><span> </span>that can quickly bootstrap a worker into a knowledge space. So, these technologies will help Citizens with task-switching.</p>
<p class="reader-text-block__paragraph">Task contexts might include things like a set of parameters or instructions to set up a data flow via<span> </span><a href="https://www.getcensus.com/blog/what-is-reverse-etl">reverse-ETL into Salesforce</a>, or the meaning of a Snowflake model or, the perennial problem of where to find data.</p>
<p class="reader-text-block__paragraph">Indeed, at a certain point, the emerging data semantic layer (e.g. see<span> </span><a href="https://www.getdbt.com/product/semantic-layer/">Semantic Layer for DBT Cloud</a>) will be totally subsumed into LLM-based apps to make data more amenable to semantic search.</p>
<p class="reader-text-block__paragraph">Apps like ChatGPT will not replace data analysts, but will increase the productivity of those attempting analysis who do not spend all their time with the data, like a Citizen Coder.</p>
<p class="reader-text-block__paragraph">Eventually, Generative AI tooling will allow business folks to<span> </span><a href="https://productcoalition.com/the-real-difference-between-metrics-and-insights-ad1d57ce3bda">operationalize the generation of insights</a><span> </span>&#8220;at the speed of thought&#8221; &#8212; an oft-mentioned phrase, but seldom achieved.</p>
<p class="reader-text-block__paragraph">In terms of allowing easier access to data in the first place, the challenge for data-tools vendors will be to make their wares &#8220;LLM-indexable&#8221; &#8212; i.e. easily accessible via natural language querying. (This will be part of the emerging<span> </span><a href="https://www.gartner.com/smarterwithgartner/data-fabric-architecture-is-key-to-modernizing-data-management-and-integration">Data Fabric</a><span> </span>architecture, although many vendors are seemingly behind the ball with this realization.)</p>
<h2 class="reader-text-block__heading1">Gig Citizens</h2>
<p class="reader-text-block__paragraph">In working with Generative AI tools, Citizens will learn to adopt a new mode of documentation, or knowledge curation. We can&#8217;t say what this will look like, as it hasn&#8217;t been invented yet &#8212; we are still adjusting to the way interfaces like ChatGPT work and very few of us have had the opportunity to fine tune them to include our own knowledge.</p>
<p class="reader-text-block__paragraph">I predict that Citizens will learn to write &#8220;notes to self&#8221; that are actually &#8220;notes to self via bot&#8221; &#8212; i.e. document my work in a way that I can pick up the thread again quickly by asking my bot.</p>
<p class="reader-text-block__paragraph">Powerful new techniques in data sharing, as available via tools like Snowflake, and new modes of AI in data permissions and privacy protection will empower Citizens to work from anywhere and become specialists: RevOps Citizens, SalesOps Citizens, Growth-Hacking Citizens etc.</p>
<p class="reader-text-block__paragraph">This will slowly pave the way for what some CIOs are already wondering: what are the opportunities for a &#8220;Gig-Citizen&#8221; economy?</p>
<p class="reader-text-block__paragraph">One has to believe that the opportunities are vast and that new tools will emerge just to power that economy. Generative AI will play a central role.</p>
<h2 class="reader-text-block__heading1">Impacting The Bottom Line</h2>
<p class="reader-text-block__paragraph">Citizen Coding is not a new idea. It is already well underway thanks to the rise of automation and expressive tools like no- and low-code apps makers.</p>
<p class="reader-text-block__paragraph">The so-called Citizen Technologist, or Business Technologist, is just an extension of the Citizen Coder concept: someone outside of IT who knows how to drive technology to get results. A good example would be someone who knows how to set up a workflow using<span> </span><a href="https://zapier.com/">Zapier</a>.</p>
<p class="reader-text-block__paragraph">How will this impact the bottom line?</p>
<p class="reader-text-block__paragraph">Any Citizen ought to be working towards a set of KPIs that roll-up via a<span> </span><a href="https://www.petra-wille.com/blog/kpi-trees-how-to-bridge-the-gap-between-customer-behavior-product-metrics-and-company-goals">KPI hierarchy</a><span> </span>to business outcomes directly related to revenue.</p>
<p class="reader-text-block__paragraph">An insight is any data-driven realization that is actionable against improving a KPI. So, if workers are empowered to generative insights for themselves and make workflow adjustments for themselves, this ought to have a direct positive effect upon related KPIs with knock-on improvement to the bottom line.</p>
<p class="reader-text-block__paragraph">This is just one way in which AI, in this case Generative AI, will help to accelerate dividends from digital technology investment.</p>
<hr class="reader-divider-block" />
<p class="reader-text-block__paragraph">Article written with inputs from<span> </span><a href="https://www.linkedin.com/in/irzanagolding">Irzana Golding</a><span> </span>about data democratization and insights operationalization via tools like DBT Cloud.</p>
</div>
</div>
<div class="reader-ugc-post-bar reader-ugc-post-bar--expanded"></div></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Role of Data Scaling in AI Strategy</title>
		<link>http://paulgolding.com/2023/02/15/the-role-of-data-scaling-in-ai-strategy/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Wed, 15 Feb 2023 19:32:42 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=1069</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_6 et_pb_fullwidth_section et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_module et_pb_fullwidth_image et_pb_fullwidth_image_1">
				
				
				
				
				<img decoding="async" loading="lazy" width="1000" height="667" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/oberon_artwork_1970_digital_transformation_big_data_AI_drawing__d0e05558-d369-4d37-8330-e84875bfee9e-copy.png" alt="" title="oberon_artwork_1970_digital_transformation_big_data_AI_drawing__d0e05558-d369-4d37-8330-e84875bfee9e copy" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/01/oberon_artwork_1970_digital_transformation_big_data_AI_drawing__d0e05558-d369-4d37-8330-e84875bfee9e-copy.png 1000w, http://paulgolding.com/wp-content/uploads/sites/2/2023/01/oberon_artwork_1970_digital_transformation_big_data_AI_drawing__d0e05558-d369-4d37-8330-e84875bfee9e-copy-980x654.png 980w, http://paulgolding.com/wp-content/uploads/sites/2/2023/01/oberon_artwork_1970_digital_transformation_big_data_AI_drawing__d0e05558-d369-4d37-8330-e84875bfee9e-copy-480x320.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) and (max-width: 980px) 980px, (min-width: 981px) 1000px, 100vw" class="wp-image-943" />
			
			</div>
				
				
			</div><div class="et_pb_section et_pb_section_7 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_9">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_9  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_5  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><h2 class="reader-text-block__heading1">Introduction</h2>
<p class="reader-text-block__paragraph">What passes for AI strategy these days has a very low bar that tells us little to nothing about the unique potential of AI.</p>
<p class="reader-text-block__paragraph">We see this often from folks pivoting to become &#8220;AI experts&#8221; with little insights into the unique capabilities of AI.</p>
<p class="reader-text-block__paragraph">My first AI design (<a href="https://patents.google.com/?inventor=paul+golding&amp;assignee=motorola&amp;oq=paul+golding+motorola">patented &#8217;96</a>) was used in a commercial product (for improving capacity of Motorola cellular networks). I have used AI ever since (see<span> </span><a href="https://paulgolding.com/projects/">example projects</a>), so I have genuine insights into its unique<span> </span><em>commercial possibilities</em>, many of which have only manifested in recent years.</p>
<p class="reader-text-block__paragraph">What many posit as an AI strategy is often &#8220;Digital Transformation&#8221; with a bit of AI thrown in. This might be appropriate for tactical deployments, but insufficient for strategy.</p>
<p class="reader-text-block__paragraph">Very few leaders have asked:<span> </span><em>How do I transform my business<span> </span></em><strong><em>via the unique capabilities of AI?</em></strong><span> </span>Or,<span> </span><em>how is corporate strategy fundamentally impacted by AI?</em></p>
<p class="reader-text-block__paragraph">Indeed, there is a precursive question:<span> </span><em>How does the availability of &#8220;unlimited computation&#8221; impact corporate strategy?</em><span> </span>In many ways, AI provides the framework for answering that question.</p>
<p class="reader-text-block__paragraph">Very few leaders are asking these questions because they don&#8217;t know how to interpret them within the context of corporate strategy. To assist with interpretation, I am currently documenting the unique principles of AI and how to deploy them as strategic tools.</p>
<p class="reader-text-block__paragraph">I plan to share soon &#8212; to be honest, this is a kind of teaser post (and to explore if posting on LI is a useful medium).</p>
<p class="reader-text-block__paragraph">As a precursor, let me share one example.</p>
<h2 class="reader-text-block__heading1">Recipe: AI Scaling &amp; Innovation Networks</h2>
<p class="reader-text-block__paragraph">A recent discovery of AI is the so-called<span> </span><a href="https://arxiv.org/abs/2001.08361">Scaling Law</a>. It says that performance scales (as a power law) with the size of the dataset and AI model. Model complexity isn&#8217;t so important. Indeed, most of the &#8220;beyond human&#8221; capabilities of AI have come about from scaling.</p>
<p class="reader-text-block__paragraph">With this in mind, a business might ask:<span> </span><strong><em>How do we scale data as a strategic advantage?</em></strong></p>
<p class="reader-text-block__paragraph">Remarkably, many orgs do not ask this question.</p>
<p class="reader-text-block__paragraph">They lack strategies specifically related to the accumulation of &#8220;data capital&#8221;. Data is still seen as a by-product of current processes versus an investment into future gains.</p>
<p class="reader-text-block__paragraph">But there are some great examples from folks who really get it.</p>
<p class="reader-text-block__paragraph">Recently, I encountered an alliance of farmers who understood that certain types of automation are inevitable in order to overcome common challenges, such as<span> </span><a href="https://www.fb.org/news-release/labor-challenges-increase-farm-economy-pressures">labor pressures</a>.</p>
<p class="reader-text-block__paragraph">They<span> </span><strong>strategically</strong><span> </span>identified the need for scalable data and so collaborated in the curation of a pooled repository of high-quality agricultural data.</p>
<p class="reader-text-block__paragraph">Operationally, the alliance are organizing this effort via an &#8220;Innovation Network&#8221; in which resources and information are pooled and managed via a separate body.</p>
<p class="reader-text-block__paragraph"><img decoding="async" src="https://media.licdn.com/dms/image/D5612AQHsBRsvlEvnRQ/article-inline_image-shrink_1500_2232/0/1676491275175?e=1684972800&amp;v=beta&amp;t=AaObWWWb4vvH8fwU8OQ_QZM4g8gQb3DRH-BRyhGaOdQ" loading="lazy" alt="No alt text provided for this image" id="ember2071" class="ivm-view-attr__img--centered reader-image-block__img lazy-image ember-view" /></p>
<p class="reader-text-block__paragraph">The network taps into technology partners and key expert individuals who form what might be called<span> </span><a href="https://hbr.org/2004/04/a-network-of-invention">&#8220;A Network of Invention&#8221;</a>. This is a smart move because it leverages organic &#8220;network effects&#8221; from knowledge without the need to develop lots of in-house AI expertise, keeping in mind that such expertise is often hard to come by.</p>
<p class="reader-text-block__paragraph">The alliance also works with an incubator to co-opt start-ups into the network. The data pool is planned to be a major input into the incubator program.</p>
<p class="reader-text-block__paragraph">I have been involved with setting up several such networks, including for O2, the UK&#8217;s premier telco, via a loose collaboration of internal innovation teams (&#8220;labs&#8221;), strategic partners, corporate customers, incubators and informal &#8220;inventor&#8221; networks.</p>
<p class="reader-text-block__paragraph">The farmer&#8217;s alliance doesn&#8217;t yet know for which applications the data will prove most valuable. But they realize that without the means to scale data they will have little chance of exploiting scaling laws. There is a prior expectation of utility from the data given the existence of such laws and related patterns, like<span> </span><a href="https://ai.googleblog.com/2022/11/characterizing-emergent-phenomena-in.html">Emergence</a>.</p>
<p class="reader-text-block__paragraph">In the alliance&#8217;s case, a key expectation from the data is reduced cycle time from AI experiments to in-field benefits, which is critical in agriculture where innovation cycle times are often overly long.</p>
<h2 class="reader-text-block__heading1">Summary</h2>
<p class="reader-text-block__paragraph">AI strategy must be formulated along the axes of AI&#8217;s unique attributes, such as Scaling Laws. This law alone gives rise to various strategic considerations, such as the accumulation of &#8220;data capital&#8221;.</p>
<p class="reader-text-block__paragraph">Noting that scaling is hard, a possible strategic approach is the recipe:</p>
<blockquote class="reader-text-block__quote"><p>Pooled data capital + innovation network</p></blockquote>
<p class="reader-text-block__paragraph">But there are many other recipes for accumulating &#8220;data capital&#8221; for strategic competitive advantage. More to come in later posts 🙂</p>
<p class="reader-text-block__paragraph"></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>ChatGPT &#8212; An Unblocking Tool</title>
		<link>http://paulgolding.com/2023/01/09/chatgpt-great-for-unblocking/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Tue, 10 Jan 2023 07:13:40 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=907</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_8 et_pb_fullwidth_section et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_module et_pb_fullwidth_image et_pb_fullwidth_image_2">
				
				
				
				
				<img decoding="async" loading="lazy" width="800" height="533" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy.png" alt="" title="1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c copy" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy.png 800w, http://paulgolding.com/wp-content/uploads/sites/2/2023/01/1000_artwork_1970_business_man_on_computer_asking_robot_to_ge_7ee904a4-fd4c-469f-96b3-00524aa3e00c-copy-480x320.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 800px, 100vw" class="wp-image-923" />
			
			</div>
				
				
			</div><div class="et_pb_section et_pb_section_9 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_10">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_10  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_6  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><h1>Summary</h1>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400"></span></p>
<p><strong>TL;DR: ChatGPT killer use case is mental unblocking.</strong></p>
<p>I participated in a number of discussions with &#8220;knowledge workers&#8221; about their experiences using <a href="https://openai.com/blog/chatgpt/">ChatGPT</a> in the workplace via the OpenAI UI.</p>
<p><span style="font-weight: 400">Many non-technical users are finding it sufficiently useful as a writing aid that they imagine continuing its use, especially as it gets better. Use cases varied, from outlining video scripts to helping with product descriptions to generating job descriptions to general re-writes of documents. Many found it useful as an ideation tool due to its content diversity, having been trained on a truly super-massive set of examples. </span></p>
<p><span style="font-weight: 400">The primary use case for ChatGPT is accelerating writing tasks where it’s easier to hack a prompt than think of what to write &#8212; or do &#8212; from scratch. <strong>It is a useful mental unblocking tool</strong>. The knack is learning how to write productive prompts as a kind of &#8220;programmatic writing&#8221; vs. &#8220;chatting&#8221;.</span></p>
<p><span style="font-weight: 400">All those I spoke with encountered issues, such as hallucination. But this did not seem to present enough of a barrier to prevent usage. </span></p>
<p><span style="font-weight: 400">Those competent at writing often did not rate the outputs as good enough to wholesale replace entire writing tasks, such as writing finished blog posts. It often has a tendency to produce unappealing unoriginal prose unless manipulated via an excess of prompt hacks. However, some claimed that such outputs still provided a better-than-nothing starting point.</span></p>
<h1><span style="font-weight: 400">Introduction</span></h1>
<p><span style="font-weight: 400">There&#8217;s much hype about ChatGPT, so I asked various folks what they have found it good at. Mostly I spoke with &#8220;knowledge workers”, including quite a few in creative professions (e.g. digital agencies). I did not speak to any who struggled to write to begin with.</span></p>
<h1><span style="font-weight: 400">What is ChatGPT?</span></h1>
<p><span style="font-weight: 400">For those unfamiliar, ChatGPT is a specially trained version of another AI model built by OpenAI, called <a href="https://openai.com/blog/gpt-3-apps/">GPT-3</a>, which is a class of model called a Transformer. If you are a coder and want to know more, I provided a <a href="https://github.com/pgolding/transformer/blob/main/Transformers%20Demystified.ipynb">detailed code-level annotation on Github</a>. For the rest of us, a Transformer is a giant computer program that spends a mind-blowing amount of time trying to guess how to complete sentences that it finds online.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">This guessing game is a brute force attempt to learn language structure. It turns out that if done long enough over a big enough corpus, the program gets stupendously good at guessing which words go with which over very long spans of text, or long enough to subsequently generate quite productive prose. This is known as self-supervised learning because the program doesn&#8217;t need to be told (supervised) how to complete the task because the words it is trying to guess were already there: it simply masked them out in order to guess them.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">What has surprised seemingly everybody, including AI experts and linguists, is how good this brute force method is when carried out a super-massive scale. And <a href="https://arxiv.org/abs/2001.08361">scale is seemingly what matters most</a>, not any particular architectural sophistication.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">The scale of the model allows it to learn a tremendous number of structural patterns over large spans of text. Once it has finished guessing, the trained model can be used to carry out other language-related tasks. </span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">One such fine-tuned task is language generation. As you can imagine, a program that knows so much about which words go with which can complete entire sentences just by “guessing” the next most probable word, one at a time, over and over. This isn&#8217;t just a testament to the model&#8217;s scale, but a feature of language, namely that longer sequences are composed from smaller ones, which are composed from smallers ones (and so on), sometimes called <a href="https://en.wikipedia.org/wiki/Recursion">recursion</a>.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">Language has this amazing creative aspect: take any starting word and there will be dozens, sometimes hundreds, or more, of potential next words. Over the span of a sentence, this gives rise to a mind-boggling number of novel combinations of words. Indeed, perhaps we don&#8217;t realize that many of the sentences we utter in our lifetime are entirely novel in the history of language. This creative aspect of language is considered by experts to be one of the great mysteries of the mind that is very <a href="https://philpapers.org/archive/ASOLAS.pdf">hard to explain with a scientifically tractable theory</a>. This is perhaps why it has surprised so many that a brute-force computer program has been able to detect enough structure in its training data to allow it to complete various language tasks with human or beyond-human performance.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">The truly surprising aspect of pre-trained transformers is that not only can they generate a passage of text, but they can interpret an initial seeding sequence, called a prompt, such that it can be used as a kind of instruction as to how to generate the text. This makes the AI a kind of &#8220;programmatic writing&#8221; machine. (We shall return to this perspective.)</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">This seems like magic and has led many users to feel an illusory sense that the model &#8220;understands&#8221; what&#8217;s being requested. Here&#8217;s an example:</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400"><img decoding="async" loading="lazy" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-12.03.29-PM.png" width="831" height="190" alt="" class="wp-image-914 alignnone size-full" srcset="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-12.03.29-PM.png 831w, https://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-12.03.29-PM-480x110.png 480w" sizes="(min-width: 0px) and (max-width: 480px) 480px, (min-width: 481px) 831px, 100vw" /></span></p>
<p><span style="font-weight: 400">What’s happening here? It looks as if the program has understood the sentence and looked up the answer, like it might a database of questions. But that is not what’s happening. The program is generating the green text as a highly probable follow-on text to the prompt &#8212; highly probably in the sense of all the vast swathes of text it has seen whilst training and which of those in which combinations might plausibly follow from the prompt. Think of it this way: if we were to look at all the text on the web and find that prompt, or something similar, then most likely we would see a list of sales methods afterwards.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">Let me be clear: <em>it is generating that green text</em>. It is not trying to look it up from the training corpus. It doesn’t know that this is the correct answer &#8212; there is no “correctness” or “truth” inherent in the program. Indeed, it does not understand any of the words in the prompt either. It is blindly performing a function: given the input sequence, what is a highly probable output sequence. The output is only plausible, and potentially truthful, because the words it saw (in relation to the prompt) during training are plausible and contain truths, facts (or not) and useful explanations and descriptions.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">And this is where scale comes into the picture. If the program sees enough examples of text, as in a truly super-massive corpus, then it will have seen enough examples of this question, or what it represents, such that its answers are so probable as to overlap with common experience or useful information as available on the web and in various books.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">Representation is key, as this is what the model learns. For example, the model will learn the pattern “What…[plural nouns]&#8230;?” has a high probability of being followed by a numerical list. We can tell this by repeating the pattern and swapping out the meaning:</span></p>
<p><span style="font-weight: 400"><img decoding="async" loading="lazy" src="https://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-9.49.03-PM.png" width="371" height="173" alt="" class="wp-image-919 alignnone size-full" srcset="http://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-9.49.03-PM.png 371w, http://paulgolding.com/wp-content/uploads/sites/2/2023/01/Screen-Shot-2023-01-09-at-9.49.03-PM-300x140.png 300w" sizes="(max-width: 371px) 100vw, 371px" /></span></p>
<h1><span style="font-weight: 400">What is it good for? Unblocking.</span></h1>
<p><span style="font-weight: 400">Now we know what it is and how it works, what is it good for? Here I participated in a number of discussions, in-person and via forums, with various knowledge workers who had tried to use the tool in earnest to do useful work. I only asked about uses of the raw OpenAI interface, not any specialist application like the AI Copywriter tools built using GPT. </span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400"><em>Everyone I spoke to found it useful for something</em>. This is not surprising given its super-massive scale. It is hard to stump it, even, apparently with very esoteric questions of the kind that philosopher and quantum physicist David Deutsch <a href="https://twitter.com/DavidDeutschOxf/status/1611996984918810624?s=20&amp;t=bBqwYZ8yuEDyouC6ogt_xA">has been posting on his twitter account</a>.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">Many folks were surprised by how good the tool was in getting a response that proved useful. A senior content manager at a high-end digital agency told me how he had experimented with generating script outlines for corporate videos. Having recently done some consulting related to online retail in the Metaverse, I tried something similar with a prompt: <em>Write me a corporate video script outline for a person whose shopping experience is enhanced by shopping in the metaverse</em>. The output followed a coherent structure with a meaningful narrative. For someone unrehearsed in script writing, it provided a useful starting point. For professional script writers, it probably isn&#8217;t so useful, except, perhaps, for ideation of scenes.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">A consultant friend of mine tried to write blog posts about innovation, his area of expertise. He found the outputs coherent as a blog post, but a bit flat, lacking any originality. Indeed, it looked trite. This is to be expected if the tool is basically trying to find the most probable text.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">But, like all probabilities, they can be constrained, or made conditional. This is the art of refining the prompt. In that same conversation, another innovation consultant reworded the prompt to generate something more convincing. We all concluded that it was still not good enough to warrant publication (in the name of the original consultant) but some of the participants believed that it was far enough along to get there via editing, even using ChatGPT to suggest new edits on a paragraph or sentence basis.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">Despite some believing that this prompt-manipulation will become unnecessary with new versions of the tool, this seems an unlikely prospect. This is because language is, no matter how it is generated, ambiguous and able to convey many different voices, tones and ideas with even subtle rewording.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">It was interesting to watch a live session of a Gen-Z user “hack” the prompts back and forth to generate a LinkedIn bio for a friend. It was interesting to watch this younger user take to the tool like it was how writing had always been done. Of those involved in the process, we all agreed that the final version was better worded than the author’s original bio. This hacking on a theme to get a final result seems like it could be skill, but one that is more powerful in the hands of someone with a good command of language to begin with, but nonetheless unaware of what a good LinkedIn bio might look like until they saw it, versus thinking of it entirely from scratch.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400"><strong>And this seemed to be the killer use case: hacking on a prompt to converge upon productive content without the pain of filling the void of blank paper &#8212; a kind of mental unblocker, as it were</strong>. Indeed, a professional writer friend of mine told me how some of his fellow fiction writers were using it often to “fill the gaps” with hard to find narrative turns or adjectives or scene ideas.</span></p>
<h1><span style="font-weight: 400">Using ChatGPT: Programming words with words</span></h1>
<p><span style="font-weight: 400">I am not going to offer a long list of tips and instructions, but rather an insight into how to use ChatGPT productively. And it&#8217;s an insight that all of the correspondees discovered for themselves, also widely discussed on the Web &#8212; e.g. see <a href="https://oneusefulthing.substack.com/p/how-to-use-chatgpt-to-boost-your?utm_source=profile&amp;utm_medium=reader2">this article by Wharton Professor Ethan Mollick</a>.</span></p>
<p><span style="font-weight: 400">The trick is to understand what I explained above about how ChatGPT works, as in it is able to consult a truly super-massive corpus to understand how to construct plausible, or highly probable, sequences. This means it is capable of detail and nuance, but only if you instruct it with specific detail and nuance to begin with. </span></p>
<p><span style="font-weight: 400">Users learn that ChatGPT is a kind of &#8220;programming words using words&#8221; versus a mind-reading &#8220;Oracle&#8221; &#8212; and yes, some folks use it in that latter way and get disappointing results. Part of the programming can be the inclusion of lots of specific detail, even as bullet points, without attempting to construct a prompt that another human might find meaningful, per se.</span></p>
<p><span style="font-weight: 400">Consider a generic prompt like: </span><em>Write me an AI strategy</em>. This will likely produce the kind of generic or trite response expected, like identify a problem, find the data, select a model, train it and use it etc. But something more elaborate might produce more interesting responses, like: <em>Why should a CEO of a mature company that is struggling in the marketplace care about the manifold hypothesis?</em></p>
<p><span>You&#8217;d have to know that the manifold hypothesis is related to AI, but this is a more nuanced prompt that generates a more nuanced response related to the nature of high-dimensionality data.</span></p>
<p><span>The best way to discover this &#8220;programming words with words&#8221; is to experiment. Take you baseline prompt and mix-in a lot of details, perhaps quirky and tangential ones, just to understand how the machine works.</span></p>
<h1><span style="font-weight: 400">Issues</span></h1>
<p><span style="font-weight: 400"></span>Unsurprisingly, given that we know how Transformers work (brute force probability machines), we expect some of the outputs not to meet the mark. To recap, the model does not attempt to &#8220;make sense&#8221; of the prompt, but rather looks at the sequence and, based upon patterns found it it, at many levels (morphologically, syntactically, semantically) computes the most probable follow-on sequence based upon a super-massive scale of prior examples (of the morphological representations, not necessarily the exact wording of the prompt).</p>
<p>The model can easily produce incoherent output, which has been documented by many. They can also hallucinate facts or fail to generate prompts that are well reasoned based upon the prompt.</p>
<p>However, I didn&#8217;t find anyone who was attempting to use it as a general knowledge bot, like a Siri. Indeed, it was well understood that this wasn&#8217;t a wise use case, nor anything that would involve having to subsequently check all factual claims in the generated text.</p>
<p>Also, that friend of mine who tried to write an innovation post, also used another AI model to detect if the text had been generated. It detected with high confidence that the text had been generated. Whilst it&#8217;s not clear how these detection scripts work, the potential for having content detected as fake could be worrying, especially if it gets dinged by Google Search.</p>
<p>Nonetheless, as said, most of the users I spoke to were not attempting to generate wholesale content for publication online, but these various writer&#8217;s companion tasks to accelerate production of content, knowledge, ideas and format etc.</p>
<h1>Where Next</h1>
<p><span style="font-weight: 400">Of course, pre-trained Transformers can do lots of other things, like summarize text or answer questions from a text. They can be used to analyze different document types, like invoices or business contracts, or even generate them. Perhaps they can find clauses in contracts that are likely to lead to poor value or lack of compliance, and so on. When it comes to the technical use of the underlying technology, suddenly many business processes become amenable to automation. However, this wasn’t the subject of my inquiry,  but the use of ChatGPT as a writer’s companion. For that, it appears to have sufficient value that it might be expected to enhance knowledge-worker productivity over time.</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">And no &#8212; none of this article was written using ChatGPT (except the indicated examples).</span></p>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">If you&#8217;d like to know more about how to use GPT-like technology in your company to uniquely add value, <a href="https://paulgolding.com/#contact_paul">feel free to reach out.</a></span><span style="font-size: 16px"> </span></p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Transformers (Code-level Commentary)</title>
		<link>http://paulgolding.com/2022/11/18/851/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Fri, 18 Nov 2022 16:17:02 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<category><![CDATA[random]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=851</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_10 et_animated et_pb_with_background et_pb_fullwidth_section et_section_regular section_has_divider et_pb_bottom_divider" >
				
				
				
				
				
				
				<section class="et_pb_module et_pb_fullwidth_header et_pb_fullwidth_header_0 et_pb_text_align_center et_pb_bg_layout_dark">
				
				
				
				
				<div class="et_pb_fullwidth_header_container center">
					<div class="header-content-container center">
					<div class="header-content">
						
						<h1 class="et_pb_module_header">Transformers</h1>
						
						<div class="et_pb_header_content_wrapper"></div>
						
					</div>
				</div>
					
				</div>
				<div class="et_pb_fullwidth_header_overlay"></div>
				<div class="et_pb_fullwidth_header_scroll"></div>
			</section>
				
				<div class="et_pb_bottom_inside_divider et-no-transition"></div>
			</div><div class="et_pb_section et_pb_section_11 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_11">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_11  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_7 et_animated  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p>I seldom, if ever, write technical posts. There are just too many excellent resources out there and my goal isn&#8217;t to explain technical things to a technical audience, as much as I enjoy teaching from time to time, or giving the odd guest lecture.</p>
<p>But I made an exception whilst perusing the literature regarding Transformers, as in Large Language Models. This was motivated by research into the potential use of Transformers for a novel non-language use case. I wanted to experiment with a few ideas, so I decided to build a base of various notebooks and code libs from which to set out on my research path.</p>
<p>Secondarily, I wanted to ground myself in sufficient code-level mechanics to explore a set of questions I have about what, exactly, are Transformers doing in relation to semantics. This was for the purposes of attempting to understand how world models might get incorporated into LLM schemas, beginning with Transformers. Of course, this isn&#8217;t a novel question, but I need my own research bed.</p>
<p>There is also the interesting question as to why it is that depth (of Transformers) is a key indicator for modeling language performance (against the various benchmarks out there). I felt that by playing with the transformations, layer by layer, I might find insights and/or confirm various similar explorations already in the literature.</p>
<p>A question that also piqued my interest is to ask what would the opposite of a Large Language Model (LLM) look like, as in a &#8220;Small&#8221;, yet nonetheless still performant, LM. This was from the point of view of not only attempting to understand what a more parsimonious system might look like, but from the point of view of understanding the scope of independent research that might be achievable without Google or OpenAI resources ($$$$$).</p>
<p>Anyway, these are not the subjects of this blog post. Rather, this post is merely <a href="https://github.com/pgolding/transformer">a link to a notebook</a> that I published that is a more fine-grained explanation of an existing code-level explanation of an Encoder-Decoder Transformer (i.e. BERT-like).</p>
<p>Per the notebook intro:</p>
<blockquote>
<p><span>I wanted to add some missing details to standard transformer annotations often found in courses and texts. But not wanting to reinvent the wheel with yet another explanation, the accompanying notebooks are mostly an expansion of </span><a href="https://d2l.ai/chapter_attention-mechanisms-and-transformers/index.html" rel="nofollow">Chapter 11 of the Dive Into Deep Learning</a><span> open source book &#8212; a valuable resource for learners. (We are all learners.)</span></p>
</blockquote>
<p>&nbsp;</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Don&#8217;t Ignore AI in a Downturn</title>
		<link>http://paulgolding.com/2022/11/01/859/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Tue, 01 Nov 2022 13:52:56 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<category><![CDATA[Uncategorized]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=859</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_12 et_animated et_pb_with_background et_pb_fullwidth_section et_section_regular section_has_divider et_pb_bottom_divider" >
				
				
				
				
				
				
				<section class="et_pb_module et_pb_fullwidth_header et_pb_fullwidth_header_1 et_pb_text_align_center et_pb_bg_layout_dark">
				
				
				
				
				<div class="et_pb_fullwidth_header_container center">
					<div class="header-content-container center">
					<div class="header-content">
						
						<h1 class="et_pb_module_header">Don't Ignore AI in a Downturn</h1>
						
						<div class="et_pb_header_content_wrapper"></div>
						
					</div>
				</div>
					
				</div>
				<div class="et_pb_fullwidth_header_overlay"></div>
				<div class="et_pb_fullwidth_header_scroll"></div>
			</section>
				
				<div class="et_pb_bottom_inside_divider et-no-transition"></div>
			</div><div class="et_pb_section et_pb_section_13 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_12">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_12  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_8 et_animated  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p>I recently wrote a guest blogpost (<a href="https://stephenshapiro.com/5-ai-misconceptions/">&#8220;Five Misconceptions about AI That Are Slowing Down Your Business&#8221;</a>) on the wonderful Stephen Shapiro&#8217;s Innovation Insights website. It was motivated by two experiences. The first was an indirect encounter with a CEO who declared that AI is a kind of luxury, as if to say that different technologies fall into different &#8220;accounting budgets&#8221;, not unlike the mental accounting claim made by behavioral economists, if their data is to be believed.</p>
<p>The second was an impression via a consulting engagement wherein the CEO of large corporation seemed to feel that AI had some special status, also assigned to &#8220;nice to have&#8221; buckets of activity.</p>
<p>There is a third observation, which is the tendency of many CxOs to believe that AI is somehow exotic that it&#8217;s completely outside the reach of direct innovation. This only leaves indirect innovation, namely the ticking of AI boxes via the use of vendors with AI-supercharged products. This can tick the box: &#8220;We are investing in AI.&#8221;</p>
<p>I will leave you to read the post, if so inclined, but the thrust is that AI is a technology to be taken more seriously during a downturn than many other technologies or transformation initiatives. Hopefully the article explains why.</p>
<p>&nbsp;</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Manifold and Mantra Hypotheses: Why AI works and Why it Doesn&#8217;t.</title>
		<link>http://paulgolding.com/2022/09/09/ai-the-manifold-and-mantra-hypotheses/</link>
		
		<dc:creator><![CDATA[paulgolding]]></dc:creator>
		<pubDate>Fri, 09 Sep 2022 20:12:48 +0000</pubDate>
				<category><![CDATA[AI]]></category>
		<category><![CDATA[cognitive]]></category>
		<category><![CDATA[data]]></category>
		<category><![CDATA[Innovation]]></category>
		<guid isPermaLink="false">https://paulgolding.com/?p=799</guid>

					<description><![CDATA[]]></description>
										<content:encoded><![CDATA[<p><div class="et_pb_section et_pb_section_14 et_animated et_pb_with_background et_pb_fullwidth_section et_section_regular section_has_divider et_pb_bottom_divider" >
				
				
				
				
				
				
				<section class="et_pb_module et_pb_fullwidth_header et_pb_fullwidth_header_2 et_pb_text_align_center et_pb_bg_layout_dark">
				
				
				
				
				<div class="et_pb_fullwidth_header_container center">
					<div class="header-content-container center">
					<div class="header-content">
						
						<h1 class="et_pb_module_header">The Manifold &amp; Mantra Hypotheses: Why AI works, and why it doesn't.</h1>
						
						<div class="et_pb_header_content_wrapper"><p>Business leaders: pay attention.</p></div>
						
					</div>
				</div>
					
				</div>
				<div class="et_pb_fullwidth_header_overlay"></div>
				<div class="et_pb_fullwidth_header_scroll"></div>
			</section>
				
				<div class="et_pb_bottom_inside_divider et-no-transition"></div>
			</div><div class="et_pb_section et_pb_section_15 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_13">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_13  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_9 et_animated  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><h3 style="text-align: center"><span style="font-weight: 400"></span></h3>
<p><span style="font-weight: 400"></span></p>
<p><span style="font-weight: 400">AI works because of a fact of nature that is sometimes expressed by the Manifold Hypothesis &#8212; that complex things can be explained via simpler mechanisms (that an AI might find from samples of the complex thing). However, AI often fails because of what I call the Mantra Hypothesis, which is the tendency of business leaders to declare, mantra style, that the business will “Put AI First” or some such moniker, but without the means to translate that intent into operational realities that might lead to a meaningful ROI.</span></p>
<p><span style="font-weight: 400">In this post, I will outline <em>for a business audience</em> why many AI projects in enterprises either fail, or are likely to fail. Reasons for failure can be summarized as:</span></p>
<ol>
<li><em><span style="font-weight: 400">Blind adoption of AI without a clearly defined set of outcomes that can be explained beyond the level of pithy directives like “We need to adopt AI”. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Assignment of AI adoption to folks who lack experience in mapping advanced tech into tangible business outcomes. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Avoidance of vanity success measures in adoption of AI. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Avoiding the trap of hoping that hiring AI technical experts will translate into business outcomes via some kind of osmosis. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Failure to understand the data requirements to make AI training viable. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Underestimation of the operationalization of data and AI required to translate AI experiments (often by scientists) into repeatable and robust products. </span></em>
<p><em></em></p>
</li>
<li><em><span style="font-weight: 400">Lack of a “Data as a Product” approach to AI, including the tendency to want to “supercharge” existing product features vs. finding new AI-first interpretations.</span></em></li>
</ol>
<h1><span style="font-weight: 400">AI: The Hype is Real, Yet Insufficient</span></h1>
<p><span style="font-weight: 400">Right now, AI-hype is effervescent. The truth is that no-one knows where we are in the </span><a href="https://www.gartner.com/en/research/methodologies/gartner-hype-cycle"><span style="font-weight: 400">hype cycle</span></a><span style="font-weight: 400"> &#8212; with many voices calling from the trough of disillusionment whilst others holler from the peak of Mount Olympus where they gather with sentient-like AI gods. The naysayers aren’t necessarily theorists like </span><a href="https://nautil.us/deep-learning-is-hitting-a-wall-14467/"><span style="font-weight: 400">Gary Marcus who claims that Deep Learning has &#8220;hit a wall&#8221;</span></a><span style="font-weight: 400"> due to its lack of compositionality &#8212; a blurry technical term that could mean many things (whole conference about it </span><a href="https://compositionalintelligence.github.io/"><span style="font-weight: 400">here</span></a><span style="font-weight: 400">). Put simply, </span><a href="https://compositionalintelligence.github.io/pdfs/Marcus.pdf"><span style="font-weight: 400">Marcus is saying</span></a><span style="font-weight: 400"> that any AI that builds a language model merely by predicting which words come next, or thereabouts, cannot retrieve deeper semantic structures (“meaning”) in the data &#8212; a capability that he argues is required to make AI usefully “intelligent” in most real-world applications.</span></p>
<p><span style="font-weight: 400">The meaning of “intelligent” is hotly debated, often by those who have apparently never read Turing’s original paper in which he suggested that the concept of thinking machines was an “absurdity”. He meant that any agreeable definition of “thinking”, if we might construct one, is commonly understood to be a biological capability, nothing to do with machines. </span></p>
<p><span style="font-weight: 400">However, for this discussion, I prefer to lean upon the moniker </span><i><span style="font-weight: 400">“usefully intelligent”</span></i><span style="font-weight: 400"> wherein the word “usefully” is far more interesting and ought to be related to business outcomes, not arcane debates about intelligence. And, in this regard, there are all kinds of brick walls that business leaders will face when attempting to adopt AI.</span></p>
<p><span style="font-weight: 400">Other voices are calling from within the trough of disillusionment, namely business leaders who don’t yet see any clear relationship between “adopting AI” (whatever that means) and tangible ROI. And I don’t mean leaders who picked up the airport best-seller on AI and tasked their CIOs to come up with an “AI strategy”. I mean leaders who already placed their bets, although probably not the whole farm, via real hard cash spent investing in AI, acquiring or aqui-hiring AI companies, yet nonetheless failing to translate such investments into meaningful outcomes.</span></p>
<p><span style="font-weight: 400">This often stems from a kind of corporate arithmetic: </span></p>
<p><em><strong>Business + AI = Better Business. </strong></em></p>
<p><span style="font-weight: 400">It apparently works for any business: </span></p>
<p><strong><i>Air-Conditioning Vendor + AI = Better Air-Conditioning Vendor.</i></strong></p>
<p><span style="font-weight: 400">And so on. Belief in this formula rapidly translates into some kind of mantra for the company: “Put AI First”. However, whilst this formula has the potential to be true, it is not an operational formula &#8212; merely believing in the hype of AI &#8212; the formula &#8212; is insufficient. It tells us nothing about how the addition of AI could make a better business. Yet, it is surprising how many business leaders have adopted this formula without asking probing questions. Moreover, many make the fatal mistake of going about acquiring technical AI expertise as the first step as if the presence of that expertise will permeate into the business and bring about the realization of the formula. Far from it.</span></p>
<p><span style="font-weight: 400">This happened repeatedly with the adoption of “Big Data” and “Data Science” only to find mediocre ROI, if not a negative. It is not uncommon for an enterprise to acqui-hire an AI start-up that is basically a team of research scientists, much closer to academia than industry, with scant knowledge of applied innovation and little to no experience of translating their various AI Python libraries into meaningful modifications of the current and future enterprise roadmap.</span></p>
<p><span style="font-weight: 400">To dig further, we might consider certain useful generalities as to why the adoption of shiny new tech can easily fail. Any business initiative that proclaims to “Put AI First”, or some such declaration, is possibly doomed from the start. The lack of a meaningful scope and outcomes is the elephant in the room, condemning the program to inevitably proceed along predictable lines. One such line will be suffering a long lag between flipping the switch to “Put AI First” and eventually concluding, perhaps wrongly, that it maybe wasn’t worth it.</span></p>
<p><span style="font-weight: 400">Without a practically useful interpretation of AI into provable customer benefits and business outcomes, AI can easily become just another fad, like when many a CEO bought into “Re-engineering” the company only to find that no one had ever really understood in tangible terms what “engineering” a company meant, never mind “Re-engineering”.</span></p>
<p><span style="font-weight: 400">This blind approach can be radically worsened when organizations offer “sexy AI” roles to loyalists, often efficient program managers with a track record in other areas, along with their hired-in friends who desperately want the shiny-new-toy badge on their resumes. Yet they lack the depth of analytical expertise and product synthesis to translate shiny tech into outcomes &#8212; a task far harder than many imagine. Desperate to show progress, they create “AI departments” and the like, just like they created “Data Science” departments, and then set about constructing various vanity measures of success to present in monthly reviews. This is a common anti-pattern.</span></p>
<p><span style="font-weight: 400">The antidote is to ensure that your product teams and key stakeholders throughout the business can articulate clear outcomes and benefits of AI adoption in their roadmaps, present and future. Indeed, I have advocated what I call the </span><a href="https://paulgolding.com/2022/04/17/100-of-10-rule-for-ai/"><span style="font-weight: 400">100-of-10 rule</span></a><span style="font-weight: 400">. It states, at least as a starting point, that your entire leadership team and their direct reports ought to be able to clearly articulate how the entire 10% of effort within the 70-20-10 rubric could be devoted to AI-related programs. Given the massive disruption potential of AI-first start-ups and the emergence of large-scale models, the potential impact of AI is worthy of such attention and prioritization. But this must be interpreted into meaningful outcomes with detailed arguments before letting rip on the AI Adoption cord.</span></p>
<p><span style="font-weight: 400">Many readers might want to stop here and address the above issues first. However, determined readers might read on as we explore the biggest AI elephant in the room, namely the “data”, or “lack of data” challenge.</span></p>
<p>&nbsp;</p>
<h1><span style="font-weight: 400">The Real Brick Wall: Data</span></h1>
<h2><span style="font-weight: 400"></span></h2>
<h2><span style="font-weight: 400">What is the data for?</span></h2>
<p><span style="font-weight: 400">There is often a hidden field in the AI adoption equation:</span></p>
<p><strong><em>Business <span style="text-decoration: underline">(+ Data)</span> + AI = Better Business.</em></strong></p>
<p><span style="font-weight: 400">Many business leaders have at least heard that AI needs data. And, so, with a warm feeling that the enterprise is awash with data, AI-adoption confidence goes up. However, this is often a mistake.</span></p>
<p><span style="font-weight: 400">Let’s begin with why AI works at all. If this basic plank isn’t well understood, then everything else rapidly becomes froth. By now, we all know that AI works by digesting data. The “works” bit can be plainly stated: a working (trained) AI can take new, previously unseen, data and reliably make some prediction that coheres with a ground truth. In other words, what it learned from specific data samples can be generalized to other samples. If you trained your AI to understand how to spot sales opportunities based upon a history of sales data, then it must </span><a href="https://developers.google.com/machine-learning/crash-course/generalization/video-lecture#:~:text=Generalization%20refers%20to%20your%20model's,Develop%20intuition%20about%20overfitting."><span style="font-weight: 400">generalize</span></a><span style="font-weight: 400"> to new sales opportunities &#8212; and it ought to generalize in reliable and </span><i><span style="font-weight: 400">actionable</span></i><span style="font-weight: 400"> ways.</span></p>
<p><span style="font-weight: 400">AI can do this thanks to a strange mathematical reality, one that is sometimes expressed by theoreticians as the </span><a href="http://colah.github.io/posts/2014-03-NN-Manifolds-Topology/"><span style="font-weight: 400">Manifold Hypothesis</span></a><span style="font-weight: 400">. For those really interested, you can look up the math, but I will explain it in much simpler terms: things in the natural world tend to obey parsimonious laws. These laws are compact whilst the things they can explain are vast. Imagine, for a moment, that we study thousands of objects falling from the sky until we can reliably predict the force with which they hit the ground &#8212; i.e. we take all our measurements, perhaps via a camera, and feed them into an AI. </span></p>
<p><span style="font-weight: 400">Now let’s say we want to predict the next hundred objects, which could be of any material and size (i.e. mass). It would be a useless system if we could only predict a particular object because we have already seen an identical one. Indeed, such a system has no predictive power at all &#8212; it is merely looking things up from history. It would also not be so useful if we could only predict something when we see an object so similar that our predictive power is too weak to be useful. However, what if our system had somehow figured out Newtonian Mechanics: F=ma. This single “piece of information” (a formula) could be used to determine the behavior of an infinite number of objects. In effect, we have “compressed” the original data space from lots of examples to a single parameter (or maybe three: F, M and A).</span></p>
<p><span style="font-weight: 400">The problem is that most of these kinds of symbolic understandings of the world have been discovered by humans via mysterious creative processes that we don’t know how to program, nor explain. Hence we rely upon a different approach, which is to see if an AI can take lots of examples of a behavior and condense them down into a function that is good enough to generalize to lots of unseen examples. Our hope, which we impose via design constraints within the AI architecture, is that the parameters of that function are far fewer than the size of the input space (in totality) such that we can use a reasonably sized set of examples to estimate the parameters well enough to generalize to a class of objects without having to observe every single possible example.</span></p>
<p><span style="font-weight: 400">To put it simply, most things in the world could, in theory, be described by some underlying set of equations, if only we knew them. This is how nature is. Although there are trillions of unique leaves on all the trees, they all arise from a broadly similar set of underlying mechanisms that could be condensed into a relatively small set of information (e.g. equations) that is far, far, far smaller than trillions of unique data points. If this were not the case, no AI could learn anything because when we say that an AI is learning, we mean that it is attempting to search for that far smaller approximating function that lies beneath the data. If there were no such smaller underlying function, or principle, then the AI could not easily find it with the clues it has (some training data) and within the confines of a reasonable amount of computational effort.</span></p>
<p><span style="font-weight: 400">What do I mean by finding? Well, that is what an AI does. It “searches” for a reduced set of data, called parameters, that not only explains the training data, but could characterize the underlying system from where the training data comes from &#8212; e.g. how leaves grow, or how objects fall, or how cancer presents in a biopsy sample image.</span></p>
<p><span style="font-weight: 400">This principle applies to digitized systems despite the fact that digitization can give rise to mind-blowingly large datasets. For example, take a tiny image sensor that is only 16 x 16 pixels, and can only encode each pixel as black or white. There are 2^256 possible images, which is 1.15 x 10^7, known by numberphiles as 115 </span><a href="https://integers.info/large-numbers/quattuorvigintillion"><i><span style="font-weight: 400">quattuorvigintillion</span></i></a><span style="font-weight: 400">. Imagine an AI had to process each combination to characterize images. Say it could process 1 Billion images per second. That would still take far longer than the age of the universe to complete.</span></p>
<p><span style="font-weight: 400">This seems absurd, but those numbers are correct, even though our intuitions can’t relate to that much data from such a tiny image space. But, here’s the really cool thing &#8212; most of those images are unrecognizable &#8212; they are just noise. Any pixel in one of those noisy pixels doesn’t have any meaningful relationship with any other in terms of forming a recognizable structure, and so, each &#8212; and all &#8212; do not “say” much.</span></p>
<p><span style="font-weight: 400">The actual number of useful images is ridiculously smaller. Take, say, recognition of hand-written digits. It only requires about a </span><a href="https://www.kaggle.com/c/digit-recognizer/data"><span style="font-weight: 400">few tens of thousands of images</span></a><span style="font-weight: 400"> of various styles to train an AI to recognize more or less any human-recognizable digit that it hasn’t seen before, no matter how varying the hand-writing style. This is because alphabets have stable morphology &#8211; i.e. the underlying rules to make a number 1 are always a vertical line with a top-left diagonal stroke no matter how a particular person writes it. What we want then is for an AI to see enough images to figure out an underlying function that ends up enacting a rule something like: pixels arranged in vertical lines with a diagonal top-left downward stroke predict a digit number 1.</span></p>
<p>&nbsp;</p>
<h2><span style="font-weight: 400">What kind of data?</span></h2>
<p><span style="font-weight: 400">So this is how AI works &#8212; it exploits the natural tendency for things in the world, even when digitized, to be generated by a set of underlying parameterized principles that could be explained with relatively little data compared to all the things those principles could generate. It turns out that the so-called </span><a href="https://www.ibm.com/cloud/learn/neural-networks#:~:text=Neural%20networks%2C%20also%20known%20as,neurons%20signal%20to%20one%20another."><span style="font-weight: 400">Neural Network</span></a><span style="font-weight: 400"> is a machine that can find an approximation of those underlying parameters sufficient enough to be usefully generalizable in recognizing classes of data of which it has only seen relatively few samples during training.</span></p>
<p><span style="font-weight: 400">So far, so good. But why does this also present a problem? The problem is still having enough data, of the right kind, to allow an AI to converge upon a useful approximating function. In many business situations, this data is really hard to come by, despite the common notion that we are all awash with “Big” data.</span></p>
<p><span style="font-weight: 400">The “right kind” is perhaps more critical. For an AI to learn, it has to be able to observe clues about the function it’s trying to approximate. How does this happen? An easy example should suffice. Consider the following sequence:</span></p>
<p><strong>1 → 1</strong></p>
<p><strong>2 → 4</strong></p>
<p><strong>3 → 9</strong></p>
<p><strong>4 → 16</strong></p>
<p><strong>5 → X</strong></p>
<p><span style="font-weight: 400">What is X? </span></p>
<p><span style="font-weight: 400">Of course, you will think 25, which is correct. </span></p>
<p><span style="font-weight: 400">You have figured out that the mapping function here is the squared operator. But if I gave you only the sequence: 1, 2, 3, 4 or </span><i><span style="font-weight: 400">only</span></i><span style="font-weight: 400"> the sequence 4, 16, 9, 1, would you have guessed that these numbers came from a squared-number generator? No. You need the inputs </span><i><span style="font-weight: 400">and outputs</span></i><span style="font-weight: 400"> to guide you &#8212; or supervise &#8212; when guessing the function. These outputs are called labels. Together with their relevant inputs they are the ground truth of our system. The approach is called </span><a href="https://www.simplilearn.com/tutorials/machine-learning-tutorial/supervised-and-unsupervised-learning#what_is_supervised_learning"><span style="font-weight: 400">Supervised Learning</span></a><span style="font-weight: 400">.</span></p>
<p><span style="font-weight: 400">There are many challenges to finding this kind of data in a way that it tangibly supports the intended business goals. For example, a CMO might think that the existence of millions of sent emails ought to present a rich-enough dataset to </span><a href="https://blogs.oracle.com/marketingcloud/post/write-better-email-subject-lines-with-machine-learning-and-oracle-responsys"><span style="font-weight: 400">deploy AI in the selection of subject lines that improve email open rates</span></a><span style="font-weight: 400">. However, it often turns out not to be so easy. In this case, there might not be enough variance in the data to be useful &#8212; i.e. many of the data points are very similar and so they don’t offer enough clues to the AI as to how to find the underlying “generator” function.</span></p>
<p><span style="font-weight: 400">Another common problem is finding the ground truth. Lots of datasets in the enterprise might be large, but they lack the supervisory labels to be useful as training data. This might just be because the necessary observations were never recorded. For example, there are plenty of e-commerce datasets wherein the products lack sufficiently rich meta-data, such as keyword tags, to provide sufficient dimensionality in the input data.</span></p>
<p><span style="font-weight: 400">It is important to recognize that the labels must represent the ground truth you are interested in. For example, if the true goal of the CMO is to map email behaviors to churn, then you need not just the email interaction data, but the churn data. Of course, in a typical enterprise, churn data might well be available, but you get the idea. Take medical applications as another example, such as the use of AI to analyze tissue samples (via slides). One ground truth might be annotations from a pathologist as to the presence of disease. However, what if the intended goal of the AI is to predict prognosis or probability of a certain treatment’s effectiveness for the patient in question. This requires a different set of ground-truth data, namely historical records of treatment outcomes.</span></p>
<p><span style="font-weight: 400">There’s also the nature of the data itself. AI works because the underlying function-approximation assumes that the training samples are highly representative of all the possible samples there could be. This is a certain statistical assumption that does not hold in some cases, such as the so-called “Long Tail” cases where infrequent samples that might not appear in the training data, can have overwhelming effects on the outcome.</span></p>
<p><span style="font-weight: 400">Of course, there might be solutions to these problems. For example, in the case of finding useful email headings, data outside of the enterprise can be used to train an AI on aspects of language that evoke a certain emotional response. This kind of data might be found in places like online reviews, or public datasets. Once trained upon this data, the AI can then be tuned to solve the specific AI problem, as seen here in the </span><a href="https://blogs.oracle.com/marketingcloud/post/write-better-email-subject-lines-with-machine-learning-and-oracle-responsys"><span style="font-weight: 400">Oracle email platform.</span></a><span style="font-weight: 400"> Generally, this technique is called Transfer Learning.</span></p>
<p><span style="font-weight: 400">However, the specific solution to data readiness, if a solution exists, is not as important as being able to recognize that the reality of the data part of the AI formula can be very hard to achieve. We will discuss more practical solutions below.</span></p>
<p>&nbsp;</p>
<h2><span style="font-weight: 400">Data in Practice: Operationalization</span></h2>
<p><span style="font-weight: 400">I could go on and on about the various data requirements needed to make AI effective. Above I have only mentioned a few of those that are somewhat related to the theory of how AI works. </span></p>
<p><span style="font-weight: 400">There is the much more practical concern of finding the data, preparing the data, labeling the data (if required) and so on. Both experimentally (when training the AI) and operationally (when deploying in a product), these data processing issues can present a significant amount of the overall effort in bringing an AI-enhanced solution to the market. This operationalization can include the very real challenge of updating the AI with sufficient regularity to adapt to new data samples or better approximations. The gap between the experimental AI program and an operationalized one can be large. This has given rise to a whole new field called Machine Learning Operations (MLOps). </span></p>
<p><span style="font-weight: 400">An additional key limitation might be gaining access to data that belongs to the customer. There are all kinds of privacy issues and related audit and compliance concerns to verify what data went into which AI model in order to prove data lineage etc. This might also be important in achieving a certain quality metric from the AI, like avoiding certain kinds of negatively-impacting decisions, such as denying someone a loan because of skin color.</span></p>
<p><span style="font-weight: 400">This is related to the wider concern of “AI safety” or </span><a href="https://openai.com/alignment/"><span style="font-weight: 400">“AI alignment”</span></a><span style="font-weight: 400"> and data governance. Put simply, if your data governance is already lacking pre-AI-adoption, as it often is in enterprises with large technical debt, then you might be in for a rough ride.</span></p>
<h2><span style="font-weight: 400">Data as a Product and AI-First Mindset</span></h2>
<p><span style="font-weight: 400">It is almost shocking to see how often the adoption of data-related disciplines, like Data Science and AI, seem to throw out everything we have learned  in the modern software era about how to translate commercial intentions into commercial realities, as if the lessons of Agile and the like were never learned. Indeed, this might well be the case, but let’s be more charitable in our estimations of baseline product operations.</span></p>
<p><span style="font-weight: 400">There is a tendency in data-related projects to plow straight into complex costly programs with the assumption they are going to work. If we have learned anything from modern software, it is that this assumption is typically incorrect, as the product won’t quite work in the way intended or expected. If those programs are largely staffed by AI researchers and technicians, then it’s often the case that very few of them understand how to manage technical product programs.</span></p>
<p><span style="font-weight: 400">It is not inconceivable that an AI team will go about making some kind of tool, say a method to enhance sales, only to find that the sales team doesn’t use it. No stakeholders were consulted, no user tests were done, no demos were conducted of the product in progress, no tests were conceived to assure quality of its outputs. And, often worse of all, no experiments were done to verify scope etc.</span></p>
<p><span style="font-weight: 400">It is all too common for rogue data to upset the entire apple cart, again due to lack of data governance. Sales folks changed the way they rated sales prospects via some undocumented hack in the CRM and suddenly the reality of garbage-in, garbage-out is dramatically amplified by an AI that is ingesting inverted or incorrect data.</span></p>
<p><span style="font-weight: 400">There is no exception for data not to be treated as a product. All of what we have learned about incremental development, testing, stakeholder involvement, demoing, design-thinking, and so on, still applies, even if you’re making a dashboard.</span></p>
<p><span style="font-weight: 400">But there is another product perspective that is often overlooked, namely the absence of critical interpretation of novel AI-first capabilities. There is a tendency for AI to be used as a means to supercharge existing product capabilities. For example, an e-commerce product recommender might be enhanced using AI, as might a website personalization module. This is the “low-hanging fruit” approach. There might not be anything wrong with such an approach, but it should be undertaken with caution.</span></p>
<p><span style="font-weight: 400">With any such amplification, one first has to ask how far it really translates into ROI. As you might have already understood, the most precious resource in AI adoption is the AI expertise that ultimately generates the core of the innovation, notwithstanding all of the above constraints in translating that innovation into meaningful ROI. In many sectors, you will be competing, if not now, then soon, with so-called AI-first start-ups who plan to tackle your sector using a radically different approach. For example, in my own work for an e-commerce platform giant, I proposed an AI-first approach called “No Design” web authoring &#8212; the conversion of English sentences into finished web pages. The point is that this is uniquely possible because of recent AI advances (called Transformers) and potentially far more impactful within a 1-3 year ROI horizon than tinkering with, say, personalization or search.</span></p>
<p><span style="font-weight: 400">Put simply, the “No Design” approach has the potential to radically alter how retailers think about their website strategies, opening up new ideas due to “superhuman” scaling of creative effort etc.</span></p>
<h1><span style="font-weight: 400">Summary</span></h1>
<p><span style="font-weight: 400">In going about AI adoption, you should believe the hype. AI really does have transformational potential, perhaps radical in some cases. However, as with all advanced tech, the mere adoption of it does not guarantee success. Enterprises tend to be very amnesic, so it is not unexpected that all the lessons learned from modern product development are suddenly forgotten in the face of a new paradigm like AI.</span></p>
<p><span style="font-weight: 400">Naive approaches should be avoided. These include assigning AI as “just another program” to company loyalists who have shown proficiency with running programs. As an antidote to the mere circulation of mantras and hype, business leaders should insist upon well argued AI enhancements to the roadmap articulated by stakeholders with the backing of experiments where possible.</span></p>
<p><span style="font-weight: 400">The greatest caution should be applied to any assumptions about data. We have explained the magic of AI (the Manifold Hypothesis) in searching for approximate functions that explain the training data, hopefully in generalizable ways, but this magic needs enough data of the right kind to be effective. It is too easy to mistake existing meta-data as a useful basis for the actual ground truth required for the magic to work in the way intended as a useful outcome.</span></p>
<p><span style="font-weight: 400">The operationalization of AI is not to be underestimated, potentially overwhelming the scientific work to the point of rendering it ineffective in practice. Careful attention must be paid to effective data governance and AI quality assurance.</span></p>
<p><span style="font-weight: 400">The product mindset must prevail, which could be interpreted as mapping the benefits of Agile, Lean, and so on, to all data programs. The ease with which AI researchers might write Python scripts can easily create a massive bias about how easy it is to build AI products. But these are not products. They are prototypes, or often not far from demos.</span></p>
<p><span style="font-weight: 400">The innovation mindset must also prevail, thinking anew about product strategy in light of novel AI capabilities. This is the so-called “AI First” approach often overlooked by enterprises due to an eagerness to attain some validation of AI investments via “low hanging fruit” product enhancements. It is also overlooked when using existing program experts to manage AI, thinking as they might along familiar product lines.</span></p>
<p>&nbsp;</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div></p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
