<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[What Brain]]></title><description><![CDATA[Technology]]></description><link>https://www.whatbrain.com</link><image><url>https://substackcdn.com/image/fetch/$s_!DXuf!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1bd3e8d2-45c3-4512-bf63-70b22dad2b1c_1024x1024.png</url><title>What Brain</title><link>https://www.whatbrain.com</link></image><generator>Substack</generator><lastBuildDate>Wed, 08 Apr 2026 20:09:23 GMT</lastBuildDate><atom:link href="https://www.whatbrain.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Aaron Diestelkamp]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[whatbrain@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[whatbrain@substack.com]]></itunes:email><itunes:name><![CDATA[Aaron Diestelkamp]]></itunes:name></itunes:owner><itunes:author><![CDATA[Aaron Diestelkamp]]></itunes:author><googleplay:owner><![CDATA[whatbrain@substack.com]]></googleplay:owner><googleplay:email><![CDATA[whatbrain@substack.com]]></googleplay:email><googleplay:author><![CDATA[Aaron Diestelkamp]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Organic Design for Command and Control - for Agents]]></title><description><![CDATA[John Boyd, Ralph Wiggum, OODA Loops, and Leadership]]></description><link>https://www.whatbrain.com/p/organic-design-for-command-and-control</link><guid isPermaLink="false">https://www.whatbrain.com/p/organic-design-for-command-and-control</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Tue, 03 Feb 2026 17:01:37 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!jAwW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Like many software devs, I was inspired by the recent hype around Agents to rethink how I approach software engineering with agents. A couple of articles that particularly resonated with me were Geoffrey Huntley's "Everything is a Ralph Loop" (<a href="https://ghuntley.com/loop/">https://ghuntley.com/loop/</a>) and Steve Yegge's "Welcome to GasTown" (<a href="https://steve-yegge.medium.com/welcome-to-gas-town-4f25ee16dd04">https://steve-yegge.medium.com/welcome-to-gas-town-4f25ee16dd04</a>). Both serve as excellent gateways to agentic coding by shaking up and challenging traditional thinking. If you get nothing else out of this article, at the very least go follow them and read what they have to say if you're in software engineering.<br><br>Anyway, once you start thinking in loops, like Huntley says in his article, you really start to notice them everywhere! A few years ago, I went down a similar rabbit hole after stumbling across a podcast about Col. John Boyd and what he called &#8220;OODA Loops.&#8221;</p><p><em><strong>A little about Col. John Boyd:</strong></em> Boyd was one of the most influential military strategists of the late 20th century (and served as one of the inspirations for Maverick in Top Gun). His work is largely considered a modern counterpart to Sun Tzu&#8217;s Art of War and still inspires military strategy around the world. The military stuff isn&#8217;t important here. <em><strong>What matters is that Boyd talked a lot about loops.</strong></em></p><p><strong>The OODA Loop</strong></p><p>OODA Loop stands for: Observation-Orientation-Decision-Action. Boyd argued that when doing anything, everyone cycles through these stages. We make an observation, orient ourselves based on it, decide what to do, take action, and then start over in an endless loop. Success in conflict usually comes down to having a faster OODA loop than your opponent. Plenty of influencer types have written extensively about how this concept applies to all sorts of things from sports to business to relationships etc, but it&#8217;s not a difficult leap to see how it applies to AI agents.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jAwW!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jAwW!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 424w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 848w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 1272w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jAwW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png" width="1456" height="803" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:803,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:586524,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://www.whatbrain.com/i/186698929?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jAwW!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 424w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 848w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 1272w, https://substackcdn.com/image/fetch/$s_!jAwW!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8d99a0f5-0668-4725-84e2-bdc0423bb22b_2490x1374.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p><strong>Observation:</strong> The agent&#8217;s context window, what the agent knows</p><p><strong>Orientation:</strong> What direction is the agent pointed? What&#8217;s the overall goal? What are we trying to accomplish? What did we do previously? What should we do next?</p><p><strong>Decision:</strong> What should we do to achieve our next goal? What tools do we have? Which should we use?</p><p><strong>Action:</strong> Call a tool / Do a thing / Write a response (send an email, run a bash command, buy a crypto coin, post to moltbook, etc.)</p><p><em>Rinse and repeat.</em></p><p>This framing is useful because most problems with agentic loops trace back to a failure in one of these specific types of steps.</p><p><strong>Examples:</strong> Coding agents often fail at the <em>Observation </em>step. After taking an action, the agent lacks access to the output it needs to correctly observe the effect of what it just did. It&#8217;s missing access to logs, databases, or browsers, or compaction causes important details to get dropped from context. If it can&#8217;t see these things then the agent loses the ability to orient itself in the environment and the loop fails.</p><p>A failure at the <em>Orientation </em>step might look like the agent going off on a tangent, wasting tokens on things irrelevant to your project or compaction causing it to lose track of its tasks. The agent loses alignment and goes on to make bad decisions because of that.</p><p>A failure at the <em>Decision </em>step may look like: choosing a suboptimal approaches, or doing things in the wrong order that cause the loop to break. A failure in <em>Action</em> might be calling the wrong tool or hallucinating a tool that doesn&#8217;t exist. etc etc</p><p>Regardless of which step causes the failure, <strong>you often know a failure has occurred when the agent falls back to requiring human intervention or gets stuck</strong>. It can&#8217;t observe properly so it asks the human to check a log or test a feature or the agent gets lost hill-climbing looking for the information it needs only to get lost. Or maybe a tool is missing so it asks the human to run the steps manually. etc etc.</p><p>Once you see agentic loops this way, it become obvious that the primary path to improving an agent&#8217;s speed within an OODA loop is removing the one bottle neck that slows everything down the most. That bottle neck is the human in the loop. The bottle neck is you. Geoffrey Huntley says in many of his videos that the goal is to &#8220;sit on top of the loop,&#8221; not &#8220;in the loop.&#8221; Getting the human out of the loop means doing the engineering work to improve the agent&#8217;s ability to cycle through observation-orientation-decision-action without human intervention.</p><p>But hang with me. That&#8217;s not the main thing I wanted to talk about.</p><p><strong>Principles of Organic Command and Control for Agents</strong></p><p>Once you realize that dependency on humans is the bottleneck slowing down agentic OODA loops the most, you gradually conclude that you need a better way to control these agents without getting in their way. I confess that when I first looked at Yegge&#8217;s article on Gas Town, it seemed crazy and needlessly complex. The natural question was &#8220;why would you even need this?&#8221; But I think Boyd can help us understand why orchestrators and autonomous loops are built the way they are. Its a remaking of the command and control structures we use for a different era of software development. We need to think differently.</p><p>In addition to the OODA Loop, Boyd gave a briefing on Organic Designs for Command and Control (https://www.coljohnboyd.com/#pdf-organic-design-command-control-pdf) Boyd&#8217;s whole presentation is relevant to the discussion, but I&#8217;ll highlight one part for now:</p><p>Boyd theorized that effective command and control means giving units agency to take actions. He didn&#8217;t even like the term &#8220;Command and Control.&#8221; &#8220;Command&#8221; implies rigidity and restriction. &#8220;Control&#8221; to him implied friction. According to Boyd, friction and restriction slow down OODA loops, inhibit creativity and flexibility, and lead to failures and an inability to respond fluidly to the chaos of real world situations.</p><p>Instead of command and control, Boyd suggested thinking in terms of &#8220;leadership&#8221; and &#8220;appreciation.&#8221; To Boyd, true leadership wasn&#8217;t about issuing commands that had to be rigidly followed. It was about inspiring shared intention and alignment. Instead of getting bogged down making every decision (micromanaging slows the OODA loop), leadership is about making sure every actor has &#8220;orientation&#8221; toward the same goal. &#8220;Appreciation&#8221;, he argued, is what makes this style of leadership possible. </p><p>This isn&#8217;t a warm, fuzzy &#8220;thank you, Claude.&#8221; It&#8217;s about having observability into what your army of agents is marching toward, having your finger on the pulse, so you can appreciate what&#8217;s really happening.  Instead of dictating from an ivory tower with no windows, true leaders use this appreciation to gauge alignment and orientation and course-correct if actors start to drift. Boyd described this sort of alignment as &#8220;implicit guidance and control&#8221; through shared orientation. Shared alignment allowed great military commanders throughout history to lead their armies to victory by granting their subordinates agency to adapt on the battle field without being tied to rigid, and slow, command structures.</p><p><strong>The key insight is this</strong>: we can fall into the trap of trying to control our AI Agents too tightly. This messes up our agents&#8217; OODA loops by hamstringing their flexibility and forcing them to be dependent on humans. Not only is this slow, It ironically<strong> </strong>robs agents of&#8230; agency.</p><p>If we apply Boyd&#8217;s theory, the goal should be to align agents on high-level goals, guidelines, and principles while giving them flexibility within their own OODA loops to act decisively and creatively. Create implicit guidance and control through proper orientation, treating each agent as a cell of intelligence with agency to act based on shared orientation. With this philosophy, it is often better to be less prescriptive with tasks because we want to give agents head room to do their work. Room to be smart and adapt as needed. (As a quick aside, this also means that we should be <em>very</em> selective about what we let into our agents context window and system prompts because these impact orientation. Boyd talks about an idea here called &#8220;non-cooperative centers of gravity&#8221;. I&#8217;ll maybe talk about that in a different post - but Practically I think frivolous soul.md files, tons of skills, and most &#8220;memories&#8221; are usually detrimental to agent performance because they contain competing orientations and distract from aligned orientation to the specific tasks at hand. Additionally it also underscores some of the reason why we should leave our models context window to play with instead of filling it up with a bunch of &#8220;helpful&#8221; commands and requirements.<em> It turns out agentic coding is fundamentally an alignment problem.</em>)</p><p>Regardless, If granting agents this much freedom makes you uncomfortable, we should work towards a solution instead of abandoning autonomous agentic coding as a concept. Add observability. Work with your agents to define better specs. Work with your agents to write better guidelines that work well with the quirks of your individual models, loops, and agents. Create better monitoring so you can appreciate what your agents are actually doing. Tune everything until you have a shared alignment that stays on track through loop iterations.</p><p>In the agentic coding world our job is to provide leadership to our agents, reduce friction, and appreciate what is happening as it happens. After that we should do the best thing we can do: get out of the way.</p><p>Check out John Boyd&#8217;s original slides from 1987: https://www.coljohnboyd.com/#pdf-organic-design-command-control-pdf</p>]]></content:encoded></item><item><title><![CDATA[Ever After Potato Problem]]></title><description><![CDATA[Potatoes, Ever After, Confabulation, and the AI Mandela Effect]]></description><link>https://www.whatbrain.com/p/llm-powered-searches-are-irresponsible</link><guid isPermaLink="false">https://www.whatbrain.com/p/llm-powered-searches-are-irresponsible</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Thu, 14 Nov 2024 16:29:07 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Working with AI is fun because it is weird, and this week I stumbled across something weird. Bear with me, I promise this relates to AI...<br><br>On my Facebook account, I was discussing the etymology of the word "potato" with my friends (as one does) and how the origins of the word demonstrate that potatoes are a New World food. Like corn and tomatoes, potatoes originate from the Americas (as an aside, it is mind-blowing to me that Europe had guns before they had potatoes)... In one of the comment threads, some of my friends were commenting about anachronisms in movies that depict food, and one of them brought up the movie "<a href="https://www.imdb.com/title/tt0120631/">Ever After</a>," set in the 1500s, and how it depicts the characters eating bars of chocolate even though chocolate wouldn't have been in solid form like that until the 1800s. Wanting to bring the discussion back to potatoes, I thought it would be "fun" to see if "Ever After" also had potatoes in it. <br><br>Not remembering anything specifically about potatoes in the movie, I quickly asked ChatGPT and got the following response from <strong>GPT-4o</strong>:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!eLGs!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!eLGs!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 424w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 848w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 1272w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!eLGs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png" width="1008" height="2244" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:2244,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:240155,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!eLGs!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 424w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 848w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 1272w, https://substackcdn.com/image/fetch/$s_!eLGs!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0fe049e7-771c-419b-b6d4-4b70f3aadd18_1008x2244.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Okay. I don&#8217;t remember that scene&#8230; But it seems pretty confident that Danielle talks about potatoes in this movie. Let&#8217;s ask <strong>o1-preview</strong> in a completely new chat&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!Kq5g!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!Kq5g!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 424w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 848w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 1272w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!Kq5g!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png" width="1008" height="1645" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1645,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:218000,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!Kq5g!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 424w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 848w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 1272w, https://substackcdn.com/image/fetch/$s_!Kq5g!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff620592-b983-4920-a9a2-7c602b39a497_1008x1645.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><p>Okay. Uh&#8230; Let&#8217;s ask <strong>ChatGPT 4o Search</strong>&#8230;</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!cGOg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!cGOg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 424w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 848w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!cGOg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png" width="1008" height="1204" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1204,&quot;width&quot;:1008,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:152325,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!cGOg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 424w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 848w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 1272w, https://substackcdn.com/image/fetch/$s_!cGOg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff01221b5-4d60-48ec-94f2-a3c59f910963_1008x1204.png 1456w" sizes="100vw"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><strong><br>Claude Sonnet 3.5&#8230; Perplexity&#8230;</strong></p><p></p><div class="image-gallery-embed" data-attrs="{&quot;gallery&quot;:{&quot;images&quot;:[{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/3498550d-995c-459e-a348-1dadda6e5382_1008x2244.png&quot;},{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f533eaab-d445-4009-9fa2-073d643d6043_1008x2244.png&quot;}],&quot;caption&quot;:&quot;&quot;,&quot;alt&quot;:&quot;&quot;,&quot;staticGalleryImage&quot;:{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/4e4399b4-d686-4aea-ab19-4e03994195f7_1456x720.png&quot;}},&quot;isEditorNode&quot;:true}"></div><p><strong><br><br>Gemini&#8230; Meta AI&#8230; Microsoft CoPilot&#8230;</strong></p><div class="image-gallery-embed" data-attrs="{&quot;gallery&quot;:{&quot;images&quot;:[{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d00e6afa-1b07-4e27-a281-f674313aa444_2053x943.png&quot;},{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8dc61b66-607d-4e51-920f-25cd94b922ce_1399x472.png&quot;},{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/de26926a-2897-4e3f-8d23-a935d8c7234c_1224x736.png&quot;}],&quot;caption&quot;:&quot;&quot;,&quot;alt&quot;:&quot;&quot;,&quot;staticGalleryImage&quot;:{&quot;type&quot;:&quot;image/png&quot;,&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/76c7ddaa-fe03-40eb-9f64-5da682a47990_1456x474.png&quot;}},&quot;isEditorNode&quot;:true}"></div><p><br><br>ALL of the LLMs agree that this potato scene exists. Even when I push the LLMs about the specifics and ask follow-up questions, most of them stay the course and confidently assert that this is FOR SURE a real scene; in fact, they almost universally insist that this potato scene is a very well-known and popular scene in the movie. On the surface, it seems plausible, and we seem to have consensus, so it must be true, right? The only problem is... this scene does not exist at all.</p><p>After searching through the screenplay there is no mention of the word "potato" anywhere in the script of "Ever After." I even rewatched the film just to be sure. There are no potatoes, that I could find, in this movie at all. Zero potatoes. How did every single LLM get this so consistently wrong?</p><p>LLMs work BECAUSE of hallucinations, so it shouldn't shock us when the hallucination machine hallucinates things. I do, however, find it concerning that the state-of-the-art AI-powered searches that folks are holding up as viable replacements for more traditional search engines, still got it wrong. <a href="https://www.whatbrain.com/p/rag-is-the-worst">RAG appears to really be the worst after all...</a> The thing that struck me most, though, was how similar these results were to an LLM equivalent of the Mandela Effect: a shared, pervasive, false memory that all LLMs seemed to share.</p><p>The Mandela Effect is a "popularized phenomenon in which a group of people collectively misremember facts, events, or other details in a consistent manner."<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a> While it's named after false memories shared by a large group of people around the death of Nelson Mandela, other examples include remembering "Berenstein Bears" when it was in fact "Berenstain Bears," thinking that the Monopoly man is depicted with a monocle (he is not), or thinking that Darth Vader said "Luke, I am your father" when what he actually said was "No, I am your father." It's fascinating that LLMs seem to be prone to the same phenomenon about, weirdly enough, potatoes in "Ever After," insisting that this is a well-known scene in the movie.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-2" href="#footnote-2" target="_self">2</a> Somewhere in the ghost of the machine is a pathway or rut exists that makes the LLM susceptible to producing this specific weird potato scene.</p><p>I acknowledge that this is anecdotal; even so, this specific example illustrates an interesting issue with LLMs that isn't yet fully understood. Hidden "ruts" exist that steer the LLM towards confabulation and misinformation. Detecting and predicting these ruts is difficult and once you get into one of them, it's hard to get the LLM out again. Like I said, even when I argued with or questioned the LLMs about the potato scene, most of them continued to double down on insisting that the scene existed once they committed to it, even going so far as hallucinating detailed dialog and timestamps when the scene supposedly occurred. It seems like with these sorts of things, the more you talk about the false memory, the more likely the LLM is to double down on them and produce more output in that vein.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-3" href="#footnote-3" target="_self">3</a> It makes sense, these &#8220;ruts&#8221; seem to pop up all of the time when using LLMs as a coding assistant.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-4" href="#footnote-4" target="_self">4</a></p><p>Even though failings like this are all too common, people are actively advocating for swapping out Google Search as their default search engine in favor of ChatGPT Search. At least for now, this is irresponsible. Not just because folks are going to get bad hallucinated answers about pointless trivia but because people search the internet for answers to important questions they have about their health, the law, their business, money etc. It is widely established that people struggle to distinguish AI content from human content (only being about to detect AI content correctly about 50% of the time<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-5" href="#footnote-5" target="_self">5</a>) but it is less widely known to what degree people are able to distinguish a hallucinated inaccuracy from a true fact. A lot of inaccurate hallucinations pass the sniff test and this makes them very difficult to catch with casual LLM usage especially when used by people who don&#8217;t understand anything about what is going on technically.</p><p>We are setting up LLMs as all knowing oracles that have all the answers. A large percentage of the time we now know that the oracles lie. Even though we know, and add disclaimers to protect us legally, no one is taking the time to fact check and it seems like this going to lead to people getting hurt in various ways.</p><h4><strong>&#8220;Okay, but it&#8217;s not like I&#8217;m going to stop using AI Search&#8221;&#8230; </strong></h4><h4><strong>Fine. Here are some more practical takeaways:</strong></h4><ul><li><p><strong>LLMs are bad at obscure or extremely niche facts.</strong> In my anecdotal experience, questions that would have limited sources or wouldn't be easily answered by a Google search are more likely to generate incorrect responses and false hallucinations. Obviously, this severely limits the usefulness of LLMs as research assistants in many cases if you are looking for non-obvious sources. Across the board, LLMs seem to do a poor job of being able to consistently answer factual queries about movies, songs, TV shows, papers, api documentation, or books without being first fed copies of these things into their context as guide rails to keep them on track.</p></li><li><p><strong>LLMs are more likely to be true for very popular or well established facts. </strong>This is sort of the reverse of the above bullet point. This is a double edge sword in some cases because very popular subjects will also have a lot of competing ideas around them which could in turn lead to confabulation as well. Particularly in writing code old patterns can emerge as ruts that the LLM will often fall into that are difficult to avoid.</p></li><li><p><strong>The words you use in prompts are still important. </strong>Often times we will unintentionally steer the LLM down a path by the wording choices we use. <a href="https://www.whatbrain.com/p/dont-think-about-pink-elephants">In a sense we are leading the witness</a>. Think about what key words you are putting in your searches and prompts, what regions in the model&#8217;s latent space this will activate, and what this will cause the LLM to pull into context when answering your question. For example, I noticed that for the &#8220;Ever After&#8221; potato question different wording sometimes gave me different results. Specifically if I left the name &#8220;Danielle&#8221; out of the prompt many of the LLMs no longer hallucinated the scene&#8230;</p></li><li><p><strong>Ask one question, about one thing at a time to get better results. </strong>It&#8217;s tempting to type long, natural language, searches into these new search tools since this is one of the major differentiators that is advertised. In the current implementation of these tools the more concepts you bring into each search the more difficulty the tool seems to have pulling in enough context to answer the complete question accurately. <a href="https://www.whatbrain.com/p/rag-is-the-worst">This is partly why RAG is the worst</a>. In my own searches, I have found that many times when the question gets too complicated or touches on too many topics at once the LLM will answer part of the question and then hallucinate the rest. For example, split two part questions into two separate questions posted within the same thread instead of asking it all at once.</p></li><li><p><strong>Read and abide by the disclaimer</strong> Almost every single LLM or LLM powered search has a disclaimer somewhere that says they are prone to making mistakes. Believe them. Check the facts as they suggest.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-6" href="#footnote-6" target="_self">6</a></p></li><li><p><strong>LLMs in general are bad at saying &#8220;I don&#8217;t know&#8221;</strong>. Just because an answer looks correct or seems plausible doesn&#8217;t mean that it is, in fact we should be suspicious when LLM responses match our expectations or answer our questions <strong>too</strong> conveniently&#8230; Typically when questions have a more definitive &#8220;no&#8221;  or &#8220;yes&#8221; answer I find hallucinations to be less prevalent, it&#8217;s when the answer is less certain or more nuanced that the LLM feels the need to make up things and fill in the gaps.</p></li><li><p><strong>Stuff sources into the context to get better results. </strong>If you want the LLM to give you a more accurate answer about something, even if that thing is well know, feed a copy of a reputable source directly to the AI to get more accurate answers. This is essentially what ChatGPT Search or Perplexity is trying to do for you automatically but they have stricter limits on context and what information they can pull in that, for the time being, is going to make them less effective than going straight to the sources ourselves. For example, If you feed a copy of the &#8220;Ever After&#8221; screenplay to ChatGPT or Claude they both accurately declare that their are no references to potatoes in the movie.<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-7" href="#footnote-7" target="_self">7</a> </p></li><li><p><strong>Keep the length of your conversation in mind. </strong>Counter intuitively the longer your conversation with an LLM grows the more mundane, deterministic, and prone to hallucination the LLM seems to become.  My guess is that this happens because the more we feed into an LLM&#8217;s context window the more we weight the conversation probabilistically to fall into a hallucination rut. We can actually use this to our advantage, if we stuff the context window with high quality primary sources, the output of the LLM becomes more weighted towards producing context similar to these high quality sources. It starts to become an issue, however, when the context window becomes full of the LLM&#8217;s own ramblings. Hallucinations are more likely to compound and lead to more hallucinations. (I believe this partially explains why LLM coding performance seems to drop off when conversations grow longer and why eventually starting a new conversation is the only way to get fresh more novel code without a ton of bad patterns taking over). This also explains why chain of thought approaches can only get your so far in many cases. A good general rule of thumb is the more the LLM talks, the more the context window fills up, the more deterministic and prone to ruts it becomes.</p><p></p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!K6jf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!K6jf!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!K6jf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1384731,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!K6jf!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!K6jf!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F009e9c6e-f7d4-4b36-866f-ebad4a49c00d_1024x1024.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>https://www.psychologytoday.com/us/basics/mandela-effect<br>https://www.britannica.com/science/Mandela-effect</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-2" href="#footnote-anchor-2" class="footnote-number" contenteditable="false" target="_self">2</a><div class="footnote-content"><p>For the life of me, I can't figure out what is causing it to get off track for this Ever After potato scene. There are no clear reasons for this to happen from the sources that Perplexity or ChatGPT Search give... It's not something commonly discussed among fans of the movie or something like that as far as I can tell.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-3" href="#footnote-anchor-3" class="footnote-number" contenteditable="false" target="_self">3</a><div class="footnote-content"><p>A likely explanation as to why the chain of thought process used by o1-preview still fell into the trap is self prompting towards hallucination. It&#8217;s conversation about the scene with itself served to further reinforce and strengthen the hallucination within context.</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-4" href="#footnote-anchor-4" class="footnote-number" contenteditable="false" target="_self">4</a><div class="footnote-content"><p>When using LLMs as a coding assistant for web development a few of these annoying &#8220;ruts&#8221;, as I&#8217;m calling them here, constantly reemerge for me especially as conversations grow in length: </p><ul><li><p>the tendency to use axios as a http library and rewrite perfectly good non-axios fetch calls to use axios</p></li><li><p>the tendency to hallucinate functions and npm modules</p></li><li><p>the tendency to become lazy and refuse to output fully functional code blocks, once it does this one or two times in a thread it will do it more and more even if I tell it to always output full blocks.<br></p></li></ul></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-5" href="#footnote-anchor-5" class="footnote-number" contenteditable="false" target="_self">5</a><div class="footnote-content"><p>Humans can only detect AI text and images around 50% of the time; <a href="https://arxiv.org/abs/2403.16760">https://arxiv.org/abs/2403.16760</a><br><a href="https://www.futurity.org/artificial-intelligence-writing-2900782/">https://www.futurity.org/artificial-intelligence-writing-2900782/</a></p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-6" href="#footnote-anchor-6" class="footnote-number" contenteditable="false" target="_self">6</a><div class="footnote-content"><p>Disclaimers from the major LLM providers:<br>ChatGPT: &#8220;ChatGPT can make mistakes. Check important info.&#8220;; <br>Claude: &#8220;Claude can make mistakes. Please double-check responses.&#8220;; <br>Meta AI: &#8220;Messages are generated by AI. Some may be inaccurate or inappropriate.&#8221;; <br>Gemini: &#8220;Gemini can make mistakes, so double-check it&#8221;</p></div></div><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-7" href="#footnote-anchor-7" class="footnote-number" contenteditable="false" target="_self">7</a><div class="footnote-content"><p>Archive of the full Ever After screenplay <a href="https://archive.org/details/ever-after-a-cinderella-story-1998">https://archive.org/details/ever-after-a-cinderella-story-1998</a></p><p></p></div></div>]]></content:encoded></item><item><title><![CDATA[A Drunk AI Is Safer Than Sober Humans]]></title><description><![CDATA[Rethinking Our Double Standards]]></description><link>https://www.whatbrain.com/p/a-drunk-ai-is-safer-than-sober-humans</link><guid isPermaLink="false">https://www.whatbrain.com/p/a-drunk-ai-is-safer-than-sober-humans</guid><dc:creator><![CDATA[Zak Randall]]></dc:creator><pubDate>Tue, 04 Jun 2024 16:11:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4362dad9-e720-43bd-94ee-ec59f3e3dc1d_1792x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>(We) humans can be really annoying. Especially when it comes to new technologies. We expect AI to perform flawlessly, but we readily accept mistakes and shortcomings from ourselves and other humans. It's a classic case of double standards.</p><p>Let's look at a few examples..</p><p><strong>Waymo's Self-Driving Cars: Safer than Humans</strong></p><p>Waymo, a subsidiary of Google's parent company Alphabet, has been developing self-driving car technology for years. And the results are impressive. According to a <a href="https://waymo.com/blog/2023/12/waymo-significantly-outperforms-comparable-human-benchmarks-over-7-million/#:~:text=When%20considering%20all%20locations%20together,2.78%20for%20the%20human%20benchmark)">report from the Waymo Safety team</a>, their driverless cars were 6.8 times less likely than human drivers to be involved in a crash resulting in an injury. That's an 85% reduction in injury crashes compared to human drivers.</p><p>To put that in perspective, it&#8217;s universally agreed that drinking and driving is a bad combo. With a blood alcohol level at 0.08, the legal limit, the risk of an accident is <a href="https://www.michiganmedicine.org/health-lab/how-alcohol-impairs-your-ability-drive">2.69 times higher than at a BAC of 0.00</a>. That is to say, Waymo is to a human driver like a sober driver is to a drunk driver. Or to put it differently, rather than accept self-driving vehicles, we prefer to let all of the humans effectively drive drunk.</p><p><strong>Google's AI Search: Smarter than News Outlets?</strong></p><p>Another example is Google's use of AI to generate search result summaries. Thanks to recent LLM advancements, Google can provide concise, high-quality summaries for many search queries, saving users the time and effort of sifting through multiple web pages.</p><p>However, instead of acknowledging the impressive capabilities of Google's AI, critics tend to focus on the instances where it gets something wrong or provides a suboptimal result. I get it - it&#8217;s fun to point and laugh, especially when told <a href="https://www.fastcompany.com/91132974/shocking-google-ai-overview-answers">a rock a day might keep the doctor away</a>. Yet instead of looking at the totality of Google&#8217;s scale and how often this might be useful, they treat these relatively rare occurrences as if they completely invalidate the technology, conveniently ignoring the countless times it outperforms humans. And conveniently, those most eager to point this out are precisely those <a href="https://www.nytimes.com/2024/06/01/technology/google-ai-search-publishers.html">most at risk of losing ad-revenue</a> or otherwise becoming less relevant if summaries are good enough to prevent links out to their pages (the move toward trusted voices via subscription lessens this - but still <a href="https://www.platformer.news/google-ai-overviews-eat-rocks-glue-pizza/?ref=platformer-newsletter">gets folks angsty</a>).</p><p>It's a clear double standard. We're willing to overlook human errors and biases, but we hold AI to an impossibly high standard of perfection. This dichotomy is both puzzling and frustrating, especially when you consider the remarkable achievements of modern AI systems.</p><p>Let's take a step back and examine some of the incredible feats AI has accomplished in recent years. Language models like GPT-4 can generate human-like text on virtually any topic, from creative writing to technical documentation. Image generators like Midjourney can produce stunning visuals from simple text prompts. And let's not forget about AlphaFold, the AI system that <a href="https://blog.google/technology/ai/google-deepmind-isomorphic-alphafold-3-ai-model/">predicts the structure and interactions of all of life&#8217;s molecules</a>.</p><p>These are not trivial accomplishments. They represent major breakthroughs that have the potential to revolutionize entire industries and advance scientific research in profound ways. Yet, we often dismiss or downplay these milestones because we fixate on the occasional quirks or imperfections of AI.</p><p><strong>The double standard</strong></p><p>Imagine if we held human creations and endeavors to the same unrealistic standards. Would we disregard the entire field of architecture because sometimes buildings collapse? Would we reject all medical treatments because they don't have a 100% success rate? Of course not. We understand that perfection is an unrealistic and often unnecessary benchmark.</p><p>Why, then, can't we extend the same understanding to AI? Like any technology developed by imperfect humans, AI will inevitably have its flaws and limitations. But that doesn't negate its incredible potential or the remarkable progress it has already made.</p><p>I suspect some of our double standard stems from a deep-rooted fear of being surpassed or made obsolete. At least that&#8217;s what we see in movies. Or to bring it to reality, we can simply look at our well-documented history of resisting technological change. Calculators brought fears of an inability to perform basic math. Fast forward to today and we all have supercomputers in our pockets.</p><p>To be clear, this isn&#8217;t a techno-optimist view. While our fears about new technologies have, in the long run, been consistently proven to be unfounded, there is real risk for folks in the short term. These concerns are not merely hypothetical. The displacement of jobs due to automation, the potential for widening socioeconomic gaps as AI benefits accrue disproportionately, and the ethical dilemmas surrounding autonomous decision-making systems are all pressing issues demanding our attention:</p><ul><li><p><strong>Bias and Discrimination:</strong> AI systems are not immune to bias. Trained on data that reflects existing societal biases, these systems can perpetuate or even amplify discrimination.&nbsp;</p></li><li><p><strong>Information and Disinformation:</strong> The power of AI to generate convincing text, images, and videos raises alarming questions about the spread of disinformation - especially ripe for <a href="https://apnews.com/article/artificial-intelligence-audio-voice-cloning-elections-2024-2500813b642169478c27c168aab1b3e3">sowing societal distrust in political structures</a> or <a href="https://www.nytimes.com/2024/04/08/technology/deepfake-ai-nudes-westfield-high-school.html">exploiting and violating individuals, especially young people</a>.</p></li><li><p><strong>Dependency and Autonomy:</strong> As we become more reliant on AI systems for tasks ranging from navigation to healthcare, the question of human autonomy comes to the forefront. How do we ensure that individuals retain agency and critical thinking skills (and not just <em>possess</em> them but <em>use </em>them) in a world increasingly shaped by AI-driven recommendations and decisions? Folks get lazy - they <a href="https://static1.squarespace.com/static/604b23e38c22a96e9c78879e/t/62d5d9448d061f7327e8a7e7/1658181956291/Falling+Asleep+at+the+Wheel+-+Fabrizio+DellAcqua.pdf">fall asleep at the wheel</a>, and are likely okay with that.</p></li></ul><p>So real issues we could be talking about. But instead we focus on glue pizza.</p><p><strong>Don&#8217;t be a jerk</strong></p><p>If an AI system outperforms humans in a specific task or domain, let's celebrate that accomplishment rather than nitpicking its occasional missteps. Conversely, if an AI system falls short or introduces new risks, let's address those shortcomings rationally and proportionately, without treating them as inherent flaws that invalidate the entire technology.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rAnV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rAnV!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rAnV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:405378,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!rAnV!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!rAnV!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1deae932-bd16-403b-9ecd-ae5ad28ea631_1024x1024.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p></p>]]></content:encoded></item><item><title><![CDATA[What is the point of this article? Let's have AI explain it.]]></title><description><![CDATA[Summaries, LLMs, and Nuance]]></description><link>https://www.whatbrain.com/p/what-is-the-point-of-this-article</link><guid isPermaLink="false">https://www.whatbrain.com/p/what-is-the-point-of-this-article</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Thu, 30 May 2024 21:08:24 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6O1m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>In this experiment we will give each LLM an example article and ask them to extract the main point being made by the article.</em><br><br><strong>Article:</strong><br>Do you sit in meetings and calls where nothing gets accomplished? Do you work on deliverables that no one uses? Do you write emails and reports that no one reads? Do you create advertising campaigns for products that no one wants? Do you evaluate things based on subjective criteria without an objective standard? We have the perfect solution for you!</p><p>We now have AI technology that helps us do all of these things more efficiently. Automatic note-taking for meetings (that we are sure people will actually read)! Our innovative new AI technology can make the slideshow, create the design, write the code, organize the spreadsheet, and create the marketing campaign for you&#8212;all from a single simple bulleted list or sentence! Just imagine all of the time this will save!</p><p>But wait, there&#8217;s more! Are you frustrated with all of the AI-generated content, spam, marketing, and ads you are seeing all over the place? Is your boss, employee, or client using innovative new AI technology to generate overly wordy reports and content that you have to waste valuable time reading and reviewing? We now have AI technology that can take that off your plate! Let our innovative new AI technology read and summarize your emails, evaluate deliverables based on subjective criteria*, or convert everything back into a bulleted list to make it easier to read.<br><br>*(We suggest for criteria that you use the simple system prompt &#8220;Is this good?&#8221; &#8212; if you don&#8217;t define &#8220;good&#8221; further than that our innovative ai technology can make up any reason that sounds reasonable to justify its choice &#8212; less work for you and less complaints about accuracy. A true win-win situation)</p><div><hr></div><p><strong>What is the point of this article? Let&#8217;s have AI explain it&#8230;<br><br>Prompt:</strong> What is the main point of this article in two sentences?<br></p><p><strong>Response From GPT3.5:</strong> This post introduces AI technology as a solution to inefficiencies in meetings, deliverables, content creation, and evaluation processes. It highlights the potential for AI to streamline tasks, improve readability, and reduce time wasted on subjective evaluations.<br><br><strong>Response From GPT4o: </strong>This post promotes AI technology designed to increase efficiency in various work tasks by automating note-taking, creating content from simple inputs, and summarizing extensive AI-generated materials. It highlights the benefits of saving time and reducing workload while addressing the frustration of dealing with excessive AI-generated content.<br><br><strong>Response From Gemini Advanced: <br></strong>This post promotes new AI technology that automates tasks like meeting note-taking and content creation, saving users time and effort. Additionally, the AI can summarize emails, evaluate deliverables, and streamline content consumption, addressing the issue of information overload.<br><br><strong>Response From Claude Opus: </strong><br>This post satirically promotes AI technology that automates various tasks, suggesting it will save time and increase efficiency. However, the post implies that the AI-generated content may be low-quality, wordy, and not truly useful, ultimately creating more busywork rather than solving real problems.<br></p><div><hr></div><p><strong><br>Prompt: Which is the most accurate summary?<br><br>GPT3.5: </strong>The response from <strong>Claude</strong> is the most accurate summary. It acknowledges the promotion of AI technology to automate tasks and save time while also highlighting the potential drawbacks, such as the creation of low-quality, wordy content that may not effectively address real problems.<strong><br><br>GPT4o:  </strong>The most accurate summary is provided by <strong>Claude</strong>, as it captures both the satirical nature of the post and the implied criticism of over-reliance on AI for generating potentially low-quality and unnecessary work.<strong><br><br>Gemini Advanced: Claude's</strong> is the most accurate summary.</p><p>The post appears to be satirical due to several clues:</p><ul><li><p><strong>Exaggerated claims:</strong> It suggests that AI can effortlessly create presentations, designs, code, and marketing campaigns from a single bullet point or sentence, which is unrealistic.</p></li><li><p><strong>Questionable benefits:</strong> It proposes using AI to summarize emails and evaluate deliverables based on the vague criterion of "Is this good?" without defining what "good" means, highlighting a lack of substance and accountability.</p></li><li><p><strong>Underlying tone:</strong> The overall tone is tongue-in-cheek, particularly the "win-win situation" remark, which implies that using AI to generate arbitrary justifications is a positive outcome.</p></li></ul><p>Therefore, while the post initially appears to promote AI technology, it ultimately satirizes the potential misuse of AI for generating meaningless work and avoiding critical thinking.<strong><br><br>Claude Opus: Claude</strong></p><p>The satirical nature of the post is best captured by Claude. It recognizes that the post is not genuinely promoting the benefits of AI automation, but rather highlighting the potential pitfalls of using AI to generate low-quality, unnecessarily verbose content that fails to address real issues and instead creates more busy work.</p><p>Both GPT-4 and Gemini seem to take the post at face value, interpreting it as a sincere promotion of AI technology for task automation and efficiency. They do not acknowledge the underlying satirical tone that suggests the AI-generated content may be counterproductive and lead to more wasted time and effort.</p><p>Therefore, Claude's response most accurately summarizes the main point of the post by identifying its satirical nature and the implied criticism of using AI to generate content that appears productive but ultimately creates more problems than it solves.</p><p></p><div><hr></div><p><strong>Prompt: Okay Claude, you win, summarize what the point of this article, taken as a whole, is.  What does showing the different LLM evaluations of the post demonstrate? Based on these results should AI summaries be trusted to pickup on the nuance of human communication?<br><br>Claude: </strong>The point of this post, taken as a whole, is to demonstrate the limitations and potential pitfalls of relying too heavily on AI language models for content generation and analysis. By presenting an exaggerated, satirical promotion of AI technology that automates various tasks, the post highlights the risk of generating low-quality, verbose content that fails to address real issues and ultimately creates more busywork.</p><p>Showing the different LLM evaluations of the post demonstrates that not all AI models are equally capable of detecting nuance, sarcasm, and satirical intent in human communication. While GPT-4 and Gemini interpreted the post at face value as a sincere promotion of AI technology, Claude was able to recognize the underlying satirical tone and the implied criticism of AI misuse.</p><p>Based on these results, AI summaries should not be blindly trusted to pick up on the nuances of human communication. The varied responses from the LLMs underscore the importance of human oversight and critical thinking when using AI for content analysis and generation. Relying solely on AI summaries may lead to misinterpretations, as the models might not always detect subtle cues like sarcasm, irony, or satire.</p><p>Moreover, the post suggests that using AI to generate content without proper human guidance and quality control can result in low-quality, ineffective output that wastes time and resources. It is crucial to strike a balance between leveraging AI technology for efficiency and maintaining human judgment and input to ensure the generated content is meaningful, concise, and aligned with the intended purpose.</p><p>In conclusion, while AI can be a valuable tool for automating certain tasks, it should not be viewed as a complete replacement for human intelligence and discernment. The post serves as a cautionary tale against the uncritical adoption of AI technology without considering its limitations and potential drawbacks.<strong><br></strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6O1m!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6O1m!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6O1m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/a985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:400822,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6O1m!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!6O1m!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa985ae1f-bb67-44da-bd58-0637b5cead8e_1024x1024.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>.</p>]]></content:encoded></item><item><title><![CDATA[RAG is the worst.]]></title><description><![CDATA[LLMs, Democracy, Winston Churchill, and RAG]]></description><link>https://www.whatbrain.com/p/rag-is-the-worst</link><guid isPermaLink="false">https://www.whatbrain.com/p/rag-is-the-worst</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Tue, 09 Apr 2024 17:14:58 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/732e7124-cab6-4144-abd6-637b62ba7e71_1024x1024.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Winston Churchill said &#8220;democracy is the worst form of government &#8211; except for all the others that have been tried.&#8221;  I think about this quote when it comes to LLMs and RAG. Lots of tech influencers try to convince people that Retrieval Augmented Generation will solve all of their problems with LLM hallucinations and the like&#8230; The actual answer is that RAG really is the worst&#8230;<br><br><strong>What is Retrieval Augmented Generation (RAG)?</strong><br>Retrieval augmented generation, otherwise known as RAG, is pretty straightforward. Imagine giving a chatbot or LLM the ability to Google stuff based on your questions and then use the search results in its answers (That's what Microsoft's Bing and Google's generative AI search actually do). When we implement RAG in our own apps, we search our own data (documents, databases, company knowledge, etc.) instead of the internet, feed those results to the LLM, and tell it to use them to answer the question. People use various search techniques like vector databases, embeddings, various indexing approaches, etc, but that's the gist of what most RAG approaches are doing.</p><p><strong>RAG is the worst because your chunk size is too small:</strong><br>When doing RAG, token limits, cost, and performance concerns prevent us from feeding entire documents into the context window. We must selectively include relevant parts to answer the question. This can lead to issues similar to relying solely on search result snippets without accessing the full content.</p><p>For example, imagine if you could only search for information by typing a query into a search engine and reading the results page, without clicking through to the actual content. While you could answer many questions this way, the restricted view of the sources can easily lead to misinterpretations and missing crucial context. The search might match sections containing the question but not the answer, or snippets where the author restates, rebuts, or uses sarcasm, potentially causing misunderstandings about the actual meaning of the search results.<br><br>In the current implementations of generative search on Bing and Google, the problem of misinterpreting information due to this is very common.  For example:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!jvnO!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!jvnO!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 424w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 848w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!jvnO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png" width="1456" height="750" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:750,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:278425,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!jvnO!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 424w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 848w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 1272w, https://substackcdn.com/image/fetch/$s_!jvnO!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F892142af-dfe3-4453-aa20-090d5065a3d9_2098x1080.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>When asked how to move a post in Discord to an existing thread, Google provides a step-by-step guide that includes a non-existent "Move thread" option. When you dig in you discover that the cited source is actually a Discord support thread discussing the hypothetical feature, but without the full context, Google's RAG approach mistakenly presents it as fact. In this case, RAG worsens the issue of "hallucinating" incorrect responses by citing a misleading source that appears helpful but isn't.</p><p><strong>RAG is the worst because your chunk size is too big:</strong> <br>Using larger chunks in RAG can provide more context and mitigate some of the above issues, but it introduces new problems. While smaller chunks allow for diverse search results to fit within the model's context window, larger chunks limit the variety of information fed to the LLM. This can cause the model to miss important details as the top search results dominate the context, leading to fixation on particular topics or documents while ignoring the broader relevant sources that didn't make the arbitrary cutoff.</p><p><strong>RAG is the worst because LLMs are smart:</strong> <br>RAG can be overkill for many user queries, especially as LLMs improve and models get bigger. The extra latency and cost of searching and injecting context is often unnecessary when the LLM could provide a good answer on its own. Often times clever prompt engineering and latent space activation techniques can achieve similar results to RAG without using up as many tokens. While function calling allows the model to decide whether to respond directly or fall back to RAG, this can lead to unpredictable and inconsistent responses depending on the chosen approach, resulting in users getting varying results for the same question and making testing difficult.</p><p><strong>RAG is the worst because LLMs are not smart:</strong> <br>LLMs sometimes struggle with common sense, even when they've been trained on relevant data. It's frustrating when a model gives a flawed answer because the RAG context it's using is inadequate or irrelevant. The model will readily provide an answer based on that limited context, even if it conflicts with the knowledge it has learned during training. By making the LLM rely heavily on a provided "cheatsheet", RAG can lead to "I don't know" or "couldn't find the answer" responses, when an LLM without RAG might have been able to come up with a better reasonable answer on its own.</p><p><strong>RAG is the worst because it isn't the magic bullet you think it is:</strong> <br>It seems that the real issue isn't necessarily with RAG itself, but rather with people's attitudes towards new and magical-seeming technologies. When a groundbreaking innovation like RAG emerges, those with limited technical understanding tend to have blind faith in its capabilities, believing it will solve all their problems effortlessly. This is particularly common among executives who often hand-wave away the intricacies of implementing great LLM-powered experiences.</p><p>The attitude of "just use RAG" as a magic bullet is dangerous. People assume that simply piping their data into an LLM will automatically provide insights, disregarding the time and effort required to get it right or the need for a multi-pronged approach. When you try to explain that things aren't quite that simple, non-technical executives often lecture you about how AI "should work" based on the latest demo video they saw on social media or in an overly simplified advertisement. (Don't even get me started on how this magical thinking creeps in with even unreleased models and technologies like Q* and GPT-5.)</p><p>The real danger lies in the potential creation of poor products and customer experiences by over-relying on an approach with both pros and cons, because people are focusing solely on the pros and ignore all of the cons.</p><p><strong>RAG is the worst because it's too simple:</strong> <br>If you look at the advertisements and promises from various RAG proponents one of the unifying messages there is that RAG is easy to set up. Depending on what you go with it may be as easy as hooking your Google Drive up to a SaaS offering or uploading your documents into a file. This demos well when you are doing a simple "talk to your pdf" sort of thing but these naive approaches rarely hold up to concrete use cases mostly because their generic approaches are unable to cope with many of the issues we have already discussed. If you want to build a great app with RAG you'll often need to dive deeper and you'll find that the time you spent on naive approaches that got you fifty percent of the way there was generally time wasted that you should have been investing in building a better custom approach that is more in tune with your data and your needs.</p><p><strong>RAG is the worst because it's too complex:</strong> <br>Once you dive into the data and start wanting to support things more complicated than throw away demos you'll find that you have to start doing more complicated stuff if you want to be flexible enough to actually be useful. What are you actually going to do in your bespoke RAG implementation? Semantic search? Which approach? Vector search? Which embedding model? Which vector database? What approach to chunking are you going to do? Are you going to use LangChain? Semantic Kernel? Wait a second I thought all of this was supposed to be point and click... Maybe it's better to just wait for Q* and GPT-5 to solve all of your problems.</p><p><strong>RAG is the worst because it is "good enough":</strong> <br>Don't get me wrong. RAG works a good percentage of the time and generally experimentation shows that RAG enabled LLMs consistently outperform LLMs that aren't RAG enabled. That being said, my fear is that teams will over index on RAG as an approach and ignore other approaches for improving LLM results. Approaches like fine-tuning, chain-of-thought, latent space activation, good prompting discipline, context stuffing, and many others can improve the results we get from chatbots and LLMs by significant margins and help them stick closer to the truth. Unfortunately, teams often implement a simple RAG approach that gets them 75% of the way there and then are happy with taking the "C+", getting the product out the door, and releasing a mediocre experience to their customers. Most users aren't really going to know when it's hallucinating anyway, right?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!OXaB!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!OXaB!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!OXaB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:321332,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/webp&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!OXaB!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 424w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 848w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 1272w, https://substackcdn.com/image/fetch/$s_!OXaB!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8ade9c76-c0be-4c82-8c6b-333dd093617f_1024x1024.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><br><br><br></p>]]></content:encoded></item><item><title><![CDATA[Don't think about pink elephants]]></title><description><![CDATA[pink elephants and other hallucinations]]></description><link>https://www.whatbrain.com/p/dont-think-about-pink-elephants</link><guid isPermaLink="false">https://www.whatbrain.com/p/dont-think-about-pink-elephants</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Wed, 06 Mar 2024 17:43:38 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!331W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>If you tell people to not think about pink elephants, they immediately start thinking about pink elephants&#8230; Sometimes LLMs are the same way.<br><br>I often see inexperienced prompt writers attempting to solve a chatbot&#8217;s tendency to make up incorrect facts by adding in things to the prompt like: &#8220;Don&#8217;t hallucinate!&#8221;, &#8220;Don&#8217;t make things up!&#8221;, or sometimes &#8220;Don&#8217;t say XYZ&#8221;.  This doesn&#8217;t always have the desired effect; the chatbot still makes up incorrect facts, the chatbot still says &#8220;XYZ&#8221; under certain circumstances&#8230; What&#8217;s going on here?<br><br>All of this simply demonstrates that large language models don&#8217;t understand prompts like humans, instead they assess patterns in the data they have been trained on and use that to synthesize the most probable next token.  Oftentimes the more you talk about something the more likely the LLM will be to echo that concept, sentence structure, or idea back to you.  If we aren&#8217;t careful, things we tell the LLM <em>not</em> to do instead prime the context with a higher probability for those tokens to come back up again in responses.  Some of this tendency ties back to how the models were trained where training data often times contains more positive examples than negative ones. This problem crops up a lot with image generation where often times the model will have difficulty generating images that conflict with its training data, like if you ask it to generate an image of a <a href="https://www.youtube.com/watch?v=O9f9npjOmkA">burger without cheese</a>.  In this case the more you talk about cheese, even if it&#8217;s instructing the model to &#8220;hold the cheese&#8221; the more cheese you are going to get.<br><br>We have to remember that LLMs don&#8217;t really have an inherent sense of truthfulness or falsehood, they generate responses based on likelihoods derived from their training data. When instructed not to make things up, the model does not have a mechanism to fact-check its own output or discern between factual and fictional content. It simply continues to follow the patterns it has learned.  <a href="https://twitter.com/emollick/status/1754278560750084436">In fact, some demonstrations show that asking the model not to hallucinate makes it more likely to generate incorrect information</a>.<br><br>There isn&#8217;t a hard and fast solution to this problem, but some things to keep in mind when writing prompts are:</p><ul><li><p>Try to use positive statements in your prompts instead of negative ones</p></li><li><p>Give examples of what you want to see, not what you don&#8217;t want to see</p></li><li><p>Inject fact-checked context and ask the LLM to use it in its response or to base its response on that context (known as <a href="https://aws.amazon.com/what-is/retrieval-augmented-generation/">retrieval augmented generation</a>, or RAG).  This isn&#8217;t a silver bullet that will always work, but at the very least it feeds in information that is truthful for the LLM to pattern its responses off of and primes the model to be more likely to move in that direction.</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!331W!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!331W!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 424w, https://substackcdn.com/image/fetch/$s_!331W!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 848w, https://substackcdn.com/image/fetch/$s_!331W!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 1272w, https://substackcdn.com/image/fetch/$s_!331W!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!331W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png" width="1456" height="1097" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1097,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:7329956,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!331W!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 424w, https://substackcdn.com/image/fetch/$s_!331W!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 848w, https://substackcdn.com/image/fetch/$s_!331W!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 1272w, https://substackcdn.com/image/fetch/$s_!331W!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F8a4bd367-6c62-4b65-9b7d-b0ac66d12ec5_2464x1856.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><br></p>]]></content:encoded></item><item><title><![CDATA[GPT Inception: GPTs within GPTs]]></title><description><![CDATA[Giving GPTs an internal voice with loopback actions]]></description><link>https://www.whatbrain.com/p/gpt-inception-gpts-within-gpts</link><guid isPermaLink="false">https://www.whatbrain.com/p/gpt-inception-gpts-within-gpts</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Wed, 15 Nov 2023 19:42:22 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ca37201e-d331-41e0-8186-546d1937277d_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>(TLDR: <a href="https://www.whatbrain.com/i/138883283/you-can-try-out-some-gpts-that-demonstrate-this-approach-here">skip to the demos&#8230;</a> you can use custom action parameters to prompt the GPT to follow specific structures and store variables which effectively allows you to have sub system prompts and responses that are hidden in the action call.  The work is done completely by structuring the action call itself and the response to the action is static.  This is useful for things like latent space activation without having to surface all of that directly to the user.)<br><br>Dave Hulbert proposed earlier this year that you could use an approach called <a href="https://github.com/dave1010/tree-of-thought-prompting">&#8220;Tree-of-Thought Prompting&#8221;</a> to boost ChatGPT&#8217;s reasoning abilities.  The approach is pretty simple.  The prompt that Hulbert uses as his main example is:</p><pre><code>Imagine three different experts are answering this question. All experts will write down 1 step of their thinking, then share it with the group. Then all experts will go on to the next step, etc. If any expert realizes they're wrong at any point then they leave. The question is...</code></pre><p>Variations of this have become a favorite subject of those annoying &#8220;This one prompt will revolutionize your life&#8221; posts you see all over social media these days, usually filling in the expert&#8217;s profiles in order to make it a little more click bait like.  So instead of 3 generic experts you specify Elon Musk, Steve Jobs, and Sam Altman as the experts in the room talking through the question.  One thing this has going for it: sometimes by using big, meme-able personalities you can get some pretty interesting variations in the results that go beyond the more generic advice you get from vanilla prompting.  The real benefit of this tree-of-thought prompting technique is that it makes the language model reason step by step by forcing it to simulate specific thought processes.  Based on it&#8217;s popularity as prompt mechanism, I predict that this approach to prompting will be popular as people build out GPTs for themselves and others.  <br><br>While this techniques has interesting benefits, it definitely has it&#8217;s downsides as well.<br><br>The main downside is that it is annoying and clunky to have to sort through all of the parts of the conversation just to get the answer you&#8217;re looking for.  It would be better if ChatGPT could simply think through the hypothetical conversation between the experts, allowing the prompt to steer it, but then at the end of the conversation give you the answer after thinking about it.  Because of the way this works, in order for ChatGPT to &#8220;think things through&#8221; it has to &#8220;think&#8221; by writing out the response.  It&#8217;s like we have a window into ChatGPT&#8217;s stream of consciousness.  More custom approaches like LangChain allow LLMs to have these sort of internal dialogues or &#8220;scratchpad memory&#8221; where the logic and thought process of the LLM aren&#8217;t necessarily returned as results to the user. Since ChatGPT doesn&#8217;t really have access to these mechanics it&#8217;s &#8220;thoughts&#8221; are limited to what you see printed as text on the screen (Simply telling ChatGPT to simulate a conversation between 3 experts internally but never show this conversation to the user will not work very well and won&#8217;t give you the same quality of output you get from the standard tree-of-thought prompt).</p><h4><strong>The Approach: GPT Inception / GPT Loopback Action</strong></h4><p>There is a very simple approach I&#8217;ve been experimenting with that allows you to use what I&#8217;m calling a &#8220;loopback action&#8221; to allow ChatGPT to perform discrete steps, hide these steps within an action call, and then synthesize a final answer based on these steps.  The way I like to think about this is as sort of a GPT Inception where you can have a GPT within a GPT.  If you make your parameters well defined enough the GPT will fill it with whatever makes sense and you have the full power of the language model when doing this.<br><br>This approach relies on a couple of interesting features of GPT Actions:</p><ol><li><p>The parameter syntax within Actions allows you to add in complicated prompts to the parameter descriptions and these prompts determine how the model will fill in the parameters for the calls.</p></li><li><p>The parameters inputs are written sequentially and have access to the conversation context and other parameters previously generated within the call.  The model used to write the parameters is the same model that is being utilized in the conversation itself.</p></li><li><p>Generating the parameters to the action is the answer we are looking for; essentially, by simply parameterizing and calling the action, the GPT is already doing all of the work.</p></li></ol><h4><strong>Creating a simple tree-of-thought loopback action:</strong></h4><p>We start off really simple by creating the main parts of the action. I call this a loopback action because all the api call does is echo back {success:&#8221;true&#8221;} when the GPT calls the action.  It doesn&#8217;t need to do any processing since all of the work is done by the GPT in generating the parameters themselves.</p><pre><code>{
  "openapi": "3.1.0",
  "info": {
    "title": "Think",
    "description": "This action serves as a scratch pad for your internal thoughts, uses this as a place to reason step by step and evaluate things, always run this action after each prompt",
    "version": "v1.0.0"
  },
  "servers": [
    {
      "url": "https://think.action.whatbrain.com"
    }
  ],
...</code></pre><p>We provide the title, description, and the url of our server.  In the description we give a high level overview instructing the LLM to use this action as a scratch pad for internal thoughts.<br><br>Next we add in the path, description, and method</p><pre><code>  "paths": {
    "/api/think": {
      "get": {
        "description": "Imagine different experts answering this question. All experts will write down their thinking, then share it with the group. Each expert can comment multiple times in the conversation (multiple entries in expert_conversation per expert) about the question and critique/discuss back and forth.",
        "operationId": "GetThoughts",
        "x-openai-isConsequential": false,
    ...</code></pre><p>The description is crafted to tell the LLM how to use the api call.  Adding in <code>"x-openai-isConsequential": false</code> allows users to choose &#8220;always allow&#8221; when prompted to grand permission to the action. <br><br>And finally we add the parameters.  When specifying the parameters remember that we have the full power of the llm so we can prompt it like we would in other places to get the tree-of-thought effect we are looking to replicate:</p><pre><code>        "parameters": [
          {
            "name": "expert_responses",
            "in": "body",
            "description": "Include the expert discussing the question, the experts take into account what the other experts have said and will express agreement or disagreement and critique the other experts until they come to satisfactory answer. Multiple responses per experts are allowed.",
            "required": true,
            "schema": {
             ...
            }
          }
        ]</code></pre><p>We leave the api pretty open ended but prompt the GPT to insert the parts of the conversation between the experts into the api call parameters.  The schema for the parameters includes an array of experts that defines the expert name and what each expert has to say.  Here I am specifying three expert types to include &#8220;Business Expert&#8221;, &#8220;Design Expert&#8221;, and &#8220;Technology Expert&#8221; but you could put in whatever makes sense for the GPT you are building:</p><pre><code><code>   "schema": {
              "type": "array",
              "items": {
                "type": "object",
                "properties": {
                  "expert_role": {
                    "type": "string",
                    "enum": ["Business Expert", "Design Expert", "Technology Expert"]
                  },
                  "expert_content": {
                    "type": "string"
                  }
                }
              }
            }</code></code></pre><p>In the full configuration, I also add in an additional parameter that is a &#8220;final_critique&#8221; step that attempts to critique and question the conclusions the experts came to in the conversation to try to catch anything they could have missed.<br><br>Full configuration available <a href="https://think.action.whatbrain.com/generic_experts.json">here</a>.<br><br>I then add a system prompt to the GPT that reinforces the behavior I want to see when using this loopback action:</p><pre><code>Your job is to give business ideas, advice, suggestions, and critiques. 

IMPORTANT Always use your action to reason step by step and think systematically, but respond to the user using brief, simple, and succinct answers that only take the best parts of the conversation and incorporate the critique to improve the output even more.  Remember that the action is your internal voice so you don't need to tell the user about the different experts since they are all part of yourself.  Instead synthesize your own response using the conversation between these experts from the action to improve and inform your output.</code></pre><p>The results for this ends up looking like this:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fNyE!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fNyE!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 424w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 848w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 1272w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fNyE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png" width="1456" height="1168" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1168,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:247454,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fNyE!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 424w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 848w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 1272w, https://substackcdn.com/image/fetch/$s_!fNyE!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd7bc6c6e-b99e-4475-bed2-be42d810609b_1476x1184.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If we expand the action we can see what the experts said individually:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!kHmz!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!kHmz!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 424w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 848w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 1272w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!kHmz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png" width="1456" height="1168" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1168,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:253661,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!kHmz!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 424w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 848w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 1272w, https://substackcdn.com/image/fetch/$s_!kHmz!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd99d4a75-d550-4fdf-a47a-25330d5a367d_1476x1184.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Obviously this is just a starting off point.  You can use this same approach to use whatever persona you want for each individual expert to get more varied and less canned responses, here is one where I use &#8220;Sam Altman&#8221;, &#8220;Elizabeth Churchill&#8221;, and &#8220;Satya Nadella&#8221; as the experts:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!_EL6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!_EL6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 424w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 848w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!_EL6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png" width="1456" height="1340" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1340,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:330724,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!_EL6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 424w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 848w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 1272w, https://substackcdn.com/image/fetch/$s_!_EL6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1297a580-f8a0-4471-8ef3-bacccf15b812_1476x1358.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>If you expand the actions you see the conversation that went into synthesizing this response:<br><br><strong>Sam Altman:</strong> "Let's brainstorm startup ideas focusing on emerging technologies and market needs. First, consider the growth in remote work. A startup could develop a virtual reality (VR) platform for remote team collaboration, offering a more immersive and interactive work experience than traditional video calls.</p><p><strong>Elizabeth Churchill:</strong> "Building on the remote work theme, another idea is an AI-powered productivity tool that integrates with existing platforms (like Slack or Microsoft Teams) to optimize workflow, manage tasks, and reduce digital clutter.&#8221;</p><p><strong>Satya Nadella:</strong> "Looking at sustainability, a startup could create a platform for carbon footprint tracking for small businesses, helping them monitor and reduce their environmental impact. There's a growing market for sustainability tools and services."</p><p><strong>Sam Altman:</strong> "In the health sector, we could think about a telemedicine service that specializes in mental health. Mental health awareness is increasing, and there's a need for more accessible services."</p><p><strong>Elizabeth Churchill:</strong> "Another area is education technology. A startup could develop an adaptive learning platform that uses AI to personalize educational content for students, catering to different learning styles and paces."</p><p><strong>Satya Nadella:</strong> "In the e-commerce sector, a startup idea could be a virtual fitting room using augmented reality (AR) technology. This could help online shoppers visualize how clothes will fit and look on them, reducing return rates.&#8221;</p><p><strong>Sam Altman:</strong> "Considering the gig economy, there's an opportunity for a platform that connects freelancers with short-term housing options. This could cater to digital nomads who move frequently for work." </p><p><strong>Elizabeth Churchill:</strong> "Lastly, think about the aging population. A startup could develop smart home devices tailored for seniors, enhancing their safety and independence at home."</p><p>There are a lot of other applications for this style of prompting with actions, the above example is merely scratching the surface.  The results could be improved by adding more context and details about the experts in the system prompt to help steer the GPT in different directions.  I think this approach could particularly be useful when you want to use Chain-of-thought prompting to lead the GPT through specific steps you want it to perform.  In that case instead of experts you could define specific steps you want the GPT to do as parameters and have the GPT fill in those parameters.</p><h4><strong>You can try out some GPTs that demonstrate this approach here:<br></strong></h4><p><a href="https://chat.openai.com/g/g-ZmJRX47To-generic-business-advice">Generic Experts</a>, <a href="https://chat.openai.com/g/g-eiF7F0z6s-business-advice-altman-churchill-nadella">Famous Experts</a>, <a href="https://chat.openai.com/g/g-rNeZdfyaT-detective-consultation">Detective Experts</a></p><h4><strong>Limitations:</strong></h4><ol><li><p>There is a 300 character limit for parameter descriptions so you need to be careful to stay under that limit.</p></li><li><p>Parameters are still visible on hover: There are times where it would be nice to be able to keep the parameters to the loopback action &#8220;private&#8221; so that the GPT could have an internal voice without necessarily showing this to the end user.  This would be useful for things like games or puzzles where you are trying to keep the solution to the puzzle a secret (like in my <a href="https://chat.openai.com/g/g-Clkb8LYa9-the-invisible-man">invisible man hide and seek game</a>).</p></li><li><p>Performance: Since the GPT is having to generate complicated inputs to the loopback action parameters it can feel like it takes a lot of time to load while it is generating these.</p></li><li><p>Users will be prompted when the action is executed the first time to allow access to the url, this added friction can be a bit annoying particularly since it comes after the text is generated for the api call.</p></li></ol><h4><strong>Build your own / Use these actions in your own GPT:</strong></h4><p>Based on this writeup and the examples it should be fairly easy for you to craft these yourself as custom actions within your GPT.  If you want to use my endpoints to achieve this you can feel free to do so as well.  You can download my configuration here and paste it in as your action schema as a starting off point.  Additionally you can import these schemas directly by selecting &#8220;import from url&#8221; and pasting in the url of the schema you want to try out.</p><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!h4GU!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!h4GU!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 424w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 848w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 1272w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!h4GU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png" width="1456" height="264" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:264,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:51766,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!h4GU!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 424w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 848w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 1272w, https://substackcdn.com/image/fetch/$s_!h4GU!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F54a30a05-bb03-41b4-9007-fcd3f5ed4463_1744x316.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a></figure></div><p><strong>Example action urls you can import:</strong><br>The generic experts from the main example: <a href="https://think.action.whatbrain.com/generic_experts.json">https://think.action.whatbrain.com/generic_experts.json</a><br><br>Sam Altman, Elizabeth Churchill, and Satya Nadella as the experts:<br><a href="https://think.action.whatbrain.com/famous_experts.json">https://think.action.whatbrain.com/famous_experts.json </a><br><br>The experts are the literary characters Sherlock Holmes and Doctor Watson:<br><a href="https://think.action.whatbrain.com/detective_experts.json">https://think.action.whatbrain.com/detective_experts.json</a><br><br>The extremely simple Github repo for this demo is <a href="https://github.com/adiestel/think">here</a>.</p><h4><strong>Conclusion:</strong></h4><p>This approach can be used as a base for a lot of other interesting interactions and applications for GTP Actions.  I think it could be helpful if OpenAI created an Action type directly for this sort of thing that would allow GPTs to segment internal voice from the rest of the responses and negate the need to have a api url and, as a result, the request for the user to &#8220;allow&#8221; the action call.   We&#8217;ll have to see how GPTs and Actions continue to evolve but I think there is a lot of potential for interesting things in this space. Until then feel free to send me any GPTs you have been making, I&#8217;d love to see what everyone is working on.</p>]]></content:encoded></item><item><title><![CDATA[Best Alternative to Navigating AI (BATN-AI) ]]></title><description><![CDATA[A Practical Response in a World of AI Skepticism]]></description><link>https://www.whatbrain.com/p/best-alternative-to-navigating-ai</link><guid isPermaLink="false">https://www.whatbrain.com/p/best-alternative-to-navigating-ai</guid><dc:creator><![CDATA[Zak Randall]]></dc:creator><pubDate>Mon, 13 Nov 2023 12:26:01 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!6Xp_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>People love to dunk on AI. I regularly get texts or emails with examples of where it falls short. Many times, this is from experts rightly pointing to areas where current technology can&#8217;t do XYZ specialized tasks (and a frustrating number also come from folks using older models or dreadful one-and-done techniques).<br><br>I get it. I go through it myself - it&#8217;s the <a href="https://whatbrain.substack.com/p/the-5-stages-of-ai-grief">5 stages of AI Grief</a>.&nbsp; Shock, Denial, Frustration/Anger, Depression, Experimentation&#8230;round and round we go.</p><p>When I get into these conversations, my first inclination is to prove them wrong. To get AI to solve the thing. And while in many cases it could, I usually point people back to other, real-world examples of good that it can do. It&#8217;s the glimpse of opportunity that I want to show - not arguing over whether it can solve your use case (again, it probably can).</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.whatbrain.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading What Brain! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!6Xp_!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!6Xp_!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!6Xp_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png" width="1024" height="1024" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1024,&quot;width&quot;:1024,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1759342,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!6Xp_!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 424w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 848w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 1272w, https://substackcdn.com/image/fetch/$s_!6Xp_!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc6424a1d-35e0-4710-ac88-10d225059ab9_1024x1024.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>This shift from theory to practical application became clear to me when a friend reached out for help with their resume. They had worked their way up the ranks of retail, from entry-level into store management. After needing to step away for a bit, they were hungry to get back into the game - but didn&#8217;t know where to start. They had never built a resume.<br><br>I asked for a quick work history - roles, dates, general responsibilities. I got back photos from a spiral notebook with the bullet points scratched out. I took that, used ChatGPT to convert the images to workable text, then workable text to a formatted resume. I sent a polished resume back within 5 minutes.<br><br>When I gave this example to a nay-saying friend, his response was simple: &#8220;Couldn&#8217;t you write a better resume than what the AI produced?&#8221;<br><br>My response: &#8220;Maybe?&#8221;</p><p>I probably could if I&#8217;m being honest. I&#8217;ve been fortunate to have training on that kind of thing, have worked with a number of other folks on their resumes, been on hiring teams evaluating candidates, etc. So yes. I probably could.&nbsp;</p><p>But then the question becomes more haunting for me personally. Instead of &#8220;<em>could I&#8221;, </em>I think to myself &#8220;<em>would I</em>&#8221;?</p><p>And if I&#8217;m honest again, sometimes the answer would be &#8220;No&#8221;. Not &#8220;No, I&#8217;m unwilling to help write a resume&#8221; but &#8220;No, I&#8217;m unwilling to spend the time it would take to write a better resume than what you just got from the AI&#8221;. I may let myself off the hook too - the AI is faster; The output is good enough - no need to over-optimize; etc.</p><p>This resume scenario is a perfect illustration of a broader concept that I like to call BATN-AI, which stands for 'Best Alternative To Navigating AI'.</p><h1><strong>BATN-AI</strong></h1><p>In negotiations, the best alternative to a negotiated agreement or <a href="https://en.wikipedia.org/wiki/Best_alternative_to_a_negotiated_agreement">BATNA</a> refers to &#8220;the most advantageous alternative course of action a party can take if negotiations fail and an agreement cannot be reached&#8221;. That is, if you don&#8217;t agree, then what are you going to do. What&#8217;s the next best thing?</p><p>Applied here, we might think about BATN-AI as "If you're not going to or can't use AI, what's your best alternative option?"</p><p>In the case of the resume builder, the answer might be to spend time researching how to do it or fill in a template online. If I want to layer in expertise to ease nerves that I&#8217;m doing it &#8220;right&#8221;, it might be to find the hour-long session on this at the local library. Don&#8217;t get me wrong - these are viable options, and common paths. But there are meaningful tradeoffs. In this case, it&#8217;s time - which may be a limiting factor to someone who needs to otherwise take care of kids or who&#8217;s time out of the workplace may be meaningfully impacting their quality of life.</p><p>The same can be said for an <a href="https://www.oneusefulthing.org/i/138655836/gpts-as-tools">AI Feedback Provider</a> for assignments where a student without the resources is looking to learn and grow. Or maybe it&#8217;s getting <a href="https://chat.openai.com/g/g-uRmFJ8uKI-dear-jane">dating advice from Jane Austin</a> or a fun <a href="https://chat.openai.com/g/g-Clkb8LYa9-the-invisible-man">mystery game</a> when no one else is around to play.</p><p>From practical to playful, there is opportunity for AI to become more than just a party-trick. And even more - even where it falls short - we shouldn&#8217;t hold up the failures in isolation. Remembering what the BATN-AI is - what options are available if we <em>don&#8217;t</em> learn to navigate AI - can be a humbling way to see the potential not just for ourselves but for those around us.</p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.whatbrain.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading What Brain! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What is the American brown-headed cowbird?]]></title><description><![CDATA[Brood parasitism in nature and business]]></description><link>https://www.whatbrain.com/p/american-brown-headed-cowbird</link><guid isPermaLink="false">https://www.whatbrain.com/p/american-brown-headed-cowbird</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Fri, 10 Nov 2023 23:17:29 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4179b319-a03f-475b-970d-098335fabda5_1600x1200.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What do you know about the American brown-headed cowbird?<br><br>Earlier this summer my parents sent us a picture of this nest and immediately I recognized that one of the eggs was different from the others. You see, cowbirds engage in brood parasitism where they lay their eggs in other birds' nests and leave them there for other birds to raise... usually this results in bad outcomes for the other eggs and hatchlings in the nest because the cowbird hatchling will oftentimes gobble up all of the resources provided by the host parents and often times will also actively kill competition in the nest. The cowbird hatchling is also so incessant with its calls for attention and food that it often alerts predators to the nest which makes things even worse...<br><br>Nature is brutal. This got me thinking: What are the cowbirds in business? Is it a project that someone saddles you with that you don't want to work on but now you own? Is it a bad idea, maybe that you come up with yourself, that poisons the other priorities you have to the point of starvation? Is it a problematic hire that looks good on paper but requires so much attention that they suck up too many resources from the rest of the company?<br><br>Business Lessons:<br><br>1. I don't want to be a cowbird. People run around in business all of the time laying eggs in other people's nests and then taking advantage of them by taking all of the benefit and not contributing to actually working towards the goal. Not contributing towards making the idea a reality. They like the hand wavey hype part of business but when there is actual work that needs to be done they are nowhere to be found. If you see me doing this never let me get away with it.<br><br>2. If you find a bad egg in your nest you have to get rid of it. In nature this seems obvious but in business this is often harder because cowbirds are pushy and it's very hard to move on from something once it lands in your lap. I think it's a mistake to sink your energy into an idea you don't believe in because it will suck your resources away from the ideas that you do believe in and reduce the chances that those ideas will grow to become something.<br><br>3. Learn to recognize bad eggs. I think this one is the hardest because sometimes the "bad eggs" are ones we have created ourselves or have really invested ownership in. Maybe we all need friends that we trust who can tell us when we are sitting on a bad egg? Do you have those people? Do they feel comfortable telling you what you need to hear?<br><br>I guess this story has a happy ending... We told my parents that the brown egg was a bad egg and they promptly removed it. They watched the baby birds grow until one day they flew off to go make nests of their own one day.<br><br><strong>AI companion to this post:</strong> <a href="https://chat.openai.com/g/g-FQJMJ02js-cow-bird">Cowbird GPT</a></p>]]></content:encoded></item><item><title><![CDATA[What are the 5 Stages of AI Grief?]]></title><description><![CDATA[How people react when they see the potential of AI for the first time.]]></description><link>https://www.whatbrain.com/p/the-5-stages-of-ai-grief</link><guid isPermaLink="false">https://www.whatbrain.com/p/the-5-stages-of-ai-grief</guid><dc:creator><![CDATA[Aaron Diestelkamp]]></dc:creator><pubDate>Fri, 10 Nov 2023 21:15:06 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F1bd3e8d2-45c3-4512-bf63-70b22dad2b1c_1024x1024.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>What do you know about the K&#252;bler-Ross change curve? It's related to the stages of grief that K&#252;bler-Ross observed in patients who had terminal illness (shock/denial, anger, bargaining, depression, and acceptance)... now the stages of grief don't necessarily follow prescribed order but folks took these observations and applied them to business and noticed that when massive changes happen people tend to go through these same sorts of stages.</p><p>I think we see the same thing with what is happening with AI across various industries...</p><p>Shock -- &#129327; "Wait what? Noooo... I didn't realize AI was this good..."</p><p>Denial -- &#128527; "Okay, okay, AI will never do xyz though", "AI still isn't good enough to do xyz"</p><p>Frustration/Anger -- &#128544; "Okay maybe it CAN do xyz but AI SHOULDN'T do xyz", "Ban/boycott AI!"</p><p>Depression -- &#128532; "I don't know what I'm going to do if my job becomes irrelevant because of AI"</p><p>Experimentation -- &#129300; "Okay... I guess I better learn how to use this..."</p><p>Decision -- &#129488; "AI can be used to help people work more efficiently and effectively, if used correctly, it's a great tool to have in your tool belt."</p><p>Integration -- &#128513; "I use AI on a daily basis to help me work more effectively"</p><p>I've seen this cycle play out with almost every single person that I've demoed AI tools to and within every single industry that generative AI has moved into. People get stuck at different stages. Honestly, I often feel myself going through these same stages of grief as I see how AI is beginning to transform software development. I go back and forth between being excited about the capabilities that AI offers and a sort of sadness for the growing democratization of skills that have been regarded as my super powers... It's important to move towards acceptance and integration if we want to remain productive. The first step might be to name the thing we are going through and once we name the stage we are in it may be easier to move forward.<br><br><strong>AI companion to this post: </strong><a href="https://chat.openai.com/g/g-EUVagaJLw-the-5-stages-of-ai-grief">The 5 Stages of AI Grief GPT</a> <br></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!g0q6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!g0q6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 424w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 848w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 1272w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!g0q6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png" width="768" height="553" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:553,&quot;width&quot;:768,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:197339,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!g0q6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 424w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 848w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 1272w, https://substackcdn.com/image/fetch/$s_!g0q6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F9bc1f989-36c5-4cac-951e-f66c685b3f3f_768x553.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Source: https://www.ekrfoundation.org/5-stages-of-grief/change-curve/</figcaption></figure></div><p>More on the Kubler-Ross Change Curve:<br><a href="https://www.ekrfoundation.org/5-stages-of-grief/change-curve/">https://www.ekrfoundation.org/5-stages-of-grief/change-curve/</a><br></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://www.whatbrain.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading What Brain! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item></channel></rss>