<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Tim Suchanek]]></title><description><![CDATA[Writing about life, engineering, and AI.]]></description><link>https://suchanek.co</link><image><url>https://suchanek.co/img/substack.png</url><title>Tim Suchanek</title><link>https://suchanek.co</link></image><generator>Substack</generator><lastBuildDate>Thu, 23 Apr 2026 12:22:48 GMT</lastBuildDate><atom:link href="https://suchanek.co/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Tim]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[tim279@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[tim279@substack.com]]></itunes:email><itunes:name><![CDATA[Tim Suchanek]]></itunes:name></itunes:owner><itunes:author><![CDATA[Tim Suchanek]]></itunes:author><googleplay:owner><![CDATA[tim279@substack.com]]></googleplay:owner><googleplay:email><![CDATA[tim279@substack.com]]></googleplay:email><googleplay:author><![CDATA[Tim Suchanek]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[Building on weekends]]></title><description><![CDATA[&#8220;What the smartest people do on the weekends is what everyone else will do during the week in ten years.&#8221;]]></description><link>https://suchanek.co/p/building-on-weekends</link><guid isPermaLink="false">https://suchanek.co/p/building-on-weekends</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Mon, 13 Nov 2023 02:47:45 GMT</pubDate><content:encoded><![CDATA[<blockquote><p>&#8220;What the smartest people do on the weekends is what everyone else will do during the week in ten years.&#8221;</p></blockquote><p>&#8212; <a href="https://twitter.com/cdixon">Chris Dixon</a></p><p>A good friend of mine, Marc Mengler, just shared this fantastic quote with me. It blew me away, because it&#8217;s so true. A variation of this is &#8220;What people in SF build on the weekends will be used by everyone else during the week&#8221;. This is so true for many developer tools, hacking projects and generally startup ideas.</p><p>Given that <a href="https://www.nytimes.com/2023/06/07/technology/ai-san-francisco-tech-industry.html">SF is so back</a> given the developments in AI, I&#8217;m curious what people here, but also curious minds around the world will hack on, which will influence the lives of pretty much everyone else.</p><p>This is a short and sweet post, and will be the last of my daily posts for a while. This is post number 48. Having written daily for a while now was an amazing experience, I&#8217;ll write about what I learned another day.</p><p>However, I&#8217;ll take a break from writing daily, as I&#8217;ve decided to work on a deep implementation effort, that will need my full attention. I&#8217;ll share soon what it is about.</p><p>No worries, I&#8217;ll get back to writing, just not now.</p><p>Thanks for following my posts so far and see you soon!</p>]]></content:encoded></item><item><title><![CDATA[Unpacking Thoughts]]></title><description><![CDATA[While it can be dangerous to anthropomorphize LLMs too much, as we don&#8217;t want to end up like in the movie Her, LLMs might have more similarities with us than we think.]]></description><link>https://suchanek.co/p/unpacking-thoughts</link><guid isPermaLink="false">https://suchanek.co/p/unpacking-thoughts</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Sat, 11 Nov 2023 16:55:49 GMT</pubDate><content:encoded><![CDATA[<p>While it can be dangerous to anthropomorphize LLMs too much, as we don&#8217;t want to end up like in the movie <a href="https://en.wikipedia.org/wiki/Her_(film)">Her</a>, LLMs might have more similarities with us than we think.</p><p>In working with LLMs, there is a powerful technique called &#8220;<a href="https://www.promptingguide.ai/techniques/cot">chain-of-thought prompting</a>&#8221;. The idea is, that instead of telling the LLM to do something for us, we give it an objective and &#8220;talk out loud&#8221;, how it would go about reasoning about it. It breaks things down to smaller steps which are a chain of logical steps it follows. It turns out, that when prompting an LLM with this technique, the reasoning capabilities are improving. Just using this technique, <a href="https://arxiv.org/abs/2201.11903">researchers were able to reach state-of-the-art </a>performance with LLMs in various assessments in arithmetic, commensense and symbolic reasoning tasks. In different words - we&#8217;re forcing the LLM to use more tokens to &#8220;think&#8221;. It&#8217;s in a sense speaking out loud while trying to solve a task.</p><p>And the same approach works for us humans. There are several approaches where using more words to describe something helps us to think:</p><ul><li><p><a href="https://www.nytimes.com/2018/10/25/style/journaling-benefits.html">Journaling helps</a> us to deal with difficult emotions, but also clarify our thinking</p></li><li><p><a href="https://en.wikipedia.org/wiki/Rubber_duck_debugging">Rubber Duck Debugging</a> is a technique used in software development, where you describe a problem to a rubber duck, and just by speaking it out loud, you enable yourself to get distance from the problem and break it down to its essential parts to be able to solve it</p></li><li><p>Psychotherapy &#8212; any form of cognitive therapy involves the patient talking a lot, which helps to uncover trauma, hidden thought patterns, and get perspective</p></li><li><p>Just talking to friends can be very powerful, and hearing back from them what they think - stating the situation from a different view</p><p></p></li></ul><p>Quoting Paul Graham:</p><blockquote><p>Once you publish something, the convention is that whatever you wrote was what you thought before you wrote it. These were your ideas, and now you've expressed them. But you know this isn't true. You know that putting your ideas into words changed them. And not just the ideas you published. Presumably there were others that turned out to be too broken to fix, and those you discarded instead.</p></blockquote><p>From <a href="http://www.paulgraham.com/words.html">http://www.paulgraham.com/words.html</a></p><p>Talking thoughts out loud can be a powerful tool, not just to increase our reasoning power, but also the ones of LLMs.</p>]]></content:encoded></item><item><title><![CDATA[Semantic Caching]]></title><description><![CDATA[Imagine you're operating an application using LLMs.]]></description><link>https://suchanek.co/p/semantic-caching</link><guid isPermaLink="false">https://suchanek.co/p/semantic-caching</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Fri, 10 Nov 2023 15:23:45 GMT</pubDate><content:encoded><![CDATA[<p>Imagine you're operating an application using LLMs. If you're successful, the LLM bill can go high very quickly.</p><p>Some of the sources for a high LLM bill are:</p><ol><li><p><strong>Repeated regression tests</strong>. Although prompts didn't change, a naive implementation would send the same requests over and over to the LLM.</p></li><li><p><strong>Prompt tuning</strong>: You might want to try out new prompts with existing test data, potentially having hundreds of tests per iteration that you test.</p></li><li><p><strong>Complex workflows</strong>: Most chatbot applications don't just directly talk to the LLM, but involve external APIs, external datasources to fulfill the query.</p></li><li><p><strong>Maxing out the context</strong>: Depending on the application, you might need to use the context window fully. With GPT-4 (non-turbo), passing in 30,000 tokens (which might just be one long document) costs $1.80.</p></li></ol><p>A naive approach to solve this is using a traditional key-value cache. You take the input that you'd send to OpenAI, be it a completion or a whole conversation to a given point and do an exact string match against it. This especially helps for regression tests, which run the same prompt over and over and can both speed the tests up and save costs. However, what if we have the following prompts:</p><ol><li><p>"How to get started with Effect"</p></li><li><p>"Effect getting started"</p></li><li><p>"Getting started with Effect"</p></li></ol><p>They all denote the exact same intent. I ran them all through one of the state-of-the-art embedding models, <a href="https://huggingface.co/BAAI/bge-large-en-v1.5">bge-large-v1.5</a>, and got these results:</p><p><code>Cosine Similarity between 1 and 1: 1</code></p><p><code>Cosine Similarity between 1 and 2: 0.9178054403912372</code></p><p><code>Cosine Similarity between 1 and 3: 0.981772531149283</code></p><p><code>Cosine Similarity between 2 and 2: 1</code></p><p><code>Cosine Similarity between 2 and 3: 0.9321289492871373</code></p><p><code>Cosine Similarity between 3 and 3: 1</code></p><p>Although it&#8217;s the exact same intent, it&#8217;s surprising to see the gap in cosine similarity between 1 and 2. I was wondering - how we could &#8220;nudge&#8221; the embeddings to be more similar, as what we care about is if they have the same intent. Instead of directly embedding the words as they are, I instead took the following variations:</p><ol><li><p>The intent of saying "How to get started with Effect"</p></li><li><p>The intent of saying "Effect getting started"</p></li><li><p>The intent of saying "Getting started with Effect"</p></li></ol><p>which yields these results:</p><p><code>Similarity between 1 and 1: 1</code></p><p><code>Similarity between 1 and 2: 0.9351850391263539</code></p><p><code>Similarity between 1 and 3: 0.9798117351520749</code></p><p><code>Similarity between 2 and 2: 1</code></p><p><code>Similarity between 2 and 3: 0.9662336980291407</code></p><p><code>Similarity between 3 and 3: 1</code></p><p>While the distance between 1 &amp; 3 is slightly lower, now 1 &amp; 2 moved together. I can also imagine specialized embeddings fine-tuned for this specific use-case of semantic caching.</p><p>Instead of simply doing a K/V lookup, we can get the embedding for the input vector first, and do a lookup against a vector store. If the similarity is above a certain predefined threshold, return the result &#8212; we have a cache hit.</p><p>As we see more applications of the <a href="https://suchanek.co/p/one-of-the-most-powerful-concepts">Voyager</a> approach, such a fuzzy or semantic cache can also be used as a way to look up skills. Is this a situation we've seen, just with very little difference? Let's take the tools out of our tool library we already created before to deal with this situation.</p>]]></content:encoded></item><item><title><![CDATA[Framing]]></title><description><![CDATA[In old-school economics, it was assumed that humans are rational actors, acting based on selfish motives, and that by understanding this, we would be able to predict economic outcomes.]]></description><link>https://suchanek.co/p/framing</link><guid isPermaLink="false">https://suchanek.co/p/framing</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Thu, 09 Nov 2023 15:53:51 GMT</pubDate><content:encoded><![CDATA[<p>In old-school economics, it was assumed that humans are rational actors, acting based on selfish motives, and that by understanding this, we would be able to predict economic outcomes. However, this mentality baffled <a href="https://en.wikipedia.org/wiki/Daniel_Kahneman">Daniel Kahneman</a>, who together with <a href="https://en.wikipedia.org/wiki/Amos_Tversky">Amos Tversky</a> researched this topic. Kahneman, who from years in<br>psychology knew that, in his words,</p><blockquote><p>&#8220;[I]t is self-evident that<br>people are neither fully rational nor completely selfish, and<br>that their tastes are anything but stable.&#8221;</p></blockquote><p>Through decades of research with Tversky, Kahneman demonstrated that all humans suffer from cognitive biases&#8212;unconscious, irrational brain processes that distort how we see the world.</p><p>Kahneman and Tversky discovered more than 150.</p><p><br>The <strong>Framing Effect</strong> is one of them, which demonstrates that people respond differently to the same choice depending on how it is framed (people place greater value on moving from 90 percent to 100 percent&#8212;high probability to certainty&#8212;than from 45 percent to 55 percent, even though they&#8217;re both ten percentage points).</p><p>This made me think of the latest framing around OpenAI&#8217;s announced <strong><a href="https://techcrunch.com/2023/11/06/app-store-for-ai-build-your-own-gpt-and-sell-it-on-openais-gpt-store/">GPT Store</a>,</strong> which will be the app store to build your chatbots, which you can publish and monetize. The part that I found particularly interesting was the framing of &#8220;Revenue Share&#8221;. That&#8217;s a very interesting framing, almost a euphemism for nothing else than &#8220;Platform fees.&#8221;</p><p>Apple charges 30% of all revenue coming from iOS apps in fees. It is, after all, a fee &#8212; people pay money to use the creation of someone who sat down and spent time creating it. Don&#8217;t get me wrong - without Apple and in this case, without OpenAI it wouldn&#8217;t be possible for anyone to even create something like this. They make the platform. It&#8217;s interesting to see how the framing here is focused on the positive part - &#8220;Revenue&#8221; - yes, people want revenue, and currently, being in SF, the epicenter of AI developments, I can feel the gold rush atmosphere. There&#8217;s almost a fever going on (for sure non-zero cognitive biases playing a role), people frantically trying to keep up with OpenAI and the opportunities that open up. And this little word &#8220;share&#8221;, which denotes the existence of a fee, is so innocent on the side. The messaging seems to work. People are excited, and in the last 48 hours, authors created chat bots to chat with their books. Some people try to make their personal assistants. Others replace their hand-crafted RAG system with a custom GPT. Let&#8217;s see how much revenue OpenAI will share, but I wouldn&#8217;t be surprised if it&#8217;s 70%, as with Apple.</p>]]></content:encoded></item><item><title><![CDATA[Closing the loop]]></title><description><![CDATA[I recently asked a friend working at OpenAI what kind of startup he'd found if he were to build one today.]]></description><link>https://suchanek.co/p/closing-the-loop</link><guid isPermaLink="false">https://suchanek.co/p/closing-the-loop</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Wed, 08 Nov 2023 15:18:59 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!60nm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>I recently asked a friend working at OpenAI what kind of startup he'd found if he were to build one today. His answer was simple: I'd solve a specific problem where I can collect training data from users, with which I can fine-tune a model on top of OpenAI, which increases the quality of the product, which brings in more users, which gets you more training data, which lets you further fine-tune the model, etc.</p><p>Midjourney is a staggering example of this. Its quality is still way superior to Dalle-3. How is that possible? One thing is that they're focused but also able to constantly increase the quality by getting user feedback.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!60nm!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!60nm!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 424w, https://substackcdn.com/image/fetch/$s_!60nm!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 848w, https://substackcdn.com/image/fetch/$s_!60nm!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!60nm!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!60nm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg" width="1456" height="766" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:766,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:171486,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!60nm!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 424w, https://substackcdn.com/image/fetch/$s_!60nm!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 848w, https://substackcdn.com/image/fetch/$s_!60nm!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!60nm!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1d272ad-ec4b-44b7-8c81-0347a9ba436f_1882x990.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p><a href="https://twitter.com/DrJimFan/status/1643279641065713665">https://twitter.com/DrJimFan/status/1643279641065713665</a></p><p>By choosing which picture to upscale, you give Midjourney feedback on which option you prefer, from where they can further train their models.</p><p>How would such a loop look like for LLM-based applications? What's the journey from starting with a prompt to a closed loop, which can even tune itself?</p><ol><li><p>Create a prompt, by using one of the prompt playgrounds out there, such as <a href="https://baserun.ai/">baserun.ai</a>'s and test it with a few test cases that you believe users will have</p></li><li><p>Deploy the prompt to production, get the first users</p></li><li><p>With the first users providing actual input, collect that input and turn it into evaluation data</p></li><li><p>With this evaluation data, you can now do a couple of things:</p><ol><li><p>Further improve the prompt</p></li><li><p>Fine-tune a model</p></li><li><p>Adjust your <a href="https://suchanek.co/p/rag-an-old-idea-in-a-new-coat">RAG</a> approach (reordering etc)</p></li></ol></li><li><p>Deploy the changes</p></li><li><p>Get more user feedback, which adds more evaluation data</p></li><li><p>Profit</p></li></ol><p>This is one of the ways to build a moat. By constantly improving on a hard problem after a few iteration cycles, you can already be miles ahead of where you started. What if there was a tool that could do this tuning for you automatically, even deploying the changes? We'll soon see this tooling, and I'm excited to see applications improve and become more useful for us humans!</p>]]></content:encoded></item><item><title><![CDATA[One of the most powerful concepts in AI agents: Voyager]]></title><description><![CDATA[Not even half a year ago, a revolutionary paper was released in the field of autonomous agents research that, in my opinion, hasn&#8217;t gotten enough attention given the magnitude of impact it will have.]]></description><link>https://suchanek.co/p/one-of-the-most-powerful-concepts</link><guid isPermaLink="false">https://suchanek.co/p/one-of-the-most-powerful-concepts</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Tue, 07 Nov 2023 14:46:41 GMT</pubDate><content:encoded><![CDATA[<p>Not even half a year ago, a revolutionary paper was released in the field of autonomous agents research that, in my opinion, hasn&#8217;t gotten enough attention given the magnitude of impact it will have. I&#8217;m talking about <strong><a href="https://arxiv.org/pdf/2305.16291.pdf">VOYAGER: An Open-Ended Embodied Agent with Large Language Models</a>.</strong></p><p>The authors from NVIDIA, Caltech, UT Austin, Stanford, UW Madison, including <a href="https://jimfan.me/">Jim Fan</a>, have built an agent that can teach itself to play Minecraft. How does it work?</p><ol><li><p>Explore the world, and whenever you find yourself in a new situation, reason about the situation and come up with an idea of what tool to build or use.</p></li><li><p>Try to use or build the tool. At this point, they have an iterative algorithm, which tries over and over, with the agent self-verifying if the tool was successfully created or the action successfully performed. Example for this: <em>Reasoning: Since you have a wooden pickaxe and some stones, it would be beneficial to upgrade your pickaxe to a stone pickaxe for better efficiency. Task: Craft 1 stone pickaxe.</em></p></li><li><p>Once the agent verifies that it successfully performed the task, it will save it in a skill library.</p></li><li><p>The next time the agent faces a similar situation, it can take the skill from the library instead of having to create it again.</p></li></ol><p>It uses GPT-4 as its reasoning engine. With this simple mechanism, Voyager was able to beat the state of the art by orders of magnitude. It allows the agent to explore the world in a self-supervised fashion without the need for human intervention.</p><p>Now the obvious question is - where else can one apply the Voyager approach? Robotics, Healthcare, Education, Environmental Tracking, Smart Home assistants - the possibilities are endless.</p><p>For a good reason, this paper won the <a href="https://neurips.cc/virtual/2022/awards_detail">NeurIPS Outstanding Paper</a> Award.</p>]]></content:encoded></item><item><title><![CDATA[OpenAI's T+1 expansions]]></title><description><![CDATA[Any product idea that you&#8217;ll ever see is always coming from a T+1 expansion.]]></description><link>https://suchanek.co/p/openais-t1-expansions</link><guid isPermaLink="false">https://suchanek.co/p/openais-t1-expansions</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Tue, 07 Nov 2023 03:28:31 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!fMjN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Any product idea that you&#8217;ll ever see is always coming from a T+1 expansion. Given the world today, including its problems and solutions that already exist, what can a better solution look like? We humans always combine existing things. Nobody can just "skip" a step. However, when doing the iteration quickly, you can have impressive progress in a short amount of time that from outside is often hard to follow and looks like a big jump.</p><p>So is the progress of OpenAI &#8212; it&#8217;s a chain of T+1 expansions, and by being incredibly fast in bringing one expansion after another, a few years in - the territory that OpenAI covers is vast.</p><p>I'm imagining this like a game of Warcraft or Settlers. You have a fog of war, which is area that is dark, you can't see it and with that you also don't have conquered it. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!fMjN!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!fMjN!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 424w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 848w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 1272w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!fMjN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png" width="1050" height="673" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:673,&quot;width&quot;:1050,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:804607,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!fMjN!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 424w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 848w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 1272w, https://substackcdn.com/image/fetch/$s_!fMjN!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F5ccfda1c-e432-47e9-a226-60b5984cfc08_1050x673.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Now, let's put ourselves into the shoes of a gen ai founder who just looked at OpenAI's offerings this year after GPT-4 came out and tried to find an opportunity, in other words, trying to find dark spots on the map, which OpenAI doesn&#8217;t control yet. Some of the obvious places that were still "dark spots" on OpenAI's map and how OpenAI just expanded on them today:</p><ol><li><p><strong>Customized models</strong> for enterprises that can run on-prem. <a href="https://en.wikipedia.org/wiki/Emad_Mostaque">Emad Mostaque</a>, CEO of StabilityAI observed this and wanted to collaborate with companies all around the world, counter-positioning OpenAI with decentralized models. This is now covered by custom models you can train with OpenAI for $2-3M. Yes it's a lot of money, but for the value these models will provide, OpenAI understands that this price is probably even cheap.</p></li><li><p><strong>Saving costs:</strong> Finding a way to get the same quality of service for less money. <a href="https://openpipe.ai/">OpenPipe</a> is one example - training smaller models based on GPT-4 responses to save costs. Some use-cases here are now less urgent, as GPT-4 Turbo has a 3x cost reduction.</p></li><li><p><strong>Chat with your data:</strong> OpenAI just released a whole SDK what makes it easy to chat with your data. You can now upload any file (they're supporting most mime types) which OpenAI will turn into a useful chat bot for you. You can connect any internal functions and call them with the same assistant. While this is far from an autonomous agent, it's covering the most popular use-case of LangChain and LlamaIndex: Talk to your documents.</p></li><li><p><strong>Well-structued output:</strong> Getting a reliable JSON output from OpenAI was hard until recently, and you needed to use threatening language like "I'll take a life" - this is not necessary anymore and we can now use GPT-4 Turbo reliably as an API.</p></li><li><p><strong>Long Context:</strong> I used Claude 2 quite a bit to chat with larger PDFs such as books. This is not needed anymore with Assistants but also GPT-4 Turbo 128k context in general.</p></li></ol><p>Given, that OpenAI has now expanded further in all these directions today, what will come next? What is another T+1 step from here? It's clear that when building a startup today if the startup is within the T+1 reach of OpenAI for its next iteration cycle, there's a significant risk involved.</p><p>So, what are some ways to be protected against that? How can you build a startup today that's not prone to being "eaten" or "killed" as people like to say on <s>Twitter</s> X by the next round of OpenAI announcements?</p><p>I'll write about that in the coming days. For now, let's just acknowledge the speed at which OpenAI is shipping. This is truly inspiring.</p>]]></content:encoded></item><item><title><![CDATA[Inevitable AI developments]]></title><description><![CDATA[Three of the most impactful inevitable trends shaping the future are cognifying, tracking, and questioning.]]></description><link>https://suchanek.co/p/inevitable-ai-developments</link><guid isPermaLink="false">https://suchanek.co/p/inevitable-ai-developments</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Mon, 06 Nov 2023 04:39:52 GMT</pubDate><content:encoded><![CDATA[<p>Three of the most impactful inevitable trends shaping the future are <strong>cognifying, tracking, and questioning</strong>.</p><p>Cognifying refers to making previously "dumb" objects and processes smarter through technology like AI. As more objects and systems become cognified, they will be able to operate with less human input and oversight. While this could lead to greater efficiency, it also raises concerns around transparency and control. How can we ensure cognified systems act ethically if we don't fully understand how they make decisions?</p><p>Tracking is also increasing rapidly with new technologies. While customized services and experiences can benefit individuals, invasive tracking threatens privacy. There is a difficult balance between utilizing data to improve lives versus respecting personal boundaries. The onus should be on companies and governments to collect only essential data and be fully transparent about how it is used.</p><p>Finally, questioning has become ubiquitous due to the overwhelming amount of contradictory information online. While healthy skepticism is positive, pervasive questioning of facts and expertise threatens to undermine institutions and spread misinformation. We must find ways to separate truth from fiction through critical thinking and responsible gatekeeping.</p><p>Overall, we must thoughtfully guide these technological changes to create an equitable and enlightened future society. With care, AI and related advances can enhance our lives without dehumanizing us in the process. As outlined in Kevin Kelly's book <a href="https://www.amazon.com/Inevitable-Understanding-Technological-Forces-Future/dp/0143110373">The Inevitable</a>, these trends will shape the future whether we want them to or not, so it is up to us to direct them wisely.</p>]]></content:encoded></item><item><title><![CDATA[The potential of GPT4-V]]></title><description><![CDATA[As we head toward the OpenAI Dev Day on Monday, people are curious about what they&#8217;ll release.]]></description><link>https://suchanek.co/p/the-potential-of-gpt4-v</link><guid isPermaLink="false">https://suchanek.co/p/the-potential-of-gpt4-v</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Sat, 04 Nov 2023 15:46:57 GMT</pubDate><content:encoded><![CDATA[<p>As we head toward the <a href="https://devday.openai.com/">OpenAI Dev Day</a> on Monday, people are curious about what they&#8217;ll release. Some of the common predictions I&#8217;ve seen so far:</p><ol><li><p>GPT4-Vision API</p></li><li><p>Stateful API</p></li><li><p>GPT4 fine-tuning</p></li><li><p>GPT4 turbo</p></li><li><p>Dalle3 API</p></li><li><p>Code interpreter API</p></li><li><p>Agent framework</p></li></ol><p>I see GPT4-Vision API as the most likely option right now, as they already demoed the API at the AI Engineer Summit. While at the summit, OpenAI had a neat little demo around blog post generation &#8212; what can it unlock in the real world?</p><p>Some ideas on how GPT4-V could be used for good:</p><ol><li><p><strong>Build your own rewind.ai:</strong> <a href="https://www.rewind.ai/">Rewind AI&#8217;s</a> interface to gather context about your life is through vision. It does regular screenshots and analyses the data. Through that, Rewind becomes very useful over time, being able to ask questions about anything we saw. While you can build this on your own with OCR, OCR wouldn&#8217;t allow you to ask semantic questions about an image, such as &#8220;What can we see here?&#8221; &#8220;What&#8217;s important in this screenshot given my context?&#8221;. While typical RAG applications vectorize text from websites, APIs, and databases and do similarity searches on it, it&#8217;ll be hard to have a fully comprehensive AI agent without tapping into the most general API of them all: Vision. There&#8217;s more to building your own rewind.ai, but the components to do so are being commoditized as we speak.</p></li><li><p><strong>Improved browser automation</strong>: Ultimately, browser engines render a website to be consumed by a human, not a machine. While we can read HTML code with machines and try to make sense of it, at some point it&#8217;ll always hit limitations, such as z-index calculations. Being able to use text and image will allow folks to build powerful scrapers and browser automation moving forward.</p></li><li><p><strong>Generative UI:</strong> When asking GPT4-V to implement a UI from a screenshot, it&#8217;s already surprisingly good at it. While the results might not be production-ready yet, it&#8217;s showing us the potential of the technology. Instead of starting at 0 as with <a href="https://v0.dev/">v0.dev</a>, what if you can provide your favorite website designs as a mood board to start with and let the AI do the rest?</p></li><li><p><strong>Turn images into APIs:</strong> When having an image of groceries, GPT-4V was able to return perfect JSON that contains a list of items, including their calories. With GPT4-V, we&#8217;ll be able to turn any image or video into an API that can be programmatically accessed and used as a building block for a new frontier of apps. It&#8217;s obvious that Llamaindex and Langchain will also add support for this.</p></li></ol><p>The potential of GPT4-V is endless. However, as with any new technology, it&#8217;s not yet perfect. A recent study found that <a href="https://arxiv.org/pdf/2310.16809.pdf">OCR models still beat GPT4V</a>, and a <a href="https://evren.ninja/gpt4-vision-prompt-injection.html">certain level of prompt injections</a> is possible through image text. It&#8217;s also clear, that it&#8217;s not affordable to run GPT4-V over millions of images. A blockbuster has about 300k-600k frames. Turning a whole movie into an API might still be a bit far out. However, we also know that costs will always decrease, and models get faster.</p><p>What can one build today that&#8217;s just not yet possible but will be possible with this technology soon?</p>]]></content:encoded></item><item><title><![CDATA[Signs of shortsightedness]]></title><description><![CDATA[In our busy lives, we often find ourselves caught up in the &#8220;now,&#8221; losing sight of the bigger picture.]]></description><link>https://suchanek.co/p/signs-of-shortsightedness</link><guid isPermaLink="false">https://suchanek.co/p/signs-of-shortsightedness</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Sat, 04 Nov 2023 04:22:04 GMT</pubDate><content:encoded><![CDATA[<p>In our busy lives, we often find ourselves caught up in the <strong>&#8220;</strong>now<strong>,&#8221;</strong> losing sight of the bigger picture. This myopia can cause us to stray from our long-term goals.</p><p>Here are five signs of shortsightedness and how to correct them.</p><ol><li><p><strong>Unintended Consequences</strong>: Acting hastily can lead to unforeseen outcomes. Solution? Take a moment to think through the potential fallout before deciding. Assign a skeptic in your group to play devil&#8217;s advocate, analyzing the possible consequences of a decision.</p></li><li><p><strong>Tactical Hell</strong>: Getting caught up in minor battles can distract you from your major goals. If you find yourself stuck in petty squabbles, step back, calm your ego, and refocus on your long-term objectives. Remember, actions speak louder than words.</p></li><li><p><strong>Ticker Tape Fever</strong>: In the age of real-time updates, we often react impulsively. Slow down, take a breath, and give yourself time to process information. Like Abraham Lincoln, exercise patience to see the bigger picture and make better decisions.</p></li><li><p><strong>Lost in Trivia</strong>: Overloading with too much information can muddle your thinking. Establish a hierarchy of priorities and focus on what truly matters. Delegate lesser tasks and avoid drowning in trivial details.</p></li><li><p><strong>Fear of Missing Out (FOMO)</strong>: The anxiety of not keeping up with trends or others' activities can lead to rushed, ill-considered decisions. Combat this by reminding yourself of your personal goals and values. It's okay to march to the beat of your own drum, prioritizing what's crucial for your long-term vision over fleeting trends.</p></li></ol><p>By recognizing and addressing short-sighted thinking, you'll foster a clearer, broader perspective that will steer you closer to your long-term ambitions. With a balanced view and patient approach, you'll navigate through life's challenges more adeptly, making decisions that resonate well into the future.</p><p>Some of these come from <a href="https://www.amazon.com/Laws-Human-Nature-Robert-Greene/dp/0525428143">Laws of Human Nature</a>, which I highly recommend to read.</p>]]></content:encoded></item><item><title><![CDATA[Do the best ideas win?]]></title><description><![CDATA[How can it be that although Stoicism is becoming more popular today, Christianity has been the predominant belief system for so long?]]></description><link>https://suchanek.co/p/do-the-best-ideas-win</link><guid isPermaLink="false">https://suchanek.co/p/do-the-best-ideas-win</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Fri, 03 Nov 2023 04:41:27 GMT</pubDate><content:encoded><![CDATA[<p>How can it be that although Stoicism is <a href="https://www.elle.com/uk/life-and-culture/culture/a42993347/stoicism-keeping-calm/">becoming more popular today</a>, Christianity has been the predominant belief system for so long? Wouldn't the best ideas always prevail?</p><p>Obviously not. There are a few reasons why Christianity gained dominance over Stoicism historically, even though Stoic ideas are resonating more in the modern era:</p><ol><li><p>Christianity gained an early boost by piggybacking on the infrastructure of the vast Roman Empire. Once Constantine converted, Christianity could spread rapidly through imperial decrees and resources. This early infrastructure advantage was critical.</p></li><li><p>Christianity appealed more to the heart than the mind. Its focus on Christ's narrative and the promise of personal salvation had greater emotional draw than Stoicism's rational emphasis on virtue and discipline.</p></li><li><p>Christianity aligned better with human psychology. Its black and white moral code and threat of divine punishment played upon innate cognitive biases in a way the nuanced ethics of Stoicism did not.</p></li></ol><p>In short, the dominance of Christianity historically does not necessarily mean it offered the "best ideas." It spread for institutional, emotional, and psychological reasons. The revival of Stoicism today shows that when ideas are considered on merit, the old Greek philosophy still has tremendous wisdom to offer the modern world.</p>]]></content:encoded></item><item><title><![CDATA[Confirmation Bias]]></title><description><![CDATA[Confirmation bias is like a pair of tinted glasses that colors the way we perceive the world around us.]]></description><link>https://suchanek.co/p/confirmation-bias</link><guid isPermaLink="false">https://suchanek.co/p/confirmation-bias</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Thu, 02 Nov 2023 03:39:50 GMT</pubDate><content:encoded><![CDATA[<p>Confirmation bias is like a pair of tinted glasses that colors the way we perceive the world around us. It's a psychological shortcut that often leads us astray, prioritizing information that aligns with our existing beliefs while ignoring the rest. This cognitive glitch can warp our understanding and decision-making. Here are three instances to paint a clearer picture:</p><ol><li><p><strong>The Cold Fusion Controversy (1989)</strong></p></li></ol><p>When Martin Fleischmann and Stanley Pons announced their discovery of cold fusion, it was like a ray of hope for a clean energy future. However, this hope made some overlook the inconsistencies in their data. Many, swayed by confirmation bias, claimed to replicate their results, choosing to ignore counter-evidence. The bubble burst when further rigorous testing failed to reproduce the same results, extinguishing the initial spark of excitement.</p><ol start="2"><li><p><strong>Political Echo Chamber</strong></p></li></ol><p>The political sphere often mirrors a large echo chamber where our beliefs bounce back at us, reinforced. A conservative might find solace in the familiar rhetoric of conservative news channels, while a liberal might resonate with the liberal ones. This selective tuning in, driven by confirmation bias, cocoons us from opposing viewpoints, reinforcing our existing beliefs while shielding us from any information that might challenge them.</p><ol start="3"><li><p><strong>Investment Decisions</strong>:</p></li></ol><p>The financial markets are a rollercoaster ride, and confirmation bias can sometimes make us cling to the wrong stocks. An investor, swayed by a single positive forecast, might ignore the storm clouds of adverse market trends, holding onto declining stocks with a false sense of hope.</p><p>These examples of confirmation bias in action show how it can subtly yet significantly skew our perception and decisions. By acknowledging the presence of this bias, we can strive towards a more balanced and evidence-based approach to our judgments and decisions, navigating through life with a clearer lens.</p><p>One way to tackle confirmation bias is with a clear <a href="https://suchanek.co/p/how-to-make-better-decisions">decision-making process</a>.</p>]]></content:encoded></item><item><title><![CDATA[AI's socratic journey]]></title><description><![CDATA[Many years ago&#847;, the Greek philosop&#847;her Socrates was to&#847;ld by &#847;the &#847;Oracle of Delphi that he was the sma&#847;r&#847;test m&#847;an alive.]]></description><link>https://suchanek.co/p/ais-socratic-journey</link><guid isPermaLink="false">https://suchanek.co/p/ais-socratic-journey</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Tue, 31 Oct 2023 13:51:45 GMT</pubDate><content:encoded><![CDATA[<p>Many years ago&#847;, the Greek philosop&#847;her Socrates was to&#847;ld by &#847;the &#847;Oracle of Delphi that he was the sma&#847;r&#847;test m&#847;an alive. Socrates wasn't so sure&#847;, so he&#847; went on a journey to &#847;talk to o&#847;ther wise people&#847; in di&#847;fferent professions.</p><p>During his conversations, he &#847;notic&#847;ed&#847; tha&#847;t these so-called wise peop&#847;le often ha&#847;d&#847; str&#847;o&#847;ng opinions about things they d&#847;id&#847;n't&#847; know much about. They had&#847;n't &#847;really &#847;thought things through. This made &#847;Socr&#847;ate&#847;s realize he was&#847; wise becau&#847;se he kn&#847;ew wh&#847;at he didn't know.</p><p>&#847;Today w&#847;e have L&#847;a&#847;rge&#847; L&#847;anguage Model (LLM). T&#847;hese are com&#847;puter programs ca&#847;pabl&#847;e of genera&#847;ting te&#847;xt that appears t&#847;o be wr&#847;itten by&#847; a human. They h&#847;ave be&#847;en trained on &#847;a lot o&#847;f &#847;dat&#847;a, which allows them to talk about&#847; a variety of topics.</p><p>Ho&#847;wever, lik&#847;e the w&#847;ise&#847; men wi&#847;th whom S&#847;ocrates spok&#847;e&#847;, LLMs sometimes giv&#847;e an&#847;swer&#847;s that se&#847;em &#847;c&#847;onfid&#847;e&#847;n&#847;t&#847; but&#847; a&#847;re in f&#847;act&#847; wron&#847;g&#847; or incomplete&#847;. Our &#847;cha&#847;l&#847;le&#847;nge now is t&#847;o help LLMs under&#847;stand what they know and what they&#847; don'&#847;t&#847; kn&#847;ow. The&#847;y <a href="https://www.beren.io/2023-03-19-LLMs-confabulate-not-hallucinate/">confa&#847;bulat&#847;e&#847;</a>.&#847;</p><p>If we can do this, LLMs will b&#847;e much &#847;more useful bec&#847;ause the&#847;y will &#847;pro&#847;v&#847;ide us with better and&#847; m&#847;ore reliable inf&#847;ormati&#847;on. In short&#847;, like Soc&#847;rates, LLMs must&#847; lear&#847;n the limits of their&#847; kn&#847;owle&#847;dge.</p><p>Any LLM that c&#847;an&#847; do &#847;th&#847;is will be t&#847;ruly w&#847;ise and&#847; w&#847;ill stand out&#847; f&#847;rom the rest. Remembering the word&#847;s o&#847;f Socrates&#847;: &#8220;I know th&#847;a&#847;t &#847;I am intelligent, bec&#847;a&#847;use&#847; I know &#847;that I know nothing&#8221; wi&#847;ll be th&#847;e key to&#847; ach&#847;ievi&#847;ng this.&#847;</p><p>The&#847; LLM t&#847;hat ca&#847;n &#847;und&#847;ers&#847;tand an&#847;d l&#847;ive by this phil&#847;osophy will&#847; indeed be &#847;the &#847;w&#847;ises&#847;t.</p>]]></content:encoded></item><item><title><![CDATA[Everything Diffusion]]></title><description><![CDATA[While Transformers are seen as the most significant innovation in AI in the last years, powering the AI summer that we&#8217;re finding ourselves in, there&#8217;s another extremely powerful technology, which is discussed less: Diffusion.]]></description><link>https://suchanek.co/p/everything-diffusion</link><guid isPermaLink="false">https://suchanek.co/p/everything-diffusion</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Mon, 30 Oct 2023 14:42:33 GMT</pubDate><content:encoded><![CDATA[<p>While <a href="https://arxiv.org/abs/1706.03762">Transformers</a> are seen as the most significant innovation in AI in the last years, powering the AI summer that we&#8217;re finding ourselves in, there&#8217;s another extremely powerful technology, which is discussed less: Diffusion.</p><p>One of the hardest and most essential parts in training neural networks are high-quality evaluations. By what metric should we train the model? Before stable diffusion, vision models would be trained on recognizing objects in images through bounding boxes, semantic segmentation, or mere instance classification. Researchers trying to create the best models measure themselves with the benchmarks that were available at the time, and the benchmarks available were not yet enabling to train models that are creating &#8220;beautiful art&#8221;.</p><p>Enter diffusion. The idea behind diffusion is quite simple. To create an evaluation function for the neural net we&#8217;re training here, we take an image as it is, apply progressively stronger random distortion, and now teach the neural net to go from a more distorted option to a less distorted version. This has been hooked up with the CLIP model, which allows vectors and text to be in the same latent space &#8594; now we can describe text and, with a multi-iteration approach, get to the point that we have a high-quality image output.</p><p>This idea of starting with something of lower quality and increasing its &#8220;resolution&#8221; or quality to train a neural net can&#8217;t just be applied to vision models. We&#8217;ll see this idea find application in all kinds of areas:</p><ul><li><p><a href="https://arxiv.org/abs/2310.17680">CodeFusion</a>: An LLM using diffusion for better code generation</p></li><li><p><a href="https://arxiv.org/abs/2208.06193">Diffusion-QL</a>: Diffusion-based policies in reinforcement learning</p></li><li><p><a href="https://arxiv.org/abs/2112.10752">Stable Diffusion</a>: Image synthesis through diffusion</p></li></ul><p>There are dozens of more applications of diffusion &#8212; a more comprehensive list can be found <a href="https://github.com/YangLing0818/Diffusion-Models-Papers-Survey-Taxonomy">here</a>. The results of the recently released <a href="https://arxiv.org/abs/2310.17680">CodeFusion</a> are very impressive. With just 75M parameters, it&#8217;s able to compete with models up to 175B parameters. This is very promising news because that means you won&#8217;t need hundreds of A100 GPUs to make an impact in AI research. What else will we be able to disrupt with diffusion? And more importantly, what will be the next big idea fully changing the game again?</p>]]></content:encoded></item><item><title><![CDATA[Breeding prompts]]></title><description><![CDATA[Nativ&#847;e AI app&#847;lica&#847;tions such as chatbots, voice a&#847;ssistants, or&#847; a&#847;gents are power&#847;ed by &#847;L&#847;LMs.]]></description><link>https://suchanek.co/p/breeding-prompts</link><guid isPermaLink="false">https://suchanek.co/p/breeding-prompts</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Mon, 30 Oct 2023 02:23:15 GMT</pubDate><content:encoded><![CDATA[<p>Nativ&#847;e AI app&#847;lica&#847;tions such as chatbots, voice a&#847;ssistants, or&#847; a&#847;gents are power&#847;ed by &#847;L&#847;LMs. Their performance depends heav&#847;ily on how the in&#847;puts or "prompt&#847;s" are formulated in the m&#847;odel.&#847; Small changes &#847;to prompts can have &#847;a sign&#847;i&#847;ficant i&#847;mpact on mo&#847;del out&#847;put.</p><p>Deep&#847;Mind research&#847;ers have fo&#847;und that providi&#847;ng re&#847;asoning h&#847;ints, w&#847;h&#847;ich guide th&#847;e model through logi&#847;cal steps, si&#847;gnifi&#847;cantly i&#847;mpr&#847;oves performance in mathematic&#847;s, i&#847;n&#847;f&#847;ormal reasoning, a&#847;nd com&#847;plex&#847; tasks.</p><p>For example, a prompt like &#8220;Let's b&#847;re&#847;ak &#847;this prob&#847;lem down ste&#847;p by st&#847;ep&#8221; will yield a more prec&#847;ise solution &#847;t&#847;han&#847; simply s&#847;tating th&#847;e problem.</p><p>T&#847;he&#847; chall&#847;enge &#847;is that&#847; ef&#847;fective remi&#847;nders must &#847;be designed &#847;manually &#847;for e&#847;ach t&#847;ask.</p><p>This ne&#847;w researc&#847;h introduce&#847;s &#847;a &#847;method called Si&#847;milar Reminding, which aut&#847;omatical&#847;ly devel&#847;ops better &#847;sugg&#847;esti&#847;ons &#847;for a given problem.</p><p>Here's h&#847;ow i&#847;t wor&#847;ks:</p><ol><li><p>The &#847;reminde&#847;r is&#847; initialized with a general des&#847;c&#847;ription of the ta&#847;sk (e.g., solving mat&#847;h word problems)&#847;, a&#847; "th&#847;inki&#847;ng style" prompt (e.g.&#847;, &#847;di&#847;viding min&#847;or problem), &#847;and "mutant prompts" that c&#847;reate variati&#847;on&#847;s of prom&#847;p&#847;ts&#847;.</p></li><li><p>I&#847;t &#847;generate&#847;s a set of variations on &#847;the fly a&#847;nd tests th&#847;em&#847; on sample problems &#847;to dete&#847;rmine their &#847;"&#847;fit" or performance.</p></li><li><p>Th&#847;e &#847;bes&#847;t-p&#847;erformi&#847;ng prom&#847;pts a&#847;re &#847;varied &#847;and combi&#847;ned to crea&#847;te a new g&#847;enerat&#847;ion of prompt&#847; varia&#847;tions. This cycle repeats, r&#847;eta&#847;ining good &#847;mu&#847;tations and eliminating less effective o&#847;nes.</p></li><li><p>Imp&#847;o&#847;r&#847;tantl&#847;y, the system&#847; also evolved its&#847; &#847;mutant prompt&#847;s over several ge&#847;neratio&#847;ns, continually improv&#847;ing the&#847; way&#847; it f&#847;ind&#847;s&#847; better promp&#847;ts.</p></li></ol><p>They call this system "<a href="https://arxiv.org/abs/2309.16797">P&#847;r&#847;omptbr&#847;e&#847;eder</a>".&#847; I&#847;n tests, it &#847;ou&#847;tperformed advanced&#847; hand-crafted reminder &#847;techniques in mathemat&#847;ics, commo&#847;n reasoning, &#847;a&#847;n&#847;d oth&#847;er tasks. It was able to generate prompts tailored to each problem area automatically.</p><p>Thi&#847;s represents an exciting new ab&#847;ility fo&#847;r AI to r&#847;ecursive&#847;ly&#847; improve themselves by cont&#847;inuously cr&#847;eating and tes&#847;ting&#847; var&#847;iat&#847;ions&#847; on the &#847;fly. As mo&#847;dels become bigger, prompt breeding may beco&#847;m&#847;e&#847; a&#847;n increasingly importan&#847;t t&#847;ech&#847;nique for exploiting their &#847;full potenti&#847;al withou&#847;t &#847;the&#847; need&#847; fo&#847;r costly retraining. The prompts we provide to AI, and how they shape reasoning may end up being just as important as the underlying algorithms.</p><p>However, keep in mind that when prompt "b&#847;reed&#847;ing", the &#847;O&#847;penAI bill c&#847;an&#847; quickly expl&#847;ode.</p>]]></content:encoded></item><item><title><![CDATA[Not-invented-here syndrome]]></title><description><![CDATA[For most startups, the priority is to leverage proven solutions when it comes to fundamentals like databases, authentication, and storage.]]></description><link>https://suchanek.co/p/not-invented-here-syndrome</link><guid isPermaLink="false">https://suchanek.co/p/not-invented-here-syndrome</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Sat, 28 Oct 2023 13:13:08 GMT</pubDate><content:encoded><![CDATA[<p>For most startups, the priority is to leverage proven solutions when it comes to fundamentals like databases, authentication, and storage. The focus should be on meeting customers' specific needs and solving the unique problems they face, rather than getting bogged down in reinventing the wheel.</p><p>In the early stages of a startup, time and resources are precious. Choosing to use a custom parser or cache may offer a slight performance advantage, but at what cost? The focus should be on getting the product to market and iterating based on customer feedback, rather than looking for marginal performance gains.</p><p>It is important for startups to identify and establish their core competencies, investing resources in these areas to create a solid foundation. However, for non-essential functions, existing libraries and tools should be used to avoid unnecessary waste of time and resources.</p><p>Once a startup has achieved product-market fit and has the financial resources to support further growth, it can consider innovating specific parts of its technology. This should only be done if it promises to deliver significant improvements, on the order of 10x, that will have a significant impact on the business or user experience.</p><p>The truly innovative and revolutionary work comes later. Startups should first focus on getting to market quickly by reusing existing code and solutions. This approach ensures the bases are covered, allowing innovation to take center stage where it will have the most significant impact.</p><p>In short, while there is a time and place to reinvent the wheel, it should not be the first course of action for most startups. Leveraging existing solutions to validate business ideas should be a priority. Once a startup has identified what truly sets it apart from the competition, it can and should boldly reimagine and reshape the wheel where it delivers the most strategic value.</p><p>Engineers are often not able to see this treasure of interesting problems hiding at the other end of the rainbow. Engineers want to solve hard problem, I get it. However, it&#8217;s the job of strong leadership to guide everyone there, using the boring tools so that later the really interesting problems can be solved.</p>]]></content:encoded></item><item><title><![CDATA[Tit for tat]]></title><description><![CDATA[The Prisoner's Dilemma is a famous thought experiment that illustrates the challenges of cooperation and competition.]]></description><link>https://suchanek.co/p/tit-for-tat</link><guid isPermaLink="false">https://suchanek.co/p/tit-for-tat</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Fri, 27 Oct 2023 13:18:19 GMT</pubDate><content:encoded><![CDATA[<p>The <a href="https://en.wikipedia.org/wiki/Prisoner%27s_dilemma">Prisoner's Dilemma</a> is a famous thought experiment that illustrates the challenges of cooperation and competition. Imagine two prisoners are arrested for a crime. The police offer each prisoner a deal: if one denounces the other and the other stays silent, the rat will be released, and the other will receive a long sentence. If both accuse each other, they both receive an average sentence. If both are silent, only a short sentence will be received.</p><p>The dilemma is that each prisoner is incentivized to denounce the other, even though both would benefit from cooperating and remaining silent. This simulates real-life situations like the Cold War arms race, where each side built weapons out of fear the other would do the same, even though both would be safer with fewer weapons.</p><p>In the 1980s, scientist Robert Axelrod organized a computer tournament where different strategies for solving the prisoner's dilemma competed. People submitted programs with names like Jesus, Lucifer, and Tester. Axelrod has received submissions ranging from a few lines of code to tens of thousands.</p><p>The simplest strategy, called Tit for Tat, won. It starts with cooperation, then simply copies what the opponent did last time. It therefore rewards cooperation but punishes betrayal, while leaving room for forgiveness. This "nice but retaliatory" strategy has proven effective and parallels real-world strategies such as the "live and let live" system that emerged in World War II trench warfare. The story illustrates how cooperation, reinforced by measured retaliation, can arise, even in difficult conditions.</p>]]></content:encoded></item><item><title><![CDATA[Scenius]]></title><description><![CDATA[As I&#8217;ve been visiting the AI Pioneers Summit today and seeing the high caliber of people on the stage, representing GitHub Copilot, Salesforce AI, Walmart, LlamaIndex, Langchain, Cloudflare, Mosaic, Replit, Nvidia, AutoGPT (and so on), I&#8217;m reflecting about the powerful network effects of San Francisco has and how all these companies are neighbors in the same town.]]></description><link>https://suchanek.co/p/scenius</link><guid isPermaLink="false">https://suchanek.co/p/scenius</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Thu, 26 Oct 2023 23:55:52 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!i8gM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As I&#8217;ve been visiting the <strong><a href="https://lu.ma/cna2nkex">AI Pioneers Summit</a> </strong>today and seeing the high caliber of people on the stage, representing GitHub Copilot, Salesforce AI, Walmart, LlamaIndex, Langchain, Cloudflare, Mosaic, Replit, Nvidia, AutoGPT (and so on), I&#8217;m reflecting about the powerful network effects of San Francisco has and how all these companies are neighbors in the same town.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i8gM!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i8gM!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 424w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 848w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i8gM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg" width="1280" height="960" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:960,&quot;width&quot;:1280,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:316376,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/jpeg&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i8gM!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 424w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 848w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 1272w, https://substackcdn.com/image/fetch/$s_!i8gM!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F7a3d98d9-b5ff-49e1-b26b-af3c0f7cd3ae_1280x960.jpeg 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Brian Eno introduced the concept of "scenius" to describe the collective intelligence and creativity of a cultural scene or group. The term combines "scene" and "genius" to highlight a community's power in fostering innovation and excellence. Scenius is the idea that genius doesn't just exist in individuals, but can also arise from the collaborative efforts of a group.</p><p>There are several factors that can help to nurture a scenius, according to Kevin Kelly:</p><ol><li><p>Mutual appreciation: Recognizing and valuing the contributions of each member of the group.</p></li><li><p>Rapid exchange of tools and techniques: Sharing knowledge and skills to help everyone improve.</p></li><li><p>Network effects of success: Celebrating and building on the successes of the group as a whole.</p></li><li><p>Local tolerance for novelties: Being open to new and different ideas.</p></li></ol><p>Some examples of scenius include:</p><ol><li><p><strong>Xerox PARC</strong>: A research and development company that has produced many innovative technologies.</p></li><li><p><strong>Paypal Mafia</strong>: A group of former Paypal employees who have gone on to found successful companies like Tesla and LinkedIn.</p></li><li><p><strong>Silicon Valley</strong>: A hub for technology and innovation in California.</p></li><li><p><strong>YCombinator</strong>: A startup accelerator that has helped launch successful companies like Dropbox and Airbnb.</p></li><li><p><strong>Burning Man</strong>: An annual event that fosters creativity and community.</p></li></ol><p>To create a scenius, it's important to find the right balance between openness and structure. A too open space can lack focus, while too much structure can stifle creativity. It's also important to embrace a certain amount of inefficiency and wastefulness, as these can be necessary for the group's vitality.</p>]]></content:encoded></item><item><title><![CDATA[Who benefits from OSS LLMs?]]></title><description><![CDATA[As use cases become clearer for how to use LLMs, companies look to save costs.]]></description><link>https://suchanek.co/p/who-benefits-from-oss-llms</link><guid isPermaLink="false">https://suchanek.co/p/who-benefits-from-oss-llms</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Wed, 25 Oct 2023 13:17:12 GMT</pubDate><content:encoded><![CDATA[<p>As use cases become clearer for how to use LLMs, companies look to save costs. Startups like <a href="https://openpipe.ai/">OpenPipe</a>, a company from the last YC batch, provide model distillation as a service. Record all your GPT-4 or Claude 2 interactions and teach them to a smaller model. Not only can you save a lot of money with this, but also potentially even gain higher quality than before. Simple tasks requiring GPT-3.5 can now be delegated to a smaller model, such as Mistral 7b.</p><p>How did we get here? Since the Llama release in March, we experienced the Cambrian explosion of open-source LLM models. Hundreds of fine-tuned options of Llama 1, but since then also, Llama 2 have been released, fighting for the first place in the <a href="https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard">LLM leaderboard</a>. I believe that Mark Zuckerberg did this very intentionally. He knew that Meta wouldn&#8217;t have a model on the GPT-4 level yet, which they&#8217;d be able to monetize, but at least he&#8217;d be able to &#8220;eat the bottom&#8221; of OpenAI&#8217;s revenue by kickstarting this revolution. In other words, this can be seen as asymmetric warfare. Instead of directly attacking someone, I give the power to the people, so that the delta of the OpenAI offering compared to what&#8217;s out there shrinks. The game is on, OpenAI needs to deliver to stay in the spot of number one.</p><p>So who benefits from this? Everyone besides the few big AI labs.</p><ul><li><p>Enterprises win, as they can now safely deploy legit GPT 3.5 alternatives in their own datacenter, without having to transmit PII</p></li><li><p>Startups win, as OSS models create a variety of opportunities around inference, training and fine-tuning</p></li><li><p>Everyone wins, as these models are able to run locally</p></li><li><p>Everyone wins, because we&#8217;re now enabling anyone to do AI research at home on their laptop by finetuning these models</p></li></ul><p>Just OpenAI, well&#8230; <a href="https://www.theinformation.com/articles/openais-corporate-sales-come-under-pressure-as-ai-customers-eye-cheaper-options">They might get into trouble</a>.</p>]]></content:encoded></item><item><title><![CDATA[How agents acquire capabilities]]></title><description><![CDATA[Currently, LLM-powered autonomous agents can solve simple tasks like sending an email, filling in the blanks in a spreadsheet, or even making phone calls.]]></description><link>https://suchanek.co/p/how-agents-acquire-capabilities</link><guid isPermaLink="false">https://suchanek.co/p/how-agents-acquire-capabilities</guid><dc:creator><![CDATA[Tim Suchanek]]></dc:creator><pubDate>Tue, 24 Oct 2023 12:58:43 GMT</pubDate><content:encoded><![CDATA[<p>Currently, LLM-powered autonomous agents can solve simple tasks like sending an email, filling in the blanks in a spreadsheet, or even making phone calls. However, as soon as it gets more complex, like implementing a whole database from scratch, controlling a robot in real life, or being able to control the browser universally, current agents fall short.</p><p>One of the main reasons for that is the limited reasoning and planning capabilities, but also simply being able to use tools.</p><p>There are numerous cases on the internet where someone mentions that GPT-4 can&#8217;t do something, but later someone can make it happen. How can an LLM-powered autonomous agent acquire new capabilities it doesn&#8217;t have today?</p><p><a href="https://arxiv.org/pdf/2308.11432.pdf">A recent survey</a> about LLM-powered autonomous agents provides an excellent overview:</p><h4><strong>Finetuning</strong></h4><ul><li><p><strong>With Human Annotated Datasets: </strong>The dataset to fine-tune is constructed through human feedback, which is turned into natural language, directly used to fine-tune the model. An example of this is the <a href="https://arxiv.org/abs/2207.01206">WebShop Dataset</a>, where researchers set up an artificial webshop that lets 1,600 humans use and collect the usage data.</p></li><li><p><strong>With LLM-Generated Datasets: </strong>As recruiting humans to annotate data can be time-intensive and costly, an alternative promising approach is to automate the dataset generation. An example is the recently released <a href="https://laion.ai/blog/strategic-game-dataset/">Strategic Game Dataset</a> from Laion, which includes 3.2 billion chess games, 236b Rubik cube moves, and 39b maze moves.</p></li><li><p><strong>With Real-world Datasets: </strong>Besides building up artificial labeling processes, collecting data from real-world applications can also be very powerful. This can be real-world product usage, which doesn&#8217;t have an explicit human labeling component but yields sufficient data to build a dataset. An example is the <a href="https://arxiv.org/abs/2306.06070">Mind2Web</a> dataset, which tries to create a generalistic dataset for the web.</p></li></ul><p></p><h4>Prompting Engineering</h4><p>One can provide some few-shot examples to the LLM in the same prompt to improve planning or reasoning. Some agent implementations also utilize prompts around the agent&#8217;s beliefs and state of mind.</p><p></p><h4>Mechanism Engineering</h4><ul><li><p><strong>Trial and Error: </strong>In this method, an agent performs an action and is then invoked again to judge the action just performed. If the action is deemed unsatisfactory, the feedback is incorporated, and the agent iterates.</p></li><li><p><strong>Crowd-sourcing: </strong>The prompt will be delegated to different agents. If they don&#8217;t respond with a consistent answer, solutions from other agents are used to update the response. In this method, (sub-)agents can incorporate other agents' opinions to improve the overall outcome.</p></li><li><p><strong>Experience Accumulation: </strong>The agent here does a look-up if it saw a similar task or prompt before and uses that to improve the result.</p></li><li><p><strong>Self-driven Evolution: </strong>The agent sets its own goals and adjusts the approach based on feedback from the environment and a reward function, gradually moving towards the overall goal. With that, the agent can acquire new capabilities in a self-driven manner.</p></li></ul>]]></content:encoded></item></channel></rss>