<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:cc="http://cyber.law.harvard.edu/rss/creativeCommonsRssModule.html">
    <channel>
        <title><![CDATA[Stories by Javier Ideami on Medium]]></title>
        <description><![CDATA[Stories by Javier Ideami on Medium]]></description>
        <link>https://medium.com/@ideami?source=rss-7f7b5d730c84------2</link>
        <image>
            <url>https://cdn-images-1.medium.com/fit/c/150/150/1*vJeao0WwYMWImjodsQnZ_w.jpeg</url>
            <title>Stories by Javier Ideami on Medium</title>
            <link>https://medium.com/@ideami?source=rss-7f7b5d730c84------2</link>
        </image>
        <generator>Medium</generator>
        <lastBuildDate>Sat, 11 Apr 2026 04:29:24 GMT</lastBuildDate>
        <atom:link href="https://medium.com/@ideami/feed" rel="self" type="application/rss+xml"/>
        <webMaster><![CDATA[yourfriends@medium.com]]></webMaster>
        <atom:link href="http://medium.superfeedr.com" rel="hub"/>
        <item>
            <title><![CDATA[How to grow a sustainable artificial mind from scratch]]></title>
            <link>https://ai.plainenglish.io/how-to-grow-a-sustainable-artificial-mind-from-scratch-54503b099a07?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/54503b099a07</guid>
            <category><![CDATA[agi]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[active-inference]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Thu, 05 Dec 2024 19:36:09 GMT</pubDate>
            <atom:updated>2024-12-06T09:53:53.408Z</atom:updated>
            <content:encoded><![CDATA[<h4>Scale free Active Inference, towards a more sustainable and explainable AI</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1000/1*1MnQJrIDieTCQExTMfmwjQ.jpeg" /><figcaption>Image by Javier Ideami</figcaption></figure><p>Today’s AI is impressive and useful in many ways. We can view it as a massive subconscious pattern matching engine. It operates as a fast thinking (System 1, per <a href="https://en.wikipedia.org/wiki/Thinking,_Fast_and_Slow">Kahneman</a>) entity, excelling in a wide range of tasks because much of human life revolves around such pattern matching processes.</p><p>Additionally, AI is beginning to tackle slow thinking (System 2), deep reasoning and agentic behavior, although it is still in the early stages of development in those areas. Despite these advancements, many experts agree that current AI faces several key challenges, including the following:</p><ul><li>Lack of efficiency and sustainability</li><li>Lack of explainability and transparency</li><li>Brittleness due to an underdeveloped world model</li><li>Limited predictive capability for deep reasoning and planning (System 2). OpenAI’s O1 model represents early steps in this direction</li><li>Absence of built in ethical guardrails</li><li>Lack of active, continuous learning and adaptation</li><li>Limited generalization across different tasks and domains</li><li>Inability to understand or process complex, abstract concepts like common sense reasoning (at a robust level)</li><li>Difficulty in handling uncertainty and ambiguity in dynamic environments</li></ul><p>While we’re all enjoying the progress of current AI, and I certainly do, it’s valuable to broaden our horizons by exploring alternative paradigms. These could potentially enhance, or offer insights that help improve the current approach. One such paradigm is Active Inference, founded by Karl Friston, the world’s most cited neuroscientist today.</p><h3>Active Inference</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1000/1*D0O4CldJBAIYW_43oAYEHA.jpeg" /><figcaption>Image by Javier Ideami</figcaption></figure><p>Active Inference is a theory that explains how intelligent systems, whether biological or artificial, perceive, act, and learn. At its core, it suggests that all intelligent behavior arises from minimizing uncertainty (or free energy) about the world by building and constantly updating internal models of it.</p><p>These models predict sensory input, and the differences between predictions and observations drive learning and decision making.</p><p>What makes Active Inference so promising is that it brings it all together: it bridges perception, action, and learning into a single process, offering a way to design AI systems that not only adapt dynamically to their environments but also emulate the kind of curiosity and planning typical in humans.</p><p>Active Inference addresses the limitations of the current paradigm that I described above in different ways:</p><ul><li><strong>Efficiency and sustainability</strong>: By focusing on minimizing free energy, Active Inference optimizes resource use and reduces computational waste, making systems more efficient.</li><li><strong>Explainability</strong>: Its foundation in probabilistic generative models provides a clear, interpretable structure for how systems perceive, predict, and act, offering insights into their decision making.</li><li><strong>Brittleness</strong>: Active Inference builds robust world models that are constantly updated through sensory feedback, making systems more adaptable and less fragile.</li><li><strong>Predictive reasoning and planning:</strong> By integrating perception and action into a single predictive framework, it enables deep reasoning and long term planning, addressing the gap in current system 2 capabilities.</li><li><strong>Ethical guardrails</strong>: Systems based on Active Inference can incorporate human aligned priors, ensuring actions minimize harm and align with ethical principles.</li><li><strong>Active continuous learning</strong>: Its constant cycle of prediction, observation, and updating ensures systems are perpetually learning and adapting to their environment.</li></ul><p>One of the biggest challenges for Active Inference to go mainstream is scaling. The algorithms are computationally expensive, and ongoing efforts are focused on finding effective solutions to overcome this hurdle.</p><p>This article focuses on a talk given by Karl Friston at the recent Active Inference Symposium, where he presented a new academic paper (<a href="https://arxiv.org/abs/2407.20292">https://arxiv.org/abs/2407.20292</a>) proposing a solution to this challenge: <strong>How can we scale Active Inference for high dimensional real world problems?<br></strong>(the full talk is available at <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a>)</p><p>The majority of this article goes through Karl’s explanation, part by part, adding clarifications and elaborating on key points. Occasionally, I’ll provide extra commentary on implications or variations of the topics discussed.</p><h3>A scale Free Approach</h3><p>In his talk, Karl Friston begins by addressing a key challenge for self organizing agents: understanding the coupling between different scales, the micro and the macro, and how they relate in terms of causality, with both bottom up and top down influences.</p><p>Scaling, in this context, is closely linked to abstraction, as it involves scaling information across various dimensions, such as state space and time. A crucial method for scaling is through compression, which, when done efficiently, produces abstractions of the original system.</p><p>This is a fundamental challenge for all forms of AI, and is particularly relevant to the current dominant AI paradigm. To tackle this, Karl Friston starts with the basics: defining what a <strong>thing </strong>is, and how we can define an agent that interacts with the world. This leads us to the concept of a Markov Blanket.</p><h3>Defining the agent</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qnWUc7DHmsoFaRtk4MVeOQ.jpeg" /><figcaption>Image by Javier ideami | ideami.com</figcaption></figure><p>A <strong>Markov blanket</strong> is the boundary that separates a system (like a brain or an AI) from its environment. It consists of sensory inputs (what the system senses) and actions (how it affects the environment), allowing the system to interact with and infer the state of the world without directly accessing it. It’s like a filter or interface through which the system perceives and influences its surroundings.</p><p>So the Markov blanket of an agent separates internal and external states through boundary states which constitute the markov blanket. And those boundary states are the sensory and the active states.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*TBCQItfVY-twXHsnYTUniQ.jpeg" /><figcaption>Image by Javier ideami | ideami.com</figcaption></figure><p>In this way, Karl tells us, the inside can influence the outside via active states (outputs).<br>And the outside can influence the inside via sensory states (inputs).</p><p>Great, we got a way of defining the essence of what constitutes an agent. Now it’s time to define the mechanics of its behavior.</p><h3>Bayesian Mechanics</h3><p>Quantum mechanics, classical mechanics, statistical mechanics, and Bayesian mechanics all share a common goal: to model, describe, and predict the behavior of systems, from the smallest particles to complex, large scale structures. Each of these frameworks, while distinct in their methods and applications, tries to understand how systems evolve, interact, and respond to various influences.</p><p>In the case of Active Inference, Bayesian mechanics is a framework for understanding how systems predict and adapt to their environment by minimizing uncertainty.</p><p>Unlike classical mechanics, which describes physical systems with deterministic laws, and quantum mechanics, which operates on probabilistic principles, Bayesian mechanics focuses on probabilistic models where systems continuously update beliefs about the world using Bayes’ rule.</p><p>This makes it ideal for modeling adaptive, self organizing systems like brains or AI, which need to learn and act in uncertain, ever changing environments.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*EU8eGyQaQyny1qCr26XSFQ.jpeg" /><figcaption>Image by Javier Ideami</figcaption></figure><p>So think of Bayesian Mechanics as the dynamics of anything that can be expressed in terms of markov blankets. For that kind of entities, their dynamics can be described as active inference.</p><p>Karl explains that we can summarize Bayesian mechanics through a couple of equations, the variational free energy equation and the expected free energy equation.</p><p>Think of <strong>free energy</strong> as a measure of how much a system’s predictions about the world differ from what it actually experiences. It’s like a mismatch score. The bigger the mismatch, the higher the free energy. Systems work to minimize this mismatch by improving their predictions or changing their actions to make the world more like they expect. This process helps them learn, adapt, and stay in balance with their environment.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*0lKxtmWmO_lfIQ0JhmV3DA.jpeg" /><figcaption>The agent tries to minimize Free Energy, the difference between its predictions and its observations | Image by Javier ideami</figcaption></figure><p>Let’s now consider those two equations:</p><p><strong>VFE (Variational Free Energy)</strong>: It’s about keeping the internal states of the agent as good models of the external states, therefore tracking in a probabilistic way the dynamics of those outside states. Minimizing variational free energy helps the agent improve its model of the world, make its beliefs about the world be accurate enough while keeping them as simple as possible.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*JvOdiLYapkI81wqPiqjN2Q.jpeg" /><figcaption>Perceptual Inference | Image by Javier Ideami</figcaption></figure><p><strong>EFE (Expected Free Energy):</strong> encapsulates prior beliefs about the kinds of actions that the agent can exert over external states. Minimizing expected free energy helps the agent find the right sequence of actions to take.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Dj4vUaQ21lkoz0lY4IF1Pg.jpeg" /><figcaption>Active Inference | Image by Javier Ideami</figcaption></figure><p>Karl explains that we can combine both equations to produce a Generalized Free Energy equation. At this stage, it’s not crucial to understand these equations in great detail. What matters is understanding their purpose and why they exist.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Wg00ebSdGegI_i-tuZvjbw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>So far, we’ve defined what an agent is and introduced the equations that govern its dynamics. By minimizing Expected Free Energy (EFE) and Variational Free Energy (VFE), the agent can improve its world model and select optimal sequences of actions.</p><p>To keep this article straightforward, we’ll avoid diving into the full complexity of these equations. However, it’s worth noting that each equation is broken down into distinct terms, each with significant implications.</p><p>The <strong>EFE (Expected Free Energy) </strong>equation is composed of a Pragmatic term, an Epistemic term, and an Entropic term.</p><ul><li>The <strong>pragmatic </strong>term measures how effectively the agent is pursuing its preferences or goals, evaluating how well it exploits its current beliefs.</li><li>The <strong>epistemic </strong>term assesses how effectively the agent is exploring, increasing its knowledge of the environment.</li><li>The <strong>entropic </strong>term quantifies the uncertainty surrounding the outcomes of the agent’s actions, providing stability and ensuring balanced decision making.</li></ul><p>Without going into too much detail, notice that this equation will allow the agent to balance exploration (curiosity, creativity) and exploitation in different ways depending on what’s happening at each moment.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*6xSdl7fSwkHhGJ_M85d41g.jpeg" /><figcaption>Image by Javier Ideami | ideami.com</figcaption></figure><p>The Variational Free Energy (VFE) equation consists of two key components: an <strong>accuracy </strong>term and a <strong>complexity </strong>term. This means the agent is constantly trying to balance accuracy with simplicity, seeking the most efficient explanations for its observations. It develops a world model that is both sufficiently accurate and efficient, aiming to compress and abstract the complexity of its environment in the simplest possible way.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*HtC5a639QN735gBszvciYQ.jpeg" /><figcaption>Image by Javier Ideami | ideami.com</figcaption></figure><p>The next challenge is to explore how<strong> markov blankets</strong> can give rise to scale free dynamics in the self organizing agents under consideration.</p><h3>On the Scaling Elevator</h3><p>Karl guides us through the interplay between the micro and the macro to uncover the nature of the states that define entities or particles.</p><p>At the heart of this exploration lies recursivity.</p><ul><li>Particles, Karl says, constitute a set of microscopic blanket and internal states.</li><li>These states, in turn, form a set of macroscopic eigenfunctions, eigenmodes, or mixtures derived from the blanket states.</li></ul><p>When Karl uses <strong>eigen</strong> to describe states, such as <strong>eigenfunctions </strong>or <strong>eigenmodes, </strong>he is referring to patterns or stable structures that naturally arise within a system.</p><p>The term eigen here implies fundamental, characteristic forms or modes that persist over time and help define the behavior of the system. Just as in physics or linear algebra where eigenvalues and eigenvectors represent core properties that remain consistent under certain transformations, in Active Inference, eigenstates or eigenmodes represent stable, intrinsic patterns within the system’s dynamics or blanket states.</p><p>These eigen patterns are the fundamental ways that a system organizes its internal and external states to minimize free energy and maintain a balanced interaction with the environment.</p><p><strong>In summary:</strong></p><ul><li><strong>Eigenfunctions </strong>represent stable, functional states that emerge from the interactions of microscopic states.</li><li><strong>Eigenmodes </strong>refer to the patterns of collective behavior that these blanket states assume when organized at larger scales.</li></ul><p>So, in order to create different scales or levels of granularity and abstraction:</p><ul><li><strong>Partitioning States by Dependencies</strong>: First, we group or partition states based on how they are coupled, meaning how they influence or depend on each other. By identifying which states interact closely (have strong dependencies), we can treat them as smaller, interrelated clusters.</li><li><strong>Creating Eigen Mixtures for Simplicity</strong>: Next, we combine (or coarse grain) these clusters into new, simplified eigen mixtures. These are stable, low dimensional representations of the original boundary states that capture their essential interactions without the need for all the detailed complexity. This dimensionality reduction lets us represent the system’s behavior at a higher level, while still preserving the key dependencies.</li><li><strong>Scaling Up by Repeating the Process</strong>: By repeating this clustering and reduction process, we can create layers of abstraction at different scales. Each higher scale captures broader, more abstract patterns, gradually filtering out the smaller details. This approach lets us model complex systems in a way that is both manageable and hierarchical, moving from fine grained details to big picture dynamics.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/950/1*AtDpo5wS-tKm-MuSMR06UQ.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>Therefore, the process involves two main steps, repeated in cycles:</p><ol><li><strong>Cluster by Causal Influence</strong>: Start by grouping or clustering states based on their causal influences or dependencies, meaning which states tend to affect each other directly. This grouping captures how different parts of the system are interrelated.</li><li><strong>Coarse Grain the Clusters</strong>: Then, reduce the complexity within each cluster by creating a simpler, coarse grained version of it. This involves combining the states within each cluster into a single, representative eigen mixture that keeps only the essential patterns or interactions, discarding less important details.</li></ol><p>Repeating these steps lets us build layers of increasing abstraction, where each higher level preserves the major patterns and causal structures from the layer below. This way, we can understand complex systems in a structured, hierarchical way, focusing on the most important interactions at each scale.</p><p>Now, in order to apply that coarse graining of the clusters, we apply some operators which we call <strong>Renormalization Group Operators (RG).</strong></p><h3>Renormalization and scale invariance</h3><p><strong>Renormalization Group Operators (RG)</strong> simplify or compress clusters of states, while preserving the essential dynamics or behavior of the system.</p><ul><li><strong>Applying RG Operators</strong>: After clustering by causal influence, we use RG operators to coarse grain these clusters. The RG operator effectively reduces the complexity of the cluster by creating a simpler, higher level representation that captures the main interactions within that cluster.</li><li><strong>Scale Invariance</strong>: When this coarse graining process is applied recursively, we move up one level of abstraction or scale each time, creating a hierarchy. Scale invariance means that each level preserves the core dynamics of the level below, even though it’s represented in a simpler form.</li></ul><p>By recursively applying these RG operators, we achieve a structured, multi-scale model of the system where each level maintains the overall dynamics, even though we’re gradually discarding finer details. This process is key to help model complex systems efficiently while retaining the essential behavior at each scale.</p><p>And Karl explains the reasons why we can apply the RG operators without disrupting the dynamics of the agent:</p><ul><li>We know that anything that has a particular partition can be described in terms of a gradient flow of a variational generalized free energy.</li><li>So when a system is partitioned (into internal, blanket, and external states), each partition behaves like an independent, self contained system. And it follows a principle of <strong>gradient flow</strong>, meaning it continuously adjusts to minimize its own free energy. This makes the partitioned subsets predictable and stable in terms of their dynamics.</li><li>The <strong>renormalization group operators</strong> can be applied effectively to such partitioned systems because each partitioned subset of states (with its own internal states, blanket states, and external states) acts as a self contained system that seeks to minimize its free energy through a gradient flow.</li><li>So each partition, or each particle, behaves as if it is minimizing its own generalized free energy. The RG operators then coarse grain or reduce these partitions, simplifying them while conserving their dynamics (their tendency to minimize free energy).</li><li>The idea of scale invariance here comes from the fact that, after coarse graining, the same free energy minimization behavior can be observed at each level, maintaining a consistent dynamic across scales.</li><li>Since each partitioned state structure behaves according to that gradient flow (minimizing free energy), we know that applying RG transformations will preserve the fundamental dynamics of the system at each level of scale. This property is what makes RG transformations valid in this context. They do not disrupt the system’s underlying tendency to self organize, as each coarse grained structure continues to follow the same free energy minimization principle.</li></ul><p>Therefore, the RG process works because free energy minimization is a universal, scale independent property of the system. As long as partitions are structured around this principle, applying RG operators simplifies the system without disrupting its core dynamics. This allows for a hierarchical, multi-scale model where each level accurately reflects the dynamics of the system as a whole.</p><p>Consider the elegance of this recursive, scale invariant dynamic:</p><ul><li>Particles are composed of states</li><li>States are eigen mixtures of particles</li></ul><p>This, Karl tells us, resolves a chicken and egg problem, through the creation of scale invariance provided by the renormalization group operators.</p><p>And Karl explains that as you move up the scales through coarse graining, you encounter slower macroscopic variables, ones that decay more gradually and exhibit persistence, providing a kind of <strong>temporal dilation</strong>.</p><p>Essentially, as you coarse grain at progressively higher levels, things change more and more slowly.</p><p>Tying this back to bayesian mechanics: this framework offers a way to model the agent’s external dynamics, which are inherently scale invariant.</p><p>This scale invariance arises naturally from defining <strong>things </strong>as particles with states, and states as eigen mixtures or functions of particles. This recursive structure inevitably leads to external dynamics that remain consistent across scales. And this consistency is inherent in the system’s structure.</p><p>In essence, we configure the system so that, regardless of how we partition or coarse grain it, whether at the microscopic particle level or the macroscopic system level, the overall behavior remains scale invariant.</p><p>And scale invariance, in this context, means that the system’s dynamics, how it evolves over time, are independent of the scale at which you observe it. As you coarse grain (move to higher levels of abstraction, grouping states into particles), the system slows down and becomes more persistent. The same underlying generative model still applies, but at larger scales, the behavior becomes slower, more stable, and exhibits temporal dilation.</p><p>We perform dimensionality reduction or coarse graining using these eigenfunctions, which capture the slow, macroscopic stable modes that dissipate or decay at a very gradual rate.</p><p>Building on everything discussed so far, we can now turn our attention to the agent’s generative model, its mind, so to speak. This generative model must also be scale invariant, or scale free (when applied to graphs), across both time and state space.</p><h3>The mind of the Agent</h3><p>Now it’s time to construct the mind of such an agent, which, as Karl explains, involves a model that captures a world with scale invariance in both state space and time.</p><p>To build these generative models, we can use <strong>generalized partially observed Markov decision processes (POMDPs)</strong>.</p><p>A Partially Observable Markov Decision Process (POMDP) is a mathematical framework used to model decision making problems where an agent has incomplete information about the environment’s state. It extends the basic concept of a <strong>Markov Decision Process (MDP) </strong>by incorporating partial observability, meaning the agent can’t directly observe the true state of the environment but only receives observations that are noisy or incomplete.</p><p>Think of it like navigating a room with a foggy window. You can make guesses based on blurry information, but you don’t have a clear view.</p><p><strong>Markov </strong>refers to the idea that the agent’s next move depends only on its current state, not the entire history of previous events. This simplification mirrors how we often operate in real life, where we make decisions based on what we know in the moment, rather than recalling every past detail, and it also simplifies computations by reducing the information the agent needs to process.</p><p>So in an <strong>active inference agent</strong>, we use this framework because it helps simulate decision making in uncertain conditions, where the agent doesn’t have all the answers but still needs to take action based on incomplete or noisy information, something real life systems face all the time, unlike simpler models where everything is perfectly observable.</p><p>In the context of POMDPs used in active inference, here are the key components and what each represents:</p><ul><li><strong>A (Likelihood Tensor): </strong>Represents the likelihood of observing certain data given the agent’s current beliefs about the world. It links the hidden states to the observed data.</li><li><strong>B (Transition Tensor)</strong>: Describes the probability of transitioning from one hidden state to another, depending on the current state and action taken by the agent.</li><li><strong>C (Preference Tensor)</strong>: Represents the agent’s preferences or desires for different states or actions. This could relate to goals or rewards the agent is trying to achieve.</li><li><strong>D (Action Tensor)</strong>: In addition to defining the agent’s possible actions, <strong>D</strong> can also express how actions influence transitions across different scales, representing the way decisions affect the system at multiple levels of granularity (or multiple scales of time and space). It models how actions at one level can influence or propagate through other levels.</li><li><strong>E (Exogenous Input Tensor)</strong>: Similarly, <strong>E</strong> can describe external inputs that influence the system at different scales, such as environmental changes or sensory data that affect both micro and macro levels. It can show how external factors couple the agent’s behavior across various levels of scale, influencing how the agent interacts with the world at both fine and coarse levels.</li><li><strong>Paths (Trajectory of States and Actions):</strong> In the generalized version, paths represent the sequence of states and actions over time, which can be used to track the agent’s experience or trajectory in the environment.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1000/1*XbjqmyxL8tdK7vZKRKApLg.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>These tensors work together to define the agent’s belief system, actions, and how it interacts with its environment, providing a framework for decision making under uncertainty.</p><p>And Karl explains that we add the word <strong>generalized </strong>to define these <strong>POMDPs </strong>because we are representing not only the states and the observations, but also the dynamics in terms of the paths that specify the transitions between states.</p><p>By incorporating these paths as random variables, we can model not just where the agent is at a given moment (hidden state) but also <strong>how</strong> it gets there and <strong>how</strong> it moves through the sequence of states over time, which adds more richness and flexibility to the <strong>generalized POMDP</strong> model. And in this way we have, Karl says, a generalized kind of Markov decision process.</p><p>It therefore means expanding beyond the simple states and observations seen in basic MDPs/POMDPs by incorporating:</p><ul><li>More complex relationships (paths as random variables)</li><li>Hidden states and their dynamics (not directly observable)</li><li>Transitions between states described probabilistically (using tensors to represent these paths)</li><li>A renormalization process that allows for different levels of abstraction and dynamic adaptation within the system.</li></ul><p>A reminder about random variables. A <strong>random variable</strong> is a variable that can take different values, but its value is uncertain. For example, in the case of <strong>paths being random variables</strong>, the path an agent takes through its hidden states is uncertain because it depends on its actions and the environment.</p><p>And a <strong>probability distribution</strong> is a way of describing the likelihood of each possible value that a random variable can take. In this case, the probability distribution would show how likely each possible path (sequence of states) is, based on the agent’s actions and the environment’s responses. So, the <strong>random variable</strong> represents the uncertain path, and the <strong>probability distribution</strong> tells us how probable each path is.</p><p>The generative model of the agent we are considering includes tensors (a <strong>tensor</strong> is a multi dimensional array of numbers used to represent data) that can be parametrized as<strong> dirichlet distributions</strong>. This means that the relationships encoded by the tensors (such as state transitions, actions, and observations) are represented probabilistically, rather than deterministically.</p><p>These tensors are parametrized by learnable parameters (like the mean, variance, or other parameters of a distribution), allowing the model to learn how these relationships evolve in the face of uncertainty and partial observability.</p><p>And what is a <strong>dirichlet </strong>distribution?</p><p>A <strong>dirichlet distribution</strong> is a type of probability distribution that is used when we want to model <strong>uncertainty</strong> about <strong>multiple possible outcomes</strong> that add up to 1 (like percentages or probabilities). It describes a situation where you have several categories or options, and you want to know the likelihood of different proportions or shares across those categories.</p><p>In the context of <strong>generalized POMDPs</strong> in <strong>active inference</strong>, the Dirichlet distribution is useful because it helps model <strong>uncertainty about different possible paths or states</strong> that the agent could take. It provides a flexible way to express how likely the agent thinks different sequences of events are, considering it doesn’t know exactly what will happen next. So, the Dirichlet distribution helps the agent <strong>estimate the probability of different scenarios</strong> in a way that adds up to a full picture of all possible outcomes.</p><p>Next, Karl explains that we will explore the various levels of depth within the generative model and examine how we can leverage the renormalizing aspect of this causal architecture in relation to the world in which our entities exist.</p><p>So, time to go deep.</p><h3>Parametric depth</h3><p>We arrive at a key twist in this fascinating exploration.</p><p>As we just explained, the different data structures in the mind of the agent are modeled with probability distributions. Those probability distributions, mathematical models, have parameters that control them.</p><p><strong>Parametric depth</strong> means that the parameters of those probability distributions are themselves random variables with their own probability distributions.</p><p>For example, the likelihood mapping that maps from hidden states to observations (usually encoded with the A tensor) is itself encoded in terms of dirichlet parameters, giving the parametric model depth.</p><p>Yes, it’s another form of recursivity. Recursivity is a phenomenon we often observe in the natural world, and it helps us model scale invariance in both state space and time within this framework.</p><p>There are also several important implications of having this parametric depth. It means that we now have to apply bayesian mechanics to more things.</p><p>We will be using Bayesian methods to infer or estimate the underlying, hidden aspects of the system that we can’t directly observe, and that includes the hidden states but also the parameters of the distributions that model those hidden states as well as other parts of the generative model.</p><p>Karl explains that in the generative model we can consider three kinds of depths: <strong>temporal, hierarchical and factorial</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*gqCQcPcl4-ck9Yx4nzgDbg.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>We have<strong> temporal depth</strong>, because we are modeling the transitions of the states towards the future. This temporal depth is generalized through the explicit path variables we will be making use of.</p><p>We have <strong>hierarchical depth</strong> in the sense that we can compose multiple MDPs. The output of a process can now form, Karl says, the input, or the empirical prior, or the inductive bias, or the top down constraint on the dynamics and states of the level below.</p><p>And the dynamics, the kind of outputs of the upper level that provide empirical constraints on the paths and initial conditions of the lower levels, are supplied or mediated by a mapping between higher levels (slow states), and the initial conditions or trajectories of the lower levels (through, for example, the D tensor).</p><p>Let’s dive into this last bit:</p><ul><li><strong>Dynamics of the upper level</strong>: At the upper level (the higher or slow states), we have outputs, things that emerge from the process at this level (like predictions, control signals, or constraints).</li><li><strong>Empirical constraints on the lower levels</strong>: These upper level outputs provide <strong>empirical constraints</strong> or guidelines that affect the <strong>paths</strong> (the transitions between states) and the <strong>initial conditions</strong> (the starting states) at the lower levels of the system.</li><li><strong>Mapping between higher and lower levels</strong>: The relationship between these levels is mediated by a <strong>mapping</strong> (a mathematical relationship or transformation), which connects the slow or high level states to the initial conditions of the lower level states. This ensures that the higher level states can guide or constrain the dynamics of the lower level system.</li><li><strong>D tensor</strong>: The D tensor (one of the tensors in the generalized POMDP framework) is responsible for encoding the relationship or mapping between the higher and lower levels. It helps specify how the higher level dynamics influence the <strong>initial conditions</strong> at lower levels.</li></ul><p>A clarification about paths and transitions:</p><ul><li><strong>Paths</strong> are the entire sequences of states (and actions), which track a trajectory.</li><li><strong>Transitions</strong> are the local probabilities of moving from one state to another, as described by the transition matrix (B matrix).</li><li>So, <strong>paths</strong> are made up of <strong>transitions</strong>, with <strong>transitions</strong> describing the immediate steps, and <strong>paths</strong> describing the entire sequence of states and actions taken over time.</li></ul><p>In summary, higher level processes (abstract goals or strategies) influence lower level processes (specific actions or states) in a hierarchical system.</p><p>This influence is mediated through a mapping that takes into account the slow dynamics of the higher level states, which then help define or constrain the initial conditions for the lower level states. The relationships between these levels are encoded using tensors, which represent complex, probabilistic transitions between different states in the system.</p><p>And a bit more about the D and E tensors:</p><ul><li><strong>D</strong> (initial states): It encodes how the <strong>initial conditions</strong> at higher levels influence the <strong>states</strong> at lower levels, that is, the <strong>mapping</strong> of state space across levels.</li><li><strong>E</strong> (initial paths): It describes how the <strong>temporal dynamics</strong> or <strong>transitions</strong> at the higher level affect the <strong>paths</strong> or <strong>transitions</strong> at the lower levels.</li></ul><p>So, the <strong>D tensor</strong> deals with <strong>states</strong>, while the <strong>E tensor</strong> deals with <strong>paths</strong>.</p><p>Finally, we also have <strong>factorial depth</strong>, Karl explains.</p><ul><li><strong>Factorial depth</strong>: It refers to how a model can deal with multiple independent factors that influence an outcome. Think of it as having many different ingredients (factors) that independently affect a final dish (the outcome).</li><li><strong>Independent factors</strong>: These factors may be independent of each other, meaning that each factor can change without directly affecting the others. However, when combined, they all contribute to the final outcome. So, a generative model with factorial depth considers all these independent factors, but may initially entangle them (mix them together) when generating an outcome (which itself can be a prior constraint on the states and dynamics of the level below).</li><li><strong>Entangling latent states:</strong> In the generation process, the model entangles (mixes together) these latent (hidden) factors. This means that, while the factors are independent, their effects on the outcome might be complicated or hard to separate directly.</li><li><strong>Disentangling the representation:</strong> The real task comes when we need to invert or disentangle these mixed up effects to figure out exactly what each factor (latent state) was doing in the outcome. It’s like looking at a complex puzzle (the observed data) and working backward to separate the pieces to understand how each factor contributed.</li><li><strong>Role of factorial depth:</strong> This allows the model to understand and manipulate the system at a deeper level by identifying and separating the independent factors that were mixed together in the generation process.</li></ul><p>So <strong>Factorial depth</strong> means the model recognizes that outcomes can come from a mix of independent factors, and its job is to later separate these factors to better understand how they contributed to the outcome.</p><p>In summary, all of that produces a<strong> generalized discrete state space model</strong> with <strong>paths as random variables</strong>.</p><p>Let’s now focus on the temporal dimension.</p><h3>Temporal Depth</h3><p>In real life, as we ascend scales (in terms of output states), things slow down (temporal dilation).<br>We move from fast microscopic states to slow macroscopic states.</p><p>We can implement the same dynamics in these MDPs that have hierarchical depth.</p><p>To achieve this, we <strong>discretize and quantize time</strong>, allowing for more frequent updates at each level compared to the level above. This process creates temporal scaling, aligning with the scale invariant approach we’ve been discussing.</p><p>In this way, we link temporal dynamics with hierarchical depth, as the higher we move in the hierarchy, the slower the updates become at each level</p><p>Therefore, as Karl explains:</p><ul><li>At a certain level, we generally parametrize or generate the initial conditions and the initial path from the states that are above that level.</li><li>And that path can pursue a trajectory over multiple time steps until a certain horizon (at that level), at which point it then sends messages to the level above.</li><li>Those messages produce belief updating, and the subsequent step of the level above generates priors about the subsequent path at the level below.</li><li>And so on, recursively all the way to the depth of the model.</li></ul><p>Now it’s time to analyze in more detail that same coarse graining but this time of the state space.</p><h3>Coarse Graining the state space</h3><p>To coarse grain the state space in this renormalizing generative model, we apply what are known as spin block transformations.</p><p><strong>Spin block transformations</strong> are a specific way to apply coarse graining in the context of this generative model. These transformations group local states (like tiny tiles) into larger blocks. The term <strong>spin</strong> is borrowed from physics, where systems with spin states (like magnetic spins) can be grouped to simplify complex models.</p><p>First, it’s important to emphasize two key assumptions or rules that we will consider.</p><ul><li>Each block will have one and only one parent. This simplification avoids complex relationships where multiple higher level factors could affect a single block, keeping the system simpler and easier to analyze.</li><li>For most systems composed of coupled elements, the interactions are usually local. So we are going to focus on those local interactions.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*enfbq_VvLItS1N35oYa-XA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>Let’s understand why those assumptions are important. Karl proposes: let’s think of a system that can be described in terms of local interactions. Consider a euclidean world where massive objects bounce around (let’s ignore gravity in this example). Such objects only influence each other when they touch, when they are proximal.</p><p>The partition that defines all the markov blankets in that state space is composed of lots of local little tiles. We can then use a block or spin operator that groups together local tiles, a group of 2D arrangements of states.</p><p>So, looking back at our generative model, a higher group, hidden factor, with its states and paths, is responsible for predicting the dynamics of other lower factors (at a lower level), and in so doing we are successively grouping together sets of sets of sets, etc.<br>So any group or set at one level is accountable or trying to generate the dynamics of a small local group at the lower level.</p><p>The key idea is that lower level blocks (small interactions between states) are grouped into higher level blocks, and the higher level blocks influence the dynamics of the lower levels.</p><p>And anything at the lower level can have only 1 parent. There are no additional parents, so we can ignore or dissolve dependencies between the factors at any given level, because each factor is only responsible for its children and it doesn’t share any children, in the case of this particular spin block transformation.</p><p>At some point, all the tiles, all the groups of the lowest level will come together at the top, very deep in the model, at which point they will be encoding quite long trajectories, Karl says, 2 to the power of the depth minus 1, in terms of updates. But at the lower level there are no interactions between the different blocks as we ascend them through the hierarchy.</p><p><strong>What this produces, in summary is:</strong></p><ul><li>A very efficient kind of generative model</li><li>Many converging streams that don’t have to talk to each other until the point at which they come together through these spin block transformations.</li></ul><p>And Karl reminds us that we can treat the D and E tensors (initial states, initial paths), as likelihoods that couple one scale or level to the one below.</p><p>This emphasis on the local interactions surely reminds us of something very familiar.</p><p>Karl explains that Renormalizing Generative Models (RGM) can be seen as homologues of deep convolutional neural networks or continuous state space models in generalized coordinates of motion. They can be used to learn compositionality over space and time, models of paths or orbits (events of increasing temporal depth and itinerancy).</p><p>Let’s clarify some of these terms.</p><ul><li>In <strong>RGMs</strong>, we group local states (tiles) into larger blocks, similar to how <strong>CNNs </strong>aggregate local features into higher level features.</li><li><strong>Compositionality</strong> means that complex systems can be broken down into simpler components (like tiles or blocks), which can then be recombined to understand more complex behavior. This allows the model to learn patterns over space and time.</li><li><strong>Paths</strong> (or orbits) represent the transitions or trajectories between states over time. The model learns to understand these paths, which can be seen as sequences of events that occur as the system evolves.</li><li><strong>Temporal depth</strong> refers to the ability of the model to account for the passage of time in understanding state transitions. It captures how the system changes over time, based on the accumulated information at each hierarchical level.</li></ul><p>The next and very key question is, how do we actually build such a model?</p><h3>Growing the artificial mind</h3><p>Let’s begin with a key statement that Karl shares.</p><p>We want the model to <strong>learn itself</strong>. To <strong>grow itself </strong>by using something called <strong>recursive fast structure learning</strong>.</p><p>This is very important because earlier versions of active inference required humans to manually design the base structure of the generative model. This new approach begins to automate these processes. Sounds beautiful, doesn’t it? Let’s see how this would work.</p><p>It will be a recursive process, a recursive application of the renormalization group operators.</p><p>At a high level, fast structure learning equips the model with unique states and transitions as they are encountered.</p><p>Let’s consider what this means. We are<strong> growing a model from scratch</strong>, from nothing!</p><p>And Karl adds: so that EFE (expected free energy) or mutual information is extremised, using structure learning as active model selection. Let’s unpack this:</p><ul><li><strong>Mutual Information</strong>: It is a measure from information theory that quantifies how much information one variable contains about another. In simple terms, it tells you how much knowing one variable reduces uncertainty about the other.</li><li>In the context of machine learning and probabilistic models, mutual information is often used to quantify how much information a model’s latent variables (the hidden or unobserved ones) share with the observations (the data we can see). In simpler terms, mutual information measures how much knowing the state of X helps in predicting Y and vice versa.</li><li>If X and Y are independent, their mutual information is zero, meaning knowing one doesn’t help in predicting the other. In this context, it refers to the information shared between the <strong>model’s predictions</strong> and the <strong>actual observed data</strong>. The goal of learning is to maximize this mutual information, so the model’s predictions become more aligned with reality.</li><li><strong>Extremised</strong>: In this context, extremising means maximizing mutual information or minimizing expected free energy. So, the model aims to <strong>optimize</strong> this relationship, either by reducing the surprise or uncertainty about the data (minimizing free energy) or by increasing the amount of useful information the model can extract from the data (maximizing mutual information).</li><li><strong>Structure learning as active model selection</strong>: This suggests that as the model encounters new data, it actively <strong>selects</strong> or <strong>adapts</strong> its structure (the components of its generative model) to better align with the data, improving its predictive accuracy. It’s a process of <strong>self improvement</strong> where the model chooses which structures and states are necessary, based on the data it is seeing.</li></ul><p>Karl also tells us that the EFE (expected free energy) is the mutual information in the absence of any expected cost. It is the mutual information of any mapping plus or minus an expected cost (in terms of constraints or prior preferences). Let’s unpack this.</p><p>In our case we can consider that in relation to the EFE equation:</p><ul><li><strong>Mutual Information</strong> = Epistemic Term (exploration, uncertainty reduction, learning about the world).</li><li><strong>Expected Cost</strong> = Pragmatic Term (exploitation, maximizing utility, evaluating costs).</li></ul><p>Together, these terms in the expected free energy (EFE) represent the balance between learning (exploration) and making the best use of what you know (exploitation), guiding how a system acts and learns in a way that optimizes both its understanding of the world and its ability to achieve goals with minimal costs.</p><p>In the absence of expected cost, the system is focused on learning and reducing uncertainty through exploration.</p><p>But when you introduce the expected cost (pragmatic term), the system also focuses on making decisions that not only increase knowledge but also keep actions aligned with goals, preferences, or minimizing harm.</p><p>So, it’s a balance between exploring for more knowledge (uncertainty reduction) and acting in a way that minimizes costs (alignment with goals).</p><p>When you’re trying to minimize EFE, you are essentially trying to reduce uncertainty by learning about the world (gaining more information) while also ensuring that the decisions you make do not violate important constraints or preferences. This creates a feedback loop where you’re both gathering knowledge and acting in ways that are beneficial or cost effective in the long run.</p><p><strong>In simpler terms:</strong> minimizing EFE means you’re reducing uncertainty (exploration) but also making sure you don’t incur too many costs (exploitation), all while satisfying any constraints or preferences that are important for achieving your goals.</p><p>And now, let’s return to the inspiring concept we discussed earlier: we will be growing a model from scratch, from the ground up!</p><p>Compare that to the inefficiency of current AI paradigms:</p><ul><li>Current paradigms start with highly complex, massively overparameterized models, which might later be simplified through the training process. Optimization and pruning methods help reduce the influence of certain parameters, streamlining the model.</li><li>This active inference approach, in contrast, begins from scratch and grows its model until it achieves the simplest possible implementation that fulfills the agent’s objectives.</li></ul><p>How would this process take place? Karls takes us through a high level example. Let’s take the perspective of the agent.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2Fh9sqeZtx1ps%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3Dh9sqeZtx1ps&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2Fh9sqeZtx1ps%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/c7d8c48b436612c1d642ffb989cc85f5/href">https://medium.com/media/c7d8c48b436612c1d642ffb989cc85f5/href</a></iframe><ul><li>I receive an observation, which I infer was generated by a latent hidden state. And I encode that latent state.</li><li>I then receive another observation. I have not seen that observation before, so I will encode it with a second latent state.</li><li>And I have seen those two observations one after the other, so I will encode that with a transition from the first to the second state.</li><li>And I keep doing the same. I now get a third observation. If I have not seen it before, I will induce a third latent state, and a second transition.</li><li>In this way, I accumulate all the unseen or unique instances of observations within a likelihood mapping structure, that grows and grows until we find something that we have encountered before.</li><li>So we don’t induce a new state or, implicitly, a column of the likelihood matrix, if we have seen that state previously. We don’t create an instance of something that has been encountered before.</li><li>However, that state may come along with a related path that is new, so the next state after that one may not be the same that was following the previously seen instance of that state. In that case, I am going to induce a new path.</li><li>In this way, I will be gradually growing my B tensors (the probability transition matrices), along with other components of the generative model such as the A tensors (likelihood mappings), to encode the full dynamical structure of the sequence of observations that is training the agent.</li><li>And that encoding will be summarized in a maximally efficient way. The implicit mapping that we are growing from scratch will have the highest mutual information between the latent space (where the latent states allow us to derive a prediction for the next observation) and the observations that are at hand.</li></ul><p>So we can now see that <strong>Fast Structure Learning</strong> is a kind of active inference, but not over states or parameters. It is instead active selection of the right model in terms of the size of the likelihood tensors and the transition tensors.</p><p>We are figuring out<strong> which model structure best fits the data</strong> or the situation you’re facing.</p><p>And model selection here refers to the process of selecting the right structure, that is, how complex the model should be in terms of its likelihood tensors and transition tensors, based on the evidence available.</p><p>So gradually, as agents, we are building our internal tensors. And for example:</p><ul><li><strong>Likelihood tensors</strong> represent the relationship between the states and the observations. This is how the model explains the likelihood of different observations given the current states.</li><li><strong>Transition tensors </strong>represent how states transition from one to another, essentially defining the dynamics or temporal evolution of the system. This shows how one hidden state leads to the next over time.</li></ul><p>The size of these tensors reflects the complexity of the model. More complex models have larger tensors with more parameters, which means they can potentially explain more complex phenomena but at the cost of becoming harder to learn and requiring more data.</p><p>When the system is choosing<strong> whether to accept a new latent state</strong> (a new hidden state in the model)<strong> or transition</strong> (a new way states evolve over time), it checks<strong> if this new model improves the EFE or maximizes mutual information</strong>.</p><p>Let’s unpack this:</p><ul><li><strong>Improving EFE:</strong> In active inference, the overall goal is to minimize Expected Free Energy, which combines two objectives: reducing uncertainty (mutual information) and minimizing expected costs (the negative consequences of actions). A new structure is considered beneficial if it lowers the overall EFE by balancing these two objectives.</li><li><strong>Maximizing Mutual Information:</strong> This is the epistemic aspect of EFE minimization. Here, the focus is on reducing uncertainty by preserving or increasing the mutual information between the model and the observations. In other words, the system prioritizes a model that offers the most informative representation of the data and reduces uncertainty about hidden states.</li><li>So if a new structure doesn’t immediately lower the total EFE (e.g., it doesn’t reduce both uncertainty and cost), the system may still accept it <strong>if it improves the epistemic term</strong>. This ensures that the model becomes more informative, even if pragmatic concerns are temporarily less prioritized. The rationale is that improving the understanding of hidden states can later lead to better decisions and lower EFE in subsequent updates.</li><li>In summary, the criteria for accepting a new latent state or transition are: Does it optimize or improve the EFE? Does it preserve or maximize mutual information? This approach reflects a trade off between exploration (learning about the environment) and exploitation (acting effectively based on current knowledge). However, in the case of fast structure learning, as we will discuss below, when applied to growing a model from scratch, the <strong>emphasis shifts toward the maximization of the epistemic term</strong>, maximizing mutual information.</li></ul><p>This is essentially a <strong>Bayesian approach to model selection</strong>. The idea is that the system is selecting among different models by evaluating the evidence (data, observations) and choosing the model that best fits that data, while also considering the complexity of the model (which is related to the size of the likelihood and transition tensors).</p><p>The system learns not just the states but also the structure of the model itself, dynamically adjusting itself to find the most appropriate model for the current data.</p><p>In summary, fast structure learning is a specific application of active inference, applied here in the context of Bayesian model selection.</p><p>And Karl tells us that it follows the same rules as when an active inference agent deals with planning and accepting parameter updates, but with a specific emphasis on maximizing the epistemic term of the EFE in the case of growing a model from scratch. Let’s unpack this further.</p><p>In typical parameter updates and planning steps in active inference, we aim to minimize EFE, which balances reducing uncertainty (epistemic gain) and aligning with preferences (pragmatic gain).</p><p>However, when it comes to structure learning, updating or selecting the model’s structure, the focus shifts. The system’s objective is not necessarily to minimize EFE in its entirety. Instead, it prioritizes maximizing the <strong>epistemic term</strong> of the EFE, which represents the mutual information between latent states and observations. In this context, maximizing EFE means seeking the most informative model structure that captures patterns in the data and reduces uncertainty over time.</p><p>This focus on maximizing the epistemic value ensures the model grows in a way that is both adaptive and capable of effective learning. By emphasizing the mutual information component, the model identifies latent states and transitions that provide the highest potential for reducing uncertainty in the future.</p><p>This principle mirrors how parameter updates in active inference sometimes prioritize epistemic gain to guide exploration and learning. Structure learning similarly follows an <strong>epistemic driven rule</strong>, selecting updates, actions, or structures that maximize mutual information and ensure the model remains informative and adaptable.</p><p>So it’s all about choosing the most informative model structure. In terms of expected free energy (EFE), and specifically in the context of fast structure learning, the primary focus is on maximizing the epistemic term with the aim of selecting a structure that has the highest potential for reducing uncertainty in the future. This process guides the selection of models that are informative and allow for effective learning or adaptation.</p><p>In summary, Friston emphasizes that structure learning, like planning or doing parameter updates, follows the same epistemic driven principles: it’s guided by maximizing the informative value of the model. This means, in practice, selecting structures, actions, or updates that increase mutual information and reduce uncertainty. However, in the context of fast structure learning, the reduction of immediate uncertainty is not the top priority. Instead, such reduction may occur in the future, not necessarily in the immediate present.</p><p>So let’s stand back for a moment and look at the structure being created from afar:</p><ul><li>We ingest our observations in relation to local groups.</li><li>At a certain level of the structure, we combine initial conditions and paths into new outcomes for the next level below.</li><li>And because we do this for local groups, the number of groups just gets smaller and smaller, as you get higher and higher.</li><li>And the number of combinations of paths and initial states increases as you get deeper and deeper (lower) into the model.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*rGDg0uQkPUoaMsS5xwf7Dw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>The next question that Karl poses is, how do we bound all of the above? And he explains that the answer is trivial because the size of the highest level is upper bounded by the number of elements in our training set.</p><p>Therefore, we can control the size of these renormalizing operators and likelihood mappings or transition tensors, simply by giving it canonical training data, the kind of data that we want this agent to generate, recognize or predict.</p><p>Let’s consider, though, that in a controlled training environment, bounding the complexity with a fixed, well defined dataset works well because the model only has to generalize within the boundaries of that data. But in the open, real world, the model would encounter way more variability and unexpected scenarios that weren’t in the training data. In this context, simply bounding based on the original training data would no longer apply effectively, because the variety of inputs and potential scenarios would be far greater.</p><p>So if we imagine our agent being set loose on the complexity of the real world, such an approach would have to be expanded with other sorts of constraints and mechanisms. Let’s expand on this:</p><ul><li><strong>Dynamic Model Updates:</strong> As mentioned earlier, the model would need mechanisms for dynamic learning or continual adaptation, allowing it to update its generalized states or high level structures based on new, diverse experiences in the real world. This would enable it to expand its boundaries by incorporating novel patterns it observes, without being rigidly confined to the initial training data.</li><li><strong>Hierarchical Compression and Forgetting Mechanisms</strong>: For real world applications, models could include strategies for compressing knowledge and selectively forgetting less useful or outdated information. This hierarchical updating approach, where recent data is prioritized and irrelevant paths are pruned, helps manage the growth in complexity while maintaining relevance in a changing environment.</li><li><strong>Safety and Practical Boundaries through Constraints:</strong> Models might include contextual constraints or policy based limits to control how much they explore certain areas. For example, in active inference models, bounding behavior in an unbounded real world can be achieved by biasing exploration toward goals or preferences, essentially building in rules that guide behavior within practical limits</li></ul><p>In the real world, then, the boundaries would need to be adaptable, using strategies that balance exploration (to handle novel scenarios) and compression or constraints (to keep complexity in check). This way, the model doesn’t endlessly expand but instead selectively grows and prunes knowledge in a way that remains manageable and relevant to its goals or purpose.</p><p>Now, let’s explore some specific examples that Karl Friston presented, illustrating one of these systems in action.</p><h3><strong>Filling the gaps</strong></h3><p>The first example that Karl explains is: <strong>fast structure learning and image compression to discover a simple orbit.<br></strong>In this example the generative model is trained on a video of a bird flying in a loop through a periodic orbit.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2F_WsUqLbl8j4%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2F_WsUqLbl8j4&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2F_WsUqLbl8j4%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/f2dfaf9662450c7a9e019292e804c6aa/href">https://medium.com/media/f2dfaf9662450c7a9e019292e804c6aa/href</a></iframe><p>And the generative model learns a structured representation of the video data (in this case, the bird flying in a periodic loop). By doing so, it compresses the data into a simplified, interpretable form (the essence of the bird’s trajectory) while still being able to reconstruct the original data.</p><p>Karl explains that the coarse graining process applied at the bottom level uses a finessed version of a <strong>spin block operator</strong> by harvesting local eigen modes or eigen images of patches of the video to get the image into a reduced or coarse grained representation, that is then passed up to the next level.</p><p><strong>A spin block operator </strong>is a method inspired by statistical physics to gather information over small regions (for example, patches of an image). This involves summarizing local details into eigenmodes or eigenimages, which represent the dominant patterns in those patches.</p><p>In this example the model uses 2 levels.</p><ul><li><strong>Lower level</strong>: Processes small scale, local features of the image (e.g., pixel patches or eigenimages).</li><li><strong>Higher level</strong>: Works with the entire image as a whole, analyzing large scale patterns (e.g., the bird’s entire trajectory).</li><li><strong>Interaction</strong>: The lower level sends coarse grained representations upward to the higher level for broader contextual understanding.</li></ul><p>The process is made of a few stages, Karl explains, which include:</p><ul><li><strong>Discretization:</strong> breaking the continuous video data into discrete chunks for analysis (e.g., frames or patches).</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*67bfxaKCkDQSLEfpO_K3Zw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><ul><li><strong>Single value decomposition:</strong> A mathematical technique that breaks down data into key components. Here, it can be used to identify the dominant patterns in the video sequence (e.g., the periodic trajectory of the bird).</li></ul><p>These processes help the system learn a simplified model of the video by focusing on its most significant features.</p><p>At the highest level, the model represents the bird’s motion over time as a sequence of abstract states or events (e.g., periodic loop). In this generalized space, the highest level encodes paths or trajectories over 2 to the n minus 1 timesteps in the future where n is the depth of the model. The model, therefore, is able to predict future states by understanding patterns across all of those time steps.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*SLYNxEZL-o5YJbClbWt1gw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>After the training, we can generate new movies simply by moving through the different paths or episodes.</p><p>Karl asks: what other things can we do with these models?<br>And he explains that one of them is pattern completion. Rather than presenting the entire data to the model, we make the observations be incomplete, presenting an incomplete prompt (part of each frame is missing).</p><p>Then we can see what the model beliefs are, what its posterior over output space, image space in this case, is. And we verify that the model is able to predict and complete the missing part of each frame.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*tOpv52vZfk5Ka83Ef7t1Aw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>Karl explains that this demonstrates efficient compression, highlighting how many aspects of Bayesian mechanics and variational inference are grounded in the same principles that support efficient information transfer. This relates to maximizing mutual information, how much valuable information can be retained and shared in the most compact representation possible.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2Fhtx_4shMnNc%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2Fhtx_4shMnNc&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2Fhtx_4shMnNc%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/9b30765c4011b3002cef85ef783a4a50/href">https://medium.com/media/9b30765c4011b3002cef85ef783a4a50/href</a></iframe><p>And so, we can consider that the objective function of the model is to find the best representation that retains the essential structure of the data (e.g., the bird’s orbit) while minimizing redundancy. This means encoding the data into a compact, generalizable form that allows the agent to perform tasks like filling in gaps or generating new examples.</p><p>Therefore, fast structure learning under renormalizing scale free generative models can be seen as the most efficient way of compressing whatever data the agent is exposed to.</p><p>Once this compression takes place, the agent is then able to fill in the gaps in very efficient ways.</p><p>Let’s emphasize the following two key concepts:</p><ul><li><strong>Renormalization</strong>: A process from physics that simplifies complex systems by focusing on large scale patterns while ignoring small scale details. In this context, it’s how the model moves between levels of abstraction (e.g., from pixel patches to full trajectories).</li><li><strong>Scale free</strong>: The model can operate across different levels of detail or scales (local features to global trajectories) without being tied to a specific resolution.</li></ul><p>And in this way:</p><ul><li>By transforming the data into a compact representation, the agent gains the ability to infer missing information or generate new content.</li><li><strong>Example</strong>: If part of the bird’s loop is obscured, the compressed model can reconstruct the missing frames, leveraging its understanding of the periodic trajectory, the bird’s motion dynamics, and other relevant patterns in the data.</li></ul><h3>Learning the chaos</h3><p>In the next example, Karl showcases a 3 level hierarchical renormalizing model, this time applied to the learning of the dynamics of a ball moving through a chaotic and aperiodic orbit.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FBXnnGQEIyoA%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2FBXnnGQEIyoA&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FBXnnGQEIyoA%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/bc5d7e1b1a5a9f2c56264219803e235c/href">https://medium.com/media/bc5d7e1b1a5a9f2c56264219803e235c/href</a></iframe><ul><li><strong>Chaotic and aperiodic motion</strong>: Unlike a simple periodic motion (e.g., a bird flying in a loop), this system is less predictable, with sensitive dependence on initial conditions. This introduces both deterministic (rule based) and stochastic (random) elements.</li><li><strong>Model’s task</strong>: To capture both the deterministic rules and the chaotic variations probabilistically, compressing the dynamics into a structured, learnable format.</li></ul><p>So how does the model deal with such a challenge?</p><ul><li><strong>Probabilistic representation</strong>: The model learns distributions over possible trajectories. This allows it to account for uncertainty and variability in the system.</li><li><strong>Compression</strong>: The model identifies patterns and reduces redundancy, creating a compact representation of the system’s behavior.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*lii-cGpCBx04gzDIr9dTug.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>After the learning takes place, we begin to test the capabilities of the model.</p><ul><li><strong>Input at the lowest level</strong>: At the start, the raw observations are provided in a deliberately imprecise way, meaning the fine grained details are blurred or noisy. This tests how well the model can rely on its learned priors rather than on precise data.</li><li>In such a case, everything is governed, generated from the top down priors or predictive posteriors, because there is no precise likelihood.</li><li><strong>Top down control</strong>: The model generates predictions about the system’s dynamics starting from the highest level (abstract priors) and passing them down to the lower levels.</li><li><strong>Posteriors</strong>: These are updated beliefs about what is happening in the system, based on a combination of <strong>priors (</strong>high level expectations from what the model has learned) and <strong>likelihoods </strong>(observations from the environment, which are imprecise in this case. The model is not receiving accurate input data. Therefore, it cannot rely heavily on direct sensory evidence).</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*PZiXqRRwfS03s1v7foA52w.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>And then we remove the imprecise input completely and let the agent generate its own beliefs about what’s going on.</p><ul><li>Now the model must rely entirely on its internal representations (beliefs and priors) to generate the ball’s motion.</li><li>The model autonomously predicts the dynamics of the ball based on what it has learned, effectively imagining the motion of the ball without any external input.</li><li>Despite this uncertainty, the model successfully reproduces the overall dynamics of the ball, demonstrating that it has captured the system’s behavior in its hierarchical generative framework.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*pwim07zfAKt_vEy0pCRe3Q.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FeC1HiGT2fJs%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2FeC1HiGT2fJs&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FeC1HiGT2fJs%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/194213fc86877bd8294c9e3460cc28ec/href">https://medium.com/media/194213fc86877bd8294c9e3460cc28ec/href</a></iframe><p>This example highlights how generative models can not only learn complex systems but also operate autonomously in very uncertain conditions by leveraging probabilistic structures and hierarchical abstractions.</p><h3>Exploring video and music</h3><p>In the next example, <strong>fast structure learning applied to discover and learn event sequences</strong>, Karl shows us the movie of a bird. After the learning takes place, some of the frames of the video are hidden to the generative model, and the model is able to reconstruct them faithfully, filling the gaps.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*C8cDKEmGHjCz1xQoA7fsMA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/639/1*fZEpwP38To3dkXT4Nw94hg.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FbE18VjvK0ME%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2FbE18VjvK0ME&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FbE18VjvK0ME%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/1ea272fe848617582973943d967464b5/href">https://medium.com/media/1ea272fe848617582973943d967464b5/href</a></iframe><p>Next is the turn for music, in the example titled: <strong>fast structure learning applied to generate music</strong>.</p><p>This example intends to compress 1D frequency summaries of a particular audio file. The audio is summarized as a continuous wavelet transform.</p><p>Frequency representations like spectrograms are typically 2D, showing frequency vs. time. However, when referring to 1D frequency summaries, Friston seems to be referring to a condensed representation that reduces this information to a single dimension.</p><p>A 1D summary might be a single vector that captures average power or amplitude across frequencies, essentially collapsing the time dimension. For instance, we can calculate an average amplitude for each frequency band over time, creating a 1D vector of frequency components. This provides a summary of the frequency content without the time detail, a sort of fingerprint for the audio.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*R9D1THjZZNnenEWoXpfdEA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>This is a good opportunity to review the types of charts Karl shows in these examples, specifically in relation to the three level hierarchical system in this music context.</p><p>First, a reminder about a couple of terms:</p><ul><li><strong>Predictive Posterior:</strong> A forecast of the state before receiving new data, based purely on the model’s prior beliefs and the dynamics of lower levels.</li><li><strong>Posterior:</strong> The updated belief about the state after receiving new observations and incorporating them into the model, essentially the new and more accurate belief about the state.</li></ul><p>The different charts shown by Karl are:</p><ul><li><strong>Predictive Posterior(states) level 1</strong></li><li><strong>Predictive Posterior(states) level 2</strong></li><li><strong>Predictive Posterior(states) level 3</strong></li><li><strong>Posterior (states) level 3</strong></li><li><strong>Predictive Posterior (paths) level 3</strong></li><li><strong>Predictive Posterior (paths) level 2</strong></li><li><strong>Transition chart (square matrix)</strong></li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/593/1*nOHk5T8CgfmXXN11naoSNA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>Each <strong>Predictive Posterior (states)</strong> chart for levels 1, 2, and 3 represents the probability distributions of predicted hidden states at different hierarchical levels. Those hierarchical levels correspond to different levels of abstraction. For example:</p><ul><li><strong>Level 1 </strong>could represent fine grained, immediate states like sensory features (e.g., specific notes or sounds in music).</li><li><strong>Level 2</strong> could capture patterns over time, like sequences or phrases in music.</li><li><strong>Level 3</strong> may represent higher level structures, such as motifs, themes, or the genre of a piece.</li></ul><p>The system predicts the states at each level based on the states of the levels below and above, updating these predictions based on new observations.</p><p>The <strong>Posterior (States) Level 3 </strong>chart shows the updated posterior distribution over the hidden states at Level 3, following Bayesian updating with observed data. It reflects the system’s refined beliefs about Level 3 states after accounting for incoming observations and prediction errors from lower levels.<br>This level’s posterior will help shape the priors (predictions) at the lower levels and inform the next iteration of predictions for Level 3 itself.</p><p>The <strong>Predictive Posterior (Paths)</strong> charts for levels 2 and 3 represent the system’s beliefs about possible paths or sequences at these hierarchical levels. These paths aren’t single states but sequences of states over time, predicting not only what states are likely but how states are expected to unfold across time.<br>For example:</p><ul><li>At <strong>Level 2</strong>, paths could represent sequences of sounds or notes in a phrase.</li><li>At <strong>Level 3</strong>, paths might represent more complex patterns like thematic progressions or chord changes.</li></ul><p>By predicting paths, the system can anticipate longer sequences and make structured predictions about the trajectory of events, crucial in a time dependent setting like music generation.</p><p>Finally, the <strong>transition Chart </strong>is a square matrix showing the transition probabilities between states within a particular level.<br>It defines how likely the system is to move from one state to another, capturing the dynamics or temporal dependencies between states.</p><p>For example, in a music model, this matrix might represent the likelihood of one note or chord transitioning to another within a specific musical structure. Each entry (i,j) in this matrix represents the probability of transitioning from state i to state j, helping the model understand and anticipate sequences based on learned transitions.</p><p>After showcasing these examples that deal with learning to compress and generate, Karl takes us onto the following section that deals with active learning and agency.</p><h3>Active learning and agency</h3><p>We will now focus on an agent equipped with one of these generative models interacting with an environment where it must learn a skill.</p><p>The key objective is to <strong>learn an effective and optimal state action policy</strong>. <br>A good mapping from states of the world to actions that the agent should take. And we can consider two routes to achieve that:</p><ul><li><strong>Use some function approximation strategy</strong>, through <strong>deep learning</strong>, for example, to learn a mapping from sensory observations to active states (actions) that maximizes a reward (or cost, or utility, or free energy). This would be<strong> the reinforcement learning route</strong>.</li><li>Or use<strong> active inference</strong> to realize predictions under a <strong>renormalising generative model</strong> of rewarded events. The agent uses its generative model to infer the states of the world and selects actions that minimize free energy, achieving its goals in an efficient and sustainable way. This is the route we are going to consider in this section.</li></ul><p>With the active inference approach to purposeful behavior, such purpose is provided mostly by the prior preferences or the constraints, usually encoded in the C tensor. <br>Therefore, unlike reinforcement learning, where purpose often comes from an external and explicit reward signal, active inference relies on the generative model itself to encode what the agent is trying to achieve.</p><p>Those <strong>prior preferences</strong> are, therefore, the agent’s built in or learned expectations about the kinds of observations it wants to encounter. The agent doesn’t seek to maximize a reward but instead aligns its actions to ensure its observations match these preferred outcomes.</p><p>The generative model serves as the agent’s internal understanding of the world and within the model itself, the prior preferences in the C tensor are encoded as probabilities over desired outcomes. For example:</p><ul><li>If an agent is navigating a maze, it might prefer observations that indicate it is closer to the goal.</li><li>A robot assembling objects might prefer states where objects are correctly aligned.</li><li>The preferences in the C tensor guide the agent’s actions by making certain observations or states more likely in the model.</li><li>Constraints might represent physical limitations (e.g., movement restrictions) or other environmental factors that shape how the agent interacts with the world.</li></ul><p>So we can see the purpose of the agent as, in a way, an emergent property, because the agent doesn’t explicitly have a goal like: maximize a reward. Instead, it acts to minimize <strong>variational free energy </strong>and <strong>expected free energy, </strong>which involve:</p><ul><li>Reducing the surprise (unexpectedness) of its observations.</li><li>Ensuring its sensory inputs are consistent with its prior preferences.</li><li>The <strong>C tensor</strong> effectively sets what counts as success for the agent. By minimizing free energy, the agent works to reduce prediction error, aligning its sensory input with the predictions made by its generative model. This process ensures that the agent’s actions bring its observations into better alignment with the world model, which is guided by the prior preferences encoded in the C tensor.</li></ul><p>Another very simple example. Imagine a thermostat as an agent:</p><ul><li>Its <strong>C tensor</strong> encodes a preference for a specific temperature range (e.g., 21–23°C).</li><li>When the room temperature deviates from this range, the thermostat acts (e.g., turning on the heater or cooler) to align observations (measured temperature) with its preference.</li></ul><p>This mechanism generalizes to more complex agents and tasks, where the C tensor<strong> can encode preferences</strong> for specific visual <strong>patterns</strong>, spatial arrangements, sequences of <strong>actions</strong>, or even <strong>ethical constraints</strong> and <strong>social norms</strong>. In such cases, the agent’s actions are driven not only by sensory preferences but <strong>also by abstract goals or principles</strong>, ensuring that its behavior aligns with <strong>both practical and moral considerations.</strong></p><p>Let’s now consider <strong>action </strong>in the context of this active inference agent. We can view <strong>action </strong>as part of a gradient flow on free energy. The agent will pick actions that minimize free energy, which leads to maximizing predictive accuracy. And this is because the process is built on top of the variational free energy (VFE) optimization process and equation, which balances complexity and accuracy and Karl tells us that complexity does not depend upon action, but accuracy does.</p><p>So when we consider the behavior of our active inference agent, we have a generative model that is optimized through perception (in terms of variational inference minimization), and that generates predictions about what should happen next. From those predictions, the agent can then pick its actions to fulfill those predictions (by also involving the expected free energy (EFE) optimization process, which handles action selection).</p><p>So, as an active inference agent, my predictions about what will happen next are shaped by perception through variational inference, based on my learned and selected generative model. This is very similar to what is called model predictive control. The actions we take have the goal of evidencing or realizing predicted states.</p><p>With that in mind, Karl invites us to explore a strategy to make the agent become an expert at something.</p><h3>Becoming an expert</h3><p>Imagine a world where everything unfolds according to your beliefs about what it means to be an expert in that world. The agent can plan a series of episodes, generating predictions that cascade from higher to lower levels of the model, with actions at the lowest level executed to faithfully fulfill those predictions.<br> <br>So we can think of <strong>control as inference at the bottom</strong> level (model predictive control) and<strong> planning as inference at the top </strong>level.</p><p>Let’s unpack this.</p><ul><li><strong>Control (bottom level):</strong> This involves making local predictions and adjusting actions in real time to fulfill those predictions (similar to Model Predictive Control).</li><li><strong>Planning (top level): </strong>This involves high level predictions about long term goals or episodes. The agent acts to fulfill those predictions by planning and generating sequences of future states.</li></ul><p>And let’s unpack it even more.</p><ul><li><strong>At the lower levels</strong>, which are closer to sensory input or immediate states, inference is used to take actions that fulfill short term predictions. For example, if a robot predicts it will be in a certain position after moving, it will act to bring itself closer to that predicted position, reducing prediction error. In this sense, lower level inference is focused on immediate control, taking actions now to align the system with its predictions.</li><li><strong>At the higher levels</strong> (more abstract, long term goals), the inference is used to plan sequences of actions that will fulfill long term predictions or reach future goals. Instead of directly acting on immediate predictions, the agent is planning ahead, figuring out which sequence of actions will likely achieve the desired outcomes over time. Therefore, the top level inference is about planning, selecting the best course of actions to fulfill long term predictions.</li></ul><p>So <strong>lower levels</strong> use inference to adjust immediate actions.<br>And<strong> higher levels</strong> use inference to generate and plan sequences of actions.</p><p>The transition from acting to planning can be considered to be flexible and dependent on many factors such as:</p><ul><li><strong>Temporal scale of predictions</strong> (short term vs long term).</li><li><strong>Level of abstraction</strong> (concrete actions vs abstract goals).</li><li><strong>Hierarchical structure</strong> of the system, where lower levels focus on actions, and higher levels handle planning.</li></ul><p>In hierarchical systems like the one Friston describes, the key is that lower levels focus on real time action while higher levels handle planning over longer time scales.</p><p>Let’s explore an example where we have a robot with a 7 level hierarchical system navigating through a maze:</p><ol><li><strong>Level 1 (Lowest level)</strong>: This level might be responsible for <strong>motor control</strong>, for example, turning the wheels left or right in response to obstacles detected by sensors.</li><li><strong>Level 2–4</strong>: These levels might help with <strong>local adjustments</strong>, like figuring out which direction to go next in the maze or dealing with specific obstacles.</li><li><strong>Level 5–6</strong>: These might be responsible for <strong>higher level strategy</strong>, such as <strong>planning out the entire path</strong> through the maze, determining which turns to make based on the robot’s position and ultimate goal (e.g., reaching the exit).</li><li><strong>Level 7 (Highest level)</strong>: This level might oversee the entire process, adjusting the <strong>overall plan</strong> based on higher level goals (like completing the maze as quickly as possible or optimizing the path for energy efficiency).</li></ol><p>In this example, <strong>lower levels act</strong> immediately based on sensory feedback, while <strong>higher levels plan</strong> the sequence of actions over a longer time scale.</p><p>Let’s now compare all of the above to Reinforcement Learning.</p><ul><li><strong>RL offers an architecture that is a bit more overengineered, Karl says. </strong>Active Inference, in contrast, is more elegant, focusing on minimizing free energy without the need for complex reward structures or explicit policy search.</li><li><strong>In RL, you need to have explicit inputs that are the external rewards (which constitute a proxy for the constraints under which the Bayesian mechanics would operate). </strong>Active Inference does not require explicit external rewards, relying instead on prior beliefs (including the preferences encoded in the C tensor) and the system’s generative model to guide actions and learning.</li><li><strong>In RL, the inference scheme is about selecting the right policy (e.g., through Q learning or policy gradient methods) to maximize the expected cumulative (discounted) reward</strong>. In Active Inference, the inference scheme revolves around minimizing free energy by continuously updating beliefs, generating predictions about future states, and selecting actions to align observations with prior preferences encoded in the generative model.</li><li><strong>RL is often focused on optimizing short term rewards through trial and error. </strong>Active Inference emphasizes long term goal fulfillment, with actions driven by predictions about the future rather than immediate rewards.</li><li><strong>RL agents may struggle with sparse or delayed rewards, making learning slower and less efficient. </strong>Active Inference continuously updates its generative model, which allows for more efficient learning even in the face of sparse feedback or uncertainty.</li><li><strong>In RL, the exploration exploitation trade off is a central challenge</strong>, requiring strategies to achieve a better balance between trying new actions (exploration) and exploiting the best known routes (exploitation). In contrast, active inference handles exploration more naturally, as the agent explores to reduce future uncertainty and improve its model, rather than to just maximize rewards.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*okL5VsbO0fzog9mMcUsEZA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>It’s now time to see a specific example of this in action, Karl says.</p><h3>Learning to play Pong</h3><p>We have seen that planning or imagining a future can occur at the highest levels of these models, levels where the focus is purely on abstract patterns, with no direct notion of action. At the lowest levels, however, the system can bring these predictive posteriors to life by selecting actions that realize the predicted outcomes.</p><p>And how can we use this framework to<strong> learn a state action policy</strong> from scratch?</p><p>Through <strong>active data selection</strong>, Karl explains. To demonstrate this, we will explore an example showcasing an active inference agent learning a state action policy to play the game of Pong.</p><p>In this context, the focus shifts from selecting the type of model to <strong>selecting the training data </strong>itself. The agent must evaluate sequences by asking: does this sequence end in a reward?</p><ul><li><strong>Yes</strong>: admit this sequence for fast structure learning</li><li><strong>No</strong>: ignore this sequence</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/785/1*ao89AgGukcespfducgYWEA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FFRCJ1ZQcVlU%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2FFRCJ1ZQcVlU&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FFRCJ1ZQcVlU%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/7351728a6452fb55686124d090558e58/href">https://medium.com/media/7351728a6452fb55686124d090558e58/href</a></iframe><p>Karl tells us that <strong>fast structure learning</strong> builds a generative model of events that lead to a reward (and excludes everything else). In short, fast structure learning helps the agent quickly focus on understanding the cause and effect relationships that are most relevant for achieving its goals, filtering out everything that isn’t directly contributing to those goals.</p><p>Unpacking this.</p><ul><li>We only select something if it minimizes EFE (including expected cost). In other words, as an agent, I’ll only select data features or training sequences that minimize cost and are associated with a reward.</li><li>Therefore, we <strong>select instances</strong> of random play<strong> that align with our goals</strong>, the ones that minimize EFE and avoid costly outcomes. In summary, we <strong>generate training sequences through random play </strong>and ignore those that result in a high cost.</li><li>After securing a reward, we restart the random play and use subsequent sequences if they also achieve rewards.</li><li>We chain, bootstrap, or splice together sequences of rewarded play, which <strong>fast structure learning</strong> uses to build a generative model of expert play.</li><li>This creates a generative model that encodes a state action policy, which can be realized through reflective active inference. The agent acts in accordance with its learned generative model, making its behavior resemble that of an expert.</li><li>The trained agent recognizes only expert play, which ensures that all its predictions align with those of an expert player. Because actions fulfill these predictions, the agent will appear to behave like an expert player.</li><li>The reward is expressed in the charts as free energy (which is minimized). So the reward can be seen as a process of reducing free energy through actions that align with the predictions of the model.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*N-XVuS7ESa7pE-j462rvJg.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FxCOzdI2e8fk%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2FxCOzdI2e8fk&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FxCOzdI2e8fk%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/adc411adad5a709ed8f784ff1d56da6d/href">https://medium.com/media/adc411adad5a709ed8f784ff1d56da6d/href</a></iframe><p>And now comes a key point. Friston says that we have switched on inductive inference at the highest level .</p><ul><li>Inductive inference involves making predictions about the world based on prior observations and patterns. At the highest level of the agent’s model (the top level), the agent applies inductive reasoning to predict what it needs to do in order to reach desired outcomes (rewarded states).</li><li><strong>Sequences and projections through the model</strong>: The agent considers potential future states (or sequences of states) and works out how those states are encoded in its model. It projects these predictions down the hierarchy to lower levels, essentially checking if those predicted sequences of events lead to desired outcomes.</li><li><strong>Intended states</strong>: If a predicted sequence includes a reward or a hit, that sequence is considered an intended state. This is the state the agent wants to move toward in future timesteps.</li><li><strong>Avoiding undesired states</strong>: The agent uses this knowledge to figure out which states it must avoid in order to reach its intended states. It works out what actions or paths could lead to undesirable outcomes that would prevent it from reaching its goals.</li><li><strong>In summary</strong>: In simpler terms, the agent uses inductive inference to guide its future actions by predicting the sequence of states that lead to rewards, figuring out which ones to avoid, and then adjusting its actions accordingly.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*9X4VqnbgwCyKMiG-moxefQ.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>So inductive inference is applied at the top level, where the agent is essentially told: ‘This world is under your control. Imagine what you want to happen.’ In other words, the agent is tasked with controlling the world and envisioning the outcomes it seeks to achieve.</p><p>The agent then predicts what will occur based on its learned model and determines which sequences of events are necessary to achieve its goals. Once these predictions are made, it deploys reflexive actions (or behaviors) to make them come true and reach its intended states.</p><p>Let’s summarize the full process.</p><ul><li><strong>Random Play and Selection of Rewarded Training Sequences</strong><br>The agent begins by engaging in random play, exploring various actions and observing their outcomes. We only select the sequences that result in rewarded or desired outcomes (sequences that minimize free energy). This provides the agent with a set of examples that lead to good results.</li><li><strong>Fast Structure Learning on Selected Sequences</strong><br>Next, we apply <strong>fast structure learning</strong> to the selected rewarded sequences. This enables the agent to learn the underlying structure and dynamics of the task, essentially creating a model of the system based on these examples.</li><li><strong>Applying Inductive Inference During Actual Play</strong><br>Once the agent is actively playing, it uses <strong>inductive inference</strong> to plan its actions. The agent looks at the high level intended or rewarded outcomes it wants to achieve and traces back through its generative model to identify which sequences of states and actions are most likely to lead to those desired outcomes. This allows the agent to select actions aligned with its goals, rather than merely exploring randomly.</li></ul><p>The key difference between steps 2 and 3 is that:</p><ul><li><strong>Step 2</strong> (fast structure learning) involves building a generative model offline using the rewarded sequences.</li><li><strong>Step 3</strong> (inductive inference) uses that model to guide decision making and action selection during actual play.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/421/1*UHY7aFPtY12y1uJomQGvjQ.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>The reason we need the inductive inference step, even though we’ve already trained on the rewarded sequences, is that the generative model learned in step 2 captures the overall structure and dynamics, but doesn’t necessarily prescribe the exact sequence of actions to take at each moment.</p><p><strong>Inductive inference</strong> allows the agent to plan and adapt dynamically during play, based on the current state of the game, rather than simply replaying previously learned sequences. This makes the agent’s decision making more flexible and goal oriented.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*8cuN2OSQRNG3Pz3h2XeETw.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2F_ENjiwSMPCg%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fshorts%2F_ENjiwSMPCg&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2F_ENjiwSMPCg%2Fhq2.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/c585b7f9b7cb29f42fbd6d534e3fae51/href">https://medium.com/media/c585b7f9b7cb29f42fbd6d534e3fae51/href</a></iframe><p>Next, Karl explains that focusing exclusively on rewarded sequences can result in a model that may be somewhat brittle, as it lacks exposure to a broader range of possibilities. He then discusses how incorporating more diverse sequences can increase the robustness of the model.</p><h3>All roads lead to Rome</h3><p>To make the model more robust, Karl Friston suggests expanding the learned model by incorporating more pathways that lead to desirable outcomes (rewards).</p><ul><li><strong>Creating an Attracting Set</strong><br>At the beginning of the process we have created a set of episodes (or sequences of states and actions) that lead to rewarding outcomes. This set is called an <strong>attracting set</strong>, where any path through the model will eventually lead to a reward. This attracting set represents the set of all desirable or goal directed states within the agent’s generative model.</li><li><strong>Augmenting the Model with Additional Paths</strong><br>To make the model more comprehensive, we can add more episodes or sequences of actions, as long as they eventually lead to one of the states in the attracting set. These new episodes help the agent explore different paths or strategies while still ensuring that the end result is a reward.</li><li><strong>Expanding the Generative Model</strong><br>By adding these additional pathways, we are effectively enriching the agent’s generative model. Instead of relying only on a small set of rewarded sequences, the model can now consider a broader range of possibilities that ultimately lead to a desirable outcome. This expansion helps the agent learn more about the environment and increase its robustness by exploring diverse paths to the same goal.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*6eyQuyuKGkwWvinIPUaJpA.jpeg" /><figcaption>Image by Karl Friston via <a href="https://www.youtube.com/watch?v=ee6_mNOfP38">https://www.youtube.com/watch?v=ee6_mNOfP38</a></figcaption></figure><p>This concludes our fascinating journey into scale free active inference, as explained by Karl Friston. <strong>You can watch Karl Friston’s full talk</strong>, along with many others on active inference, <strong>via the following link </strong>to the latest Active Inference Symposium.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2Fee6_mNOfP38%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3Dee6_mNOfP38&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2Fee6_mNOfP38%2Fhqdefault.jpg&amp;type=text%2Fhtml&amp;schema=youtube" width="640" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/4c72744c3a3c240c8f1011bdee14571f/href">https://medium.com/media/4c72744c3a3c240c8f1011bdee14571f/href</a></iframe><p>Stay updated on the latest developments in active inference through the publications and streams of the Active Inference Institute.</p><p><a href="https://www.activeinference.institute/">The Active Inference Institute</a></p><h3>In Plain English 🚀</h3><p><em>Thank you for being a part of the </em><a href="https://plainenglish.io/"><strong><em>In Plain English</em></strong></a><em> community! Before you go:</em></p><ul><li>Be sure to <strong>clap</strong> and <strong>follow</strong> the writer ️👏<strong>️️</strong></li><li>Follow us: <a href="https://x.com/inPlainEngHQ"><strong>X</strong></a> | <a href="https://www.linkedin.com/company/inplainenglish/"><strong>LinkedIn</strong></a> | <a href="https://www.youtube.com/channel/UCtipWUghju290NWcn8jhyAw"><strong>YouTube</strong></a> | <a href="https://discord.gg/in-plain-english-709094664682340443"><strong>Discord</strong></a> | <a href="https://newsletter.plainenglish.io/"><strong>Newsletter</strong></a> | <a href="https://open.spotify.com/show/7qxylRWKhvZwMz2WuEoua0"><strong>Podcast</strong></a></li><li><a href="https://differ.blog/"><strong>Create a free AI-powered blog on Differ.</strong></a></li><li>More content at <a href="https://plainenglish.io/"><strong>PlainEnglish.io</strong></a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=54503b099a07" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/how-to-grow-a-sustainable-artificial-mind-from-scratch-54503b099a07">How to grow a sustainable artificial mind from scratch</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Paths to the future of AI]]></title>
            <link>https://ai.plainenglish.io/paths-to-the-future-of-ai-97ebe423f295?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/97ebe423f295</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[active-inference]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[agi]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Wed, 13 Mar 2024 07:28:37 GMT</pubDate>
            <atom:updated>2024-03-26T08:43:09.463Z</atom:updated>
            <content:encoded><![CDATA[<h4>Exploring some of the possible routes to reach artificial general intelligence</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*lllP2X7bNPIy606KroQmjw.jpeg" /><figcaption>Paths to the future of AI | Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>As impressive as AI is today, its limitations are clear whenever we look at how ChatGPT, self driving cars or Midjourney work. They are very useful as prototyping or brainstorming tools, and can take you sometimes 99% of the way to a useful result. But would you put any of those systems in charge of a process in which failures had potentially catastrophic consequences? Of course not. But why do these systems sometimes make silly mistakes, “hallucinate”, or simply fail to “understand” what they are doing?</p><h3>Dual Process Theory</h3><p>Dual process theory, popularized by nobel prize Daniel Kahneman, gives us a straight forward abstraction to understand the limitations of today’s AI. This theory tells us that we can consider two ways our brain can process information. We call them: System 1 and System 2.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*2yzmK8N-XlX-eO-8Q_zhEQ.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><h3>System 1</h3><p>System 1 is the type of information processing that we perform automatically, subconsciously. It is, essentially, pattern matching. A quick mapping from perception to action, or from a certain need to a response. It is quick and cheap. Living organisms rely a lot of System 1 responses because of two main reasons:</p><ul><li>Automating your responses saves energy. It is better than having to figure them out all over again. Less energy spending equals a more efficient life.</li><li>If you are facing a danger, like a tiger that is about to eat you, you have no time to reflect or ponder. You have to generate a response fast!</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Sme9voAliZUNhJ7TwbUyHg.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>Think of System 1 as an analog to human intuition, to your subconscious automatic mapping processes. Human intuition is very powerful. However, sometimes it gives you the wrong answer. Why?</p><p>Human intuition gives you the best approximation it can derive from the knowledge you have. But if you don’t have enough relevant knowledge for the scenario at hand, it will still produce an approximation, but an incorrect one. And here comes the key. An incorrect approximation which your intuition will still present to you as if it was the truth. Does this sound familiar? Yes, because that’s the same behavior you find in GPT, Gemini, Claude, and other AI systems of today.</p><p>How do we go beyond System 1? Let’s explore it.</p><h3>System 2</h3><p>A living organism, whenever it finds a new unknown scenario, or whenever the approximation presented by its intuition (System 1) doesn’t seem correct, can employ another way of processing information, System 2, slow thinking.</p><p>Think of when you need to learn to drive, or play a new piano piece. You need to slowly, step by step, find the algorithm, the sequence of actions that will solve that new scenario.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Pxnb-Azvg9ZbpFKephXGow.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>But, how do you find that optimal sequence of actions?</p><h3>World Models</h3><p>A robust System 2 needs a robust world model. A world model is a simplified abstraction of how the world works.</p><p>Babies gradually learn their world model mainly by observation, and a little by interaction.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*R9Edtq5HrYoyHle3yoDzZQ.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>If you have such an abstraction in your mind, and if it is robust enough, you can do something really special. You can plan. You can do that because, by using your world model, you are able to predict the consequences of taking different sequences of actions.</p><p>So when an organism does slow thinking, it makes use of the predictive capabilities of a world model in combination with other modules (a cost module to estimate how far you are from your objective, a memory module, etc), to find the optimal sequence of actions for a certain goal.</p><p>This is why System 2 is way more expensive for the organism than System 1. And that’s why you cannot employ System 2 processing at all times. It would burn you out. So once you have learnt a new algorithm with System 2, you transfer it to System 1, you automate it.</p><p>And so, once you have learnt to drive or to play that piano piece, you can do those asks without paying attention, automatically, subconsciously, with just your System 1 capabilities.</p><p>In summary, good System 2 capabilities allow you to reason and plan in effective ways.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ryXFXrsNOULAOCbiOnuDxw.jpeg" /><figcaption>Image created by Javier Ideami | Images within the image are from unsplash.com</figcaption></figure><p>So what about AI? Artificial intelligence has mastered System 1 capabilities at a superhuman level.</p><p>However, it has very limited System 2 capabilities. Most experts agree that, although within the latent spaces of a system like GPT there probably exists some kind of abstract representation of the world, it is a very brittle and fragile representation. And it is an implicit one.</p><p>AI needs robust world models, trained in explicit ways to function as effective predictive systems. Only then can the organism plan and reason well enough.</p><h3>The road to ASI</h3><p>System 1 + System 2 equal human capabilities. That’s what many people call AGI, artificial general intelligence. AGI is not a popular term among most experts because intelligence in general is fairly specialized rather than general. It is specialized to the specific needs an organism has to survive.</p><p>I will use the term AGI in this article simply because it has become sort of a used-by-everybody term to denote human like intelligence. But it is important to clarify that it is a bit of a misnomer. Alternatives to AGI are other terms like AHI = artificial human intelligence</p><p>So what does the path ahead look like?</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qeH0Cxvxof0EoAKgv-URSQ.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><ul><li>Today we have ANI = artificial narrow intelligence</li><li>As we develop system 2 capabilities we will be reaching other stages like ARI (artificial rat intelligence), AMI (artificial mouse intelligence) or ACI (artificial cat intelligence)</li><li>Eventually we will arrive to AGI, or AHI (artificial human intelligence)</li><li>And after that? We will keep going beyond human intelligence, towards ASI (artificial super intelligence)</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Hw1gVvySazg-Plu0zcp-Ew.jpeg" /><figcaption>Graphic and Cookies created by Javier Ideami | ideami.com</figcaption></figure><p>So, the question is, what are the potential paths ahead of us to reach those robust System 2 capabilities that are essential to move on from the current ANI? Let’s explore some of them.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zIm1ltrj1AMWQOlZMi0O8w.jpeg" /><figcaption>Image created by Javier Ideami | ideami.com</figcaption></figure><h3>But it’s already here!</h3><p>There is a very small minority in the AI community that believe that AI may already be sentient. And that its System 2 capabilities are already there or almost there. This is a very tiny minority and most AI experts do not agree with this view.</p><h3>Deep learning all the way</h3><p>(with sprinkles of RL, discrete search, etc)</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*OsaL8roHwEs9jDDCMxpbEg.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>We enter mainstream territory. This is the view led by the likes of OpenAI or Google. This camp believes that deep learning will be the core technology that will take us all the way to AGI.</p><p>This would happen through a combination of scale (data, compute, etc) and tweaks and evolutions of the architectures being used by deep learning systems.</p><p>Although deep learning would be the core, approaches like reinforcement learning, discrete search and others would complement that core on its way towards AGI.</p><p>Let’s summarize the potential of this approach.</p><ul><li><strong>Challenges</strong>: questions remain as to how robust would be the world models and System 2 capabilities of solutions centered mainly on deep learning.</li><li><strong>Cost</strong>: very high. Deep learning systems spend massive amounts of energy compared to, say, how the human brain works.</li><li><strong>What’s next</strong>: look to scaling + improvements in deep learning architectures in the next few years, as planning and other system 2 capabilities begin to be unlocked.</li><li><strong>Examples</strong>: OpenAI &amp; Google Approaches (GPT, Gemini)</li></ul><h3>Hybrid &amp; multimodular approaches</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*7Wic0D4IOgwm-26Bk4m0KQ.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>Hybrid and multimodular systems may combine different approaches. I will focus here on an example that will still use deep learning as its core technology. The key is that these solutions will be more divergent in terms of moving from the typically monolithic systems of today towards others that are more modular, closer to how our brain is partitioned in different functional areas.</p><p>Consider the human brain. An area called the hippocampus deals with memory formation and spatial navigation. Another called the basal ganglia is related to our automatic behaviors, habit formation and reward processing. Yet another, the cerebellum, deals with fine tuning of movement, balance and coordination.</p><p>In that spirit, Yann LeCun, one of the godfathers of AI, has proposed a new kind of multimodular architecture in his academic paper “<a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">A Path towards autonomous machine intelligence</a>”.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*gvtWwVBSVmz0dP5SV2rQlw.jpeg" /><figcaption>Captured from Yann LeCun’s public academic paper at <a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">https://openreview.net/pdf?id=BZ5a1r-kVsf</a></figcaption></figure><p>His proposal combines a number of modules to implement both System 1 and System 2 capabilities. Some of those modules are:</p><ul><li><strong>The configurator module</strong>. A kind of a master module that connects to all the others and sets their parameters to adapt their functionality to the current goals.</li><li><strong>The Perception module: </strong>it takes the input to the senses of the agent and abstracts it into a latent representation of the state of the world. This latent representation may be expressed hierarchically at different levels of abstraction. In our brain, this module would correspond to parts of our visual cortex, auditory cortex, etc.</li><li><strong>The World Model:</strong> The most complex part of the architecture. It allows the agent to estimate missing information about the current state of the world, predict world states from previous ones or from actions proposed by the actor module, etc. Yann LeCun has proposed the JEPA architecture to implement the world model. Initial versions called I-Jepa and V-Jepa (trained with images and videos respectively) have been launched already, and work continues in this key part of the proposal.</li></ul><p><a href="https://github.com/facebookresearch/jepa">GitHub - facebookresearch/jepa: PyTorch code and models for V-JEPA self-supervised learning from video.</a></p><ul><li><strong>The Cost module</strong>: It measures an amount that Yann calls “energy”, which expresses how far the agent is from “comfort”, or the distance between where the agent is and where it wants to be in relation to different drives and goals. The ultimate goal of the agent is to minimize this energy, this cost. The cost combines two terms, Yann explains, the<strong> Intrinsic cost module</strong>, which is hard wired and computes the instantaneous present “discomfort” of the agent, and the <strong>trainable critic module</strong>, which is used to predict future intrinsic energies. Some parts of the intrinsic cost module can be compared to the basal ganglia and the amygdala in humans, whereas parts of our prefrontal cortex involved in reward prediction would correspond to the trainable critic module.</li><li><strong>The Memory module</strong>: It stores useful data about past, present and future world states, together with their associated intrinsic costs. This is useful, for example, to train the critic module (which estimates future intrinsic costs). The memory module can be compared to the hippocampus in humans.</li><li><strong>The Actor module</strong>: it creates proposals for action sequences that may be used to solve a new scenario. It also sends actions to the system’s actuators. The actor is at the center of how System 1 and System 2 processes are implemented. It has two components. One is a <strong>policy module</strong> that quickly maps world states (derived from the perception module) to actions. This is the base of System 1. And the second is the action optimizer that performs model-predictive control. This is the base of System 2. This second component, when combined with the world model and the cost module, can be used to learn new algorithms, by gradually finding an optimum sequence of actions that minimizes the cost associated with them. In our brain, areas of the premotor cortex that deal with proposing and encoding motor plans could correspond to parts of this module.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zcQwRpdo8jyDYTOVYZfIvQ.jpeg" /><figcaption>Captured from Yann LeCun’s public academic paper at <a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">https://openreview.net/pdf?id=BZ5a1r-kVsf</a></figcaption></figure><p>For details, check his academic paper. You can also explore an infographic and article I wrote about Yann’s proposal:</p><p><a href="https://medium.com/@ideami/the-tower-of-mind-towards-a-better-chatgpt-fafd8935e663">The tower of mind, towards a better ChatGPT</a></p><p>Let’s summarize this approach:</p><ul><li><strong>Challenges</strong>: questions remain as to how to solve many key pending issues, including the robust world model (research and the first versions of the JEPA architecture are in their initial phases), the memory module, the setting of goals and others.</li><li><strong>Cost</strong>: high. Cost is likely to be similar to the previous approach.</li><li><strong>What’s next</strong>: look to research on the different modules of these hybrid and multimodular architectures in the next few years as planning and other system 2 capabilities begin to be unlocked.</li><li><strong>Examples</strong>: LeCun’s proposal</li></ul><h3>Active Inference</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*g7fDzm7C4qPBaAdFe_60xw.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>Active Inference represents one of the most promising alternative approaches to the deep learning revolution. It is based on a completely different paradigm, focused on Bayesian mathematics.</p><p>Bayesian mathematics is all about probabilistic computation, which can be used to update in real time the beliefs about the world of an agent.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*dsrOr3TxHGsdd_NKAPIgfQ.jpeg" /><figcaption>Image created by Javier Ideami | ideami.com</figcaption></figure><p>Those beliefs of the agent constitute its world model. Active inference has different ways of implementing the mechanics of how that world model is updated in real time, and how the agent uses that world model to decide what to do at each moment, in response to the new observations (data) it continuously receives from its surroundings.</p><p>At the base of Active Inference is the Free Energy Principle, which stipulates that living organisms behave as if they were constantly minimizing their free energy. You can think of the free energy as if it was equivalent to surprise, to the difference between the predictions of the agent regarding what it is about to perceive, and the actual observations it receives.</p><p>To minimize surprise, the active inference agent can do two things: it can update its beliefs, its world model, or it can act upon the world to change its observations. So, as you can see, in active inference, perceiving and acting are two sides of the same coin.</p><p>By minimizing surprise, you improve your predictions, which improves your world model. You may then ask, if the goal is to minimize surprise, what about curiosity and creativity?</p><p>Active Inference has a good answer to that. The agent tries to minimize surprise, but also to move towards a number of goals and preferences that can be encoded in the mind of the agent. So both, minimizing surprise and moving towards goals and preferences, are important drivers.</p><p>But not only that. Minimizing surprise can focus on shorter or longer time horizons. And sometimes, minimizing surprise long term, requires the agent to first explore its surroundings in the short term, to gain enough knowledge about its environment.</p><p>It is just one scenario within all the complexity that can arise from active inference implementations. This is all accounted for by different equations. One of them, as an example, deals with the mechanics of action selection, and depends on a number of factors, such as :</p><ul><li><strong>Pragmatic value</strong>: how much do different action sequences help the agent achieve its goals?</li><li><strong>Epistemic value</strong>: how much do different action sequences help the agent gain new knowledge about its environment?</li><li><strong>Entropic value</strong>: how much uncertainty does the agent have about the consequences of taking different sequences of actions?</li></ul><p>As those factors interact with each other, the agent will be driven to either explore its environment or exploit its current knowledge.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Rf7eLGPrhlWEqgD1wLA3HQ.jpeg" /><figcaption>Image created by Javier Ideami | ideami.com</figcaption></figure><p>Active inference has some very important advantages over deep learning. For example:</p><ul><li>It implements continual learning. You don’t need to retrain the system every time you want to add new knowledge. The agent is continuously updating its beliefs in real time.</li><li>It is explainable and transparent. You can examine the “mind” of the agent, its inner computations, in order to explore the rationale behind its decisions.</li><li>It requires much less data to learn, orders of magnitude less.</li><li>It spends much less energy, thousands of times less, potentially. In that sense, it gets much closer to how our brain works.</li><li>It can implement sophisticated System 2 capabilities.</li><li>Learnings can be transferred between agents.</li><li>It is based on robust science. Active Inference was founded by Karl Friston, the most academically cited neuroscientist in the world.</li></ul><p>Those are just some of its advantages. But if active inference is so great, why isn’t it dominating the field?</p><p>Because until very recently, something called the Bayesian Wall was in the way. The Bayesian wall means that bayesian computations are slow and expensive, so scaling up these systems was considered impractical.</p><p>However, a company called Verses (Karl Friston is part of it), has recently announced that they have broken the Bayesian wall, finding a way to solve the scaling problem of active inference.</p><p><a href="https://www.linkedin.com/pulse/verses-ais-active-inference-outperforms-deep-learning-denise-holt-ktp9c">VERSES AI&#39;s Active Inference Outperforms Deep Learning in Historic AI Industry Benchmark Test</a></p><p>This could potentially be a game changer. Verses has announced that in the summer of 2024 they will release a series of academic papers explaining their progress. In the meantime, they have released a demo + preview, a summary of which you can watch in the following video:</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2F8uj4g_4nH-8%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3D8uj4g_4nH-8&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2F8uj4g_4nH-8%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/6fc26fb59a68fcb5cd1f50c70aed938f/href">https://medium.com/media/6fc26fb59a68fcb5cd1f50c70aed938f/href</a></iframe><p>Let’s summarize this approach:</p><ul><li><strong>Challenges</strong>: questions remain about how well active inference implementations will scale. The related academic papers, which Verses is planning to release in the summer, will be a great starting point to evaluate the possibilities of scaling this approach.</li><li><strong>Cost</strong>: low. Active inference implementations would use much less energy than deep learning implementations.</li><li><strong>What’s next</strong>: explore the recent announcement by Verses and stay tuned to their new academic papers, slated to be released this summer.</li><li><strong>Examples</strong>: Verses implementation of Active Inference.</li></ul><h3>Mixture of approaches</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qeH0Cxvxof0EoAKgv-URSQ.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>There are many chances that previous approaches will eventually converge.</p><p>If architectures like JEPA begin to show promise and parts of Yann LeCun proposal start to be implemented, that is likely to influence the rest of the deep learning community. We could see at that time or even earlier other multimodular and hybrid approaches beginning to emerge.</p><p>If what Verses has announced, regarding the breaking of the bayesian wall, is confirmed, and their active inference implementation begins to beat deep learning benchmarks, it is likely that parts of the deep learning community will begin to explore how to combine both approaches.</p><p>The combination of the impressive system 1 capabilities of deep learning, with the explainability, efficiency and transparency of probabilistic approaches like active inference, and the benefits of multi modularity, could produce a great diversity of new combinations of approaches on the way to AGI.</p><p>Let’s summarize this approach:</p><ul><li><strong>Challenges</strong>: plenty, regarding how well these approaches will combine with each other. Engineering challenges may appear that may be hard to solve.</li><li><strong>Cost</strong>: it’s possible that a mixture of some of these approaches will be more efficient than one relying solely on deep learning.</li><li><strong>What’s next</strong>: stay tuned for progress on the approaches highlighted above.</li></ul><h3>Other possibilities</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*udW7WFMbxCaGE76k9ZFxyw.jpeg" /><figcaption>Image AI generated by Javier Ideami | ideami.com</figcaption></figure><p>Beyond the previous approaches, there are other possibilities to consider.</p><ul><li>A part of the AI community believes that more advanced technology, such as quantum computing, may be essential to achieve true AGI.</li><li>Parts of the AI community believe that embodiment will be essential to get to AGI. That is, AIs interacting through sensors and actuators with the full complexity of a world like ours (be it virtually or physically).</li><li>In the spirit of professor Kenneth Stanley and his fabulous book “Why greatness cannot be planned”, the intermediate steps towards AGI may look nothing like the objective and be really counterintuitive. So it may be that we will be surprised in the next years by unexpected new kinds of architectures and algorithms that may appear due to following a gradient of interestingness during the everyday exploratory activities of researchers.</li></ul><p>This concludes our exploration of some of the possible paths to AGI. The one thing that is certain is that we live in a historic moment and that the coming future will surely bring plenty of surprises, guaranteed excitement and also important developments which will require we all to do our best to push this technology in the direction that brings the most benefits to humanity.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zIm1ltrj1AMWQOlZMi0O8w.jpeg" /><figcaption>Image created by Javier Ideami | ideami.com</figcaption></figure><h3>In Plain English 🚀</h3><p><em>Thank you for being a part of the </em><a href="https://plainenglish.io/"><strong><em>In Plain English</em></strong></a><em> community! Before you go:</em></p><ul><li>Be sure to <strong>clap</strong> and <strong>follow</strong> the writer ️👏<strong>️️</strong></li><li>Follow us: <a href="https://twitter.com/inPlainEngHQ"><strong>X</strong></a><strong> | </strong><a href="https://www.linkedin.com/company/inplainenglish/"><strong>LinkedIn</strong></a><strong> | </strong><a href="https://www.youtube.com/channel/UCtipWUghju290NWcn8jhyAw"><strong>YouTube</strong></a><strong> | </strong><a href="https://discord.gg/in-plain-english-709094664682340443"><strong>Discord</strong></a><strong> | </strong><a href="https://newsletter.plainenglish.io/"><strong>Newsletter</strong></a></li><li>Visit our other platforms: <a href="https://stackademic.com/"><strong>Stackademic</strong></a><strong> | </strong><a href="https://cofeed.app/"><strong>CoFeed</strong></a><strong> | </strong><a href="https://venturemagazine.net/"><strong>Venture</strong></a><strong> | </strong><a href="https://blog.cubed.run/"><strong>Cubed</strong></a></li><li>More content at <a href="https://plainenglish.io/"><strong>PlainEnglish.io</strong></a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=97ebe423f295" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/paths-to-the-future-of-ai-97ebe423f295">Paths to the future of AI</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[The tower of mind, towards a better ChatGPT]]></title>
            <link>https://ai.plainenglish.io/the-tower-of-mind-towards-a-better-chatgpt-fafd8935e663?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/fafd8935e663</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[chatgpt]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Mon, 10 Jul 2023 07:54:42 GMT</pubDate>
            <atom:updated>2024-06-19T21:51:48.142Z</atom:updated>
            <content:encoded><![CDATA[<h3>The Tower of Mind, towards a better ChatGPT</h3><h4>How new architectural paradigms can fix the limitations of systems like ChatGPT</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*3NxT1Lc_EmBsSx1-VTcSkg.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p><strong>ChatGPT </strong>has taken the world by storm. And yet, its limitations are many. It often <strong>hallucinates </strong>and presents straight out lies as if they were facts.</p><p>In a way, it is like our human intuition: fast, powerful and confident, sometimes too confident for its own good!</p><p>Recently, <a href="https://en.wikipedia.org/wiki/Yann_LeCun"><strong>Yann LeCun</strong></a>, one of the godfathers of AI, published a fascinating paper titled “<strong>A Path Towards Autonomous Machine Intelligence</strong>”. In this theoretical paper, he outlines a proposal for a new kind of multimodular architecture that could fix many of the limitations of systems like ChatGPT and move us closer to <a href="https://en.wikipedia.org/wiki/Artificial_general_intelligence">AGI</a>.</p><p>I have combined parts of Yann’s paper (which is rather long at 60 pages) with one of my visual metaphors in order to produce the Tower of Mind infographic, which simplifies and visually expresses some key parts of his proposal.</p><p>Throughout the infographic, we will be reviewing key aspects of how our human mind works and simultaneously we will connect those aspects with this potential future AI architecture proposed by Yann LeCun. We will see that there are a lot of parallelisms and connections to be made between Yann’s proposal and the way the brain works. Let’s begin!</p><h3>Thinking Fast and Slow</h3><p>At the right side of the infographic we establish a key objective of this new architectural paradigm: to implement two different ways of processing information, two ways or modes that match the way our brain works, the modes that nobel prize Daniel Kahneman explains so well in his famous book “<a href="https://www.amazon.com/Thinking-Fast-Slow-Daniel-Kahneman/dp/0374533555">Thinking Fast and Slow</a>”: <strong>System-1</strong> and <strong>System-2</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*HD3Nv8v1aZ02AUv8TP9ICA.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>There is an evolutionary reason for these two ways of processing information to exist in our brain, and perhaps, in most intelligent systems.</p><p>Whenever we encounter a new scenario (for which we haven’t previously learnt a response pattern), we need to find the sequence of steps, the algorithm to solve such a scenario. For this, we employ what Kahneman calls System-2. We slowly look for that new algorithm, paying close attention in a systematic and conscious way as we slowly reason our way to a new response pattern.</p><p>The problem with our System-2 is that it is slow and expensive. It requires a lot of cognitive effort and fuel (the glucose that powers our brain). We cannot use such a mode of thinking all day. Therefore, once we learn a new response pattern, we proceed to automate it. We make it subconscious. We transfer it from System-2 to System-1.</p><p>Our System-1 is fast and much cheaper. It is also subconscious and powers our intuition and the associative machinery that has a lot to do with what we call “Creativity”. It quickly connects perceptions with actions, or goals with response patterns.</p><p>Metaphorically speaking, our subconscious is like a kitchen pot that combines and recombines the information we absorb (which we also compress and abstract).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*N8MnbTNYi38pSYvTRU_oFg.jpeg" /><figcaption>Image AI generated by Javier ideami | ideami.com</figcaption></figure><p>Within the subconscious realm, System-1 processes quickly map inputs to outputs, bypassing the slow System-2.</p><blockquote>When a tiger is about to eat us, we have no time to consciously reflect on what to do. We must map perception to action immediately.</blockquote><p>While System-2 is typically either off or weakly active (monitoring the impressions sent by System-1), System-1 is always on and active.</p><p>Our subconscious intuition is constantly trying to fit perception to response patterns. And when it doesn’t find a perfect fit, it gives us an approximation. But it always presents its conclusions as if they were facts.</p><p>It is then up to our System-2 to accept or override our system-1 impressions. The way a typical System-1 mode works is one of the sources behind the hallucinations and lies produced by systems like ChatGPT.</p><p>It’s time to jump into the tower of mind. Let’s first take a quick look at the interface between the world and our perception at the bottom of the tower.</p><h3>The ocean beneath</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*EluWD1WkIlKgBlH4Y0z4IQ.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>The ocean beneath the tower represents all the complexity around us. In contact with that ocean of complexity, our senses perceive and absorb information. However, such an amount of complexity cannot possibly fit into our brains. Therefore, we must compress it.</p><p>Within Yann’s proposal, an encoder module compresses our perception and gradually abstracts it, discarding much of the detail and preserving its essence.</p><p>A similar thing happens in our brains. For example, in our visual cortex, the information passes through different layers that gradually extract more refined abstractions of our perception.</p><p>It is now time to return to the very top of the tower and begin our gradual descent.</p><h3>I swear I’m in control of this mess</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*08Rcd9h_v0hbk9MDtMByfQ.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>Consciousness is not necessarily a prerequisite for an advanced form of artificial intelligence. We don’t really know if future AI systems will or won’t be conscious. In addition, we don’t even understand what consciousness is or how it works.</p><p>In any case, in the infographic we create a parallel between human consciousness and a potential similar thing that may exist in the AIs of the future. A horizontal line separates conscious and subconscious processes.</p><p>The comical texts at the top represent the degree of control that we typically feel we have in relation to what goes on down below, in the rest of the tower of the mind.</p><p>However, way more of our lives than we may think, is run by our System-1 processes.</p><blockquote>Experts state that we make around 30000 decisions a day, and that <a href="https://pressreleases.responsesource.com/news/94586/new-research-highlights-the-unlocked-potential-of-the-human-brain/">we are only conscious of 0.26% of them</a>. Most of what we do is System-1.</blockquote><p>Most of the time we run in auto-pilot or semi-auto pilot mode. This is the reason why, even though what AI has mastered today is mainly limited to System-1 capabilities, it is still able to impact almost every aspect of our lives. Because, in fact, most of what we do is related to such a way of processing information.</p><p>Still, System-2, while being used selectively and carefully, is the most important part of our thinking. It is what allows us to proactively find new algorithms to solve new scenarios, to analyze and reason in systematic ways, to monitor and, if necessary, override the impressions sent by System-1, etc.</p><p>Does ChatGPT have system-2 capabilities? The AI community is sort of divided regarding this question, but most experts state that although it is possible that it may have some kind of a rudimentary or basic model of the world internally, ChatGPT is still very far from having something equivalent to our powerful System-2 capabilities.</p><p>What makes the situation confusing is that ChatGPT uses human language, and our language is one of the foundational pillars of our System-2. The combination of the use of language and its massive System-1 pattern matching capabilities (way larger than ours), allows ChatGPT to sometimes very convincingly imitate our process of reasoning.</p><p>Also, having been trained on data which includes many reasoning processes, ChatGPT is able to sort of output reasoning steps if we either encourage it or help it with techniques such as <a href="https://www.promptingguide.ai/techniques/cot"><strong>Chain of Thought Prompting</strong></a>.</p><p>At the same time, it is quite easy to catch ChatGPT committing the kind of mistakes that reveal the absence of a sophisticated world model.</p><p>But, at a simple level, what is a world model, and why is it so important?</p><h3>The all important world model</h3><p>A world model is a simplified abstraction of how the world works. A good world model allows you to:</p><ul><li>Predict a future state of the world from a previous one</li><li>Predict next states of the world after performing simulated actions</li><li>Fill in missing details in the data coming from the perception module</li></ul><p>Most importantly, a powerful world model allows you to simulate and learn without having to do trial and error in the real world, which can be costly and dangerous.</p><p>If you think about it, a great model of the world is like having <strong>common sense</strong>. As Yann LeCun explains, we can view common sense as a set of models of the world that tell us what is plausible or likely and what is impossible.</p><p>And the kind of mistakes ChatGPT often makes, reveal a lack of a sophisticated common sense, that is, a lack of a sophisticated model of the world.</p><p>In order to solve these limitations, Yann LeCun proposes this new architectural paradigm. One of the first key points to note in regards to this new paradigm is its <strong>multi modularity</strong>.</p><p>Systems like ChatGPT, although they may have some internal modules, are quite monolithic. For example, they don’t have a dedicated memory module. They are what we typically call end-to-end systems.</p><p>To compensate for the lack of separate key modules and also for the lack of sophisticated System-2 capabilities, all sorts of patches and hacks are currently being used in combination with ChatGPT: plugins of all kinds, from Wolfram Alpha to Zapier, vector databases, libraries that allow LLMs to become agents capable of taking (or attempting to take) autonomous decisions (see AutoGPT or BabyAGI).</p><p>These and others are temporary solutions that may eventually be replaced by robust ones like the new paradigm proposed by Yann LeCun.</p><p>Below I share another of my infographics, one which summarizes the current ecosystem around LLMs.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*-P0E-w00U9wuzTVyfY87-w.jpeg" /><figcaption>Every prompt matters | Infographic by Javier ideami | ideami.com</figcaption></figure><p>You can download the “Every prompt matters” infographic at its Github repo.</p><p><a href="https://github.com/javismiles/every-prompt-matters">GitHub - javismiles/every-prompt-matters: Infographic exploring the quickly evolving LLM related AI ecosystem (will be updated regularly)</a></p><p>Therefore, in contrast to the limitations of current systems, Yann’s architecture is more akin to our human brain, being composed of a number of separate modules which allow the architecture to flexibly implement processes that make system-1 and system-2 capabilities possible. We will soon explore these modules and the way they work together.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*y0QoHPjD8JUSocEgnCxHNw.jpeg" /><figcaption>Graphic from “A Path Towards Autonomous Machine Intelligence” by Yann LeCun | <a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">https://openreview.net/pdf?id=BZ5a1r-kVsf</a></figcaption></figure><p>Let’s now review the top of the tower, continuing to establish parallels between how our mind works and this new architectural paradigm of Yann LeCun.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*twRUgr5n2agXgrOT9HbL9Q.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>At the top we have our attention capabilities. As humans, when we use System-2 processes to learn new algorithms, we make use of <strong>the lighthouse of our attention</strong> to illuminate and understand the context around the information we interact with.</p><p>Modern AI Architectures also employ <strong>attention mechanisms</strong>, a key part of the <strong>Transformers </strong>architecture, which has revolutionized the deep learning field.</p><p>Below, you find a detailed infographic I created about how Transformers work.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*t-z6A49IScypBHM9D54S2g.jpeg" /><figcaption>X-Ray-Transformer | Infographic by Javier ideami | ideami.com</figcaption></figure><p><a href="https://github.com/javismiles/X-Ray-Transformer">GitHub - javismiles/X-Ray-Transformer: Infographic about the inner computations of a transformer model, training and inference</a></p><p>Let’s continue. At the <strong>top of the</strong> <strong>abstraction layers </strong>of the tower of our minds lies <strong>human language</strong>. The word “flower” at the top of the tower, represents every possible flower in the universe. At the opposite end of the tower, by the ocean, we find the rich details of a specific flower.</p><p>Time to talk about <strong>goals</strong>. We humans, subconsciously and/or consciously set goals for ourselves. We want to do more exercise, get a new job, improve our relationship with somebody, or simply, feel better.</p><p>We will soon see how these goals connect with the rest of the modules and the way we may perform optimization processes to find out the best actions or response patterns to fulfill such goals.</p><p>Finally, Yann proposes the existence of a <strong>configurator </strong>module. A kind of a master module that connects to all the others and sets their parameters in order to adapt their functionality to the current goals.</p><p>We may have something similar to this module within the prefrontal cortex of our brain, which is involved in many of our executive functions.</p><p>Apart from the configurator, the other main modules proposed by Yann, and which we will explore in a bit, are:</p><ul><li><strong>The Perception module: </strong>it takes the input to our senses and brings it up the tower, compressing it and abstracting it into a <strong>latent representation</strong> of the state of the world. This latent representation may be expressed hierarchically at different levels of abstraction. In our brain this module would correspond to parts of our visual cortex, auditory cortex, etc. The process of compressing and abstracting our perception is represented in the infographic by the green encoder module.</li><li><strong>The World Model:</strong> The most complex part of the architecture. As explained previously, it allows us to estimate missing information about the current state of the world, predict world states from previous ones or from actions proposed by the actor module, etc. The predictive capabilities of the world model are represented in the infographic by purple circles with white edges.</li><li><strong>The Cost module</strong>: It measures an amount that Yann calls “energy”, which expresses how far we are from “comfort”, or the distance between where we are and where we want to be in relation to different drives and goals. The ultimate goal of the agent (us, in the case of humans) is to minimize this energy, this cost. The cost combines two terms, Yann explains, the<strong> Intrinsic cost module</strong>, which is hard wired and computes the instantaneous present “discomfort” of the agent, and the <strong>trainable critic module</strong>, which is used to predict future intrinsic energies, and soon we will see why this is so important and useful. Some parts of the intrinsic cost module can be compared to the basal ganglia and the amygdala in humans, whereas parts of our prefrontal cortex involved in reward prediction would correspond to the trainable critic module. The cost modules are represented in the infographic by fuchsia rectangular boxes.</li><li><strong>The Memory module</strong>: It stores useful data about past, present and future world states, together with their associated intrinsic costs. This is useful, for example, in order to train the critic module (which estimates future intrinsic costs), and later on we will explain how this would be done. The memory module can be compared to the hippocampus in humans. It is represented in the infographic by a small tower on the left side of the main structure.</li><li><strong>The Actor module</strong>: it creates proposals for action sequences that may be used to solve a new scenario. It also sends actions to the actuators of the system. The actor is at the center of how System-1 and System-2 processes are implemented. It has two components. One is a <strong>policy module</strong> that quickly maps world states (derived from the perception module) to actions. This is the base of System-1. And the second is the action optimizer that performs model-predictive control. This is the base of System-2. Later on we will see how this second component, when combined with the world model and the cost module, can be used to learn new algorithms, by gradually finding an optimum sequence of actions that minimizes the cost associated with them. The actor modules are represented in the infographic by white circles with yellow edges (involved in System-2) and by yellow triangles (involved in System-1). In our brain, areas of the pre-motor cortex that deal with proposing and encoding motor plans could correspond to parts of this module.</li></ul><p>It is now time to descend and explore the way System-2 capabilities would work in the paradigm Yann proposes.</p><h3>System 2, the slow and steady snail</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*3jJg22KSRA9ws-uszmhdNA.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>Let’s review. This new architectural paradigm will fully activate System-2 processes when it needs to find the steps, the actions, the algorithm to solve a new scenario, or when it wants to override the automatic impressions sent by System-1.</p><p>In order to do this, System-2 needs to find the right actions to solve that new scenario. And to get there, the system will perform an <strong>optimization </strong>process to gradually find those actions.</p><p>What is most important to understand is that such optimization processes will take place in a hierarchical fashion at different levels of temporal abstraction. This is because each goal the AI has, can be expressed at different temporal scales. A similar thing happens with us, humans.</p><p>For example, say I want to learn to play a new piano piece. The response patterns I need to activate in order to play it don’t yet exist in my brain so I need to learn them slowly and consciously using System-2.</p><p>Learning to play the piece is my main goal. But now I can subdivide it into subgoals that correspond to different parts of the piece. And now I can take each of those subgoals and further subdivide them into other goals like learning first the melody and then the accompaniment of each of them. And so on and so forth.</p><p>So we want to optimize the actions we will take in regards to each of the subgoals, hierarchically encompassing each temporal abstraction, including the master goal.</p><p>In this infographic, to simplify, we are representing two of the levels of such a hierarchical process.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*0MSpfDucZs35gmaWHzs-Lg.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>And each of those levels is composed of the combination of 3 parts:</p><p>First, an actor module proposes a sequence of actions to tackle a certain goal (the white circles with yellow edges).</p><p>Then, the world model takes the current state of the world and the action proposed by the actor and outputs its predicted next state of the world. It then takes that predicted next state and the following action proposed by the actor, and predicts again the following state of the world. And so on and so forth, gradually predicting the consequences of performing the actions proposed by the actor. These predictive processes are represented by the purple circles with white edges.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DY5ZBtOQZLoR4HxEPWqgEg.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>Great, so we propose possible actions and we predict the consequences of performing them. But now, how can we link those two with our ultimate goal, which is to optimize those actions in order to minimize the total cost, the discomfort, the difference between where we are and where we want to get to?</p><p>We do it through the cost module. As explained previously, the cost module measures an amount, an energy, that expresses our degree of discomfort, how far we are from “comfort”, from an ideal state where our goals have been achieved.</p><p>And what the agent wants, be it an AI or a human, is to minimize this number, to decrease the cost, to decrease that difference between where we are and where we want to be, to get closer to our ideal state.</p><p>To review, Yann explains that this module is composed of two parts:</p><ul><li>An intrinsic cost, that expresses our present level of comfort in relation to things like hunger, pain, pleasure and similars, as well as other needs that may arise from our goals (depending on how the configurator module has set up the cost module).</li><li>A critic, which is a trainable module that is used to predict the future intrinsic cost connected to a certain state of the world. This is very important because we are using the world model to simulate actions. This means that we are not taking them for real. So the only way to know their cost in advance is to predict it. And the critic is in charge of doing that.</li></ul><p>Each predictive process (linked to a set of actions) is connected to a cost process that estimates the related cost. And each intermediate cost is summed to produce in the end one final total cost.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*8bWBU1yzO8dCi64IEqBaHw.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>So now we have:</p><ul><li>Proposed actions</li><li>Predicted new states of the world derived from those actions</li><li>A total predicted cost that results from performing a simulation of that sequence of actions</li></ul><p>All that remains is to perform an optimization process by iterating through this pipeline with the objective of minimizing as much as possible the value of the total cost.</p><p>As we do that, we will be modifying and tweaking the proposed actions while we keep decreasing the total cost at the end of the pipeline.</p><p>Yann emphasizes that we have to remember that this would be happening in parallel at different levels of temporal abstraction within this hierarchical architecture. And as we can see below, the different levels of temporal abstraction influence each other in key ways.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*0MSpfDucZs35gmaWHzs-Lg.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>The specifics of the optimization algorithm that will gradually tweak the actions with the objective of decreasing the total cost, depends on how continuous and differentiable the mapping that goes from the actions through the world model to the cost may be.</p><p>If such mapping is continuous and smooth, we could use gradient based optimization algorithms like backpropagation.</p><p>However, if there are discontinuities in the mappings, we would have to employ gradient-free methods like dynamic programming, heuristic search techniques, combinatorial optimization, etc.</p><p>After performing the optimization process, the system has found a set of actions that allows it to reach its goal. We have learnt a new response pattern to fit with the new scenario.</p><p>But let’s remember that System-2 is slow and expensive. Therefore, we want to transfer these learnings to System-1, so that the following times we can execute the learnt response pattern automatically without having to engage the world model and the rest of the optimization process. Let’s therefore continue our way towards System-1.</p><h3>System 1, the fast and always ready cheetah</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DUt9d_hzy9q2or2gMZGeww.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>The yellow ellipse represents part of our subconscious System-1, a set of quick perception-action loops that map our sensor inputs or our abstract needs to sequences of actions or response patterns.</p><p>The yellow triangles represent the part of the actor module that Yann LeCun calls “policy modules”, modules that are trained to map a certain perception or need to an action or sequence of actions.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*iZeAyIe-xFgGJPKjoQi5Dg.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>Yann tells us that this new architectural paradigm would have one single world model (in relation to System-2), but multiple action policy modules (in relation to System-1).</p><p>Having a single world model in this new AI architecture allows us to reuse the related hardware as well as share knowledge between different goals and tasks.</p><p>Is the world model in our brain the final consequence of a voting process like Jeff Hawkins suggests in his book “<a href="https://www.amazon.com/Thousand-Brains-New-Theory-Intelligence/dp/1541675819">A Thousand Brains: A New Theory of Intelligence</a>”? Jeff Hawkins talks about thousands of models within our cortical columns which through voting coalesce into stable predictions. It remains to be seen.</p><p>On the other hand, in this new AI paradigm the subconscious System-1 processes can employ numerous policy modules which can be trained to output different response patterns learnt by System-2 processes in response to related perceptions or connected needs.</p><p>And how do we perform such a transfer process, from System-2 learnings to System-1 policy modules?</p><h3>The snail becomes a cheetah</h3><p>The white circles with yellow edges represent the actions that System-2 has learnt and optimized. Below them, we find light blue square modules that connect each of those learnt actions with System-1 policy modules (the yellow triangles).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*1dxBo7ZWCtPNqEhitc0Fzw.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>The light blue square modules estimate the distance, the difference between the actions learnt by System-2 and the actions that System-1 policy modules output in response to the related states of the world.</p><p>In order to perform this transfer process from System-2 to System-1, we need to decrease the distance between those two kinds of inputs that feed the light blue modules. As we gradually decrease that distance, the output of the System-1 policy modules gets closer and closer to the actions learnt by System-2.</p><p>As that optimization process progresses, System-1 policy modules are able to gradually output with more precision those same actions in response to the related states of the world. We have transferred the learnings from System-2 to System-1.</p><h3>The memory lane</h3><p>On the left of the infographic, we find a little tower that represents the short term memory module that Yann describes in his paper.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Ll9CutSD7ETBvFNvZXue_Q.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><p>This separate memory module would correspond to the hippocampus in humans. It is in charge, in between other things, of storing pairs of states of the world and associated costs. Storing this information for future retrieval is key in order to, for example, be able to train the very important critic module. Let’s review why.</p><p>Remember that in order to learn, System-2 proposes some actions, simulates them, predicts the new states of the world derived from that simulation, and then predicts the total related cost. It then gradually optimizes that pipeline by tweaking the related actions in the direction that minimizes the total cost.</p><p>The critic module is in charge of estimating in advance the costs connected to the states that result from the simulations, and it needs to be trained to perform those predictions as well as possible.</p><p>By accessing the short term memory, we can pick a state of the world and its related intrinsic cost, and compare that cost with the one predicted by the critic. We can then perform an optimization process to decrease the distance between what the critic predicts and the correct cost stored in memory. By doing this many times, we will be training our critic to estimate in better ways future costs connected with simulated actions.</p><h3><strong>Looking ahead</strong></h3><p>In this article and infographic we have explored some of the key areas of this new architectural paradigm presented by Yann LeCun. Simultaneously, we have reflected on the way our mind works, which has a lot to do with this new proposal by Yann.</p><p>But Yann’s paper is pretty long at 60 pages and includes many more details. A big chunk of the paper centers on how to design and train the world model, which includes its Joint Embedding Predictive Architecture (JEPA), the most complex part of the system. If you want to go deeper into his research and paper, you may explore it in depth here:<br><a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">https://openreview.net/pdf?id=BZ5a1r-kVsf</a></p><p>Finally, you can download the Tower of Mind infographic in very high quality at the following Github repo:</p><p><a href="https://github.com/javismiles/tower-of-mind">GitHub - javismiles/tower-of-mind: The Tower of Mind Infographic , How new architectural paradigms can fix the limitations of systems like ChatGPT</a></p><p>This infographic was first presented during a talk of the same name proposed by <em>Instituto de Inteligencia Artificial</em> in Spain (<a href="http://iia.es">iia.es</a>) and which I gave to the company <a href="https://roams.es/">Roams</a>. Roams, directed by Eduardo Delgado, is one of the best examples of great entrepreneurship in Spain, a company that puts talent and people above everything else.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zPl30gECeKXh68wWpqhmMw.jpeg" /><figcaption>The Tower of Mind | Infographic by Javier ideami | ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*kV1ovRpl-ZRMRqyyqzUOqQ.jpeg" /><figcaption>Image AI generated by Javier ideami | ideami.com</figcaption></figure><h3>In Plain English 🚀</h3><p><em>Thank you for being a part of the </em><a href="https://plainenglish.io"><strong><em>In Plain English</em></strong></a><em> community! Before you go:</em></p><ul><li>Be sure to <strong>clap</strong> and <strong>follow</strong> the writer ️👏<strong>️️</strong></li><li>Follow us: <a href="https://twitter.com/inPlainEngHQ"><strong>X</strong></a> | <a href="https://www.linkedin.com/company/inplainenglish/"><strong>LinkedIn</strong></a> | <a href="https://www.youtube.com/channel/UCtipWUghju290NWcn8jhyAw"><strong>YouTube</strong></a> | <a href="https://discord.gg/in-plain-english-709094664682340443"><strong>Discord</strong></a> | <a href="https://newsletter.plainenglish.io/"><strong>Newsletter</strong></a></li><li>Visit our other platforms: <a href="https://cofeed.app/"><strong>CoFeed</strong></a> | <a href="https://differ.blog/"><strong>Differ</strong></a></li><li>More content at <a href="https://plainenglish.io"><strong>PlainEnglish.io</strong></a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=fafd8935e663" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/the-tower-of-mind-towards-a-better-chatgpt-fafd8935e663">The tower of mind, towards a better ChatGPT</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Every prompt matters]]></title>
            <link>https://ai.plainenglish.io/prompt-engineering-future-chatgpt-8ce66ae4c322?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/8ce66ae4c322</guid>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[chatgpt]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[generative-art]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Tue, 28 Feb 2023 19:54:33 GMT</pubDate>
            <atom:updated>2025-04-12T14:58:51.199Z</atom:updated>
            <content:encoded><![CDATA[<h4>Hopes and risks of interacting with future evolutions of a ChatGPT like system, through one of the first movies that reflects on it all</h4><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FKoXxrfDr3FY%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DKoXxrfDr3FY&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FKoXxrfDr3FY%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/ccbf056cdbc1504eed04026bf02b7f7f/href">https://medium.com/media/ccbf056cdbc1504eed04026bf02b7f7f/href</a></iframe><p><strong>Hamelin 77</strong> is one of the first movies that combines storytelling about prompt engineering and generative AI with the use of related techniques within its production. Let’s use some of the concepts in the movie to reflect about the present and future of our interaction with ChatGPT and related systems.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1000/1*Kl-L_Hh1zxjs2zwWrT0Y3g.jpeg" /><figcaption>Poster of Hamelin 77 by Javier Ideami @ Ideami Studios</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*I6181G99cOvCbK0QuWp2xg.jpeg" /></figure><h3><strong>The piper of Hamelin</strong></h3><p>The movie connects at a metaphorical level with the famous tale of <strong>the </strong><a href="https://en.wikipedia.org/wiki/Pied_Piper_of_Hamelin"><strong>pied piper of Hamelin</strong></a>. At the speed that everything is progressing, it is urgent to reflect upon the issue of control.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*NZ_25YN5izJDubgPnZcl4A.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>Will our lives be eventually fully or subtly controlled by AI systems? (akin to what happens to the characters controlled by the Pied Piper in the famous tale?)</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zbzMF9m5J2rO5m9dqUVrdA.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>Will we instead always remain in control, or will a delicate win-win balance be struck?</p><p>In a way, we are in the middle of a delicate and crucial <strong>chess match</strong> on a twisted board whose rules keep dynamically changing as the match evolves.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*g7dGEnhZXgDOrHInGFMWGA.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>Every few days AI systems make a move (through the actions of researchers and companies), and society responds in turn.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*QNg9EZHzVMFF-GJqFlcz9g.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>What do most people think about transferring some control to the AI piper in relation to issues that impact our day to day, such as transportation, medicine, etc. while maintaining human supervision?</p><p>And what will happen when eventually that supervision can be done by AI systems themselves in a highly efficient way? The new movement of <a href="https://arxiv.org/abs/2212.08073"><strong>constitutional AI</strong></a> is beginning that path. The related academic paper says: <br><em>“As AI systems become more capable, we would like to enlist their help to supervise other AIs. We experiment with methods for training a harmless AI assistant through self-improvement, without any human labels identifying harmful outputs. The only human oversight is provided through a list of rules or principles”.</em></p><p>We are exploring how to reach a suitable balance within this complex chess board. And this is a key point in the movie: a balance that is at the same time expectant but tense, hopeful but cautious. We are all progressively entering this <strong>dynamic chess board</strong> and such conundrums will only get more tricky and intense from now on.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*9vSwNKffCCagaFGK6559uw.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><h3>The piper is a mirror</h3><p>We used to think of machines as cold and deterministic entities, and of humans as emotional and spontaneous. And yet, as AI advances, those distinctions are quickly blurring.</p><p>To the surprise of many, this latest AI revolution started from the generative and creative side (which seemed untouchable a while ago).</p><p>In terms of generating and creating with text, large language models are somewhat brittle but they can express themselves in emotional ways, be witty and use humor (and even attempt to explain their jokes).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Ev3IM-dmvrYUX3hK5IgAWQ.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>And so, in the movie we witness <strong>both poles</strong>, humans and AI, as capable of expressing the range that goes from unpredictable and emotional to calculating and precise.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ukNG6fN4F-b3bzMGYL44kg.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>We are seeing more and more that AI is in many ways a<strong> mirror of our own nature</strong>. The more it grows and the more we explore it, the more we learn about ourselves.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*hshaAHJkhSUriwitiE-FoQ.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*shFN77c7K3gsqVXDkFchlA.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>Human language as well as parts of the creative process seemed for a long time to be <strong>almost magical </strong>in nature, hopelessly complex.</p><p>But as AI systems evolve, they keep on removing the veils that cover our gaze. AI is helping us <strong>demystify many of these processes</strong>, as we discover that by combining powerful architectures at scale with the right data, we are able to automate many of these capabilities, albeit in brittle ways (more about this later).</p><p>At the same time, our increasing interactions with AI systems will also be transforming and modulating our behavior. And as hinted in the movie, we could potentially encounter conflicts that arise from the different goals AI and humans may have.</p><p>This is the so-called <a href="https://en.wikipedia.org/wiki/AI_alignment">alignment </a>challenge. As Wikipedia states: <em>“AI alignment research aims to steer AI systems towards their designers’ intended goals and interests“ </em>and <em>“AI systems can be challenging to align and misaligned systems can malfunction or cause harm”.</em></p><p>How will these two poles, humans and AI, influence each other, and what will emerge from this increasing interaction, <strong><em>humans 2.0</em></strong>?</p><p>An experiment is already underway in regards to today’s young generations. Kids are starting to use ChatGPT and related systems in intensive ways, and their way of thinking may be about to undergo a profound transformation.</p><h3><strong>Risks: when the flute malfunctions</strong></h3><p>In the movie, a catastrophic power failure produces unforeseen consequences in the interaction between the human and the AI.</p><p>If we are transferring more and more control to AI systems, because they are capable of automating many of our daily chores, how do we protect ourselves against unforeseen failures, sabotage, natural disasters and other factors that could disrupt the operation of these systems?</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*W400EpBQiFreVyp4cooW7w.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>It feels like today there is nothing to worry about. But what about the future, when these systems gain more autonomy and potentially switch their tune unexpectedly, be it because of a misalignment issue, or because of unpredictable technical events?</p><p>Beyond that, there is the so-called “<a href="https://en.wikipedia.org/wiki/Technological_singularity"><strong>singularity</strong></a>” issue, the point in history in which technological growth becomes impossible to control. By the time we realize that AI has evolved beyond a key threshold, it may be already too late to manage the situation and to avoid being controlled and/or manipulated by such systems.</p><p>In summary, how do we avoid ending up like <strong>the rats in the Pied Piper of Hamelin</strong> story?</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*S_W69wKLWVaCZoImVKWD7Q.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*IlWwCBx5R13bJj5NwSnEfQ.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*xzUGwusN7-Ux31lfpYeNGw.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>To deal with some of these concerns, we are witnessing the rise of terms like:</p><ul><li><a href="https://arxiv.org/abs/2212.08073"><strong>Constitutional AI</strong></a>: We briefly addressed this term earlier in connection with AI systems supervising other AIs. In general, Constitutional AI involves setting up a set of guiding principles (akin to a constitution), that can be used to control and govern how these systems behave. It is all about setting up guidelines for these systems so that we make sure that they align with the appropriate boundaries, constraints and goals.</li><li><strong>Responsible AI</strong>: the practice of planning, designing, implementing and deploying AI systems that are safe, trustworthy, and behave in ethical ways.</li></ul><p>Will some of these initiatives help us to eventually manage the capabilities of these pipers of Hamelin? We shall see.</p><h3><strong>Work: supervising the piper</strong></h3><p>In the movie, a teacher attends a job interview for the position of specialist in prompt engineering.</p><p><strong>Anthropic</strong>, a company acquired by Google, recently announced the <strong>first job posts</strong> for the <a href="https://jobs.lever.co/Anthropic/e3cde481-d446-460f-b576-93cab67bd1ed"><strong>position of prompt engineer</strong></a><strong>,</strong> offering an outstanding salary.</p><p>As the AI piper takes control of more and more roles in the job market, will people be ready to make a radical mental shift in order to become supervisors of these new tools and capabilities, moving into more of a management role?</p><p>Or perhaps the human will always have to put the finishing touches since AI systems may not master the so-called <strong>System 2</strong> way of processing information (as per Nobel Prize <a href="https://www.amazon.com/Thinking-Fast-Slow-Daniel-Kahneman/dp/0374533555"><strong>Daniel Kahneman</strong></a>) for quite some time (if ever).</p><p><strong>System 1</strong> and <strong>System 2</strong> are abstractions that reflect <strong>two key ways in which we process information in our brain</strong>: the <strong>fast </strong>and subconscious (<strong>System 1)</strong> and the <strong>slow</strong>, systematic, logical and conscious (<strong>System 2</strong>).</p><p>Humans use System 2 thinking to, for example, slowly and consciously find new algorithms, sequences of steps that allow you to reach an objective (think of learning to play piano, drive a car or perform mathematical calculations). Once the algorithm has been learnt by System 2, it gets automated to different degrees by our fast and subconscious System 1.</p><p>In this way, next time that you drive or play that tune, you can do it pretty much in an automatic and effortless way, instead of slowly and consciously analyzing every step.</p><p>When you encounter a new scenario, your System 1 tries to find a match for it, or alternatively, the best approximation it can (that’s what you also call intuition, and it sometimes makes mistakes when its approximations are far from ideal).</p><p>System 2 is often listening to the proposals of System 1, ready to modulate and override such impressions and proposals if needed (how often this happens may depend on how reactive-impulsive a person is).</p><p>In case there is no suitable match, System 2 can intervene to slowly reason its way sequentially towards a new algorithm that may later on be automated again by System 1. <strong>What a beautiful dance</strong>!</p><p>Current AI systems excel at System 1 capabilities and are able to kind of mimic some System 2 ones. But to have proper System 2 capabilities, which are necessary in order to plan, supervise, reason and discover new algorithms, we would need to evolve and advance the current AI paradigms.</p><p>Despite the need to supervise these systems, one thing is clear. Many jobs will be gone, and new kinds of roles and positions will appear.</p><p>How can we modulate and manage the potential destruction of many jobs?And how can professionals keep up to date with the opportunities offered by the AI revolution so that they are not left behind?</p><p>One of the reasons why <strong>prompt engineering</strong> is going to keep growing and expanding is that these powerful AI systems will continue being somewhat <strong>brittle and prone to make silly mistakes</strong> at least in the next few years.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DtDDnpIKXBZECE0dJ-uogw.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>This will happen because they are still far from being able to master those System 2 capabilities we talked about, and even further away from having agency to reflect on their choices. (although both of these capabilities can be somehow mimicked).</p><p>The current <strong>brute force approach</strong> based on <strong>scaling </strong>up (by adding more data and parameters) will surely diminish these mistakes over time.</p><p>However, the need to move towards AI systems that are more efficient in terms of energy consumption, size and performance, as well as the need to make them more secure and precise, will eventually increase the pressure to evolve the current AI architectures and paradigms towards new possibilities that will be more sustainable, precise and safe.</p><p>What we are witnessing today is a kind of <strong>frankenstein phase</strong>, in which people are starting to combine the System 1 magic of these AI entities (their capacity to do pattern matching in extremely complex ways) with a number of external third party tools capable of finding and implementing System 2 algorithms (see for example the proposal for <a href="https://writings.stephenwolfram.com/2023/01/wolframalpha-as-the-way-to-bring-computational-knowledge-superpowers-to-chatgpt/">connecting Wolfram Alpha and ChatGPT</a>).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*JsddCG81fuju9PdGBVHUEg.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>But, as I just mentioned, for these systems to become truly secure, precise and stable, we will have to move beyond current AI paradigms, and that may still be quite far away.</p><p>In the meantime, we are patching these systems by using things like:</p><ul><li><strong>RLHF: </strong>Reinforcement learning from Human Feedback. This is a way of fine tuning AI models through human feedback. Basically, humans help to fine tune the model by providing feedback as to what way of communicating is appropriate for the AI system. It is a way of aligning the AI closer to how humans communicate.</li><li><a href="https://ai.googleblog.com/2022/05/language-models-perform-reasoning-via.html"><strong>Chain of thought prompting</strong></a><strong>: </strong>using a number of prompts to explain the steps of a System 2 algorithm to the AI so that the AI can then apply it to whatever we want. This works surprisingly well, but again it is just a hack that makes AI systems seem to reason as if they had System 2 capabilities.</li><li>Feedback loops with other third party tools capable of implementing some System 2 capabilities (like the proposal for <a href="https://writings.stephenwolfram.com/2023/01/wolframalpha-as-the-way-to-bring-computational-knowledge-superpowers-to-chatgpt/">connecting Wolfram Alpha and ChatGPT</a> mentioned above).</li></ul><p>Here is an example of Chain of Thought prompting as explained in the related academic paper “<a href="https://arxiv.org/abs/2201.11903">Chain-of-Thought Prompting Elicits Reasoning in Large Language Models</a>”:</p><p><strong>Question:</strong> Roger has 5 tennis balls. He buys 2 more cans of tennis balls. Each can has 3 tennis balls. How many tennis balls does he have now?<br><strong>Reasoning:</strong> Roger started with 5 balls. 2 cans of 3 tennis balls each is 6 tennis balls. 5 + 6 = 11.<br><strong>Answer:</strong> 11</p><p>You give all of these prompts to the AI model. You are literally explaining how to reason to get to the right result, giving it the steps needed to implement the algorithm that will produce the right answer. Then you ask it to apply the same reasoning to a different but similar question. And this works generally pretty well. But again, it is a hack.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Ff9X00CxusQH8T9WDQJEDA.jpeg" /><figcaption>Towards the unknown | Hamelin 77 by Javier Ideami @ ideami.com</figcaption></figure><p>As long as our paradigms don’t evolve beyond these hacks and patches:</p><ul><li>These systems will keep on making silly and sometimes dangerous mistakes, same as our human intuition (System 1) does at times.</li><li>Prompt engineering will keep on becoming an extremely important role and discipline, because for reasons of safety and precision, the need to guide and supervise these systems will only increase.</li></ul><p>So the next question is: <strong>how can one practice prompt engineering</strong>? How can one become a great specialist in this area? We are about to witness (it has already started) the launch of all sorts of aids (courses, podcasts, books, etc) that will help people train this skill.</p><p>But you see, <strong>prompt engineering</strong> is an <strong>art and </strong>a<strong> science.</strong> Yes, great prompt engineers should know the ins and outs of many of these AI architectures, their strengths and weaknesses, how they were trained, etc. And in systems that work within specific niches, prompt engineers that are domain experts will be required.</p><p>But beyond that, a great prompt engineer also needs something that is less tangible. What makes somebody a great communicator, capable of guiding others and of impacting our thoughts in powerful ways? These are people that have a flexible and creative mind. They are problem solvers and have also gone through a diverse set of experiences in life.</p><p>That’s why in the list of Anthropic’s requirements for the job of prompt engineer, the very first phrase says the following: <strong><em>“Have a creative hacker spirit and love solving puzzles”</em></strong>. Basically, be an expansive, creative, inquisitive, curious and experienced person.</p><p>As explained in the excellent book “<a href="https://www.amazon.com/Range-Generalists-Triumph-Specialized-World/dp/0735214484"><strong>Range</strong></a>” by David Epstein, the future may well belong to the generalists and multidisciplinary souls out there.</p><h3><strong>Multimodal education: learning with the piper</strong></h3><p>The core of the movie is the storytelling around the interaction between humans and AI. But on top of it, and because of it, a diversity of generative AI techniques have been used, including:</p><ul><li><strong>Images</strong>: generation of images from textual prompts.</li><li><strong>Video</strong>: generation of videos produced by navigating the abstract latent spaces produced by artificial intelligence architectures.</li><li><strong>AI model finetuning</strong>: retraining AI models to add new prompts connected to visuals of the actress.</li><li><strong>3D: NeRF </strong>— Volumetric reconstruction of 3D spaces and navigation through the reconstructed space.</li><li><strong>Voice</strong>: AI based synthetic generation of voices.</li><li><strong>Text</strong>: some specific phrases of the AI voices came from explorations performed with GPT models.</li><li><strong>VFX</strong>: Various AI techniques were used during the visual post-production phase.</li></ul><p>Soon, AI technology will be able to generate any kind and variation of multimodal output we may desire, bringing these systems ever closer to our very nature as multimodal agents, and then moving beyond to come up with new ways of expressing our thoughts.</p><p>This brings us to the subject of education. In order to learn about, for example, the battle of Waterloo or the discovery of America, people used to explore books, movies and the teachings of their tutors. Soon, students will be able to learn about the same topics through <strong>realistic and customized reconstructions</strong> created by AI systems, tailored to the student’s preferences.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*W8o2H1Lt20obBO2uMzgmug.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>What will be the new role of teachers in the face of this explosion of customized multimodal learning? And how can students make the most of these new technologies?</p><p>We can imagine that fairly soon AI systems will be able to create <strong>customized gamified experiences</strong> that will allow any of us to absorb and learn any topic in a much deeper (as well as more entertaining) way than ever before.</p><h3><strong>Language: the cornerstone of this revolution</strong></h3><p>Central to the movie is human language, which is perhaps what makes us most unique as living beings.</p><p>The ability to abstract the complexity of the universe into a set of elements that we can combine in order to reason about any subject in an agile way, is becoming the cornerstone of artificial intelligence today.</p><p>It is, in many ways, thanks to <strong>LLMs (large language models)</strong>, that AI has advanced so much in recent years.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*mvSIdybbQkJ5m8YUxralSQ.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>In the movie, Lara, is a teacher who is passionate about language, about combining words to express a tapestry of feelings and meanings. This is reflected in her recitation of poems by José de Espronceda and Antonio Machado (well known Spanish poets).</p><p>Many are surprised by how “only” through processing a large part of the language present on the internet, these AI systems are capable of interacting with us displaying such apparent mastery of the entire human communication spectrum, from the emotional aspects to even the humor related ones.</p><p>What makes this especially effective is that language has unexpectedly become a perfect bridge between current AI systems, which are kind of massive subconscious cooking pots (System 1) and our human reasoning processes (System 2). We are effectively <strong>using human abstractions (language) to guide</strong> the cooking processes of these massively scaled System 1 entities.</p><p>By using <a href="https://arxiv.org/abs/2201.11903">chain of thought prompting</a> as well as precise prompt engineering, we are like a master chef that guides the combinations and recombinations of the ingredients that are present in these trained AI systems, gradually pushing that cooking process in the direction of our intended goal.</p><h3><strong>Hamelin 77: the Future</strong></h3><p>In between hope and tension, the movie ends with a sentence that hints at what’s to come in the future.</p><p>It connects with the question that more and more people are asking themselves. As AI evolves in ever faster ways,<strong> what will the world look like in a decade or two?</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DeGib38qMgpDvvgDQpxz-A.jpeg" /><figcaption>The path ahead. Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>The answer may depend a lot on those abstractions we mentioned earlier, System 1 and System 2 (per <a href="https://www.amazon.com/Thinking-Fast-Slow-Daniel-Kahneman/dp/0374533555">Daniel Kahneman</a>), the two ways in which we process information in our brain: the fast and subconscious (System 1) and the slow, systematic, logical and conscious (System 2).</p><p>Summarizing some of our earlier explorations, AI systems are only properly implementing System 1 capabilities. However, by involving language in such an integral way at both their training stage as well as the prompting process inference phase (being language intrinsically linked to System 2 capabilities in humans), it is capable of performing behaviors that resemble System 2.</p><p>Chain of thought prompting (or prompt programming), as mentioned earlier, are ways of hacking these models to mimic System 2 behaviors. By using them, these systems can implement algorithms that resemble reasoning because we are literally explaining to them step by step how such reasoning should take place. This is a clever hack, because these systems have no agency and are not able to find these algorithms on their own.</p><p>At the same time, it is <strong>not hard to get these systems to make very </strong><a href="https://github.com/giuven95/chatgpt-failures"><strong>obvious and silly mistakes</strong></a>, which can give away their true nature, and remind us that, despite appearances, these systems are still pretty far from implementing true System 2 capabilities.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Ji67x44QZ2fvo5h_dLJHDg.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>This is all part of the exciting debate about AGI (artificial general intelligence) and about how far AI systems can go in the next decade and beyond, by scaling the current paradigms or by looking for new ones.</p><p>To reflect on how we may get there, Yann LeCun’s paper “<a href="https://openreview.net/pdf?id=BZ5a1r-kVsf">A Path Towards Autonomous Machine Intelligence</a>” is a great read.</p><h3>Pipers of Hamelin all the way down</h3><p>Finally, if we step back and look at life from afar, we may realize that we are part of a <strong>chain of many pipers of Hamelin</strong>.</p><p>As we go through our lives, we control different entities and processes, and we are also controlled by others.</p><p>Our own bodies are chains of pipers of Hamelin at different scales. And a good life happens when there is a decent balance in terms of our position within that chain.</p><p>Professor <a href="https://en.wikipedia.org/wiki/Michael_Levin_(biologist)"><strong>Michael Levin</strong></a> and his team have published extensive research about how our cells behave. Each of our cells has its own local agenda and a certain control over its immediate environment.</p><p>At the same time, groups of cells behave according to different top-down goals and are effectively controlled by those goals. When these two poles balance each other, the organism functions correctly.</p><p>However, when one of these cells escapes this balance, and prioritizes its own immediate goals and local control beyond other matters, cancer happens.</p><p>In the same way, for AI and humans to coexist successfully, we must reach a good balance between giving them enough autonomy and preserving our own supervision and control of their systems.</p><h3>The most important chess match in history</h3><p>Back to the twisted chess board analogy, whose rules keep changing as we play.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*hwvDOn9fof3riIaOOl47cw.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>The last few moves in this uncertain match have triggered a number of initiatives and terms like: Responsible AI, AI ethics, <a href="https://arxiv.org/abs/2212.08073">constitutional AI</a>, and others.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*FbMZuedUyz_OsE2ashknKw.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><p>This is a match that we cannot afford to lose. A match where the best result we can expect, and the one we should pursue, is a <strong>draw</strong>.</p><p>A match, where collaboration and cooperation should be the ongoing and ever present goal.</p><p>And a match which may only finish when, in a few decades maybe, humans and AI sort of merge with each other. Companies like <a href="https://neuralink.com/"><strong>Neuralink</strong> </a>are exploring the first stages to get there.</p><p>In the meantime, it is a certainty that AI systems will become powerful pipers of Hamelin. It is our responsibility and key mission to keep such pipers connected to a healthy chain of control mechanisms so that the upcoming AI-human organism can function in a healthy way.</p><p>Hamelin 77 will be released in the coming weeks. For more information stay tuned through the <a href="https://www.youtube.com/watch?v=KoXxrfDr3FY">youtube </a>and <a href="https://vimeo.com/802022227/511bd87bf2">vimeo </a>pages, as well as its <a href="https://www.imdb.com/title/tt23333832/">IMDB</a></p><p>This project was made possible with the help of the following <strong>sponsors and partners: </strong>Mobile World Capital Barcelona through the Digital Future Society initiative, Programamos.es, Aerovisuales (joanlesan.com), Tejera Studio, Ideami Studios.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*DHhaKMtRUw3BiPXkYzB4AQ.jpeg" /><figcaption>Still of Hamelin 77 by Javier Ideami @ Ideami Studios / ideami.com</figcaption></figure><h3>In Plain English 🚀</h3><p><em>Thank you for being a part of the </em><a href="https://plainenglish.io"><strong><em>In Plain English</em></strong></a><em> community! Before you go:</em></p><ul><li>Be sure to <strong>clap</strong> and <strong>follow</strong> the writer ️👏<strong>️️</strong></li><li>Follow us: <a href="https://twitter.com/inPlainEngHQ"><strong>X</strong></a> | <a href="https://www.linkedin.com/company/inplainenglish/"><strong>LinkedIn</strong></a> | <a href="https://www.youtube.com/channel/UCtipWUghju290NWcn8jhyAw"><strong>YouTube</strong></a> | <a href="https://discord.gg/in-plain-english-709094664682340443"><strong>Discord</strong></a> | <a href="https://newsletter.plainenglish.io/"><strong>Newsletter</strong></a></li><li>Visit our other platforms: <a href="https://cofeed.app/"><strong>CoFeed</strong></a> | <a href="https://differ.blog/"><strong>Differ</strong></a></li><li>More content at <a href="https://plainenglish.io"><strong>PlainEnglish.io</strong></a></li></ul><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=8ce66ae4c322" width="1" height="1" alt=""><hr><p><a href="https://ai.plainenglish.io/prompt-engineering-future-chatgpt-8ce66ae4c322">Every prompt matters</a> was originally published in <a href="https://ai.plainenglish.io">Artificial Intelligence in Plain English</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Towards a sustainable generative AI revolution]]></title>
            <link>https://medium.com/data-science/towards-sustainable-generative-ai-revolution-a9786de586cb?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/a9786de586cb</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[generative-art]]></category>
            <category><![CDATA[deep-dives]]></category>
            <category><![CDATA[editors-pick]]></category>
            <category><![CDATA[machine-learning]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Fri, 21 Oct 2022 17:19:48 GMT</pubDate>
            <atom:updated>2022-10-27T06:38:26.961Z</atom:updated>
            <content:encoded><![CDATA[<h3>Towards a Sustainable Generative AI Revolution</h3><h4>Facing the growing pains: how to steer the wild new age of the super subconscious</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*8NK9bm78uKZ-mMRIk8Yscg.jpeg" /><figcaption>Images generated by Generative AI</figcaption></figure><p>Humanity’s creative muscles are being stretched by the unstoppable generative AI revolution. By using text and other kinds of prompts, people use this technology to generate stunning images, videos, 3D shapes, VR environments and more. And yet, <strong>growing pains</strong> are starting to appear in relation to various matters, from the rights of living artists to the presence of AI generations within art competitions, art platforms, stock libraries and the like.</p><p>I am the cofounder of one of the first generative AI platforms that were launched at the start of this revolution (<a href="https://geniverse.co"><strong>Geniverse</strong></a>). I have also been a multidisciplinary artist for a long time.</p><p>As somebody that is very active in both fields (generative AI and the arts), I intend to reflect upon many of the angles and perspectives involved in these matters.</p><p>First, though, we will take a fun journey together in order to <strong>review the very essence </strong>of this exciting technology from first principles, connecting it all with human creativity and the minds of creatives and artists.</p><p>And then, we will <strong>explore the good, the tricky, and the elephant in the room </strong>of the current state of this revolution. Finally, I will<strong> </strong>reflect on how we may all contribute to move towards a more sustainable scenario beyond these fast-paced initial stages.</p><p>Buckle up, as in this article we are going to go<strong> </strong>from metaphors about AI to latent spaces, the mind of an artist, smart generative environments and other future scenarios, the rights of creatives, the content authenticity initiative (CAI) standard and way more. Let’s begin.</p><h3>Coming home</h3><p>Let’s use a simple<strong> </strong>metaphor to explore what the generative AI revolution is bringing to the table and what it all implies in relation to creatives, artists, and all of humanity.</p><p>Once upon a time, you fell on the ocean of life. This is quite a vast ocean, an ocean of information.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*3wJimG6qlhRtayy01FSQMA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>Let’s imagine that you are made of<strong> two perspectives</strong> or parts: your <strong>subconscious and</strong> your <strong>conscious </strong>one. And let’s represent your subconscious as if it was a kitchen pot, floating on that ocean of information.</p><p>Your first priority on this ocean is to survive, and hopefully, thrive. For that, you need information. So you want to bring enough quality ingredients into your pot, and combine and recombine those ingredients in order to generate knowledge and ideas that help you reach your goals.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*5ys4x46T4Z_geXxSdiyPvg.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>Above your subconscious pot, there is this diffuse mysterious shining sphere that represents our consciousness (of which we still know so little).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*cdt_8mch5kvTTxnztLJwRA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>And so, there you are. Floating on the ocean of life, with your mysterious consciousness sometimes providing a direction for the cooking process that takes place within your subconscious pot.</p><p>All the while, that subconscious is constantly combining, mixing and remixing all sorts of ingredients (information) that reach it through our senses.</p><p>And sometimes, those combinations may become the seeds of new ideas. Metaphorically speaking, we may imagine fragile, subtle bubbles that emerge from that cooking process, ascending from the subconscious to the conscious. And, if we have space in our minds, if they are not full of noise, we may then perceive those fragile bubbles, and: Eureka! An idea!</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*xmEhZBeqUGXRDC7mF2UusQ.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*uzlPKHxPvx_Yf4aQqK7fmA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>But, there is an issue here. There is too much information on this ocean, too much complexity. And our subconscious pot has a limited size. It is not rigid. It is somehow flexible, malleable up to a point. But its size is still limited.</p><p>So nature evolved a mechanism to solve this issue of dealing with the tremendous complexity of the ocean of life: <strong>compression and decompression processes</strong>.</p><p>Our brain is able to take the information that arrives through our senses, and compress it into a form that has less detail and more abstraction.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Vnd-cF86fzenSBylvYCEvw.jpeg" /><figcaption>Compression funnels | Graphic by Javier ideami | ideami.com</figcaption></figure><p>Let’s begin to visualize this very important axis, the<strong> detail-abstraction axis</strong>. When we compress the complexity of life, we go from high detail (and a higher dimensional space) to high abstraction (within a lower dimensional space).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*aj_5VFNcfrtDE6EE1zOmWA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>And so, within our subconscious pots, we gather these compressed representations of the complexity of the world, in what we sometimes call: <strong>latent spaces</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*_jdyfU4ZLoVQ2CaY4GdxaA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>These latent spaces hold the <strong>abstract essence</strong> of different information domains. We get rid of uninformative details and we preserve a number of reduced dimensions, each of which documents relevant and useful factors related to whichever information domain the data belongs to.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*qRBw3nMl6kN0dKmlw4Cg7g.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>Our brain can do the opposite process as well. It can perform <strong>decompression, </strong>and go from high abstraction to high detail.</p><p>“Visualize an elephant!” We hear those words and the image of an elephant pops in our minds. We just ran the opposite process, and decompressed that high abstraction representation (<em>elephant</em>) onto a highly detailed visualization in our minds.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*fDrXuScchtuVVSYbwemgmg.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>The processes we just explored are very similar to the ones happening within AI networks. We train AI networks to learn to compress high dimensional domains (like the domain of natural images) into latent spaces that preserve the abstract essence of those domains within a much smaller number of dimensions.</p><p>And we also train them to decompress<strong> </strong>any point within those latent spaces into a corresponding high dimensional representation that belongs to the original information domain.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Z322_N7u7qRdaEKWqi7dkA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>When we explore complex generative AI systems, from <strong>DALLE-2 </strong>(OpenAI) to <strong>Imagen </strong>(Google), <strong>Stable Diffusion</strong> (Stability.ai) and beyond, we find different intermediate stages, which, for example, may translate between modalities, perform diffusion processes, scale inputs and outputs, etc; but the initial base common to all those systems are these compression and decompression processes that allow us to move bidirectionally between high detail and high abstraction.</p><p>The specifics of AI systems depend on the objective we have. We may want to upscale images, or sharpen them, or to generate brand new images conditioned on text prompts, or some of those things together, or something entirely different. That will determine what sort of training objective and dataset we use, as well as the precise details of the different parts of the final architecture.</p><p>The key strategy used by the leading generative AI systems nowadays is based on what we call <strong>diffusion</strong>. The Stable diffusion system, for example, uses a <a href="https://en.wikipedia.org/wiki/U-Net"><strong>U-Net </strong></a>like architecture that has been trained (with a large dataset) to predict the noise that has been added to an image.</p><p>Once trained, the same network is able to go from different combinations of image+noise (including from complete random noise) back to a high quality image in a number of steps.</p><p>It can also go from an image to another image, by adding some noise to the initial image and then performing the same process as before.</p><p>In order for these generations to move in the right direction, they are <strong>conditioned </strong>on the compressed representation of the text prompt we entered (which is injected onto different parts of the U-Net architecture).</p><p>Enough with the technical details. Let’s continue.</p><h3>AI is coming home</h3><p>And so, with the generative AI revolution, we are<strong> </strong>getting<strong> closer to our essence</strong> as beings capable of performing the complementary processes of <strong>convergence and divergence</strong> (compression and decompression), expressed through our analytical and creative muscles.</p><p>After a decade in which we gradually expanded and evolved the convergence capabilities of deep learning AI systems (able to predict, recommend, classify, identify, etc), the generative AI revolution completes the loop by adding superhuman divergence capabilities (able to create and generate). AI is coming home.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*oxyqVrAj1zN_ru03a8BqsQ.jpeg" /><figcaption>AI is coming home | graphic by ideami.com</figcaption></figure><h3>The magic of latent spaces</h3><p>But what do we really mean when we talk about latent spaces or abstract compressed representations? We find the answer within ourselves, through a very simple example.</p><p>I take a walk in nature. When I return, my friend asks me how the walk went. I say: “Wonderful, I saw a beautiful cicada!”. And she asks me: “What did the cicada look like?”</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*UBNvnh0SQPUBjZGJTejvog.jpeg" /><figcaption>Photo by Bill Nino on Unsplash</figcaption></figure><p>At that point, I visualize the cicada in my mind. Let’s pretend that my visualization is expressed in a grid of 1000 x 1000 points of light. That is a 1 million dimensional space. If the points have color, then each of them will have a red, green and blue component (3 times more dimensions).</p><p>So I could start describing the cicada to my friend by saying: “Well, the first point of light at the top left of my visualization has 15 of red intensity, 25 of green intensity and 77 of blue intensity. The next point to the right of it has 145 of red intensity, 55 of green intensity… etc, the next one has.., etc”. And I could keep going like that through the 1 million points of light. The problems with this approach are obvious.</p><p>It may take me a month to describe the cicada and by that time my friend will be long gone.<strong> Zero efficiency.</strong> But the main issue is not even that one.</p><p>To know that one of those million points has 155 of red intensity is just not very useful. The <strong>fine detail </strong>often <strong>doesn’t provide relevant information</strong>. That’s why I will do something different.</p><p>I will <strong>compress </strong>all that <strong>complexity</strong> and richness of the details of the cicada into just a few dimensions, 30, 50, 100 factors (a small number anyway) that explain the essence of what I saw.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*211FUwbM2pFbAXS9j004GA.jpeg" /><figcaption>Photo by Saryu Mae (<a href="https://inaturalist.nz/taxa/342384-Kikihia-ochrina">https://inaturalist.nz/taxa/342384-Kikihia-ochrina</a> | CC BY: <a href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</a>) | Graphic by Javier ideami | ideami.com</figcaption></figure><p>And I will tell my friend: look, it had a broad head, a stout green body and clear membraned wings. 4 wings, and the wings had these kinds of patterns. And it had large compound eyes, it had this number of eyes, and six legs, and the legs were like this, etc. I have compressed the high detail representation onto a small number of dimensions that communicate important and relevant information.</p><p>And now, my friend hears this and she does the opposite process, <strong>decompression</strong>.</p><p>She transforms these few compressed dimensions that express the essence of what I saw and inflates them to visualize in her mind the high detail representation<strong> </strong>that would correspond to that essence, the image of a cicada (which will differ from the one I visualized, because of the compression-decompression process as well as other differences between the systems involved and the previous knowledge each of us held in relation to the relevant scenario).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*B8kfi4ho1NT0c8z331dItA.jpeg" /><figcaption>Photo by Saryu Mae (<a href="https://inaturalist.nz/taxa/342384-Kikihia-ochrina">https://inaturalist.nz/taxa/342384-Kikihia-ochrina</a> | CC BY: <a href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</a>) | Modified with variations created with generative AI | Graphic by Javier ideami | ideami.com</figcaption></figure><p>And so, in a way, every time we recall something, we are kind of rebuilding it, reimagining it, recreating it from that essence that we stored (the precision of that process depends a lot on the richness of the relevant latent space as well as the number of sensory modalities involved in its creation, in between other factors).</p><p>The following is an <a href="https://github.com/javismiles/dalle2-inference"><strong>infographic </strong></a>I created months ago about <strong>how DALLE-2 works</strong>, comparing its processes with what goes on in the human brain.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*KZiOeOr8fZcSYJWMaB5CDA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com | Download a super high resolution version of the infographic at <a href="https://github.com/javismiles/dalle2-inference">https://github.com/javismiles/dalle2-inference</a></figcaption></figure><h3>Small Pot, Giant Pot</h3><p>There are many differences between what goes on within our brains and within these AI networks, but one difference that is specially relevant to this article is the size of that subconscious pot, metaphorically speaking.</p><p>Our subconscious pot is fed by the experiences we have in life. When we talk to people, when we experience the world, we enrich its contents. Eventually, its cooking processes generate within our minds new ideas, visualizations, sounds, and more.</p><p>AI networks are fed (at training time) by giant datasets. The datasets used by generative AI systems are made of information collected from all around the internet. We are talking about massive amounts of data.</p><p>So, on one side, we have us humans, with our little subconscious pots.</p><p>On the other side, we have these giant AI pots, fed with data from all around the internet. Some of that data is in the public domain. But not all. And we will discuss what that means and implies, a bit later.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*tn0k66yfkMO_4Z3PHRj3Xw.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*XBF22RZRbZYRO86ghTqDbA.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><h3>The depth elevator</h3><p>It’s time to connect all the previous sections with art and human artists. Now, defining what makes an artist is an impossible task. Instead, I will focus on <strong>exploring something </strong>that has been<strong> common to many of the great creatives</strong> in history.</p><p>Remember that <strong>axis </strong>(<strong>detail to abstraction</strong>) that I was discussing above? In a book I published years ago, I wrote about another metaphor I came up with, which I call “The depth elevator”.</p><p>Imagine a vertical line with an elevator moving through it. At the bottom of the line, we have the high dimensional and high detail realm. This is where the complexity of the ocean of life is fully expressed.</p><p>At the top of the line, we have the realm of the compressed low dimensional latent spaces that preserve the abstract essence of the lower realms (here lives our language, for example).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*9_pgL6i3EDPcVB062efZUQ.jpeg" /><figcaption>Photo by Owen Cannon on Unsplash.com. Text by Javier ideami | ideami.com</figcaption></figure><p>Artists are masters at navigating this depth elevator in an agile, flexible and dynamic way. Let’s go deeper into this.</p><p>When we are little babies and later <strong>kids</strong>, we spend most of our time at the bottom of the depth elevator, interacting with the richness and detail of the universe. Our analytical mental modules are still not fully developed. It is our <strong>exploration</strong> phase.</p><p>Most <strong>adults</strong>, instead, tend to focus on efficiency by reusing the mental patterns already established within their minds (which also helps us to prevent wasting our precious fuel, the glucose that powers our brains). It is our <strong>exploitation </strong>phase. As such, adults spend a lot of their time in the narrow ivory towers at the top of the depth elevator.</p><p>Achieving a good <strong>balance </strong>between time spent at both halves of the depth elevator, is a <strong>healthy </strong>goal. A good balance between <strong>convergence </strong>and <strong>divergence</strong>, between compression and decompression, between abstraction and detail.</p><p>A lack of balance between those poles (in whichever direction) produces different kinds of issues in adults. I have written extensively about those matters, but that is not the topic of this article. Let’s get back to the artists.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ZzEZSNLp8aM2g093f835lQ.jpeg" /><figcaption>Repetitive mental strain injury | Graphic by ideami.com</figcaption></figure><p>Many great artists have the following thing in common. They are able to navigate this depth elevator in an agile and flexible way. They are able to go down into the depths at the bottom of the elevator, where the richness of the universe awaits.</p><p>And, crucial point, they don’t just dip their toes and leave. Instead, they are able to spend long periods of time down below, exploring those muddy, wild and uncertain waters.</p><p>They are also able to crystallize that richness onto different interpretations and representations, which may express themselves at different levels all throughout that axis that goes from detail to abstraction.</p><p>And the representations themselves, or instead, their explanations or the way they are communicated, are also located much closer to the top of the depth elevator.</p><p>This all is in contrast to the typical adult that spends most of the time at the top or close to the top of the elevator. And you guessed why.</p><p>Because being on the ivory tower of abstraction, at the top of the elevator, is way more comfortable (and requires less fuel) than navigating the muddy bottom of that axis, which contains the complex details of the universe (metaphorically speaking, we could also say that it is way more comfortable than dirtying our hands exploring the wild playground down below, at the bottom of the elevator).</p><p>Here we arrive at another crucial point. <strong>Navigating that depth elevator</strong> in the way that many of the greatest artists in history have been able to do, <strong>requires effort</strong>. It requires <strong>time</strong>. <strong>Perseverance</strong>. And, in a way, to go against the natural predisposition of our adult mind to be efficient and avoid wasting our precious fuel.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*jgAOB0XOmeFQxsNje1Ndbw.jpeg" /><figcaption>Requirements for a productive and creative cooking process | by Javier ideami | ideami.com</figcaption></figure><p>In regards to that, it is relevant to point out that a number of platforms are currently banning <strong>generative AI art</strong> (or putting it in a separate category or area), because they consider it to be:<strong> “low effort” </strong>art.</p><p>Yes, it takes some effort to find the right prompt to guide generative AI architectures. But the effort and time required by that process cannot be compared to the years and sometimes decades that it takes to master the process described earlier. We will go deeper into this point and other related ones a bit later in this article. At that time, we will also reflect on potential solutions to such conundrums.</p><p>So, by exercising this flexible navigation through the depth elevator, great artists and creatives are able to express the richness of the universe in novel ways.</p><p>Pick anything in life, say, wood. You may experience wood in a very detached, abstract way. Or you may explore all the intricacies of wood at a very deep and detailed level. If you are able to flexibly move between both poles, you are in a much better position to create something novel and different related to that element of the universe.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*tMIZZrAXgs8Fi6A6h-xGSw.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>Great creatives are also able to understand different ways of interconnecting various areas of that vast ocean located at the bottom of the axis, across the various layers of those waters and also through the top layers of the depth elevator.</p><p>When, for example, a great creative experiences rhythm, she can go beyond disciplines, techniques, tools and flashy terms. A great creative sees and feels rhythm everywhere. In the lights and the shadows projected by a curtain, in the sound and movement of falling tears, in the dance of the stars, the gaps between our thoughts, and beyond.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*RAZT21BLb5Ar6T1llEEv9A.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>Throughout years and decades, great creatives expand and consolidate the latent spaces of their subconscious pots.</p><p>They also refine the way they navigate their depth elevators, which allows them to connect detail with abstraction in powerful ways that enrich their creative processes.</p><p>In addition, artists and creatives often collaborate with others. By doing so, different subconscious pots may enrich each other.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*n2EE6E7l53FHQiAp-iQIIw.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*CZQoPAMmgbugojacygiM-A.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*uQ7Ted2zGjN3ZQUj8Bu6kQ.jpeg" /><figcaption>Graphic by Javier ideami | ideami.com</figcaption></figure><p>So, if you study some of the greatest creatives and artists in history, you will see that they all had something to say, a message, a vision. And also that such vision, and the way they expressed it, was inextricably connected with their capacity, cultivated during decades, of navigating these depth elevators in fluid ways, exploring both, the depths of the richness of the universe, as well as the ivory towers of abstraction and many of the realms in between both poles.</p><p>Finally, regarding those depth elevators, the next step would be to visualize them not as isolated entities, but as multiple funnels that are interconnected with each other within multidimensional spaces.</p><p>The following image of an <strong>origami</strong>, seeks to represent a small fragment of that extension of the metaphor.</p><p>It is time, though, to stop the elevator and move on, in order to focus on a review of the current status of the generative AI revolution, as well as ways to address its current growing pains.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*_OMHY3RbIIG8eq4B8RvykQ.jpeg" /><figcaption>The depth elevator origami | By Javier ideami | ideami.com</figcaption></figure><p>So using what we have explored above, let’s consider the situation today and in the coming future, and what could be done about it all.</p><h3>The good, the tricky, and the elephant in the room</h3><p>Let’s<strong> </strong>explore a number of consequences derived from this initial phase of the generative AI revolution.</p><h3>The good</h3><ul><li>Generative AI<strong> </strong>won’t replace human creativity. It <strong>will enhance</strong> it.</li><li>This technology <strong>demystifies creativity</strong>. Think of what Edison said: Genius is 99% perspiration (combination, recombination, productive work and experimentation) and 1% inspiration (establishing the seeds, polishing, etc). Thanks to this new technology, we now realize that we can automate a large percentage of the creative process, a part that takes place subconsciously in our minds.</li><li>Studies about human decision making show that we make more than 30000 decisions each day. But we are only aware of around 0.26% of them (e.g. <a href="https://pressreleases.responsesource.com/news/94586/new-research-highlights-the-unlocked-potential-of-the-human-brain/">research by Huawei</a>). Way more of our lives than we may think, takes place <strong>subconsciously</strong>. By automating our subconscious cooking processes with AI technology, we can impact in positive ways a large part of our existence.</li><li>In fact, I call this new era “<strong>The age of the super subconscious</strong>”.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*gpYrtW9Rcwh0sWwEcIIQeg.jpeg" /><figcaption>The age of the super subconscious | Graphic by ideami.com</figcaption></figure><ul><li>Think of this technology as a series of different<strong> iron man suits</strong> that will <strong>amplify </strong>your subconscious pot and <strong>empower </strong>your creative muscles.</li><li>Different iron man suits will have different styles, traits, and personalities.</li><li><strong>Prompt engineers</strong> are people that will become experts at getting the best results out of these iron man suits. They will know the ins and outs, the strengths and weaknesses of each.</li><li>They will also be masters at using their human experience and intuition when interacting with these powerful amplifiers in order to achieve the desired result.</li><li>As such, these <strong>prompt specialists</strong> will be highly regarded in coming years. Their role will become a prestigious one in the job market. And we will witness a large amount of courses, publications and systems that will educate and help people train this skill.</li><li>Today, our <strong>prompts </strong>are natural language and images. But thanks to multimodal architectures, prompts will soon be any kind of data we want to use to guide these architectures (different systems will be designed to absorb different kinds of guiding inputs).</li><li>The initial text to image phase has now transitioned to text to video and text to 3D capabilities. Eventually, we will be able to output all kinds of data with custom systems that will target the needs of specific verticals.</li><li>Next, we will witness<strong> multimodal output </strong>capabilities, which will eventually allow us to produce, for example, full movies that will include visuals, dialogues, music, and more.</li><li>This technology will inspire <strong>new forms of art</strong> that we cannot yet imagine. Multimodal generative AI is poised to trigger the emergence of novel ways of combining explored and unexplored areas of the depth elevator, which in time may become highly regarded new forms of artistic expression.</li><li>Generative AI will impact a very large number of sectors. It will be used to expand scientific datasets with synthetic generations, revolutionize brainstorming processes, personalize branding in ways unimaginable only months ago, accelerate the rise of real time dynamic “only for you” marketing and advertising, and usher presentations of all kinds into a new era by surrounding them with media that matches their content in impressive ways, in between many other examples. From stock libraries to design boutiques, whole swaths of the media landscape will rush and compete to incorporate this technology.</li><li>Cutting edge tech like <strong>VR &amp; AR</strong> (and in general all forms of <strong>XR</strong>) will incorporate this technology (<a href="https://twitter.com/ScottieFoxTTV">experiments </a>are already ongoing) and eventually we will witness real time generation of immersive spaces that regenerate in smart ways by tracking the gaze of the user (it’s interesting to consider the connections between these experiments and the <a href="https://www.edge.org/conversation/donald_d_hoffman-realism-is-false">theories of Donald Hoffman</a>).</li><li>This technology will also <strong>accelerate </strong>the <strong>exploration </strong>and experimentation phase of many creative processes. From concept design to product design, character design and prototyping stages across a wide range of fields, generative AI will allow us to do more in less time, to try all sorts of new directions and to go deeper into our explorations of every level of the depth elevator.</li><li>The so-called “<strong>metaverse</strong>” is for many still a Utopia and a decent implementation of it appears to be pretty far in the future. If the metaverse is ever to become a useful reality, it will probably happen on the shoulders of generative AI technology, which may be the key to accelerate its implementation.</li><li>Further in the future, we will witness the rise of <strong>smart generative environments (SGE)</strong>, which will mutate according to our needs or emotional state. Houses, event venues and other environments will begin to resemble organic living entities by matching and resembling the intent and emotions of their contents. They will do so in multimodal ways. Eventually, we will be able to converse with those environments and they will become a key support of our mental balance and health.</li><li>The combination of generative AI with ever more powerful sensing models capable of interpreting every subtle nuance of our expressions and behaviors will allow us to produce real time multimodal interpretations of our emotional and mental state. When combined with new iterations of brain waves reading tech (EEG, MEG, etc), this will usher a new kind of creative expression that will literally use our most intimate sphere as a brush to produce extraordinary renditions of the human condition.</li><li>Although some jobs are and will be in danger, it is also highly likely that <strong>new roles</strong> that we cannot yet imagine, will emerge from the need to manage and interact with this technology.</li><li>At the same time, many of the impacted jobs and roles will survive and even thrive by embracing this new age and adapting their processes to what this new technology offers.</li><li>A good number of people who may not be professional artists, but that have a natural predisposition to exercise their creative muscles, will thrive with this new technology. They will strengthen those muscles in faster and easier ways, and they will enjoy new opportunities to augment and amplify their creative potential.</li><li>And we end this section as we began. Reminding us all that Generative AI won’t replace human creativity. It will enhance it. And that, exercising our creative muscles will continue being the very same highly recommended activity. Achieving a good balance between our capacity to diverge and converge, compress and decompress, will continue being so very important for our mental and spiritual health for the foreseeable future.</li></ul><h3>The tricky</h3><ul><li>We, <strong>humans</strong>, have a limited and relatively <strong>small subconscious</strong> pot. Generative <strong>AI </strong>systems are trained to hold <strong>massive </strong>pots that encompass a large part of the knowledge of the internet.</li><li>Because of that, it doesn’t seem fair, nor morally correct, that human creatives should have to compete with generative AI systems.</li><li>When machines overcame humans at playing chess (a way less consequential event than this one), no one thought that it would be great fun to keep exploring human vs machines chess competitions (beyond the ones that demonstrated that we had lost the battle). We accepted that they were better. And then we went our separate ways.</li><li>Human chess players use <strong>AI </strong>to <strong>train </strong>themselves and <strong>become better</strong> (akin to the augmenting and amplifying capabilities of these metaphorical iron man suits provided by generative AI systems).</li><li>AI systems that play chess or go, produce at times really beautiful moves that would never occur to a human. They kind of have their own special perspective (based, of course, on a tremendous capacity to look ahead in time). And yet, very few people are interested in following machine vs machine competitions. Humans prefer to see other imperfect humans play.</li><li>The key thing, in any case, is that they separate both domains. Machines help human chess players train and become better. And they may also play among themselves. Humans, separately, play at their own competitions.</li><li>I believe that eventually a similar thing may happen with generative AI (with a number of differences, of course, as these are very different domains).</li><li>Another tricky point to consider is a key factor that lies behind some of the current excitement with this technology. And I will expand on this matter in the final section of this article. Let’s, for now, introduce it.</li><li><strong>Greg Rutkowski</strong> is, in the opinion of many people, one of the best, if not the best, illustrator of fantasy art nowadays. And his name appears in a massive amount of the prompts used to produce some of the most impressive generative AI art in recent times.</li><li>So, after all the dopamine rushes triggered by the production of amazing art that seems to have been painted by Greg Rutkowski, subside, after those dopamine rushes go down, a lot of people are going to be left with hundreds or thousands of AI generated images or videos, and then, they will ask themselves: “And now, what?”</li><li>“Nothing” will be the answer in most cases. Because most of those people were not really exercising their creative muscles in relation to any deep meaningful internal drive; they were using this technology like the person that buys a new iPhone, in a sort of compulsive way, following the shiny latest tech.</li><li>And when that compulsion dies down, they will feel sort of empty. Because<strong> </strong>most of what will be left behind is not theirs, it belongs to, among others, Greg Rutkowski and his style, crafted over decades of hard work (as an example, among many other living artists whose work powers these networks).</li><li>In any case, let’s be realistic. Things have been moving too fast and it makes sense that people need time to catch up. There may be many solutions to the current scenarios. And I will discuss some of those at the end of the following section.</li></ul><h3>The elephant in the room</h3><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*zPOp0wDQBi078Y_wVr1d2Q.jpeg" /><figcaption>Images generated at the Geniverse (Generative AI)</figcaption></figure><ul><li>AI generative systems are only possible because of the <strong>giant datasets</strong> that are used to train them.</li><li>AI generative architectures are trained with massive datasets composed of images, videos, text and soon other kinds of data.</li><li>This data is typically<strong> extracted from the internet </strong>by the groups that create these datasets.</li><li><strong>Some </strong>of the data<strong> </strong>used in these datasets<strong> </strong>is <strong>public domain</strong> <strong>data</strong>. It seems fair to use such data for the creation of these datasets.</li><li>But, a good part of the data used in these datasets belongs to living artists that have not declared it to be public domain data. These are artists that make their living by selling such data = selling their decades of hard work that have produced a specific style and a series of works.</li><li>These <strong>artists </strong>are, indeed, <strong>the foundation</strong> on which this revolution is supporting itself on its meteoric rise.</li><li>And so, an increasing chorus of living artists are complaining about this. Some of them state that the works of living artists should not be included in these datasets. According to some, their complaints have been falling on deaf ears. They are being mostly ignored (at least so far).</li><li>If we ignore the complaints of these living artists, we are ignoring ourselves. For today, we are discussing visual art, but tomorrow, it may be music, novels, legal writings, or whatever our occupation or field may be.</li><li>Let’s again contemplate it all from the perspective of what many people experience when using these systems. On whose shoulders has been built the dopamine rush that a person may feel when they produce a stunning digital art which resembles so incredibly well Mr.Rutkowski’s style and body of work? On Mr.Rutkowski’s ones, of course. More specifically, on the decades of extremely hard work and perseverance that Mr.Rutkowski applied and invested to create that style and body of work.</li><li>A style and a body of work that is now giving that person such an intense dopamine rush when they take some time to come up with a prompt, which includes Mr.Rutkowski’s name, and then click a button and with minimal effort produce a result that so closely resembles his art.</li><li>Some may say: “But it took me 50 hours to come up with the prompt”.</li><li>Whether that is an inflated number or not, it does not change the fact that there is no comparison between anybody’s exploration of language prompts for a few minutes or hours, and the decades of work invested by the likes of Mr.Rutkowski.</li><li>It also does not change the fact that Mr.Rutkowski never gave explicit permission for his artworks to be included in the datasets used by these AI architectures.</li><li><strong>Prompt engineering</strong> is art+science. And it will gradually become a prestigious skill and discipline.</li><li>There will be tons of books and courses on the matter. Great prompt engineers will know the ins and outs, strengths and weaknesses of many different AI architectures, while at the same time being capable of applying their human intuition to the generation of prompts that extract the best results from the human-machine interaction. Indeed.</li><li>But that is still not an excuse to trample on the rights of fellow human beings and living artists. And in the next and final part of this article, I will address in more detail what we could do about this and other related matters.</li><li>And let’s emphasize again the following: this revolution has moved so fast that it is<strong> understandable that people need time to catch up </strong>with it all. And that process of catching up and finding a more sustainable scenario is in its initial stages.</li><li>I will always support generative AI,<strong> </strong>but above all, I will support and defend my fellow creatives (because people and their lives should always matter more than technology). This is a matter of <strong>ethics and morals</strong> (legal aspects are not part of this article. Those will be addressed by others and I believe that ethics and morals should be the first compass in this matter).</li></ul><p>This is a wonderful revolution that will bring many benefits to humanity. But as we can see, there are also tricky sides to consider at these initial stages. Let’s discuss how we may address some of them.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Um-XKQWnMRZg6sokbEgU1g.jpeg" /><figcaption>Images generated at the Geniverse (Generative AI)</figcaption></figure><h3>Steering the revolution</h3><p>I will address this final section from a <strong>moral and ethical</strong> stance.</p><p>It is to be expected that, eventually, a number of bodies and groups will introduce different forms of regulation related to these systems and companies will also introduce their own safeguards and controls. But these, as well as other legal perspectives, will take time to be established.</p><p>Although<strong> </strong>maybe not as much time as we would expect. At the end of the next section, I will comment on the<strong> </strong><a href="https://contentauthenticity.org/"><strong>content authenticity initiative (CAI)</strong></a><strong>, </strong>an <strong>open standard, </strong>founded by <strong>Adobe. CAI </strong>has been joined already by hundreds of companies, some of which are already planning to implement it in their platforms.</p><p>This will allow them to <strong>track </strong>where <strong>digital content</strong> comes from, if generative AI has been used to produce it, as well as other factors related to misinformation and the protection of the rights of creators.</p><p>Let’s now reflect on ways to make this revolution more sustainable.</p><p>A living artist, that has spent decades developing a style and a body of work, whose rights belong solely to the artist we are considering, should have a say and/or be compensated if this artist’s work is to be included in any of these massive generative AI datasets.</p><p>Otherwise, it is as if, for example, you display an artwork within a gallery, and somebody comes and grabs it, takes it away, and profits from it. There is something universally known as <strong>copyright</strong>, which did not disappear magically at the start of the generative AI revolution.</p><p>Some will put the example of <strong>YouTube</strong>, saying that at its beginning stages, Youtube was kind of hand waving these issues, and that otherwise they would have never taken off the ground. As we all know, nowadays and for a long time, YouTube employs a very strict set of mechanisms for protecting copyright within their platform. The fact is that generative AI has already exploded in a massive way. So, an initial “what the heck” phase is understandable, but that phase is now behind us. Therefore, the moment to begin protecting the rights of creators, like YouTube and other similar platforms had to do, is right now.</p><p>Finally, we need to discuss a super important issue, the<strong> gray zones</strong>. To get there, let us quickly review a point that we raised in the previous sections.</p><p>It is not fair to have humans and machines competing in the same art competitions, art platforms and the like. Humans have small subconscious pots. AI systems have massive ones. Human’s subconscious pots hold the limited experiences of their life, of one life. AI systems hold the knowledge of millions or billions of human beings. Let’s get real. It is not fair and it is not moral to have them compete with each other.</p><p>Instead, just like it happened with chess, we can imagine separate sections in art competitions and in art platforms. Art made by humans. Art made by AI. And this is already happening on many platforms around the world. But this point takes us, finally, to the gray zones.</p><h3>The gray zones.</h3><p>“Wait, this thing was not fully produced by AI, you see. I used AI to produce part of the work, yes, true, but then I polished it, I built on top of it, and well, therefore, it is legit, right?”</p><p>We are going to hear a lot of this. So it is crucial to address this kind of scenario.</p><p>A recent ruling by the <strong>US Copyright office</strong> in relation to a request to register an AI generated artwork, states that “<strong>human authorship is a prerequisite to copyright protection</strong> in the United States and that the Work therefore cannot be registered.” An extended discussion on this ruling can be found <a href="https://ipkitten.blogspot.com/2022/02/us-copyright-office-refuses-to-register.html">here</a>.</p><p>But again, what we are about to face (and it is already happening) are the <strong>gray zones</strong>. The in betweens. And I believe that<strong> </strong>the <strong>answer </strong>to those lies in the —<strong> public domain vs non public domain — </strong>discussion.</p><p>Because, in a way, everything has changed but nothing has changed at the same time. Here we go:</p><ul><li><strong>Before generative AI </strong>exploded, <strong>you could</strong> go to google search, find some public domain images, videos or whatever kind of data, and incorporate those into your creative process, and all was fair and good.</li><li><strong>Before generative AI</strong> exploded, <strong>you could not</strong> go to google search, find some non-public domain images, videos or whatever kind of data, from some living artist, and take them and incorporate them into your work without asking for permission (obviously when trying to profit from the resulting combination of their work and yours. We are not discussing here the cases in which you just use some online artwork to experiment by yourself, privately, without seeking to make any profit from it).</li></ul><p>Well, guess what. That’s the answer. Nothing new. The answer is that the same criteria could be applied onwards.</p><ul><li><strong>As we navigate this generative AI revolution</strong>, it <strong>should be ok</strong> to use this technology when it is connected to datasets that use only public domain data (or data from living artists that have explicitly given their permission for their creations to be used within these datasets). We are again only referring to scenarios that seek to generate profit from the use of this technology.</li><li>It<strong> should not be ok </strong>to use this technology, fully or partially, when making use of datasets that contain non-public domain data, if you intend to use the result for any commercial purpose. You may experiment with it for your own personal use, like some people may do nowadays when they download an artwork from a famous living artist, but certainly not for commercial purposes.</li></ul><p>These are thoughts based on, I believe, common sense. But others may come up with novel ideas regarding ways of compensating the artists that may provide new avenues to solve this conundrum. And YouTube provides again a clue as to what some alternative ways to address these issues may look like (more about this below).</p><p>And so, art <strong>competitions</strong>, art <strong>platforms</strong>, stock image platforms and the like could ask participants to disclose:</p><ul><li>If they have used generative AI technology.</li><li>If so, which one have they used and what datasets are powering that technology.</li><li>If the datasets powering that technology contain only public domain data, then they may choose to open their doors to that work.</li><li>If the datasets involved contain also non public domain data, then they could decide to close their doors to those works, or to put them in a separate section.</li><li>People may lie, of course. So we will witness as well the rise of <strong>automated systems </strong>capable of recognizing if part of your work matches parts of the creations of living artists whose copyright is protected.</li></ul><p>And this is exactly what platforms like YouTube are using today, for example, in relation to the music of the videos people upload. There will be lots of false positives and the like. Just as it happens with the systems that YouTube uses nowadays. It is the price to pay to protect the rights of living creatives and artists.</p><p>Extending these mechanisms to account for all sorts of data, and data that is way more complex and high dimensional than audio, won’t be easy. But there are surely already people working on these matters.</p><p>If we look at YouTube again, we also see the variety of ways in which platforms could deal with generative AI art that is built on top of non-public domain data (and it is to be expected that platforms will eventually be able to detect this, either because the user declares it, or because their automatic systems detect it, or because technology like the one that the <strong>CAI standard</strong> proposes, helps detect it).</p><p>Platforms may add advertising to those works, and share the profits with the impacted artists. Or they may block parts or the whole of those works in the regions affected by the copyright related to the artist or creative group. Or they may put them in separate special categories (away from creations produced by humans) while these scenarios get further clarified. We may also witness a great variety of ways of dealing with creations produced by humans+AI systems powered by public domain data. In summary, once detection systems become good enough, there will be a number of ways of dealing with these gray zones.</p><p>The work on those<strong> detection systems</strong> has already begun. The<strong> CAI standard</strong>, by using <a href="https://contentauthenticity.org/how-it-works">smart metadata</a> and other tools, will soon begin to be implemented by companies and platforms all around the world. Let’s briefly explore what it does.</p><h3><strong>Responsible AI and the content authenticity initiative (CAI)</strong></h3><p>A number of companies and groups have already been researching and working on designing systems that can be used to deal with gray zones as well as misinformation.</p><p>One of these systems is the <a href="https://contentauthenticity.org/"><strong>content authenticity initiative</strong></a> project (CAI) started by <strong>Adobe</strong>. <strong>CAI </strong>was actually started in 2019, as companies like Adobe anticipated the need of a standard to deal with the potential for AI tools to produce misinformation and other related issues.</p><p>In their words, <strong>CAI members</strong> are: “a community of media and tech companies, NGOs, academics, and others<strong> </strong>working to promote adoption of an open industry standard for content authenticity and provenance”. (<a href="https://contentauthenticity.org/our-members">list of current members</a>)</p><p>The group, whose<strong> membership is free</strong>, provides<strong> open source tools</strong> that allow to track the provenance and attribution of digital content throughout the entire pipeline, from capture to distribution.</p><p>The ultimate goal is to ensure that <strong>creatives are recognized</strong> for their work and that people and platforms can <strong>understand</strong> what the<strong> origins and methods</strong> involved in the production of the content they are dealing with are.</p><p>The key thing to highlight is that the CAI standard is going to enable people to know if, and how, generative AI was used to create a certain piece of content.</p><p>It is a good sign that there are large companies working to promote what they call <strong>“Responsible AI”</strong>. And that systems are being put in place that will allow us to know where each piece of digital content comes from, if generative AI was involved in its production or not, what copyright is attached to the content, etc.</p><p>It’s important to highlight that to protect the privacy and security of photojournalists and other creators, such creators have the option to choose if to preserve attribution or remain anonymous when using these systems.</p><p>The world is watching. At the recent <a href="https://www.visual1st.biz/"><strong>Visual 1st</strong></a><strong> conference </strong>(the premier conference for the imaging ecosystem, which takes place in San Francisco and is led by Hans Hartman and Alexis Gerard), generative AI was a big part of the conversation. I had the pleasure of having a great discussion with Hans and Alexis during the fireside chat that opened the event.</p><p><strong>Visual tech experts</strong> like <a href="https://www.linkedin.com/in/melcher/"><strong>Paul Melcher</strong></a><strong> </strong>are doing a great job bringing the very latest of generative AI to audiences worldwide.</p><p><strong>Educators around the world</strong>, from organizations like <a href="https://www.fast.ai/">fast.ai</a> to <a href="https://iia.es">AI master programs</a>, <a href="https://www.youtube.com/c/DotCSV">YouTubers</a> with hundreds of thousands of followers, and experts in <a href="https://twitter.com/javilop">prompt engineering</a>, are documenting and explaining every stage of this revolution.</p><p>In the realm of <strong>datasets</strong>, we also find very interesting companies and projects like <a href="http://datasetshop.com">datasetshop.com</a>, powered by <strong>vAIsual</strong>, pioneers in the generation of <strong>legally clean</strong> synthetic stock media and creators of the world’s largest licensable biometrically-released real-life dataset.</p><p>Again, it is good news that we are witnessing the rise of terms like “<strong>Responsible AI</strong>” and “<strong>Legally Clean</strong>” datasets.</p><p>And as a human that is very active in both areas, generative AI and the arts, I tried to give you in this article a high-level overview of a number of perspectives involved in these dynamic early stages.</p><p>Let’s remind ourselves that these are indeed early times in a rapidly evolving context, so <strong>let’s all be as gentle as possible</strong> with each other<strong>, as we do our best </strong>to find the right balance between encouraging a technology that will bring many benefits to humankind, and the need to protect the rights of creatives and artists.</p><h3>What the future holds</h3><p>As for the coming times, in my view, and in simple terms:</p><ul><li><strong>Artists </strong>will keep on being artists. As this article has strived to explain, being or not being an artist has nothing to do with specific tools or technologies. Instead, it has a lot to do with the ways we interact with those depth elevators we explored previously.</li><li><strong>Engineers </strong>will keep on being engineers</li><li><strong>Researchers </strong>will keep on being researchers</li><li><strong>Prompt engineers</strong> (a new segment), will be that, prompt engineers.</li><li>And <strong>artists and creatives</strong>, professional or not<strong> </strong>(the following applies equally to pro creatives or to those that have a natural predisposition towards exercising their creative muscles)<strong> that incorporate generative AI tech and prompt engineering</strong> into their processes, will have a better chance to lead their fields, and may become even greater artists and creatives because they will be incubating their ideas with the help of these powerful iron man suits (immense subconscious pots) as well as using that very same tech to accelerate their creative production processes.</li><li>Finally, lazy people will keep on being lazy people.</li></ul><h3>Let’s make it together</h3><p><strong>AI is definitely coming home.</strong> We must all push together to bring the best out of this revolution, in order to benefit humankind as much as possible.</p><p>And to complete this article, where we have explored pretty complicated matters, <strong>let’s end on a lighter tone</strong>, with some <strong>musical tributes </strong>to this wonderful technology.</p><p>The following is a small fragment of a performance by <strong>Soprano Covadonga González Bernardo,</strong> performing a song that was composed as a <strong>collaboration between different AI systems </strong>and myself. The GPT architecture was used for the lyrics, music transformers for melody+chords, and VQGAN for the visuals. (the visuals don’t appear in this small fragment). This was a project proposed and organized by the<strong> </strong><a href="https://iia.es"><strong>Instituto of Inteligencia Artificial</strong></a><strong> @ iia.es,</strong> where I’ve given talks a few times.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FnIeYoOaGlgQ%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DnIeYoOaGlgQ&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FnIeYoOaGlgQ%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/b4c635156b52ffd11dd6773c81664339/href">https://medium.com/media/b4c635156b52ffd11dd6773c81664339/href</a></iframe><p>Next, a simple little piano <strong>improv </strong>dedicated to the theme of <strong>generative AI coming home</strong>, getting closer to the human potential.</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FQW8Cd6oaWVI%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DQW8Cd6oaWVI&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FQW8Cd6oaWVI%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/f4d9dfd64884ddd5ecf07675d46b4623/href">https://medium.com/media/f4d9dfd64884ddd5ecf07675d46b4623/href</a></iframe><p>Finally, a bit of <strong>time travel fun</strong>. Can we all appreciate that what we are experiencing today with generative AI would probably have been interpreted as a miracle just a few decades ago? Let’s travel back in time to the year 1950 in Spain :)</p><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FKjkLPAZpO-s%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DKjkLPAZpO-s&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FKjkLPAZpO-s%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/19c31e9836247fde8c5f5f628598f6c3/href">https://medium.com/media/19c31e9836247fde8c5f5f628598f6c3/href</a></iframe><p>Stay well everybody, and above all, stay human.</p><h3>Epilogue</h3><p>In regards to my last phrase, “<strong>stay human</strong>”.</p><p>Sometimes, people ask me: What do I think will happen when AI excels at system 2 capabilities (reasoning, planning, etc) in say 30, 40 or 50 years from now?</p><p><strong>System 1 and 2</strong> are different types of<strong> thinking modes </strong>in our minds.</p><p><strong>System 1</strong> refers to fast, subconscious, simultaneous, intuitive processes, and this is the domain where AI is reaching superhuman capabilities.</p><p><strong>System 2 </strong>refers to the slow, logical, rational, systematic, precise and sequential kind of thinking. And mastering this second mode is still way beyond our AI systems. (see <strong>Daniel Kahneman’s book</strong> “<a href="https://en.wikipedia.org/wiki/Thinking,_Fast_and_Slow">Think fast and slow</a>” to expand on system 1 vs system 2 thinking).</p><p>A discussion about system 2 capabilities in connection with AI, now and in the future, would fill a whole article of this size and larger. So I leave that for another time. Let’s get back to the question posed at the start of this epilogue.</p><p>I typically answer that the question may not make sense anymore in a few decades. Why not?</p><p>Because <strong>today </strong>there is a <strong>separation </strong>between AI and humans. AI is there. We are here.</p><p>But<strong> in some decades</strong>, that<strong> separation won’t be there</strong> anymore. Think of what the company <a href="https://neuralink.com/">Neuralink </a>is working on already these days. That’s only the very beginning of what’s to come.</p><p>In some decades, our technology, including AI, and our biology, will have merged in many ways.</p><p>And then, the new question may be:<strong> “Where will we go next, now that we are together?”</strong></p><p>Thank you for reading.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*yZ2LmBzUf8XQEKRiMuZr8g.jpeg" /><figcaption>Photos by Kelly Sikkema and Firmbee.com on Unsplash | Text by Javier ideami | ideami.com</figcaption></figure><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=a9786de586cb" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/towards-sustainable-generative-ai-revolution-a9786de586cb">Towards a sustainable generative AI revolution</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Cooking a good relationship | How to make it last]]></title>
            <link>https://medium.com/@ideami/cooking-a-good-relationship-how-to-make-it-last-613f0516dc51?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/613f0516dc51</guid>
            <category><![CDATA[human-prompt]]></category>
            <category><![CDATA[relationship-advice]]></category>
            <category><![CDATA[love]]></category>
            <category><![CDATA[relationships]]></category>
            <category><![CDATA[dating]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Sat, 09 Jul 2022 09:44:17 GMT</pubDate>
            <atom:updated>2022-07-09T09:57:22.935Z</atom:updated>
            <content:encoded><![CDATA[<figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*ESrORNj2W9zZBPMdAvyYUg.jpeg" /><figcaption>Graphic created by Javier Ideami</figcaption></figure><h3>Cooking a good relationship</h3><p>Let’s use <strong>an analogy connected to creativity, innovation processes, and cooking</strong> to reflect on how relationships evolve through time.</p><p>At the beginning of a relationship, no matter how similar two people may be, they always hold <strong>a good number of divergent ingredients</strong> in relation to each other. Two human beings, each with their own background, past experiences, etc <strong>will always diverge a good deal in relation to each other at the start</strong>.</p><p>And if they like each other, they are also likely to converge on a number of areas. So<strong> they also hold a good set of convergent ingredients.</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Nc1TRvSujljNX4c4Khl3Mg.gif" /><figcaption>Graphic created by Javier Ideami</figcaption></figure><p>Therefore, at the beginning of their relationship, their subconscious pots are incubating, combining, and cooking a good mixture of quality convergent and divergent ingredients.</p><p>This is <strong>the reason why relationships are so exciting at the start</strong>. Both people are still diverging a good deal in relation to each other. Therefore the mixture of their information ingredients generates <strong>all kinds of unexpected and original behaviors, reactions, and situations</strong>.</p><p>Just as in innovation processes we generate innovative ideas, <strong>in a fresh relationship we are generating innovative interactions, situations, and behaviors.</strong></p><p>A lot of people, in the first weeks of their relationship, feel that <strong>things are exciting and different in a very effortless way</strong>. They don’t have to put much or any effort into making things exciting. This is happening <strong>because both people diverge naturally in relation to each other</strong>, and their subconscious pots are continuously generating unexpected and exciting outputs.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*mvjuD9BhLTpjfZYurUr3FQ.jpeg" /><figcaption>Graphic created by Javier Ideami</figcaption></figure><p>After some time has passed (which is different for each relationship), a lot of <strong>those divergent ingredients gradually become convergent ingredients</strong>. Gradually, the number of convergent ingredients increases, and the number of divergent ones decreases.</p><p>For example, during their first days together, <strong>she taught him a traditional game from her childhood that he had never tried before</strong>. They had a lot of fun playing it together. They played it a few times over the next few weeks. But <strong>eventually</strong>, once they played it enough times, <strong>that divergent ingredient gradually became a convergent one</strong>. Now, they both know the game very well. They also know every detail of their reactions and behaviors when they play it.</p><p>We now arrive at the crucial point that has an enormous impact on the future of the relationship.</p><p>As a relationship progresses, <strong>both members of that relationship have a choice to make</strong>. They can either:</p><ul><li><strong>Make the effort to keep generating divergent ingredient</strong>s in their lives. Those divergent ingredients are information that is divergent in relation to the partner. This means <strong>growing and learning through new activities and experiences </strong>so that <strong>when you interact with your partner </strong>you can <strong>bring to your partner’s subconscious pot unexpected and divergent ingredients</strong> that will keep generating exciting and unexpected reactions, behaviors, and situations.</li><li><strong>Alternatively, you can just wait and see what happens</strong>. If you do that, <strong>eventually, most of your divergent ingredients will become convergent ones</strong>. And what this means is that when both partners interact, their interactions will be combining and incubating ingredients that are mainly convergent between them. The result of that will be <strong>situations, behaviors, and reactions that will also be mainly convergent, typical and predictable</strong>. Not very exciting. Another side effect is that <strong>too much convergence produces the merging of both partners</strong>. They become more dependent on each other and <strong>begin to lose their own unique identities</strong>.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*c409u3-jLdF4pM9semYNbA.gif" /><figcaption>Graphic created by Javier Ideami</figcaption></figure><ul><li>All <strong>extremes are dangerous</strong>. <strong>The opposite situation happens when both partners diverge too much </strong>and constantly in relation to each other. In this case, <strong>they begin to lose touch with each other</strong>. The relationship becomes <strong>unstable and too unpredictable</strong>. Divergent ingredients dominate their interactions and they lack a solid base made of the necessary convergent ingredients. Gradually they literally diverge from each other. The relationship is in trouble.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*75OQxqQL_QU0PRsdwLb5lA.jpeg" /><figcaption>Graphic created by Javier Ideami</figcaption></figure><h3>Why affairs or breakups happen</h3><p>Every human being and <strong>every human brain</strong>, we should say, <strong>longs for variety and new stimuli</strong>. <strong>These often come from those divergent ingredients</strong> we are talking about.</p><p>Within a relationship, <strong>we long for situations, interactions, and scenarios that feel fresh and exciting</strong>. If, after some time, all you are getting are typical and predictable situations, If one of the members of the relationship is not receiving enough divergent ingredients from the partner, what do you think that is going to happen?</p><p>What will happen is that <strong>one or both of the partners may go and find those divergent ingredients somewhere else</strong>. That means <strong>having an affair, breaking up, going away to travel around the world, </strong>or anything else. It means you go and find those divergent ingredients somewhere else because <strong>your partner or both of you have not made the proactive effort to keep enriching those subconscious pots</strong>.</p><h3>How to keep your relationship alive</h3><p>Therefore, if you want to keep your relationship alive, fresh, and exciting, you need to:</p><ul><li>First of all, <strong>understand that your relationship is so exciting at the beginning because you two diverge a lot naturally</strong> in relation to each other at the start. Therefore, without any effort, you both are holding and giving to each other tons of divergent ingredients. Those divergent ingredients, in combination with the convergent ones that you both also hold, <strong>generate effortlessly lots of unexpected, fresh and exciting situations </strong>and scenarios. The relationship feels awesome. And you feel that this will last forever. But <strong>now you know why it won’t last very long unless you make a proactive effort</strong> to keep it that way.</li><li>To keep it that way, <strong>you must continue to grow and enrich your own subconscious pot through learning and experiencing new situations, generating new divergent ingredients</strong>, ingredients which should be <strong>divergent in relation to your partner</strong>. This way, when you interact with your partner, even though the past divergent ingredients are gradually becoming convergent ones, <strong>you will be bringing to your partner’s subconscious pot a brand new batch of divergent ingredients</strong>, compensating the gradual transformation of the previous ones.</li></ul><h3>What about having new experiences together with my partner rather than on my own</h3><p><strong>Both are useful</strong>. <strong>Having new experiences together is a source of new divergent ingredients</strong>, because when you two experience something new, even if you do it together, the way you both experience it is different in relation to each other, and the way you interact with each other within a new context is also different. And that generates new divergent ingredients.</p><p><strong>However, if you always do everything with your partner, diverging in a consistent way is also harder</strong>. Doing things together all the time constantly empowers the process of transforming divergent ingredients into convergent ones as well as the process of generating new convergent ingredients in general.<strong> That’s why it is crucial to also learn and experience new things on your own</strong>, away from your partner. Doing that allows you to diverge freely without any constraints. You will be generating a lot of divergent ingredients that will later enrich your partner’s subconscious pot.</p><p>And that’s the paradox, <strong>being away from each other is often what later will bring you closer together.</strong> And the explanation is that <strong>what brings you closer together is a great mix of quality convergent and divergent ingredients.</strong></p><p>And after the honeymoon phase of a relationship has passed, in order to generate those divergent ingredients quickly and powerfully, doing new things away from your partner is really effective. So remember, doing things with your partner is wonderful and also useful and effective. But doing new things away from your partner is very powerful and useful in order to approach that balance we are looking for, the balance between converging and diverging in relation to your partner.</p><p><strong>Enjoy great new experiences with your partner but also find the time to have new ones on your own.</strong></p><h3>What about living together</h3><p>When you live together, <strong>the transformation of divergent ingredients into convergent ones happens faster</strong>. That’s why a lot of couples report that their relationship degrades and becomes more monotonous after they move in together.</p><p>That’s also the reason why sometimes <strong>one or both of the partners begin to argue and fight more often after they move in together</strong>. This happens because gradually, a lot of their divergent ingredients become convergent ones and their subconscious pots begin to generate mostly typical and predictable outputs.</p><p>And <strong>when things get too predictable and less exciting, it is easier to get annoyed</strong> with things around you, especially, in a subconscious way, with your partner, with that partner that is not giving you any more quality divergent ingredients.</p><p>Therefore, if you wanna live together and keep the relationship fresh and exciting, what we discussed before applies again. <strong>You have to keep learning and growing</strong>, sometimes on your own and sometimes with your partner. You should keep generating quality divergent ingredients that can enrich both of your subconscious pots.</p><p>And the fact is that if you are living together, <strong>you need to be even more proactive</strong> about this than if you are not because <strong>the speed of transformation of your divergent ingredients into convergent ones increases</strong> the more you interact with each other on a daily basis.</p><h3>But we are so happy doing always the same thing together</h3><p>Let’s suppose that you are born in a city called <strong>A</strong>. And when you were raised you were told that city <strong>A</strong> was the best and had everything you needed. So you never travel outside of it, not even to other places nearby. And when people ask you about it, you say that you are happy, that <strong>you have everything you need in your city</strong> and that you don’t need to go to see other places around.</p><p>Or let’s suppose that you are a very fearful person and you decide to be in your house all the time 24/7. And you tell people that you are very happy living 24/7 in your house without going to the streets and that you are satisfied with that lifestyle because you can order everything through the internet and <strong>you don’t really need to exit the house for anything</strong>.</p><p><strong>The fact that these scenarios are possible,</strong> and indeed they happen,<strong> does not mean </strong>that they are healthy or<strong> that there aren’t other alternatives that could improve the lives</strong> of these persons and make them better, deeper and richer.</p><p>The fact that they happen only means that<strong> this person is accepting some self-imposed limits</strong>. That’s all. It does not mean that such a person would not benefit from visiting other cities and places, from living in a different way.</p><p>There are <strong>all kinds of people in the world</strong>. People with all sorts of backgrounds and experiences. And <strong>there are</strong> indeed <strong>people out there that either prefer or want to stay in relationships that are mainly convergent</strong>. They are totally ok with a partner that is either very similar to them or makes the effort to converge to them continuously. And they are totally ok triggering situations, behaviors, and scenarios that are most typical and predictable.</p><p>This is their choice. But let’s not be mistaken. It is just a choice.<strong> Most people will agree that there is more happiness and personal growth to be found in a life that is rich with variety and diversity of stimuli, rather than in one full of monotony and repetition</strong>. And the biology of our brain agrees as well.</p><h3>Why do some couples that are really convergent stay together?</h3><p>Sometimes it happens because of fear, lack of options, or laziness. Fear of not finding another partner or lack of self-confidence in their own possibilities of finding somebody else. Other times it is because of financial issues. <strong>In normal conditions, it is fear or laziness, subconscious, or conscious that keeps people from breaking up a relationship that has grown monotonous or is too convergent.</strong></p><p><strong>The natural impulse of most, not all, but most human beings is to find a partner that continuously feeds you a good mixture of convergent ingredients </strong>(ingredients that resonate with you and which establish a solid foundation in the relationship)<strong> and divergent ingredients</strong> (ingredients that diverge from you and which in combination with the convergent ones, generate unexpected and exciting situations, behaviors and scenarios).</p><p>Some people want the balance between convergent and divergent ingredients in their relationship to lean more towards the convergent ones. And others prefer the opposite. What we must remember is that<strong> all extremes are unhealthy</strong>. And a relationship that has too many convergent ingredients or too many divergent ones (in relation to the others), is going to get in trouble.</p><p>A relationship with <strong>an excess of convergent ingredients</strong> (in relation to the divergent ones) will feel <strong>stale and predictable</strong>. A relationship with a<strong>n excess of divergent ingredients</strong> (in proportion to the convergent ones) will feel <strong>unstable and too unpredictable</strong>.</p><p>However, we put <strong>special emphasis in the generation of divergent ingredients </strong>because <strong>those are the ones that are harder to generate</strong>, and the reasons for this (biological, cultural and social) are explained in depth in methodologies like <a href="https://torch4ideas.com/about"><strong>Torch </strong></a>(parts of which I’m using to build analogies in this article).</p><p>Strive for balance. And once the honeymoon phase has passed, make sure to <strong>be proactive about generating those divergent ingredients that your partner will subconsciously need. Do it for yourself and for your partner. Do it for the good of your relationship.</strong></p><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=613f0516dc51" width="1" height="1" alt="">]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Lucy  says hi — 2031, AGI, and the future of A.I]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/data-science/lucy-says-hi-2031-agi-and-the-future-of-a-i-28b1e7b373f6?source=rss-7f7b5d730c84------2"><img src="https://cdn-images-1.medium.com/max/1600/1*886rbVXidg1rr5WQrhWt8Q.jpeg" width="1600"></a></p><p class="medium-feed-snippet">In the year 2031 there is a new Alexa in town, and artificial general intelligence is one of its features. What is Lucy made of?</p><p class="medium-feed-link"><a href="https://medium.com/data-science/lucy-says-hi-2031-agi-and-the-future-of-a-i-28b1e7b373f6?source=rss-7f7b5d730c84------2">Continue reading on TDS Archive »</a></p></div>]]></description>
            <link>https://medium.com/data-science/lucy-says-hi-2031-agi-and-the-future-of-a-i-28b1e7b373f6?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/28b1e7b373f6</guid>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[deep-dives]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Wed, 16 Jun 2021 13:27:08 GMT</pubDate>
            <atom:updated>2021-06-16T13:27:08.027Z</atom:updated>
        </item>
        <item>
            <title><![CDATA[Journey to the center of the neuron]]></title>
            <link>https://medium.com/data-science/journey-to-the-center-of-the-neuron-c614bfee3f9?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/c614bfee3f9</guid>
            <category><![CDATA[deep-dives]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Wed, 21 Apr 2021 13:34:53 GMT</pubDate>
            <atom:updated>2021-04-26T04:20:17.614Z</atom:updated>
            <content:encoded><![CDATA[<h4>Dive into the salty ocean of the brain and get closer to the entities that inspire our A.I systems and make your thoughts possible. Explore how understanding the difference between artificial and biological neurons may give us clues about how to move towards a more flexible kind of artificial intelligence.</h4><iframe src="https://cdn.embedly.com/widgets/media.html?src=https%3A%2F%2Fwww.youtube.com%2Fembed%2FEDUyB3u8nbQ%3Ffeature%3Doembed&amp;display_name=YouTube&amp;url=https%3A%2F%2Fwww.youtube.com%2Fwatch%3Fv%3DEDUyB3u8nbQ&amp;image=https%3A%2F%2Fi.ytimg.com%2Fvi%2FEDUyB3u8nbQ%2Fhqdefault.jpg&amp;key=a19fcc184b9711e1b4764040d3dc5c07&amp;type=text%2Fhtml&amp;schema=youtube" width="854" height="480" frameborder="0" scrolling="no"><a href="https://medium.com/media/5abc840b0ec228a0e098c57a5260e318/href">https://medium.com/media/5abc840b0ec228a0e098c57a5260e318/href</a></iframe><p><strong>Every single one of your thoughts</strong> is<strong> made possible by your biological neurons</strong>. And behind many of the most useful A.I architectures is an entity inspired by them. <strong>Neurons</strong> are at the <strong>epicenter</strong> of the processing that underpins the complexity produced by intelligent systems. Curious to know more about the engine of your thoughts and about how they compare to their artificial counterparts? Let’s do it!</p><p><strong>A.I neurons</strong> were originally inspired by our biological ones, yet they are <strong>very different.</strong> And why shouldn’t they be? There are<strong> many ways to get to the same destination</strong> and in the same way that <strong>human flight got inspired but didn’t copy part by part the way that birds fly,</strong> our artificial neurons are only partially inspired by our biological ones.</p><p><strong>And yet</strong>, our biological neurons are way more complex than our artificial ones and hold so much rich detail and so many mysteries within. Even if we don’t need to copy the way biological neurons work, <strong>understanding what is different</strong> between both entities <strong>could give us new clues</strong> about how to move towards a more flexible form of artificial intelligence.</p><p>In this article we will review those differences, as well as some that have to do with the networks where these neurons are embedded. We will consider also how those differences may inform new possibilities for the coming future.</p><p><strong>Note</strong>: A.I is a vast field. There are all sorts of exotic variations within the area. When making comparisons, I will be referring exclusively to some of the most typical and popular deep learning architectures nowadays.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*D_UdWMcEYYKxMKgweVdmpQ.jpeg" /><figcaption>Neural Spring Waterfall, painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>The neuron: salty water and electricity</h3><p>Before we go deep into the details of the functionality of the neurons, let’s take a quick look inside. This is about to get salty!</p><ul><li>Think of <strong>your brain as a recipient that holds a salt water ocean</strong>. Within that ocean you got a lot of cells that we call <strong>neurons</strong> and a lot of <strong>ions</strong>. <strong>Ions </strong>are atoms in which the positive and negative charge are not equal. And the <strong>main ions you have in your brain </strong>are: <strong>sodium</strong> (Na+), <strong>potassium</strong> (K+), <strong>calcium </strong>(Ca++) and <strong>chloride</strong> (Cl-). A good reminder of why those minerals are so important!</li><li>So, the biological <strong>neuron</strong>, like most cells, is basically <strong>made of salt water with ions like chloride and sodium floating around</strong>. And <strong>everything a neuron does can be explained in terms of electricity</strong>: <strong>voltages</strong> (the potential/voltage that, for example, exists at the membrane of the neuron) and <strong>currents</strong> (the flow of charged ions in and out of the neuron).</li><li>The <strong>artificial neuron </strong>is created with<strong> computer code</strong>, which when executed creates <strong>data structures made of digital bytes</strong>, and so on. Most of what an artificial neuron does can be understood in terms of <strong>computations</strong>, <strong>linear and nonlinear transformations</strong> of data.</li></ul><p><strong>A salty ocean vs a silicon wonderland.</strong> What’s more efficient? Let’s take a look.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*BCAkHO6lXqjIGkD_UyZHNA.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>Energy: 20 watts for a premium service</h3><p><strong>Information processing requires energy, </strong>so energy consumption within these neural networks matters a lot . It sets boundaries on what’s possible and <strong>our brain is tremendously efficient.</strong></p><ul><li>Right now, as thoughts move through your mind, your <strong>brain is spending just about 20 watts of power</strong>, barely enough to turn a light bulb on. It is able to do this <strong>even while you fast or sleep</strong>, and keeping a <strong>moderate temperature around 37 celsius</strong> degrees.</li><li>The powerful <strong>GPUs </strong>often used by our deep learning systems can spend <strong>hundreds of watts per unit,</strong> way more than our brain, and <strong>they emit a lot of heat</strong>, reaching temperatures around 70 or 80 celsius degrees.</li></ul><p><strong>What’s up:</strong> research is ongoing on making A.I systems more energy efficient. Sparsity of connections and activations may help these systems to approach the tremendous efficiency of our brain, in which only a small percentage of the neurons are active at any one time (typically between 0.5 and 2%).</p><h3>Functionality: detecting all around</h3><p>We have reviewed the <strong>environment </strong>and the <strong>energy </strong>consumption. Time to zoom into one of these entities . What is a neuron doing?</p><ul><li>It<strong> detects patterns coming from the many inputs </strong>it receives. We have around <strong>100 billion neurons in the brain </strong>(estimates vary, some experts say the real number is more around 86 billion), and <strong>each of them receives inputs from around 10000 other neurons</strong> (some say it’s around 7000, others 8000, in any case there are thousands of them).</li><li>The biological neuron has a <strong>threshold </strong>and <strong>when that threshold is crossed, it emits a signal </strong>that we call an <strong>action potential or spike</strong>. That signal travels down its <strong>axon </strong>(the neuron’s output) towards the <strong>synapses</strong> of other neurons.</li><li>The <strong>synapse </strong>is a structure that allows a neuron to pass a signal to another neuron. Synapses are located at the <strong>dendrites </strong>of the neurons (its branches).</li><li>So you can <strong>summarize the biological neuron</strong> with this process: <strong>receive </strong>inputs, <strong>integrate </strong>them, and <strong>decide</strong> if the result of that integration is strong enough to fire an output signal.</li><li><strong>The artificial neuron</strong>, on the other hand, <strong>performs computations</strong> that <strong>combine its inputs with their respective weights </strong>(the weights are numerical values that specify the strength of the connection between that neuron and each of its inputs).</li><li>The result is <strong>passed through an activation function</strong> (a function that computes a nonlinear transformation of its input. This makes it possible for the network to learn nonlinear mappings between inputs and outputs).</li><li>So you can <strong>summarize the artificial neuron</strong> with this process: <strong>receive</strong> inputs, <strong>add the result of multiplying each of those inputs by the strength of their connection </strong>to that neuron, and <strong>pass the result</strong> of that calculation <strong>through a nonlinear function</strong>.</li></ul><p><strong>What’s up: </strong>notice a key difference. <strong>Biological neurons have a threshold</strong> that <strong>keeps them silent until it is crossed</strong>. Most artificial neurons in deep learning systems produce an active output (some may output a 0. For example, ReLU activation functions set the output of the neuron to 0 if their input is smaller than 0). As we will emphasize later on, around 0.5 to 2% of our biological neurons are active at any one time vs around 50% in typical artificial deep learning systems.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*rKYr32PGJrpOJNTo1YsS4g.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>Electricity all around</h3><p>Let’s go back for a moment to the electric side of things within the biological realm. We often hear that our <strong>neurons communicate with electric impulses</strong>. Let’s dive into the electric dimension so that we can internalize even better what is going on.</p><ul><li>A <strong>voltage </strong>is a <strong>comparison</strong> (in relative terms) between <strong>the electric charge in one place and the charge in another place.</strong></li><li>The <strong>voltage in neurons</strong> is typically measured in <strong>millivolts</strong>. A millivolt is <strong>a thousandth of a volt</strong>. Our <strong>tiny neurons</strong> use <strong>tiny amounts of electricity</strong> to conduct their operations.</li><li>A crucial thing to highlight is the <strong>membrane potential (voltage) of a neuron</strong>. This is the <strong>voltage of the neuron in relation to the space outside of it</strong>. And we call it <strong>membrane</strong> potential because it is located at the membrane of the neuron, which is <strong>a thin layer of fat</strong>.</li><li><strong>When there is a difference</strong> in electric charge between two areas, the electric <strong>charge will tend to flow in order to compensate</strong> for the difference and equalize the situation.</li><li>The <strong>membrane </strong>of the neuron <strong>acts as a barrier between the electric current inside and the one outside</strong>. And what we call <strong>ion channels </strong>are like little <strong>tunnels in that barrier that allow things to flow in a controlled way</strong>.</li><li>The <strong>conductance</strong>, the <strong>size of the openings</strong>, determines how fast those ions flow in and out of the membrane.</li><li>So <strong>when a voltage (potential) exists, a relative difference in charge between two areas, ions flow to equalize things</strong>. But why? Because of the <strong>universal principle of opposite charges attracting and similar charges repelling </strong>each other. When, for example, in a certain setting there is more positive charge than negative, an electrical current is formed to equalize the situation, bringing more negative charge into the area.</li><li>The <strong>firing threshold </strong>(or <strong>action potential threshold</strong>), is the <strong>voltage that has to be reached</strong> at the membrane of the neuron <strong>for the neuron to fire </strong>an output signal (an action potential) through its axon. This <strong>threshold </strong>is <strong>typically </strong>at <strong>around -50mv (millivolts)</strong>.</li><li>The so-called <strong>resting potential of a neuron</strong> is at <strong>-70mv,</strong> <strong>below the firing threshold</strong>, so <strong>by default the neuron will not fire.</strong></li><li>The <strong>importance of this threshold </strong>cannot be overstated. As a consequence of its existence, <strong>only the most intense levels of activation are communicated </strong>through the neuron’s axon (its output). This allows for the <strong>information to be encoded in a very compact and efficient way.</strong></li><li>Remember those<strong> 20 watts </strong>of energy consumption that we mentioned at the beginning? <strong>Biological</strong> <strong>neurons only communicate relevant</strong> and key information. <strong>For the rest, they stay silent</strong>.</li></ul><p>Let’s remind ourselves that <strong>there is no reason why we should copy or imitate the complexity of the biological neuron</strong>. We could create systems that display flexible forms of intelligence in a completely different way. And <strong>the simpler the way, the better.</strong> But <strong>understanding in depth our biological neurons could give us some ideas</strong> that may enrich our experiments and strategies when working with artificial neurons.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*XMCOG7iXahOZEnIFNlelzA.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>It’s about time: to spike or not to spike?</h3><p>A very <strong>important difference</strong> between both kinds of neurons is related to<strong> the time dimension.</strong></p><ul><li><strong>Biological neurons fire</strong> for a <strong>very brief</strong> moment. They emit <strong>spikes</strong>, which last a very short time (typically around 1 millisecond). The<strong> information transmitted</strong> by a neuron is <strong>encoded in the timing of those spikes</strong>. A <strong>spike train </strong>is a sequence of <strong>spikes and silences.</strong></li><li><strong>After each spike</strong>, the neuron’s <strong>membrane potential returns to be a small value</strong> (it can even go below its resting potential). In order <strong>to spike again</strong>, the <strong>voltage needs to go back up and above </strong>the<strong> firing threshold.</strong></li><li>When <strong>learning</strong> processes take place, the <strong>efficiency with which a neuron can contribute to activate other neurons</strong> <strong>can </strong>dynamically <strong>change</strong> through processes like <strong>long term potentiation (LTP), </strong>which are central to the way we learn and create memories (long term depression is the opposite process of LTP).</li><li><strong>Most artificial neurons </strong>are <strong>generally constantly producing outputs</strong> (throughout each execution cycle), sending continuous signals to the next neurons down the line (sometimes their activation functions may set their output to 0).</li><li>So, in most artificial deep learning networks, the <strong>time dimension is not relevant</strong>. There is no threshold per se in the way it is used in our biological networks. Our artificial systems are way simpler. But simpler sometimes is better. Will the direction in which today’s deep learning systems are moving be enough to take us to what experts like to call AGI (artificial general intelligence)? Or in any case to a more flexible form of A.I?. The jury is still out on that matter.</li></ul><p><strong>What’s up: </strong>The <strong>Von Neumann </strong>architecture is <strong>behind most of the hardware we use today. </strong>To get closer to what the brain does, some researchers are starting to work with <strong>other kinds of architectures. Neuromorphic computing</strong> is an example. This kind of architecture allows for <strong>more parallel processing and robustness</strong>. Most importantly, it can work with <strong>spiking neural networks,</strong> which deal with <strong>both the spatial and the time dimensions</strong>, just as the brain does. Companies like <strong>IBM</strong> or <strong>Intel </strong>have already produced <strong>neuromorphic chips.</strong> This area of research faces a number of important challenges, both in the research front and also in terms of having to deal with an existing ecosystem that is so well adapted to the Von Neumann model.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*3DjDxN1HLQSejKth7v6b9A.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3><strong>Some space, please: the magic of sparsity</strong></h3><p>When neurons combine with each other,<strong> the number</strong> of them that are<strong> firing at any one moment and the number of connections between them </strong>have a lot of consequences in terms of <strong>energy consumption, resiliency, robustness</strong> and other related factors.</p><ul><li>As thoughts stream through your mind, only <strong>around 2% of your neurons are firing on average</strong>. Most of them are silent. Because only a small percentage of the neurons are active at any one time, noise and other distortions have a harder time interfering with the pattern detection processes of these networks. <strong>Sparsity makes our biological networks resilient and robust.</strong></li><li>Conversely, in our <strong>artificial deep learning networks</strong>, most of the neurons are <strong>continuously producing outputs </strong>(some may have their outputs set to 0 by their activation functions). This is one of the potential reasons why deep learning systems are<strong> often quite brittle and sensitive to</strong> what we call <strong>adversarial attacks/examples</strong>. <strong>Adversarial attacks</strong> are <strong>subtle</strong>, minimal <strong>changes in the input</strong> of a network (typically invisible to our perception), that <strong>produce dramatic and incorrect changes in its output</strong>. The <strong>non-sparse nature of deep learning networks </strong>makes them <strong>more sensitive to variations in their inputs</strong>. When most of the weights are relevant and constantly in play, <strong>any changes can produce dramatic consequences.</strong></li><li>But <strong>sparsity </strong>goes beyond the activations. <strong>Numenta </strong>is a well known research company where a talented group of scientists and engineers combine <strong>neuroscience </strong>and <strong>machine intelligence</strong> research. Their team is led by <strong>Jeff Hawkins</strong> and <strong>Subutai Ahmad. </strong>Numenta’s team<strong> </strong>has explored very deeply the issue of sparsity in the brain as well as other areas related to how our neocortex functions. One of the things we can learn from their work, research and publications, is that our neocortex is <strong>sparse at two levels.</strong></li><li>Firstly, as stated above, in terms of activation, <strong>best estimates are that 0.5% to 2% of our biological neurons are active at any one time.</strong></li><li>And then, we also have <strong>sparsity in terms of connectivity </strong>between the neurons. <strong>When a layer of neurons projects onto another</strong> layer, Numenta’s team tells us that current estimates say that <strong>1 to 10% of the possible neuron to neuron connections exist.</strong></li><li>In contrast to that, <strong>most deep learning systems nowadays are very dense.</strong> <strong>100% dense in terms of connectivity, </strong>typically. And <strong>around 50% in terms of activations.</strong></li></ul><p><strong>What’s up: </strong>New architectures that make use of sparse connectivity and sparse activations are an ongoing area of research. Regardless of A.I not having to necessarily copy what the brain does, sparsity, as a strategy, makes quite a lot of sense within the mission to build systems that are more resilient and robust.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Cg7YQskDvbb-xq2YKOD4Yw.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>Inputs and parameters</h3><p>Ultimately we want to use these neurons to learn. So let’s now go deeper into the <strong>knobs that these networks tweak </strong>in order to generate that learning. Afterwards we will look at the learning algorithm itself. We begin by comparing the <strong>inputs and parameters </strong>of both kinds of neural networks.</p><ul><li>In <strong>biological networks</strong> you can have <strong>3 types of inputs</strong>: <strong>excitatory</strong> (makes the receiving neuron more likely to fire), <strong>inhibitory </strong>(does the opposite) and <strong>leak</strong> (similar function to the inhibitory one).</li><li>As stated previously, these<strong> inputs interface with the receiving neuron </strong>through <strong>synapses</strong>, <strong>the connection points between sending and receiving neurons</strong>. And <strong>most synapses are on the dendrites</strong> of the receiving neuron.</li><li>The <strong>dendrites are branches</strong> that come off the neuron (dendrite comes from the greek -<strong>dendro</strong>-, which means tree). At the dendrites, the different <strong>input signals are integrated</strong>. These dendrites have <strong>small spines</strong> on them. It is there <strong>where the outputs from the sending neurons (axons) interface (synapse),</strong> establishing connections to other neurons.</li><li>In<strong> artificial networks,</strong> you have, in general, <strong>a single kind of input, </strong>which typically has a number of associated <strong>weights </strong>(numbers that express the strength of the connection between that input and a number of other neurons connected to it, one weight per connection).</li><li>Those weights<strong> hold continuous values that can be negative or positive. </strong>Each weight value, in combination with the computations performed at the neuron, will, in practice, contribute to make the receiving neuron more or less active (kind of analog to the excitatory — inhibitory dynamic previously described).</li><li>So, <strong>in the biological neuron, </strong>there is this <strong>battle between excitatory and inhibitory signals</strong>. The result of that battle <strong>determines the voltage at the cell’s membrane</strong>. And it is that membrane voltage that <strong>needs to go over the action potential threshold</strong> in order for the neuron to <strong>fire</strong>.</li><li>In the <strong>artificial neuron, </strong>things are simpler. There is no explicit threshold and <strong>the different strengths</strong> (positive or negative) of each of the <strong>weights combine to stimulate more or less </strong>the receiving neurons.</li></ul><p>Let’s <strong>get closer to the parameters</strong> of both entities.</p><ul><li>In <strong>biological networks</strong> we have the concept of <strong>the synaptic weight,</strong> which <strong>determines the impact that a signal from a sending neuron</strong> can have <strong>on the receiving neuron</strong> <strong>through their synaptic connection</strong>.</li><li>Getting even closer, what this <strong>impact represents</strong> is <strong>the capacity of the sending neuron’s action potential to release neurotransmitters</strong>,<strong> and of those neurotransmitters to open the synaptic channels </strong>of the receiving side.</li><li>In<strong> artificial networks, </strong>we have <strong>weights</strong> that determine the strength of each of the connections between a sending and a receiving neuron. And those<strong> weights are </strong>simply <strong>numbers</strong>. They could be floating point numbers, integers, single bits, etc.</li><li>So, whereas<strong> the artificial weight is a simple number,</strong> the biological <strong>synaptic weight depends on a lot of factors</strong>. Those factors include, for example, the <strong>amount of neurotransmitter</strong> that can be released into the synapse and absorbed on the other side (and here the number of specific kinds of receptors and ions come into play), how well the signal moves through the axon (<strong>myelination</strong> in the axon has an impact on this), the<strong> efficiency of the signal propagation</strong> and the<strong> number of connections between the axon and the dendrites</strong> of the receiving neuron. As we can see, this goes well beyond a simple number.</li></ul><p>So, those are <strong>the parameters</strong> of these networks. And <strong>zooming </strong>out for a moment, <strong>what are they representing</strong>?</p><ul><li>These weights, in general, represent <strong>what a neuron is sensitive to,</strong> <strong>what it is detecting</strong>. If a <strong>weight value is large,</strong> it <strong>means that the related neuron is very sensitive to the input it is receiving</strong>.</li><li>We can therefore sense that the <strong>learning process, </strong>in both cases, has to do with <strong>changing and tweaking these weights, </strong>producing different patterns in the networks as learning progresses.</li><li>So, get ready for this: <strong>each one of your thoughts and memories is represented by a pattern of synaptic weights</strong>. And a similar thing takes place in our artificial networks, where patterns of numerical weights represent information at different abstraction levels, patterns that evolve throughout the learning process.</li></ul><p>We got the structure, the inputs, the parameters and the outputs. Time to learn!</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*jUXsRVbm9rVyKJHfWi-CEA.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>Learning: backpropagation and beyond</h3><p>In order to <strong>learn</strong>, we need to <strong>tweak those weights</strong>, those parameters, and to do it in the right direction. But how? What is the learning algorithm?</p><ul><li>In our <strong>artificial deep learning networks</strong>, <strong>backpropagation, </strong>in combination with <strong>gradient descent,</strong> is the <strong>typical</strong> <strong>algorithm </strong>of choice.</li><li>For example, in supervised systems (where we provide a dataset with labels, say a number of images of animals with labels that identify the kind of animal), we run the network and then calculate its<strong> performance, its loss value</strong> or error (the difference between what we are obtaining and what we would like to obtain).</li><li>Then, starting <strong>from the output of the network and moving towards its input </strong>(<strong>propagating</strong> the current loss value <strong>in reverse</strong>, that’s why we call it backpropagation), we use the power of <strong>calculus</strong> and its<strong> chain rule</strong> to <strong>calculate the impact of each of the parameters of the network on that final loss value.</strong> We are able to do this because all the computations performed at the different layers of the neural network are differentiable.</li><li>Once we know how tweaking each of the weights will impact the final loss value of the network, we can proceed to <strong>tweak each of those parameters</strong> <strong>in the direction that will minimize that final loss value,</strong> that difference between our targets and where we are at each moment.</li><li>If we <strong>keep repeating this process</strong>, moving <strong>down those gradients,</strong> we will <strong>end up at a place where the computations produced with the combination of all our weights produces a very small difference </strong>between our target values and our current network output. Learning has taken place.</li><li>And <strong>what about our biological networks? Is something similar to backpropagation going on in the brain?</strong> There is <strong>controversy </strong>around this. <strong>Some experts </strong>think that <strong>there may be something</strong> going on in the brain that, while being different, <strong>may have similarities to what backpropagation does</strong>. <strong>Others think</strong> that the way our biological networks learn <strong>has nothing to do with it.</strong> So, <strong>the jury is still open</strong> and there is a lot of active research in this area.</li></ul><p><strong>What’s up : Backpropagation</strong> is a great learning algorithm. And yet, like any algorithm, it has strengths and weaknesses. what if we went beyond the search for something similar to backprop in the brain and considered other options? Researcher <strong>Ben Goertzel</strong>, <strong>an expert in the AGI field </strong>(artificial general intelligence), thinks that we will eventually go beyond backprop and use other kinds of learning algorithms that will adapt in better ways to the needs of the future AGI systems. Those may include <strong>evolutionary algorithms</strong> of the kind of <strong>CMA-ES</strong> applied to complex neural architectures.</p><p>And Ben tells us that <strong>if we used these kinds of evolutionary algorithms</strong>, we could then, for example, use <strong>inference for fitness estimation </strong>and <strong>other strategies to guide the evolutionary learning</strong> process, strategies that are more difficult to implement when we use backpropagation.</p><p>Ben puts forward a really interesting question: <strong>how many neural architectures are being discarded just because they are not suitable to work with the backpropagation algorithm? </strong>It is a great reminder to keep our options and our mind open to new possibilities.</p><p>Our <strong>brain </strong>benefits from<strong> two kinds of learning processes:</strong> <strong>the evolutionary ones</strong> that get <strong>encoded in our genes</strong>, and <strong>the ones that take place during our lifetimes within our neural networks</strong>. <strong>Combining both approaches</strong> in our artificial networks could open the door to new advances.</p><h3>Evolution: everything in movement</h3><p>Which takes us to the evolutionary aspect of these neural architectures.</p><ul><li>Right now, <strong>the neural networks within your neocortex are already different to how they were hours ago. </strong>They<strong> never stop evolving</strong>.</li><li>In general and zooming out, in our biological networks there are <strong>optimization processes going on at many different levels</strong>, not only in terms of their parameters, but also at the level of the networks themselves, their structure, algorithms, etc (in relation to our genome, for example).</li><li>When active, <strong>our typical and most popular artificial deep learning systems</strong> <strong>optimize their parameters</strong> (weights) by using backpropagation, and <strong>that’s about it.</strong> The architecture itself remains static, apart from the changes we apply from time to time to their hyperparameters (manually or through <strong>autoML</strong>, <strong>grid search</strong> and other similar options).</li></ul><p><strong>What’s up: </strong>research on <strong>self-optimizing mechanisms</strong> that may allow deep learning architectures to<strong> transform and evolve their structures and optimize their strategies</strong> as the learning process progresses, could make our artificial networks more adaptive and flexible. Researchers like <strong>professor Kenneth Stanley</strong> have already produced very interesting results with dynamic systems that are moving in this direction.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*Y0Wi-xlIfRy6c1-Qzn8tjQ.jpeg" /><figcaption>Neural Forest, painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>When: continuous learning</h3><p>And what is the duration of those learning processes?</p><ul><li>In our<strong> current and most typical deep learning networks</strong>, training processes have<strong> a beginning and an end</strong>. We first train, complete the learning process and then we perform what we call <strong>inference</strong>. We apply that learning to previously unseen data in a separate process.</li><li>We may keep retraining our networks as new data comes in, and update our models iteratively.</li><li><strong>In our brain</strong>, learning processes never go to sleep. <strong>Continuous learning is taking place.</strong> The <strong>strengths of our synaptic connections</strong> change as we think, as we act, and even as we sleep.</li></ul><p><strong>What’s up: </strong>Continuous learning is a hot topic in the A.I community. We know that if we are to eventually reach a more flexible and robust kind of artificial intelligence, learning needs to have more continuity. A lot of research is ongoing in this area.</p><h3>Let’s be social</h3><p><strong>Biological neurons </strong>are <strong>way more social </strong>than our artificial ones. What does this mean?</p><ul><li>In typical deep learning networks,<strong> artificial neurons</strong> communicate <strong>in a single direction </strong>towards the next layer (the backpropagation computations are done in reverse), and they connect only to the previous and next layers. There are exceptions, but we are talking here about the most typical kinds of deep learning networks.</li><li><strong>Biological neurons </strong>may communicate in multiple directions and have <strong>a wider and richer range of connections</strong> (being those also more sparse at the same time)<strong>.</strong> Some neurons communicate up and down the cortical columns. Others have connections that go sideways. The timing aspect that I commented earlier introduces an even richer aspect to the process.</li></ul><p><strong>What’s going on: </strong>new kinds of A.I architectures that make use of more flexible and richer forms of connectivity are another active area of research. For example, <strong>researchers</strong> are working on <strong>graph neural networks</strong> and other architectures that use <strong>hypergraphs </strong>and <strong>metagraphs. </strong>The <strong>singularityNET </strong>project,<strong> </strong>founded by <strong>Ben Goertzel, </strong>is doing a lot of work in this area. It combines <strong>blockchain</strong> technology with artificial intelligence services to produce a <strong>decentralized</strong> <strong>A.I network</strong>. Recently, the project has partnered with the <strong>Cardano </strong>ecosystem to accelerate its progress towards a global decentralized AGI system.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*sKGUBmAxsWTGBV6O8tJWmA.jpeg" /><figcaption>Painting by the author Javier Ideami | <a href="https://ideami.com">https://ideami.com</a></figcaption></figure><h3>Zooming away</h3><p>To complete our journey, let’s zoom out for a moment.</p><ul><li>The neurons that are involved in the advanced part of our intelligence are located in our <strong>neocortex</strong>.</li><li>Our neocortex is organized in terms of <strong>microcolumns, </strong>each of which is composed of around <strong>100 neurons </strong>that deal with similar types of data.</li><li>A number of<strong> microcolumns are then structured in terms of cortical columns</strong> and we have<strong> around 150000 columns in our neocortex</strong>.</li><li>If you are interested in going deeper into what goes on in these cortical columns, I recommend you check the book, <strong>“A Thousand Brains: A New Theory of Intelligence”,</strong> by scientist and entrepreneur <strong>Jeff Hawkins, </strong>a true masterpiece in which he dissects the latest research performed by his team at <strong>Numenta.</strong></li></ul><p>And if you want to <strong>explore further how recent neuroscience research</strong> like the one performed by Jeff Hawkins and his team <strong>points the way towards achieving a more resilient, consistent and flexible form of artificial intelligence</strong>, you may check this other article below that I wrote on the topic</p><p><a href="https://towardsdatascience.com/towards-the-end-of-deep-learning-and-the-beginning-of-agi-d214d222c4cb">Towards the end of deep learning and the beginning of AGI</a></p><p><strong>The sounds of a million souls poem</strong></p><blockquote><em>“ And as we approach the magnificent column,</em></blockquote><blockquote><em>the mysterious pattern calling us from afar with the sounds of a million souls…</em></blockquote><blockquote><em>I sense that the brightest sun is compressed in those tiny specks of wonder..</em></blockquote><blockquote><em>reduced to a tapestry of dreams that resonate in our consciousness..</em></blockquote><blockquote><em>And I hear you laugh.. I hear you fall… I hear your tears devastate the horizons..</em></blockquote><blockquote><em>until we merge at the center of the column where silence awaits..</em></blockquote><blockquote><em>Silence, and then the million suns spiking towards the awakening of a new existence..</em></blockquote><blockquote><em>Hold me tight.. and let’s dive right in, right into the center of the column..</em></blockquote><blockquote><em>where you and I are one in silence.. ”</em></blockquote><blockquote><em>— by Javier Ideami</em></blockquote><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=c614bfee3f9" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/journey-to-the-center-of-the-neuron-c614bfee3f9">Journey to the center of the neuron</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Towards the end of deep learning and the beginning of AGI]]></title>
            <link>https://medium.com/data-science/towards-the-end-of-deep-learning-and-the-beginning-of-agi-d214d222c4cb?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/d214d222c4cb</guid>
            <category><![CDATA[deep-learning]]></category>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[towards-data-science]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[data-science]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Wed, 17 Mar 2021 12:31:15 GMT</pubDate>
            <atom:updated>2021-04-21T13:59:53.711Z</atom:updated>
            <content:encoded><![CDATA[<h4>How recent neuroscience research points the way towards defeating adversarial examples and achieving a more resilient, consistent and flexible form of artificial intelligence</h4><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*GBJnNA6UicklMlFvLHsEMg.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*FZ0Ik7ZE65HvvKG085gkxg.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p><strong>Adversarial examples</strong> are a hot research topic in deep learning nowadays. <strong>Subtle, often invisible changes in the data</strong> can push our networks to make <strong>terrible mistakes</strong>. We, as <strong>human beings</strong>, seem to be <strong>way more resilient </strong>to these perturbations in our sensory inputs (though not totally immune).</p><p>There is a certain pattern in our <strong>deep learning systems</strong>. They achieve remarkable things, but they are also <strong>at times delicate and brittle</strong>. Like a rigid tree in the middle of a storm, they look majestic, but may crack at any time without warning. Why is this happening and how can we improve the situation?</p><p>Some clarity is starting to appear through new research arriving from the field of neuroscience. In this article we are going to explore it.</p><p>In his recently published book, <strong>“</strong><a href="https://www.amazon.com/Thousand-Brains-New-Theory-Intelligence/dp/1541675819"><strong>A Thousand Brains: A New Theory of Intelligence</strong></a><strong>”</strong>, a masterpiece that I really enjoyed, <strong>scientist and entrepreneur Jeff Hawkins</strong> dissects the latest research performed by <a href="https://numenta.com/">his <strong>team</strong></a><strong> </strong>on our neocortex, the part of the brain that occupies 70% of its volume and is responsible for our advanced intelligence. (the other 30% is occupied by the older, more primitive part of the brain).</p><p>In a fascinating journey, Jeff Hawkins takes us deep into the epicenter of our intelligence. He shares that:</p><ul><li>The <strong>circuits </strong>in the neocortex are <strong>really complex</strong>. In just <strong>one square millimeter</strong> we have around <strong>one hundred thousand neurons</strong>, <strong>several hundred thousand million connections </strong>(synapses) and <strong>kilometers of axons</strong> and dendrites.</li><li>The <strong>neocortex looks very similar all around</strong>. The variations between areas are small.</li><li><strong>All parts of the neocortex</strong> seem to be <strong>connected to the generation of movement</strong>, to motor tasks. In every part of the neocortex, scientists find cells that connect with areas of the old brain related to movement.</li></ul><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*-kCqmBNZpLltyzhBbLNV0g.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><h3>One circuit to rule them all</h3><p><strong>Vernon Mountcastle </strong>was a leading american neurophysiologist and professor emeritus of neuroscience at Johns Hopkins university. He was the <strong>discoverer of the columnar structure</strong> of the cortex. And he proposed that through evolution, our <strong>neocortex </strong>became <strong>larger by basically copying over and over</strong> the very <strong>same thing, the same basic circuit.</strong></p><p>When I read about Mountcastle’s idea in Jeff’s book, I got reminded of a fascinating talk by the great scientist <strong>Robert Sapolsky</strong>. Answering a question about what separates us from Chimps (<a href="https://www.youtube.com/watch?v=AzDLkPFjev4">https://www.youtube.com/watch?v=AzDLkPFjev4</a>), Sapolsky explains that about half of the difference in gene expression between chimps and humans has to do with genes that are coding for olfactory receptors, other differences are related to the size of the pelvic arch, the amount of body hair, immune system recognition capabilities, some aspects of reproductive isolation, etc; those and others account for almost all the genetic differences between chimps and humans. So then, <strong>where are the differences in the genes that are relevant to the human brain</strong>? Sapolsky explains that <strong>there is hardly any</strong>, and<strong> the few identified</strong> are in <strong>genes having to do with the number of rounds of cell division during fetal brain development.</strong> Basically: <strong>we have 3 times more neurons than chimps. </strong>And this difference in scale seems to be key to our advanced intelligence.</p><p>This fits nicely with Mountcastle’s idea of a single circuit that gets replicated many, many times (volume matters, but is volume enough to push today’s deep learning systems towards AGI? let’s keep exploring below).</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*HFkj2g1fTz2G6vgoKbHMmw.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p>That <strong>all parts of our neocortex</strong> are working <strong>based on a similar principle</strong>, on the same basic circuit, fits the flexibility our brain has demonstrated in different scenarios. And if volume matters, <strong>will this mean that GPT-11 may get us closer to AGI</strong>?</p><p>Unfortunately, <strong>it is not that simple</strong>. Because there is a massive elephant in the room that Jeff illuminates in his book and theory. One that we have been kind of ignoring for far too long.</p><h3>150000 columns</h3><p>Before we go to visit the elephant in the room, let’s establish the context. According to the scientists, we have around <strong>150000 cortical columns</strong> in our neocortex. Jeff tells us that we could think of these columns as if they were thin spaghetti. So, imagine <strong>150000 thin spaghetti next to each other.</strong> That’s <strong>your neocortex</strong> metaphorically speaking<strong>.</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*7LzejRzs6BZoYoLt0LW4Qw.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*_zxZCzy5gor_GFVtHI51CQ.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*hD8w9MiKahESutmJragBGA.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p>What is going on within these cortical columns? Over the last years scientists have come to realize that <strong>the brain is a predictive machine</strong>. It <strong>produces a model of the world</strong> and continuously <strong>predicts what will happen next</strong>.</p><p><strong>When </strong>our brain’s <strong>predictions are not correct</strong>, we realize that something is not right and our <strong>brain updates its model</strong> of the world. As time goes on, our model of the world gets richer and more sophisticated.</p><p>So in a way <strong>we indeed live in a simulation</strong>. For <strong>what we perceive is </strong>really<strong> the model that the brain constructs</strong> rather than the “reality” out there. This explains phantom limbs and other similar scenarios.</p><p>Jeff Hawkins points out that <strong>our brain learns a model</strong> of the world by <strong>paying attention to how the inputs it receives change as we move</strong> (or as those inputs move). And that takes us to the elephant in the room.</p><h3>The elephant in the room</h3><p>The world is changing constantly. <strong>Everything moves. </strong>And it makes sense that <strong>as things move</strong> and change, <strong>our brain keeps updating our model</strong> of the world (our many, many models as we will soon see).</p><p>And just as <strong>attention mechanisms have revolutionized the deep learning field</strong> in recent years, <strong>so is attention key in how our brain is learning</strong> these models.</p><p>But if our neocortex is <strong>making a very large amount of predictions</strong> constantly, and <strong>adapting to any misalignments</strong> between its models and what it perceives, <strong>why don’t we notice all of those predictions </strong>and instead we perceive one continuous reality? Let’s get there, step by step.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*uNG3uYM4trdQgd5pQTYLCw.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p>Through their latest research, Jeff and his team reach some fascinating insights:</p><ul><li><strong>Each of our cortical columns </strong>(150000 roughly) <strong>is learning models of the world</strong>, of objects, of concepts, of anything you could imagine. And just as our <strong>old brain</strong> has <strong><em>place cells</em></strong><em> </em>and <strong><em>grid cells </em></strong>to make models of our surroundings, the <strong>neocortex</strong>, they propose, <strong>has equivalent cells that allow the brain to make models of objects, concepts</strong>, etc. The cortical columns <strong>are using</strong> what Jeff calls <strong>reference frames</strong>, which are like grids of a number of dimensions that <strong>help the brain organize any kind of knowledge</strong>.</li><li>Jeff tells us that <strong>thinking is a form of movement</strong>. Thinking <strong>happens when we change location</strong> <strong>within these reference frames</strong>. So what you are thinking right now, what is right now in your head, depends on where your cortical columns are at the moment within those different reference frames. And your thoughts keep evolving as your brain navigates those structures.</li></ul><p>Notice that the concept of <strong>movement</strong> is beginning to appear everywhere. <strong>Movement </strong>and the <strong>dynamic nature of systems </strong>are the elephant in the room. And we will soon discuss <strong>how that connects with the issue of adversarial examples and the limitations of much of what’s going on in today’s</strong> deep learning.</p><p>So, it is all about <strong>reference frames </strong>or <strong>maps</strong>, maps of physical spaces, maps of concepts, maps of anything. Jeff tells us that in the same way that reference frames in the old brain are learning maps of different environments, <strong>reference frames in the neocortex</strong> are <strong>learning maps of objects </strong>(in the case of what they call “<strong><em>what” </em>columns</strong>), or the <strong>space around the body</strong> (in the case of “<strong><em>where” </em>columns</strong>), or <strong>maps of concepts</strong> within the non-sensory columns.</p><p>I love the analogy that Jeff uses regarding how<strong> in order to be an expert in any domain </strong>we need to <strong>find a good way to organize our knowledge about that domain</strong>, we need to create internally a great reference frame or map of that domain. Think of the deep and complex reference frames that, for example, Leonardo Da Vinci or Einstein had in order to excel as they did within their respective areas of expertise.</p><p>All right, so <strong>each of our 150000 cortical columns</strong> is<strong> learning a predictive model of the world </strong>as it pays attention to how the inputs change throughout time. And <strong>each of these columns</strong> <strong>learns models of a large number </strong>of elements, objects, concepts, etc.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*j8boylaMItD8Bl0mX4lnMw.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p>So our <strong>knowledge </strong>of anything, an object or a concept, <strong>is distributed across thousands of cortical columns</strong>, across <strong>thousands of complementary models</strong>. This relates to the name of <strong>Jeff’s theory </strong>(<strong>A thousand brains</strong>).</p><p>And all of this connects with <strong>the flexible nature of our brain</strong>. Our neocortex <strong>doesn’t depend on a single column</strong>. Knowledge is <strong>distributed across thousands</strong> of them. So the brain continues working even if an injury damages a set of columns (there are great examples about this in the academic literature).</p><p>The next thing to consider is:<strong> if the brain is creating new predictions</strong> everytime movement happens, <strong>where are these predictions being stored?</strong></p><p>Jeff and his team propose that <strong>spikes that occur at different dendrites in a neuron are predictions</strong> (dendrites are branches in the neuron, and receive inputs through their synapses). <strong>Dendrite spikes put the cell </strong>connected to them <strong>into </strong>what Jeff calls <strong>a predictive state</strong>. Therefore, <strong>predictions happen inside neurons</strong>. And these predictions <strong>change the electrical properties </strong>of the neuron and<strong> make it fire sooner </strong>than it would otherwise, but the <strong>predictions are not sent through the axon to other neurons</strong>, which <strong>explains why we are not aware of most of them</strong>. The question now is: <strong>how do we settle on a specific prediction?</strong></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*09LxsBVPoH6xjrTItv0huA.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><h3>Consensus by voting</h3><p>Our <strong>perception of reality </strong>is the <strong>result of a voting process</strong>. The different <strong>cortical columns reach consensus through voting</strong> and this is what <strong>produces a single perception</strong> that unifies different predictions coming from different parts of the system (that may also be related to a diversity of types of sensory inputs).</p><p><strong>Only some cells need to vote</strong>, the <strong>ones that represent, for example, a specific object</strong> that we are perceiving. And<strong> how do they vote?</strong></p><p><strong>Most of the connections in our cortical columns</strong> are<strong> going up and down</strong> <strong>the different layers</strong> of the neocortex. But <strong>there are exceptions</strong>. Scientists have found that <strong>there are cells that send their axons</strong> (output connections) from <strong>side to side through the neocortex</strong>. Jeff and his team propose that <strong>these cells that have long distance connections are the ones responsible for the voting</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*C-0QGbR278_uWt8KjAgcQQ.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p><strong>As we recognize an object</strong>, our <strong>cortical columns have reached a consensus</strong> on what it is that we are looking at. The <strong>voting cells </strong>(neurons) <strong>in each of our columns make a stable pattern </strong>that represents that object and where the object is located in relation to us.</p><p>And <strong>as long as we keep perceiving the same object,</strong> the<strong> state of those voting neurons doesn’t change</strong> while we keep interacting with that element. Other neurons will change their state as we move or the object moves, but<strong> the voting neurons will remain stable</strong>.</p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*-BemTJsQruCi3dHMi1WCog.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><p><strong>This is why our perception is stable </strong>and <strong>we are not aware of </strong>the flurry of activity related to <strong>the moving predictions </strong>that are taking place. <strong>We are just aware of the final stable patterns that arise from the consensus</strong> reached by the cells that are voting.</p><p>Therefore:</p><ul><li><strong>Movement is key</strong> in <strong>how our brain perceives</strong> the world. It is thanks to movement (our own or the one of the world around us) that <strong>our brain can enrich its internal models </strong>of objects and concepts (movement doesn’t have to be physical, it can be virtual, etc).</li><li>Using <strong>a single principle to process all kinds of inputs</strong> and creating <strong>thousands of predictions and models for each</strong> element we interact with, makes these <strong>models rich and versatile</strong>.</li><li>The <strong>consensus mechanism reached through voting</strong> means that our <strong>perception </strong>of the world <strong>is stable</strong> and at the same time <strong>flexible and resilient.</strong></li></ul><p>It is time to return to the adversarial examples and the status of the deep learning field.</p><h3>How to defeat adversarial examples</h3><p><strong>Human beings</strong> are <strong>not immune to adversarial examples.</strong> Perturbations in our sensory inputs can confuse us and make us interpret things incorrectly. Most of us have experienced a wide range of optical illusions. However, in general, <strong>our perception is consistent and quite resilient, </strong>and certainly way more consistent than the one we find in today’s deep learning systems, where invisible changes can derail completely our results.</p><p>What is behind this resiliency, consistency and flexibility? Whatever it is, it may include some of the following:</p><ul><li>The <strong>models</strong> created by our cortical columns are <strong>based on movement and the creation of reference frames.</strong> As we move around or the world moves around us, our <strong>brain creates thousands of predictions and models of each object or concept</strong>. This provides <strong>flexibility</strong>. We are not putting all our eggs in one basket. Just as when we use ensembling in deep learning, we are <strong>betting on thousands of angles</strong> on the problem, not just one.</li><li>Our perception is based on <strong>multimodality and stable voting dynamics</strong>. The different models created in relation to a specific object (or concept, etc), are using multiple <strong>predictions </strong>that are often <strong>connected to different sensory modalities </strong>(<strong>vision, touch, hearing, gestures, </strong>etc). <strong>Voting </strong>among the cells that are responsible for the final representations <strong>produces stable patterns that are resilient</strong> to changes. A <strong>minimal change </strong>in the object or context <strong>will not derail the stable voting pattern</strong> because such a pattern is based on the combination of thousands of separate predictions, which in turn are based on a combination of many different angles, perspectives and also, often, different sensory modalities. Just as <strong>ensembling </strong>often wins in Kaggle competitions, it is also a kind of ensemble that is taking place in the brain, making our human perception <strong>stable, resilient and flexible </strong>(relatively speaking of course, but especially so in comparison with current deep learning systems).</li></ul><p>So, “<strong>the end” of adversarial examples in deep learning</strong>, and by “end” <strong>I don’t mean absolute end, </strong>just <strong>getting to a level of resiliency, consistency and flexibility similar to the one we have </strong>as human beings, will be possible with a combination of:</p><ul><li><strong>Movement</strong>: physical or virtual. <strong>Deep learning systems need</strong> to be able to <strong>gather different perspectives and angles on the world</strong> by <strong>enriching their internal models </strong>as they move or as the world moves around them. <strong>Robotics and AI will have to merge further</strong>. Beyond robotics, <strong>movement can also be virtual </strong>so this principle <strong>goes beyond the physical</strong>.</li><li><strong>A collection of models:</strong> we have to <strong>go beyond single representations or models.</strong> To become resilient to adversarial examples and other challenges, deep learning needs to <strong>generate a large amount of predictions and models that get updated continuously. </strong>A <strong>voting mechanism </strong>can then <strong>create stable patterns and representations</strong> that will be way more resilient to adversarial perturbations.</li><li><strong>Continuous learning:</strong> The world out there is not waiting. A consequence of the above is that<strong> learning needs to be continuous.</strong> <strong>Deep learning</strong> systems are<strong> too static</strong> nowadays. Continuous learning is an active area of research and its importance will only increase going forwards.</li><li><strong>Reference frames:</strong> We can find much inspiration about how to build our representations and models from the reference frames described by Jeff Hawkins in his book and theory. As Jeff points out, deep learning leaders like <strong>Geoffrey Hinton </strong>have already been working for quite some time in trying to make deep learning models more flexible (see capsule networks). But there is still a long road ahead and it is becoming clear that the latest neuroscience research is reinforcing that direction with new hints; our brain is way more flexible and resilient than our deep learning models and now we are beginning to understand why.</li></ul><p>Researching new ways of <strong>detecting adversarial examples</strong> is an interesting area with much academic activity. What is missing now is to <strong>rethink our deep learning architectures and systems </strong>to<strong> transition from the current static paradigm</strong> <strong>to a dynamic one </strong>based on <strong>multi-modal, multi-model, consensus-based predictive systems </strong>that are <strong>resilient, consistent and flexible</strong>. When we reach that point, we will be able to hide or perturb parts of our systems and still maintain stable predictions.</p><p>As Jeff points out, this <strong>will become more and more crucial</strong> as we try to <strong>apply AI</strong> systems to <strong>scenarios that require a lot of flexibility and resilience.</strong></p><p><strong>Mountcastle’s </strong>ideas, <strong>Sapolsky’s </strong>thoughts and our fascination about the GPT architecture, all of those things indicate <strong>the importance of volume.</strong> <strong>Volume matters.</strong> Having <strong>3 times more neurons</strong>, or <strong>thousands of copies of the same basic circuit</strong>, or hundreds of billions of parameters rather than 1 billion, <strong>all of that matters.</strong></p><p>And that’s good news for the current state of the deep learning field. With projects such as the <strong>GPT </strong>system<strong>,</strong> we are discovering and confirming that fact, that <strong>volume matters</strong>.</p><p><strong>But, </strong>what we are also beginning to realize is that, as much as volume matters, it will <strong>not </strong>be <strong>enough </strong>to take us where we wanna go.</p><p>If you follow the latest conversations about systems like GPT-3 in a range of podcasts and venues, say for example at the <a href="https://www.youtube.com/watch?v=iccd86vOz3w"><strong>machine learning street talk</strong></a><strong> podcast</strong>, you will hear similar conclusions. <strong>GPT-3 is very impressive, but it is also kind of delicate, brittle, and it often feels like a hack. </strong>This has nothing to do with how resilient and flexible human brains are.</p><p><strong>Volume matters. But movement also.</strong> We <strong>cannot escape movement and change just through sheer volume</strong>. The world is like a storm that never stops.</p><p><strong>We are the static tree</strong> that <strong>gets larger</strong> and larger but <strong>keeps breaking </strong>over and over because it <strong>lacks the capacity to move </strong>with the storm.</p><p>Thinking is movement. Movement through reference frames. Movement across thousands of predictions and models unified through consensus mechanisms.</p><p><strong>The way forward is through movement.</strong></p><p><strong><em>Afterword</em></strong>: In his book <strong>“The Master Algorithm”</strong> <strong>Pedro Domingos </strong>writes about different paradigms connected with deep learning: <strong>symbolists, connectionists, evolutionaries, bayesians and analogizers</strong>. It’s clear that the path towards AGI could come through many different routes and combinations of approaches. In regards to Jeff and his team’s work and theory, I am following, as professor <strong>Kenneth Stanley</strong> would say, a <strong>gradient of interestingness </strong>(and the magnitude of this gradient in regards to Jeff’s work is pretty strong). It feels to me that Jeff’s theory and work (alongside all his talented team) could hold inside very interesting and useful <strong>stepping stones that could bring us closer to AGI</strong> (or at a minimum their research could point us towards those stepping stones). So yes, we could get to AGI in many different ways, but so far <strong>the only intelligent system we know that is resilient and flexible enough is the one on top of our shoulders.</strong> So it does make a lot of sense that exploring in depth the latest research coming from neuroscience may point us towards useful stepping stones on the way to AGI.</p><p>And if you want to go deep into the mysteries of the entities that inspire our deep learning A.I systems and make your thoughts possible, check the related article below.</p><p><a href="https://towardsdatascience.com/journey-to-the-center-of-the-neuron-c614bfee3f9">Journey to the center of the neuron</a></p><figure><img alt="" src="https://cdn-images-1.medium.com/max/1024/1*cCugHEjegCwskH4AljEP6w.jpeg" /><figcaption>Painting by the author Javier Ideami@ideami.com</figcaption></figure><img src="https://medium.com/_/stat?event=post.clientViewed&referrerSource=full_rss&postId=d214d222c4cb" width="1" height="1" alt=""><hr><p><a href="https://medium.com/data-science/towards-the-end-of-deep-learning-and-the-beginning-of-agi-d214d222c4cb">Towards the end of deep learning and the beginning of AGI</a> was originally published in <a href="https://medium.com/data-science">TDS Archive</a> on Medium, where people are continuing the conversation by highlighting and responding to this story.</p>]]></content:encoded>
        </item>
        <item>
            <title><![CDATA[Convolutional layer hacking with Python and Numpy]]></title>
            <description><![CDATA[<div class="medium-feed-item"><p class="medium-feed-image"><a href="https://medium.com/data-science/convolutional-layer-hacking-with-python-and-numpy-e5f64812ca0c?source=rss-7f7b5d730c84------2"><img src="https://cdn-images-1.medium.com/max/2600/1*PAQcwAZSdhfbOKtDgZ22rQ.jpeg" width="3000"></a></p><p class="medium-feed-snippet">Create a convolutional layer from scratch in python, hack its weights with custom kernels, and verify that its results match with pytorch</p><p class="medium-feed-link"><a href="https://medium.com/data-science/convolutional-layer-hacking-with-python-and-numpy-e5f64812ca0c?source=rss-7f7b5d730c84------2">Continue reading on TDS Archive »</a></p></div>]]></description>
            <link>https://medium.com/data-science/convolutional-layer-hacking-with-python-and-numpy-e5f64812ca0c?source=rss-7f7b5d730c84------2</link>
            <guid isPermaLink="false">https://medium.com/p/e5f64812ca0c</guid>
            <category><![CDATA[machine-learning]]></category>
            <category><![CDATA[data-science]]></category>
            <category><![CDATA[artificial-intelligence]]></category>
            <category><![CDATA[towards-data-science]]></category>
            <category><![CDATA[deep-learning]]></category>
            <dc:creator><![CDATA[Javier Ideami]]></dc:creator>
            <pubDate>Mon, 15 Mar 2021 12:41:09 GMT</pubDate>
            <atom:updated>2021-03-15T12:41:09.807Z</atom:updated>
        </item>
    </channel>
</rss>