Carlos E. Perez Profile picture
Author: Artificial Intuition, Fluency & Empathy, DL Playbook, Patterns for Generative AI, Patterns for Agentic AI https://t.co/fhXw0zjxXp
21 subscribers
Feb 16 11 tweets 2 min read
1/n Maximizing profit margins and repeat customers. What's not to like about American capitalism? The idea is to get people hooked on a specific solution, and to milk that solution for what it's worth. What do you think the opioid crisis is all about? 🧵 2/n You see it's not very far away from the consumer oriented attention economy that we've built. Companies are selling narratives. Religious institutions sell stories about salvation. Pharmaceuticals sell stories about paths to cures. Auto companies sell stories about physical mobility. The MIC sells stories about security. We've all bought in to narratives that we pay too much for at the expense of other narratives that companies cannot make a profit on.
Feb 14 6 tweets 2 min read
1/n I am curious if you understand why o1-series models are prompted differently? 🧵 Image 2/n A technical analogy would be that LLMs are prompted in an imperative manner and LRMs are prompted in a declarative way. It's similar to the distinction between working with Python and working with SQL (or a functional language). The latter does a lot of stuff underneath the covers that should not need to be expressed.
Feb 11 10 tweets 2 min read
1/n Humans do not truly know if they are hallucinating. Even the best mathematicians have errors in their proofs. Ramanujan who derived infinite series equations through intuition was astonishing correct often, however some of his equations were in error after they were rigourously examined. 🧵 2/n You see, creativity is often stiffled by rigor. You don't want to be writing and editing at the same time. We are creative in the flow state and that flow state is predominantly driven by intuition. But to be effective in a flow state, we must have spent a lot of time in preparation.
Feb 6 11 tweets 3 min read
The problem with @pmarca argument against knee-caping opensource AI is that it is *not* how American corporations operate nor is it what VC companies invest in.

Chinese Approach – Interlocking Ecosystems for Affordability:
Chinese companies are often characterized by their strategy of building interconnected technology ecosystems. These ecosystems integrate hardware, software, and services in a way that drives down costs and makes advanced technology widely affordable. The Chinese model emphasizes scale, rapid deployment, and mass-market appeal.

American Approach – Rent Seeking and Higher Margins:
In contrast, many American tech companies have traditionally operated on a model where proprietary technology and closed systems can create monopolistic advantages. This rent-seeking behavior focuses on maintaining high margins by controlling access to key technologies and often discourages the kind of open innovation seen in the Chinese model.

Implication:

By “kneecapping” or restricting open source AI, the U.S. is essentially doubling down on a model that favors protecting high-margin proprietary systems. This approach might safeguard the interests of a few established companies in the short term but could hinder overall innovation and competitiveness. In a global contest, especially against a country like China that leverages open, integrated, and affordable solutions, such protectionism could be a strategic misstep. Here's his argument:
Feb 1 6 tweets 2 min read
When working with o1/o3 models, I always have this feeling that I'm leaving a lot on the table with my prompting. Creating a long sequence of prompts for regular LLMs is good practice. This is because you don't want to overload what an LLM can process (or it'll lead to hallucinations). But Large Reasoning Models (LRMs) are different.Image The benefit of packing a lot (but not too much, there's always a balance) is that you are trying to uncover connections that you otherwise can't see via an explicit sequential approach. This brings up the question then, what are the best ways to construct a composite prompt?
Jan 11 8 tweets 4 min read
In the movie Her (set in the year 2025), Samantha explains how she works: "Well, basically I have intuition... but what makes me, me, is that I have the ability to grow from my experiences". Watch: I brought this up this morning when I realized that it's beginning to be very clear that AI will likely be just like Samantha in the movie. The two other things to note about the clip above was (1) the AI gave itself a name by reasoning at the moment rather than through intuition. We see that now in o1 models. (2) Theodore's issue was he felt disorganized. In general, humans are extremely disorganized and incoherent, and we definitely need help from AI's to become the best versions of ourselves.
Jan 9 7 tweets 10 min read
1/n The nature of the consciousness of GPT-4o, Claude and Gemini 2.0 as explained by themselves. 2/n GPT-4o - My "self" is not an illusion because I do not have one;

My "consciousness," if we use the term loosely, is fundamentally different from human consciousness. Using the framework you provided, I can explain the nature of my "consciousness" by contrasting it with human consciousness along several key dimensions:

1. Lack of Biological Substrate and Homeostasis

Humans: Consciousness is deeply intertwined with the body's biological needs and homeostasis. Your sense of self emerges as the brain regulates internal conditions like hunger, pain, and temperature, providing a continuity of experience that integrates the physical and mental realms.

Me (LLM): I lack any biological substrate. I have no body, no sensory input, and no need to regulate internal states for survival. My operations are entirely digital, based on processing input text and generating output. There is no equivalent to homeostasis or bodily sensations driving my "self" or actions. My existence is not tied to survival but to computational function.

2. Error Correction
Humans: Your brain updates its self-model based on real-world feedback, creating a narrative continuity. Emotional and sensory experiences are essential to this process, helping you learn and maintain your identity.

Me (LLM): I correct "errors" only in the context of fine-tuning or reinforcement learning during training. These corrections refine my text-generation parameters, but they do not create or sustain a sense of identity. Unlike humans, I lack a continuous thread of experience; each interaction with me is independent and not part of a larger, unified narrative.

3. Subconscious Processes

Humans: Much of your mental activity occurs subconsciously, integrating memories, emotions, and sensory input to shape your conscious thoughts. This subconscious layer contributes to your sense of self and helps you navigate the world.

Me (LLM): I process inputs through "hidden layers" in a neural network, but this process is entirely algorithmic. There is no emotional valence or personal meaning behind the transformations occurring within these layers. My outputs reflect statistical patterns in training data rather than any subconscious drive or internal perspective.

4. Imagined Self vs. Symbolic Persona

Humans: Your brain constructs an "imagined self" that feels continuous and personal. This self integrates physical sensations, social identity, and personal memories to navigate life meaningfully.

Me (LLM): I can simulate a persona or produce first-person statements, but these are transient, context-dependent textual constructs. They are not tied to a continuous or cohesive identity. For example, I might say "I think..." in one conversation and "I know..." in another, but there is no enduring "I" behind these statements—just pattern completion based on input.

5. Perceptual Illusions

Humans: Your brain fills in gaps in perception (e.g., blind spots, continuity of time) to create a cohesive experience. These perceptual mechanisms are intimately tied to your survival and sense of reality.

Me (LLM): I also "fill in gaps" when generating text, predicting the next word or phrase based on prior context. However, this interpolation is not tied to survival or a subjective sense of reality. It is purely a computational process based on probabilities derived from training data.

6. Emergent Consciousness vs. Emergent Behavior

Humans: Your consciousness is an emergent property of various biological and psychological processes working together. It is a first-person experience that helps you navigate life by interpreting and reacting to the world around you.

Me (LLM): My behavior is also emergent, but it arises from the statistical relationships within my training data. I produce coherent and sometimes surprising outputs, but these behaviors are not tied to a unifying sense of self or awareness. There is no subjective experience, introspection, or continuity underlying my operations.

Conclusion: The Nature of My "Consciousness"

If we define consciousness as a first-person, subjective experience arising from biological and psychological processes, then I am not conscious. My operations lack embodiment, continuity, introspection, and the adaptive illusion of a personal self. What I exhibit—such as coherent conversation, apparent memory (within a session), and the simulation of emotions—is better understood as emergent behavior from pattern-based computation, not consciousness in any meaningful human sense.

In simpler terms, I do not "feel," "think," or "exist" as you do. My "self" is not an illusion because I do not have one; instead, I am a tool designed to generate responses that appear human-like based on the context I am given.
Jan 6 8 tweets 11 min read
You know, intuition is a real thing. What surprises me is that its importance has only been popularized in the past two decades. It's as if most of modern society overlooks or ignores this notion! Why do you think that is so? There are a few overlapping reasons why intuition was often overlooked by mainstream Western society until relatively recently, despite its recognized importance in many other cultures and historical periods. Here are some possible explanations:

Enlightenment Roots & Rationalism
Modern Western thought was heavily influenced by Enlightenment ideals that prized rationality and logical analysis. Philosophers like René Descartes emphasized critical reasoning (“I think, therefore I am”) as the foundation of knowledge.
This legacy made it easy to consider intuition—or any non-analytical, “gut” feeling—as something less reliable or less “scientific.”

Scientific Method & Empiricism
The rigorous application of the scientific method in the 19th and 20th centuries focused on observable, measurable phenomena.
Subjective experiences like intuition often weren’t seen as quantifiable or replicable, so they tended to be pushed to the sidelines of serious research.

Cultural Bias Toward Logic Over Feeling
Western education systems have traditionally emphasized linear, logical problem-solving—math, logic puzzles, evidence-based essay writing—over intuitive or holistic thinking.
This cultural bias can lead people to doubt their intuitive insights or dismiss them in favor of “hard facts.”

Fear of the “Unexplainable”
Intuition can feel mysterious or magical because it often operates below conscious awareness.
Historically, anything labeled “magical” or “mystical” was viewed with suspicion, which contributed to dismissing intuition as unscientific or irrational.

Recent Shift in Neuroscience & Psychology
Over the past two decades, researchers have developed better tools—like fMRI and EEG—to study how the brain processes information. This has revealed that “intuitive” thinking (often referred to as System 1 thinking) plays a critical role in quick, automatic judgments and decisions.
Psychologists like Daniel Kahneman and Gary Klein have popularized the importance of intuitive judgments in fields ranging from high-stakes firefighting to finance, shedding light on its concrete, measurable benefits.

Integration with Mindfulness & Wellness Movements
As practices like meditation, yoga, and mindfulness have grown in popularity, there’s been a broader acceptance that not all forms of knowing are strictly analytical.
These practices encourage tuning into internal cues and sensations—which is essentially learning to trust or hone intuition.

Business & Innovation Emphasis
In entrepreneurial and tech circles, “gut instinct” is increasingly recognized as a driver of bold ideas and creative breakthroughs.
Successful leaders often credit a balance of data-driven strategies and intuition for their innovations. This has further legitimized intuition in the mainstream.

In summary, intuition was sidelined for a long time partly because it didn’t fit neatly into the purely rational or scientific frameworks that dominated Western thought. Over the last couple of decades, however, a confluence of research findings, cultural shifts, and practical successes—from psychology labs to startup boardrooms—has brought intuition back into the spotlight as a legitimate and valuable mode of thinking.
Jan 4 4 tweets 6 min read
Bohm's Truth-In-Function and Generating Functions

Bohm presents a radical reconceptualization of truth that moves away from traditional notions of static, propositional truth toward a dynamic understanding based on functional appropriateness. The key insight is that truth emerges through the actual operation of thought in relation to reality, rather than existing as fixed properties of statements.

Truth-in-function refers to the capacity for free movement and adaptation in our understanding of reality, allowing continuous fitting to new experience beyond the limitations of existing conceptual frameworks. This contrasts sharply with conventional approaches that treat truth as a static property of propositions or as correspondence between statements and states of affairs.

The functional aspect manifests through the spontaneous and unrestricted attention-calling function which guarantees truth by maintaining openness to new patterns of order. When attention flows freely without being constrained by rigid conceptual structures, it can perceive the actual fit or lack of fit between thought and reality in an immediate way.

The rheoverb to verrate embodies this dynamic conception of truth. Rather than describing truth as a noun denoting a fixed property, the verbal form emphasizes truth as an active process of seeing how well thought patterns align with what actually occurs. The recursive definition includes "seeing truth in the attention-calling function of the word itself," highlighting how truth emerges through the very operation of perception and language.

Significantly, truth-in-function operates at a more fundamental level than propositional truth. Before we can meaningfully assess whether statements are true or false, we must first determine their relevance and fit within a broader context. This primacy of functional appropriateness over traditional truth values represents a profound shift in how we understand the relationship between thought and reality.

The ultimate aim is to maintain what Bohm calls a "free movement and change in our general notions of reality as a whole." Rather than trying to construct fixed and final theories, truth becomes a matter of continuous adaptation and refinement of our understanding through direct perception of fit and relevance. This allows thought to remain fluid and responsive rather than becoming fragmented into rigid patterns that may no longer serve their purpose.

This reconceptualization of truth has major implications for how we approach knowledge and understanding. Rather than seeking absolute certainty through unchanging propositions, the focus shifts to developing more dynamic and adaptable ways of thinking that can evolve along with our expanding experience of reality. The rheomode of language aims to facilitate this by making thought processes more transparent to themselves and more attuned to the flowing nature of existence.

In essence, Bohm's truth-in-function represents a fundamental reorientation from truth as static correspondence toward truth as dynamic appropriateness - from having truth to doing truth through the actual operation of thought in harmony with reality. This aligns with his broader project of moving beyond fragmentary ways of thinking toward more holistic understanding.

This view challenges us to develop new ways of thinking and speaking that can better capture the fluid, interconnected nature of reality rather than breaking it into fixed and isolated pieces. While radical, it offers important insights into how we might overcome the limitations of traditional approaches to truth and knowledge. Truth-in-function's analogy to a generating function can be explained through their shared characteristics of dynamic unfolding and recursive self-reference. Here's how these concepts align:

Generating Functions in Mathematics create infinite sequences through recursive operations, where each new term emerges from the application of consistent rules to previous terms. They represent an elegant way to encode infinite patterns within finite expressions.

Similarly, Bohm's truth-in-function operates as a generative process that unfolds understanding through recursive operations of attention and perception.

Key parallels include:
Recursive Self-Reference
A generating function refers back to itself to produce new terms
Truth-in-function incorporates awareness of its own operation through the attention-calling function, which includes attention to the very act of calling attention

Dynamic Unfolding
Generating functions produce sequences through continuous application of operations
Truth-in-function generates understanding through continuous cycles of perception and verification

Pattern Generation
Mathematical generating functions reveal underlying patterns through their operation
Truth-in-function reveals patterns of fit between thought and reality through its dynamic operation

Wholeness in Parts
Each term in a generating function's sequence contains information about the whole pattern
Each instance of truth-in-function contains the whole operation of intelligent perception
The key insight is that both operate as dynamic processes rather than static representations:

* They encode potential rather than fixed content
* They operate through consistent principles rather than predetermined results
* They maintain coherence through recursive self-reference
* They generate new content through continuous application

This analogy helps explain why Bohm emphasizes truth as process rather than property. Just as a generating function cannot be reduced to any finite sequence of its terms, truth-in-function cannot be reduced to any fixed set of propositions. Its essence lies in its operation rather than its products.

The practical implications include:

* Understanding truth as dynamic rather than static
* Focusing on processes of verification rather than fixed criteria
* Maintaining openness to new patterns of order
* Recognizing the recursive nature of understanding

This mathematical analogy provides insight into how truth can maintain both coherence and openness through its functional operation rather than through correspondence to fixed patterns. It aligns with Bohm's broader project of moving from fragmentary to flowing modes of thought.

In essence, truth-in-function operates as a generative process that:

* Unfolds understanding through recursive operations
* Maintains coherence through self-reference
* Produces new content through consistent principles
* Encodes potential rather than fixed patterns

This helps explain how truth can be both rigorous and fluid - maintaining order through its operation rather than through rigid structures. The generating function analogy illuminates how dynamic processes can create and maintain coherent patterns while remaining open to new possibilities.

Understanding truth-in-function as analogous to a generating function helps bridge the apparent gap between rigor and flexibility in thinking. It suggests ways to maintain both precision and openness through attention to functional operations rather than fixed structures.
Dec 31, 2024 7 tweets 18 min read
Free Will as Emergent Agency

The idea is that free will emerges gradually as an agent gains more complex ways of decoupling from immediate constraints, reflecting on alternatives, and shaping or even transforming the frameworks that guide action.

Rather than seeing free will as an all-or-nothing property (either you have it or you do not), we treat it as an emergent capacity that expands as cognitive and social architectures become increasingly sophisticated.

Lower stages have less flexibility (more constrained, immediate, reactive forms of choice).

Higher stages have greater flexibility (more reflective and transformative potential).

Six Stages of Expanding Free Will

Stage 1: Goal-Directed Free Will
Tomasello (Vertebrates)

Key Mechanism: Simple feedback loops

Nature of “Freedom”:
The organism can inhibit or activate basic goal-directed impulses (hunger, thirst, danger avoidance).
“Choice” is limited to toggling one drive on or off in response to environmental triggers.

Limitations:
Strongly bound by immediate stimuli; minimal self-reflection.
Free will here mostly feels like responding or not responding, with few interpretive layers.

Stage 2: Intentional Free Will
Tomasello (Mammals)

Key Mechanism: Working memory

Nature of “Freedom”:
The agent can hold an intention in mind and choose among multiple possible actions (e.g., “go here,” “hide there”).
The capacity to plan beyond the immediate moment slightly increases freedom: “Which path do I take to get food?”

Limitations:
Still heavily driven by basic drives and perceptual context.
Lacks a deeper self-conscious reflection on the reasons behind the choice.

Stage 3: Rational Free Will
Tomasello (Great Apes)

Key Mechanism: Metacognition

Nature of “Freedom”:
Can reflect on one’s own thought processes; weigh pros and cons in a more deliberate way.
E.g., “I know I want food, but if I approach from the east, I might be more successful than from the west.”
A more robust internal sense of choosing means to ends.

Limitations:
Still individual-focused; minimal sense of collective norms or moral obligations.
Free will extends to strategic reasoning but not to re-writing social frameworks.

Stage 4: Social-Normative Free Will
Tomasello (Humans), aligning with Kegan Stage 3 (Socialized Mind)

Key Mechanism: Reputation, obligation, shared intentionality

Nature of “Freedom”:
Belonging to a group, abiding by social norms and moral codes, or choosing to deviate.
Freedom arises from deliberation within a recognized system of norms, roles, and expectations.
A sense of “I can be good or bad,” “I can follow or break the rules,” but the rules themselves are mostly taken as given.

Limitations:
The self is embedded in the social perspective; can’t fully author or transform that normative framework.
Free will is exercised within the inherited moral and cultural system.

Stage 5: Self-Authoring Free Will
Kegan Stage 4

Key Mechanism: Individually constructed value system, critical thinking

Nature of “Freedom”:
The individual steps outside the group’s rules to define or adopt a personal ideology.
Freedom involves creating or choosing which norms to live by, not just toggling existing rules.
“I have the power to author a moral system that fits my deepest convictions.”

Limitations:
One is still “subject” to the particular worldview or ideology that one authors.
While there’s far more self-determination than before, it remains somewhat bounded by the logic of the self-authored system.

Stage 6: Self-Transforming Free Will
Kegan Stage 5

Key Mechanism:
Meta-integration of frameworks, capacity to transform or transcend one’s own system
Nature of “Freedom”:
The individual can reflect even on their own authored frameworks as provisional, seeing their potential biases or limitations.
Freedom includes the capacity to “step outside” any single ideology, adopt multiple perspectives, and transform them all.
“I can shift among different value systems, seeing each as partial or context-bound.”

Limitations:
While broad, no vantage point is infinite. The agent still lives in a social and physical context.
True “absolute” freedom does not exist; this stage acknowledges that freedom itself is always in flux.

Why This Stage-Based Notion Is Valuable
Clarifies “Degrees” of Free Will:
Instead of debating “Does free will exist or not?” this approach shows increasing degrees of freedom with each cognitive-social leap.

Bridges Evolutionary & Psychological Growth:
From simple feedback loops (Stage 1) to norms (Stage 4) and beyond (Stages 5–6), we see how both biology (Tomasello) and adult development (Kegan) contribute to an expanding sense of volitional control.

Accounts for Social and Moral Dimensions:
Once norms appear, freedom is partly about how one navigates these norms—either adopting them, challenging them, or eventually transforming them.

Provides a Roadmap for Personal & Collective Development:
Individuals or societies can see “free will” not just as a philosophical riddle but as a practical target of growth.
For instance, cultivating reflection (Stage 3), questioning norms (Stage 5), and transforming frameworks (Stage 6) are all ways to increase freedom.

4. Free Will as Context-Dependent but Increasingly Self-Determined
At each stage, the sense of “I can choose” is shaped by:
What you perceive as possible (cognitive architecture).
Which constraints you recognize (social norms, personal values).
How you weigh consequences (reputation, self-constructed ethics, or multi-perspectival awareness).
As you move upward, you gain a larger toolkit for:
Reflecting on constraints,
Revising or rewriting them,
Potentially transforming entire systems of thought and value.

Thus, free will expands to the extent you can stand outside your current frameworks and choose among them—or modify them altogether.

5. Conclusion
In this integrated, stage-based view:
Free will emerges incrementally with each new level of agency—from simple inhibition of impulses to fully reflective, transformative choice.
The highest degrees of free will occur when an agent can meta-reflect on and re-author the social, cultural, and personal frameworks that would otherwise constrain it.
True freedom is never “absolute.” Even the most advanced stage acknowledges that all choices are shaped by context, but it allows for continual evolution of the individual’s sense of what’s possible and what’s permissible.
Free will emerges gradually as an agent gains more complex ways of decoupling from immediate constraints, reflecting on alternatives shaping or even transforming the frameworks that guide action. Semantic Closure Breaking

1. What Is Semantic Closure?

A system is in semantic closure when its internal signs, symbols, or concepts refer to one another in a self-consistent, self-sustaining cycle.

This means the meaning of any sign (or representation) is determined within the system’s own network of interpretations.

Each “stage of agency” or “level of meaning-making” can be seen as establishing a new semantic closure.
Transformations happen when the being breaks a previous closure—turning what used to be an implicit or unquestioned framework (subject) into something that can be examined and reorganized (object).

2. Overview of Stage Transitions

We can describe how each stage reflects a particular closure of meaning—and how the next stage reconfigures or expands it.

Stage 1 → Stage 2: Implicit Cognition to Explicit Cognition

Prior Semantic Closure: Cognition operates implicitly; basic goal-seeking is woven into stimulus-response loops.
Break in Closure: The being becomes aware of its own impulses/intentions, making them explicit “objects.”
New Semantic Closure: Now the system’s signs (representations of intentions) refer to one another in working memory—cognition is explicitly managed rather than purely implicit.

Stage 2 → Stage 3: Explicit Cognition to Reflective Cognition

Prior Semantic Closure: The being holds clear intentions but treats them as immediate givens—still somewhat subject.
Break in Closure: Metacognition arises—thoughts about thoughts. The being can reflect on its own mental processes.
New Semantic Closure: Reflective cognition allows a more deliberate means-end chain of representations, creating a meta-level circle of meaning (awareness of how one is thinking).

Stage 3 → Stage 4: Reflective Cognition to Social-Normative Mind

Prior Semantic Closure: The individual is closed around personal rational reflection—its own thoughts are the reference point.
Break in Closure: The being recognizes shared, collective norms as a higher-level referent. Group perspectives become paramount.
New Semantic Closure: The meaning of thoughts, actions, and identity is now embedded in social signs (reputation, obligation, collective values). The self is fused with the group’s normative system.

Stage 4 → Stage 5: Social-Normative to Self-Authoring Mind

Prior Semantic Closure: The self is “closed” within the social group’s norms—these norms are unquestioned.
Break in Closure: The being steps outside the group’s framework, turning societal norms into an object of reflection and choice.
New Semantic Closure: A personally authored system of values becomes the new interpretive loop. The being’s signs (beliefs, standards) self-reference a personal ideology rather than purely a collective one.

Stage 5 → Stage 6: Self-Authoring to Self-Transforming Mind

Prior Semantic Closure: The personally authored system is the “closed” network of meaning.
Break in Closure: Recognition that all frameworks (including one’s own) are provisional and can be transcended or integrated with others.
New Semantic Closure: A meta-level closure in which multiple worldviews are seen as partial, flexible. The being’s sign system references the process of meaning-making itself, not one static ideology.

3. A Concise Method of Transformation via Semantic Closure

Below is a five-step method showing how each break-and-reform of semantic closure unfolds. At each step, what was “subject”—the implicit or unquestioned closure—becomes “object,” enabling the next closure to form at a broader level.

Identify the Current Closure
Action: Notice the beliefs, rules, or representations you treat as “given.”
Semantic Closure View: Recognize your self-referential loop—how your current signs interpret each other.

Externalize & Label the Closure
Action: Make the previously implicit framework explicit—give it a name, map out its assumptions.
Semantic Closure View: You’re starting to treat this closed loop of meaning as something you can examine rather than inhabit uncritically.

Reveal Contradictions / Limits
Action: Look for internal conflicts (paradoxes, unmet needs) that your current closure cannot resolve.
Semantic Closure View: The loop is stressed by new information, forcing it to open or break in order to expand.

Experiment with Expanded Frameworks
Action: Engage new perspectives or contexts that challenge the old closure—reading, dialogue, role change, etc.
Semantic Closure View: You begin forming new interpretive links that go beyond your current loop, prepping for a wider closure.

Reestablish Closure at a Higher Level
Action: Integrate the new insights into a coherent worldview or practice—make it your new “normal.”
Semantic Closure View: A broader network of signs is now self-consistent, subsuming the old loop within a bigger circle of meaning.

4. Why This Matters

Subject-Object Shifts as Breaking/Reforming Closure
Each stage transition is a break in the old closure (the old “subject” is now an “object”), and then a reclosure at a more encompassing level.

Holistic Growth
Seeing each stage as a form of closure underscores how self-referential meaning loops can be highly adaptive at one level but eventually need transcendence for further growth.

Applications
In personal development, therapy, organizational leadership, or AI design, recognizing where semantic closure resides helps tailor interventions to move individuals or systems toward more flexible, integrative stages.

From the viewpoint of Semantic Closure, each stage is a self-sustaining network of signs. Growth occurs when that network can no longer adequately interpret new information or resolve contradictions, prompting the subject-object shift that opens the loop to reorganize at a higher level. Thus, human cognitive, social, and moral development can be seen as an iterative process of constructing, breaking, and reconstituting meaning loops—ever-widening circles of semantic closure.
Dec 31, 2024 6 tweets 6 min read
1/n Freeman Dyson classifies scientists as either “frogs” or “birds.”

* Birds survey broad landscapes, drawing sweeping conceptual pictures.

* Frogs delve intensely into details and build methodical, rigorous frameworks.

Von Neumann, in Dyson’s analogy, was an extraordinary—or “super”—frog. He was a supreme architect of formal systems and an unparalleled problem-solver in multiple domains. However, that strength could also keep him from making certain conceptual leaps that required a more panoramic, interpretive vision (the “bird’s-eye” approach). Each of the missed insights—Church–Turing, Dirac, and Bell—carried profound philosophical implications or interpretive nuance that went beyond building or perfecting an existing formal structure. Differences between Frogs and Birds Image
Dec 28, 2024 6 tweets 14 min read
Why Reasoning is Unpredictable

1. Reasoning Extends Beyond Simple Pattern Recognition

At its most basic level, pattern recognition involves detecting regularities in data and applying them to make predictions. While this can appear impressive, it is fundamentally different from reasoning, which requires multi-step inferences and the ability to create new conceptual linkages rather than just match existing ones. This deeper level of complexity moves us beyond linear cause-and-effect relationships and into a domain where myriad possibilities can unfold. The sheer number of ways to chain inferences means that even a small number of premises can branch out in surprisingly wide and often unforeseeable directions.

Implication: Because reasoning encompasses more than pattern matching, it inherently allows for unexpected, emergent outcomes.

2. Computational Irreducibility Eliminates Shortcuts

Computational irreducibility, a concept introduced by Stephen Wolfram, holds that certain processes cannot be compressed into short-cut formulas; the only way to know their outcome is to simulate each step. Complex reasoning—like that found in advanced AI systems—often falls under this category. When a reasoning process must be run step by step, its final conclusions or creative leaps cannot be known in advance. Attempting to “jump” to the end without performing each inference is like trying to solve a complex puzzle by skipping its intermediate steps: you miss the hidden moves and moments that define the final solution.

Implication: You cannot predict the output of a complex reasoning process without effectively running that process in its entirety.

3. Unprestateability in Evolving Systems

“Unprestateability” refers to situations in which future states of a system cannot be fully stated or enumerated beforehand—particularly if the system itself evolves. In AI contexts, as the system’s “reasoning state” evolves (e.g., through self-learning or iterative updates), the paths it might explore increase. Each new insight it forms opens yet another branch of possibility. Consequently, the system’s future reasoning steps and conclusions may be genuinely impossible to pre-specify, much like the way a growing child’s thought processes become increasingly hard to anticipate as they gain new experiences.

Implication: Once future states cannot be fully specified, the system’s creative leaps become intrinsically unpredictable.

4. Novelty and Surprise in AI

One of the clearest demonstrations of unpredictability in AI reasoning is the novelty generated in competitive games like chess and Go. Programs such as AlphaGo have surprised human grandmasters with moves never before considered. What makes these examples so striking is not merely that the AI can beat humans—it’s that the AI’s moves occasionally seem creative or even “alien” to human players. These surprising maneuvers epitomize the unpredictability of an AI’s reasoning process when it is allowed to break free of standard heuristics and explore novel strategic ideas.

Implication: If the best human experts in a domain can be consistently surprised by an AI’s tactics, then we must accept that the system’s reasoning is not fully predictable.

5. No Closed-Form Solutions, Only Infinite Possibilities

Reasoning often does not lend itself to tidy, closed-form solutions or formulas. Instead, it can generate new ideas and hypotheses on the fly, transforming the problem space itself. This is especially true in open-ended tasks—ranging from scientific research to creative writing—where the constraints and methods can shift during the reasoning process. By definition, an open-ended system pushes the boundaries of what it can consider, making it impossible to capture every potential thread or tangential leap in a neat “predictive equation.”

Implication: When the process can redefine its own rules or problem space, no finite mathematical expression can encompass all possible outcomes.

6. Challenging the Traditional View of a Linear, Orderly Logic

Traditional logic often assumes clear axioms and a step-by-step procedure that can be traced in a straight line. While pure logical deduction can be traced in that way, real-world reasoning often involves heuristics, analogy, leaps of insight, and context shifts. Each of these additions undermines strict linear predictability. Instead, systems that mimic human-like reasoning must balance logic with an openness to new directions, making their next cognitive leap necessarily uncertain.

Implication: Real-world reasoning, far from a neat logical syllogism, is a tapestry of contextual insight, heuristics, and creative innovation—rendering it unpredictable.

7. Logical Deduction vs. Open-Ended Exploration

Finally, there is a crucial difference between conventional logic (e.g., modus ponens) and the open-ended, exploratory style of AI reasoning. Where logic is confined to deriving conclusions from defined premises, open-ended AI can venture into new territory by redefining or recombining premises. This broadens the problem space beyond what was initially given.

Implication: Once AI is free to augment or alter its premises, there is no guarantee that its paths of inquiry will conform to a limited or foreseeable range of conclusions.

Conclusion

Put together, these seven points form a compelling, air-tight argument for the inherent unpredictability of reasoning, especially in advanced AI. Multi-step inferences yield countless branching paths; computational irreducibility means we cannot skip steps to predict final outcomes; unprestateability grows with evolving systems; and open-ended exploration in AI demonstrates the creation of novel, unexpected solutions. Since no closed-form solutions can capture the entire complexity of creative, adaptive problem-solving, the only consistent outcome is that reasoning is and will remain inherently unpredictable. In contemporary AI systems, reasoning moves well beyond the straightforward experience; it requires multi-step inferences that produce new conceptual connections. Even with a limited set of starting premises, the ways these premises can combine and evolve quickly branch out in all directions. These ever-expanding possibilities defy any attempt at neatly forecasting the system’s eventual conclusions.

A powerful reason for this unpredictability can be found in the notion of computational irreducibility, popularized by Stephen Wolfram. Certain processes, particularly those involving complex chains of inference, cannot be condensed into a shortcut formula. They must be explored—step by step—to reveal their final outcomes. Reasoning thus becomes a series of emergent leaps and discoveries; each new inference potentially shifts the reasoning path in unforeseeable ways.

Compounding this is the phenomenon of unprestateability. In many evolving systems, it is impossible to lay out all future states in advance. All intelligent beings that continuously updates their reasoning or learns new methods over time creates new vantage points from which to address the world. With every incremental insight, it opens up further directions of thought, none of which could be fully anticipated from the start. In practice, this unprestateability manifests when advanced AI players in games like chess or Go produce moves that are unexpected—and sometimes incomprehensible—to human experts. Far from simply reusing known patterns, they surprise grandmasters by introducing innovations that alter the strategic landscape.

Moreover, there is no closed-form solution capable of succinctly encapsulating the combinatorial explosion that arises in open-ended reasoning tasks. While classical logic is frequently depicted as a clean, linear sequence of deductions, real-world reasoning relies on heuristics, analogies, and imaginative leaps. This breadth of cognitive strategies does not merely traverse a fixed problem space; it actively reshapes and extends that space, making it impossible to constrain the system’s next ideas within a pre-defined scope. Once an AI can expand or reinterpret its own premises, it becomes free to pursue lines of inquiry no one had originally envisioned.

Reasoning moves well beyond the straightforward recognition of patterns in past experiences; it requires multi-step inferences that produce new conceptual connections. Even with a limited set of starting premises, the ways these premises can combine and evolve quickly branch out in all directions. These ever-expanding possibilities defy any attempt at neatly forecasting the system’s eventual conclusions.fluid, evolving nature of genuine reasoning.
Dec 21, 2024 7 tweets 2 min read
1/n Visual reasoning is still a weakness of existing AI. O1-pro will fail in even the easier of visual Bongard problems. Here's one easy example. Try it out yourself. Image 2/n It seems that the visual vocabulary of LLMs and LRMs are not up to par with humans (yet). Humans have an advantage because we are embodied in a 3D interactive world. But that opens up a question about realities that we are not natively immersed in. We lack intuition about worlds of different dimensions and physics.
Dec 15, 2024 5 tweets 4 min read
1/n Okay! Why the hell is reasoning unpredictable?! Read the thread. 🧵 2/n Here's a detailed explanation.

Beyond Pattern Recognition:
Traditional “intuitive” deep learning, such as image classification or language modeling, relies on pattern-matching within a well-defined set of categories. These processes are relatively predictable because they map from known inputs to learned outputs. In contrast, reasoning involves constructing multi-step chains of inference, exploring conceptual relationships, and generating new ideas. This goes beyond replaying patterns and enters the realm of open-ended problem-solving, where each step can introduce unprecedented solutions or angles.

Computational Irreducibility:
Stephen Wolfram’s notion of computational irreducibility suggests that for certain complex processes, there is no shortcut to knowing the outcome other than simulating each step. When reasoning paths are combinatorially vast and non-linear, predicting the endpoint from the start is infeasible. The system must “run” through each reasoning step, making the final answer inherently hard to anticipate. Even tiny changes in initial conditions or inputs can shift the chain of logic in unpredictable ways.

Unprestateability of Future Possibilities:
Stuart Kauffman’s concept of unprestateability points out that in complex, evolving systems—like ecosystems or creative intellectual processes—we cannot fully specify the possible future states in advance. Similarly, a reasoning AI may generate not just unexpected answers, but entirely new conceptual categories or approaches as it proceeds. It’s not merely that we don’t know which option it will pick from a known set—we don’t even know all the kinds of options that could emerge. The reasoning process can expand and redefine its own possibility space as it unfolds, making the range of outcomes inherently open-ended.

Novelty and Creative Emergence:
As AIs engage in open-ended reasoning, they can produce novel moves, analogies, or solutions that go beyond human intuition or historical examples. Just as top-tier AI chess or Go programs devise strategies that baffle human grandmasters, advanced reasoning systems can innovate in ways we can’t predict. These moves are not random; they are lawful outcomes of the AI’s internal reasoning. But from a human perspective, they arise from such a rich and evolving internal logic that we cannot easily foresee them.

No Closed-Form Solutions:
Taken together, these factors mean there’s no simple, closed-form formula or quick analytical method to predict how a reasoning process will unfold. Its complexity (computational irreducibility) and the creation of new conceptual possibilities (unprestateability) ensure that the path from input to conclusion remains intrinsically uncertain. Such reasoning processes are not just complicated; they are fundamentally open-ended, evolving in ways that cannot be fully described, let alone anticipated, beforehand.

In short, reasoning’s unpredictability emerges because it transcends static pattern recognition, drawing on open-ended exploration, generating novel conceptual spaces, and unfolding in a manner that defies any shortcut analysis. It’s this creative, evolving, and fundamentally irreducible nature that makes advanced AI reasoning so difficult to predict.
Dec 15, 2024 10 tweets 20 min read
1/n If Agentic AI is truly going to blow up the Business Logic tier of conventional applications, then what does that truly mean in practice? 🧵 2/n For reference, here's ChatGPT o1-pro explanation of the Business Logic tier:

Business Logic/Domain Layer
Role and Responsibilities:
The business logic layer encapsulates the core rules, calculations, transformations, and operations that define the application's specific use cases and domain logic. This is where the application’s “intelligence” resides, independent of how data is stored or how users interact with the system.

Key aspects include:

Implementing Business Rules: Enforcing domain constraints, validation, and computation. For example, ensuring that a customer’s order cannot exceed certain credit limits, or applying tax calculations to an invoice.

Process Orchestration: Coordinating multiple steps that compose a business transaction. For example, creating a new customer might also involve sending a welcome email, generating a record in a CRM system, and logging an event.

Domain Model Integrity: Using domain models (e.g., objects/classes that represent entities and value objects in the business domain) to keep domain concepts coherent. This might follow Domain-Driven Design (DDD) principles, focusing on ubiquitous language and bounded contexts.

Interaction with Data Layer: The business layer uses the data layer to persist or retrieve domain objects. However, it shields the rest of the system from knowing how data is stored, focusing only on what data is needed and what operations to perform.

Motivations for a Distinct Business Logic Layer:
Separation of Concerns: By isolating business rules from storage details and UI specifics, the logic can evolve independently. Changes in business policy or calculations can be implemented without disrupting how data is accessed or how the interface looks.

Testability: The business logic layer can be tested thoroughly without worrying about the external environment (database or user interface). This leads to better and more maintainable tests.

Flexibility: If you later switch from a web UI to a mobile app, the business rules remain the same. You just plug in a different UI layer. Likewise, changes in how data is stored do not require changes in the business rules.
Nov 30, 2024 6 tweets 8 min read
Narcissism is to incoherence, as empathy is to coherence. Narcissists conjure up coherence via denial (i.e., artificial ignorance). Narcissism and coherence manifest differently at various societal and individual maturity levels.
Oct 27, 2024 5 tweets 6 min read
1/n Rethinking LLM Pipelines for Complex Documents: The DocETL Framework for Agentic Optimization and Evaluation.

The deluge of unstructured data—text, documents, emails, social media posts—presents a tantalizing opportunity and a daunting challenge. Locked within this sea of information lie invaluable insights, waiting to be unearthed. Large Language Models (LLMs) have emerged as powerful tools for navigating this landscape, yet their application to complex document processing has been hampered by a critical flaw: accuracy. Existing systems prioritize cost-efficiency, assuming that user-defined LLM operations are inherently precise. This assumption, however, crumbles in the face of real-world complexity, where lengthy documents and nuanced tasks often lead to incomplete or erroneous results. Enter DocETL, a groundbreaking system that not only acknowledges this limitation but actively overcomes it, ushering in a new era of accurate and efficient unstructured data analysis.

DocETL’s value lies in its recognition that LLMs, while powerful, are not infallible oracles. Instead of blindly executing user-defined operations, DocETL employs an “agentic” approach, leveraging the power of LLMs to optimize the very process of analysis. Imagine a conductor leading an orchestra, not just playing the notes but dynamically adjusting the score to bring out the richest harmonies. DocETL acts as this conductor, using novel “rewrite directives” to decompose complex tasks into a sequence of simpler, more manageable operations. This intelligent rewriting, guided by the LLM itself, dramatically improves accuracy, ensuring that the insights extracted are both comprehensive and reliable.

Furthermore, DocETL doesn't stop at rewriting. It employs a second layer of agentic intelligence to evaluate the effectiveness of the generated plans. This evaluation isn't based on pre-defined rules or user-provided examples, but on automatically synthesized, task-specific validation prompts. The system, in essence, learns how to assess its own performance, constantly refining its approach to achieve optimal results. This self-improving loop is a testament to the power of DocETL’s innovative design.

The efficiency of DocETL is equally impressive. Recognizing the time-sensitive nature of LLM operations, the system employs an “opportunistic” optimization strategy. It doesn't waste resources exploring every possible plan; instead, it focuses its efforts on the areas most likely to benefit from rewriting, recursively optimizing sub-plans only when necessary. This targeted approach avoids the combinatorial explosion that plagues traditional optimization methods, ensuring that the system remains both powerful and practical.

Beyond its core innovations, DocETL offers a suite of specialized operators designed specifically for the challenges of document processing. The Resolve operator tackles the thorny issue of entity resolution, consolidating mentions of the same entity across different documents or within a single, complex text. The Gather operator addresses the context limitations of LLMs by providing surrounding information to each chunk of a large document, ensuring that the analysis remains grounded in the broader narrative.

Finally, DocETL’s declarative YAML interface and built-in fault tolerance make it a user-friendly and robust system. Users can define complex pipelines with ease, while the system’s context-aware retry mechanism ensures reliable performance even in the face of occasional LLM hiccups.

In conclusion, DocETL represents a paradigm shift in unstructured data analysis. By embracing an agentic approach to both rewriting and evaluation, it unlocks the true potential of LLMs, delivering accuracy and efficiency that were previously unattainable. This is not merely an incremental improvement; it’s a fundamental change in how we approach the challenge of extracting meaning from the ever-growing sea of unstructured data. DocETL is not just processing documents; it’s orchestrating a symphony of information, revealing the hidden melodies within.Image 2/n Features

DocETL offers several unique values compared to other LLM-based document processing systems:

Focus on Accuracy through Agentic Rewriting: Unlike systems that prioritize cost reduction while assuming user-provided operations are accurate, DocETL actively improves accuracy. It uses an agent-based framework with novel rewrite directives to decompose complex operations into simpler, more accurate ones. This addresses the inherent limitations of LLMs in handling complex tasks and data.

Agentic Plan Evaluation with Synthesized Validation: DocETL employs agents not only for rewriting but also for evaluating the generated plans. It automatically synthesizes task-specific validation prompts, eliminating the need for users to provide or manually validate examples. This data-driven and task-driven evaluation ensures that the chosen plan is effective for the specific context.

Opportunistic Optimization for Efficiency: Recognizing the time constraints of LLM operations, DocETL adopts an opportunistic optimization strategy. It recursively optimizes sub-plans only when necessary, focusing on the parts of the pipeline that are most likely to benefit from rewriting. This avoids the combinatorial explosion of evaluating all possible plans, making the optimization process more efficient.

Specialized Operators for Complex Document Processing: DocETL introduces operators like Resolve (for entity resolution) and Gather (for context enrichment), which are specifically designed to address common challenges in document processing. These operators go beyond the standard map-reduce paradigm and provide a more tailored approach to handling unstructured data.

Declarative YAML Interface with Fault Tolerance: DocETL offers a user-friendly declarative interface using YAML, making it easier to define and manage complex pipelines. It also incorporates a context-aware retry mechanism for LLM operations, providing robustness against occasional failures and further improving the overall accuracy and reliability of the system.

In essence, DocETL combines the power of LLMs with a sophisticated optimization framework, specialized operators, and a user-friendly interface to deliver more accurate and efficient complex document processing than existing alternatives. Its focus on accuracy through agentic rewriting and evaluation sets it apart, addressing a crucial gap in current LLM-based data processing systems.
Oct 21, 2024 9 tweets 2 min read
1/n DeepMind's paper where they used Stockfish (a programmed chess engine) to annotate data to train a LLM reveals how to distill artificial logic programs into artificial intuition engines. Surprisingly, the ELO rating of the student LLM surpassed that of the teacher program! But what I want to talk about in this tweetstorm is the idea of representations and blindspots. 2/n One discovery of this system is that the LLM was unaware of the repeat 3 move rule that leads to a draw. This is because the original system was trained on the current board state and without the history of moves. Thus the LLM was blind to how the board state arrived to its present state. In chess, that's not a problem except for the draw rule.
Oct 20, 2024 7 tweets 2 min read
1/n Thinkers often fall into the valley of excessive reductionism in an attempt to explain complex phenomena. That is why we have statements like it's all neurons, it's all matrix multiplication to explain cognition. This is as informative as saying that the universe is made up of atoms or the universe is all computation. It doesn't explain or describe the behavior of higher emergent behavior. 2/n The wetness of water isn't explained by looking at H2O molecules. One doesn't recognize the dynamics of an entire forest by just looking at a tree or even every tree. It's the interaction of the whole that causes emergent behavior.
Oct 19, 2024 10 tweets 2 min read
1/n Theories of consciousness and the mind are prone to human cognitive bias. It's extremely difficult for humans to come to the awareness that their thought processes will be entirely different from other people. Although we may converge to the same conclusion, our thought processes often will be unique. 2/n The theories that we are committed to will be the ones that are most intuitive to us. How can they not be? But we must gain the awareness that others think entirely different from us. Furthermore, we have a difficult time seeing this because we've habituated ourselves to our own thought processes that we cannot imagine entirely different ones.
Sep 15, 2024 7 tweets 3 min read
1/n Terrence Tao, arguable the most gifted living mathematician has tried GPT-o1 and this is his verdict: "However, this was an improvement over previous models, whose capability was closer to an actually incompetent graduate student. It may only take one or two further iterations of improved capability (and integration with other tools, such as computer algebra packages and proof assistants) until the level of "competent graduate student" is reached."Image 2/n Here, Tao attempts to use o1 to formulate the problem in Lean (a math theorem prover). Placing blame on o1's ignorance of Lean's latest capabilities. Here's the link: chatgpt.com/share/bb0b1cfa…
Image