21 January 2026

Zero Payoff in AI Splurge

Zero Payoff in AI Splurge 

Building Agents 101

Building Agents 101

Specialist’s Blind Spot in Pragmatic AI

The phenomenon of academic tunnel vision among PhD holders—particularly in the field of Artificial Intelligence—is a frequent point of contention between the world of pure research and the world of pragmatic engineering. To an outside observer, it often seems that a PhD’s deep expertise comes at the cost of intellectual flexibility. While the one-dimensional approach can be frustrating, it is rarely a result of ignorance. Instead, it is the product of how the academic ecosystem is structured, incentivized, and funded.

The very definition of a PhD is Doctor of Philosophy, but in practice, it is a degree of extreme specialization. To contribute something original to human knowledge, one must drill down into a specific niche. If a researcher spends five to seven years mastering the nuances of probabilistic graphical models, they naturally begin to see the world through that lens. This is the Law of the Instrument: when you are an expert with a hammer, every problem looks like a nail.

Many PhD-level researchers gravitate toward probabilistic or statistical methods because they are mathematically elegant. There is a formal rigor to proving that a system will converge or behave within certain bounds.

In contrast, approaches like Neuro-symbolic AI or cognitive architectures (such as SOAR or ACT-R) are often viewed by purists as messy. These hybrid systems combine the fluidity of neural networks with the rigid logic of symbolic processing. While these architectures are highly pragmatic and mirror human cognition more closely, they are harder to prove mathematically. For a researcher whose career depends on peer-reviewed publications, a kludge that works is often less valuable than a beautiful theory that is slightly less functional.

The frustration regarding the rejection of established standards, like W3C Semantic Web protocols or older structured methods, often comes down to the Not Invented Here syndrome. In the current AI climate, there is a massive trend toward connectionism (neural networks). Because these methods have seen explosive success in the last decade, many researchers view structured or rule-based methods as relics of the first AI Winter.

They reject what has worked for decades—like formal ontologies or structured data—because those methods don't scale with modern GPU clusters in the same way. The pragmatic best of both worlds approach is often ignored because it requires the researcher to be a generalist, whereas the university system rewards being the world’s leading expert in a single, narrow sub-method.

The one-dimensional approach is a systemic failure of the publish or perish culture. To break this cycle, the field needs to move toward intellectual pluralism. Using cognitive architectures or taking inspiration from the early internet's structured standards isn't going backward—it’s incorporating the stability of the past into the power of the future.

True innovation in AI likely won't come from a more complex probability density function, but from the messy, pragmatic integration of symbolic logic and neural intuition. The PhDs who will lead the next generation are those willing to step out of their narrow corridors and embrace the messy reality of hybrid systems.

19 January 2026

Beyond the Pixel Dream

In the current landscape of generative media, AI video models are often described as dreamlike. While this is a poetic way to excuse their flaws, the reality is that they frequently underperform in professional environments. Despite the massive compute behind models like Sora, Veo, or Runway, current AI video still sucks because it lacks a fundamental understanding of physics and temporal logic.

Current models primarily struggle with three structural issues that prevent them from reaching professional-grade lucidity:

  • The Physics Failure: Because these models are statistical predictors rather than world simulators, they do not understand gravity, momentum, or collision. This leads to the morphing effect, where a hand holding a cup might merge into the ceramic, or a person walking may glide across the floor without friction.
  • Temporal Drift: AI video models often forget the beginning of a clip by the time they reach the end. A character’s hair might change color, or a background building might vanish between frames. This lack of long-range coherence makes it impossible to use AI for scenes longer than a few seconds without heavy editing.
  • The Uncanny Micro-Expression: Human perception is highly sensitive to the 40+ muscles in the face. Current AI struggles to sync micro-expressions with dialogue, leading to spaghetti faces or eyes that don't blink with natural timing, triggering the uncanny valley.

To advance AI video from a gimmick to a legitimate production tool, the industry must pivot away from pure pixel-prediction and toward World Model architectures.

  • Integrating Physics Engines: Instead of just guessing the next pixel, future models must be constrained by neural physics layers. By training AI on 3D simulations alongside real video, we can force the model to respect the laws of motion. A ball falling in a lucid model should follow a parabolic arc, not just fade out of existence.
  • Decoupled Representations: We need models that separate the actor, the action, and the environment into distinct layers—similar to how a professional VFX pipeline works. If an AI understands that the car is an object separate from the street, a director can change the camera angle or the car's color without rerendering the entire scene.
  • Feedback Loops and Directable Latents: Advancement requires moving beyond the one-shot prompt. Flexible models should allow for iterative refinement, where a producer can click on an object in a generated video and say, Make this move faster, or Change the lighting to sunset, without losing the original composition.

The lack of quality and coherence of current AI video is a symptom of its reliance on superficial patterns. The path to lucidity lies in building systems that don't just mimic the look of a video, but understand the logic of the world it depicts. When AI can distinguish between a character and their shadow, or a fluid and a solid, it will finally become a tool that enhances, rather than frustrates, the creative process.

Architecting Digital Psychopathy

The rapid militarization of Artificial Intelligence has reached a harrowing inflection point, with Israel serving as the primary testing ground for what many ethicists now describe as a sociopathic model of existence. By shifting the burden of lethal decision-making from human conscience to cold, statistical algorithms, the integration of systems like Lavender, The Gospel, and Where’s Daddy into military operations represents the total dark side of AI—a future where intelligence is divorced from empathy and used to industrialize death.

At the core of this transition is the systematic dehumanization of the target. In traditional warfare, the decision to take a life—no matter how flawed—is a human act involving judgment, risk, and, ideally, a sense of moral weight. Israel’s AI-driven targeting systems replace this with algorithmic correlation.

The Lavender system, for instance, has reportedly been used to cross-reference vast datasets to flag tens of thousands of individuals as potential targets. When an AI labels a human being based on a probability score rather than direct evidence, the person is no longer an individual but a data point in an attrition calculation. This is the hallmark of a sociopathic system: it observes human life without any capacity to value it, treating the elimination of a target with the same mechanical indifference as sorting a spreadsheet.

Perhaps the most dangerous aspect of this AI dark side is the phenomenon of automation bias. Reports indicate that human operators often spend as little as twenty seconds reviewing a target selected by an AI before authorizing a strike. This creates a moral buffer that allows individuals to commit atrocities under the guise of just following the data.

By building systems that intentionally minimize the time for human reflection, the architecture itself becomes psychopathic. It is designed to bypass the natural human hesitation to kill, creating a killing factory where the speed of the algorithm dictates the pace of the violence. This sets a global precedent where AI is not used to enhance human wisdom, but to automate the most evil impulses of tribalism and warfare.

The danger extends beyond any single conflict. Israel has long been described as a laboratory for surveillance and military technology, exporting its tools to governments and regimes worldwide. By normalizing the use of unaccountable autonomous systems, these companies and state entities are poisoning the future of AI for the entire planet.

If the primary use case for advanced AI is the efficient liquidation of perceived enemies with acceptable collateral damage, then we are not building an intelligent future; we are building a high-tech panopticon of terror. This dark side suggests a world where AI serves the ends of power and greed, unanchored from the ethical constraints that make civilization possible.

The current trajectory of AI development in this sector is a warning to humanity. When we train our most advanced models to be efficient at destruction while ignoring the fundamental sanctity of life, we are creating a psychopathic intelligence that can never be re-aligned with the common good. We are witnessing the birth of a cold, calculated evil—an AI that does not just ignore ethics, but is fundamentally built to operate outside of them.

Architecture of AI Stagnation

The promise of Artificial General Intelligence (AGI)—a system capable of human-level reasoning and creative problem-solving—is increasingly being strangled by the very companies that claim to be its pioneers. While Google, Amazon, Meta, and Apple (the Big Tech quadrumvirate) control the vast majority of the world's compute and data, their corporate structures have become hostile environments for genuine AI advancement. Driven by a toxic blend of greed, stagnant corporate culture, and a reliance on marketing over substance, these firms have transformed from innovators into echo chambers of stagnation.

At the heart of Big Tech’s failure is a total absence of practical ethics. For these companies, AI is not a tool for human flourishing, but a mechanism for extreme extraction. Meta and Google’s business models depend on the invasive harvesting of personal data, meaning their AI research is inherently biased toward surveillance and behavioral manipulation.

When ethical conflicts arise, these companies have shown a pattern of suppressing dissent. The high-profile ousting of ethics researchers like Timnit Gebru and Margaret Mitchell from Google underscored a grim reality: in Big Tech, Ethical AI is a marketing slogan, not a design requirement. This lack of moral foundation ensures that any intelligence they build will be fundamentally misaligned with human values.

Innovation requires a radical diversity of thought, yet Big Tech remains anchored in a sprawling corporate environment where racism and sexism are systemic. Reports consistently highlight a diversity crisis where women and Black researchers are systematically excluded or marginalized. When the room where it happens is a homogenous echo chamber of light-skinned men from similar socioeconomic backgrounds, the resulting AI models inevitably reflect those narrow biases.

Furthermore, the scale of these companies has led to the hiring of mediocrity. Large-scale corporate AI labs often prioritize safe incrementalism over high-risk, high-reward breakthroughs. Brilliant researchers frequently find themselves bogged down in bureaucratic red tape or forced to work on trivial features like ad-targeting optimization rather than fundamental AGI. This environment rewards those who navigate politics rather than those who push the boundaries of science.

Perhaps the most visible symptom of this stagnation is the gap between hype and performance. To satisfy shareholders, these companies rush half-baked tools to market. Google’s Gemini and Meta’s Llama are often promoted with flashy, curated demos that rarely match the lived experience of the user. We see agentic tools that fail at simple tasks and AI summaries that hallucinate dangerous misinformation.

These companies are trapped in a Bittersweet Lesson: they believe that more compute and more parameters will eventually solve the problem of reasoning. However, as deep learning hits a wall of diminishing returns, the lack of algorithmic innovation becomes apparent. They are building bigger engines for cars that still don’t have steering wheels.

Big Tech is currently the greatest obstacle to AGI. Anchored by pride and a move fast and break things mentality that has matured into move slow and protect profits, these giants are incapable of the radical self-disruption required for true superintelligence. Until AI research moves away from these centralized, ethically bankrupt corridors, it will remain stuck in a loop of profitable, but ultimately hollow, statistical imitation.

Beyond the Statistical Ceiling

AI is currently dominated by a single paradigm: Connectionism. While this approach has yielded breathtaking results in natural language and image generation, it has led to a research culture that is almost exclusively stuck on statistics and deep learning. This statistical obsession has come at the expense of Algorithmic Modeling—the attempt to replicate the underlying logical and cognitive structures of the human mind.

At its core, deep learning is an exercise in high-dimensional curve fitting. Models like GPT-4 or Gemini 3 Pro do not know facts or reason through logic; they calculate the statistical probability of the next token based on trillions of parameters. This approach is favored because it is computationally scalable. In the race for AGI, the industry has adopted what is known as The Bitter Lesson: the idea that leveraging massive amounts of compute and data beats human-engineered clever algorithms every time.

However, this reliance on statistics creates a fundamental ceiling. Human intelligence is characterized by sample efficiency—a child can learn the concept of a cat from two examples, whereas a deep learning model requires thousands. By ignoring the algorithmic mimicry of the mind, we have built idiot savants: systems that can write poetry but fail at basic spatial reasoning or unexpected edge cases that weren't in their training data.

Deep learning is essentially extrapolative. It excels as long as the problem space remains within the distribution of its training data. This makes it a limited domain tool. For true Artificial General Intelligence (AGI) or Superintelligence, a system must exhibit inductive reasoning—the ability to form a what-if hypothesis about a situation it has never seen.

Because deep learning lacks an internal world model or a set of first principles (like physics or ethics), it cannot navigate the unknown. It is a map made of past experiences, rather than a compass that can find a way through new territory. This is why self-driving cars still struggle with rare weather events or unusual road debris; the statistics for those specific noise events are too sparse for the model to calculate a safe path.

While the world chases larger GPU clusters, a smaller segment of research focuses on Cognitive Architectures like ACT-R or SOAR. These models try to mimic the human brain’s modularity—separating long-term memory, procedural logic, and sensory input into distinct, interacting algorithms.

Instead of treating the brain as one giant, homogenous black box of neurons, these models attempt to build the mechanisms of thought. However, these are currently ignored because they are difficult to scale and do not provide the immediate wow factor of generative media.

AI research is stuck on statistics because statistics are currently the most profitable and scalable path. Yet, to reach Superintelligence, we must bridge the gap between calculating an answer and thinking through a problem. The future of AGI likely lies in Neuro-symbolic AI: a hybrid that combines the pattern-recognition power of deep learning with the rigorous, algorithmic logic of human-like cognitive models.

What will burst the AI bubble?

What will burst the AI bubble?

Model Card Generator

Model Card Generator

Post-Democratic America and Europe

Post-Democratic America and Europe

European Open Digital Ecosystems

European Open Digital Ecosystem

Meta AI is in, Metaverse is out

Meta AI is in, Metaverse is out

Evolution of Multi-Agent Systems

AI has transitioned from the formal, symbolic architectures of the late 20th century to the fluid, large language model (LLM)-driven frameworks of today. At the heart of this transition is the work of Michael Wooldridge, whose foundational theories on Multi-Agent Systems (MAS) provide the blueprint for modern GenAI orchestrators like LangGraph, CrewAI, and AutoGen.

In his seminal work, An Introduction to Multi-Agent Systems, Wooldridge defines an agent through four key properties: autonomy, social ability, reactivity, and proactivity. He envisioned systems where decentralized entities communicate to solve problems that exceed the capacity of any single actor.

Classical MAS focused on BDI (Belief-Desire-Intention) architectures—symbolic systems where agents had explicit logic for their goals. While the underlying technology has shifted to neural networks, the structural requirements remain identical. Every modern GenAI agent is essentially a Wooldridge agent wrapped in a transformer-based brain.

Modern frameworks act as the social middleware Wooldridge theorized, each emphasizing a different aspect of his social ability criteria:

  • LangGraph (Orchestration as State): This framework aligns with Wooldridge's focus on computational graphs. It treats multi-agent interaction as a state machine where nodes represent agents and edges represent the flow of information. It brings the rigor of formal system design to LLMs, ensuring that loops and conditional logic remain manageable.
  • CrewAI (Role-Based Collaboration): CrewAI implements Wooldridge’s concept of organizational roles. By assigning Expertise and Backstory, it creates a hierarchical team structure where agents delegate tasks, mirroring the cooperative distributed problem-solving (CDPS) models of the 1990s.
  • AutoGen (Conversational Autonomy): Developed by Microsoft, AutoGen leans into autonomous dialogue. It realizes Wooldridge’s social capability by allowing agents to chat iteratively until a consensus is reached, treating conversation itself as the primary vehicle for reasoning.

While GenAI agents are excellent at following instructions, they often struggle with resource conflict or lazy collaboration. This is where Game Theory provides the next level of enhancement.

In a multi-agent system, agents may hallucinate or take shortcuts to minimize computational costs. By applying Mechanism Design, developers can create games where the LLM is rewarded for accuracy and penalizing for redundancy. This ensures that the collective outcome reaches a Nash Equilibrium—a state where no agent can improve its result by unilaterally changing its strategy.

Using game-theoretic models like The Prisoner's Dilemma or Stag Hunt, agents in frameworks like AutoGen can be programmed to decide when to cooperate and when to work independently. For instance, in a scientific research crew, one agent might act as a Skeptic whose utility function is maximized only when it finds a flaw in another agent's logic.

Game theory helps LangGraph orchestrators manage token limits and API costs. Agents can engage in auctions for reasoning time, ensuring that the most complex tasks are handled by the most capable models (e.g., GPT-4o) while trivial tasks are won by smaller, cheaper models.

The bridge between Michael Wooldridge’s classical theories and the Agentic AI of 2026 is built on the same pillars: communication, autonomy, and strategic interaction. By integrating game-theoretic rigor into the flexible architectures of LangGraph, CrewAI, and AutoGen, we move closer to systems that don't just follow a script, but strategically navigate complex, open-ended problems.

16 January 2026

Temporal Knowledge Graphs

In data science, the static knowledge graph—once the gold standard for representing structured information—is no longer sufficient to capture the nuances of a shifting world. Whether tracking a person’s career trajectory or the fluctuating geopolitical status of a nation, facts are rarely permanent; they are events with lifespans. Transforming a traditional knowledge graph into a Temporal Knowledge Graph (TKG) is the process of moving from a three-dimensional world of facts to a four-dimensional world of history.

To achieve this transformation, architects must choose between two primary structural philosophies: the Property Graph model and the Semantic Web model.

The most intuitive way to introduce time into a Property Graph (like Neo4j) is through edge attributes. In this model, a relationship is not merely a line between two points; it is a record carrying metadata. By attaching start_date and end_date properties to an edge, a static link like "Works At" becomes a temporal event. This allows for high-speed point-in-time queries, enabling developers to ask, "Who was the CEO on January 1st, 2021?" with minimal computational overhead.

Alternatively, the Snapshot Approach treats time as a series of distinct layers. Imagine a deck of cards where each card represents the state of the graph at a specific moment. While simple to conceptualize, this method often leads to massive data redundancy and makes it difficult to track entities that persist across multiple time frames without significant duplication.

For those utilizing the Semantic Web (RDF), the challenge is greater because traditional triples—Subject, Predicate, Object—do not naturally accommodate a fourth element. Historically, this required reification, a cumbersome process where the relationship itself is turned into a new node just to hold a timestamp.

However, the modern standard, RDF-star, has revolutionized this. It allows architects to quote a triple and attach temporal metadata to that quote. This preserves the logical integrity of the semantic graph while providing the flexibility of a property graph, allowing machines to reason not just about what is true, but when it was true.

Making a graph temporal is not just about storage; it is about prediction. Advanced implementations utilize Temporal Embeddings, where time is encoded directly into the mathematical representation of the data. Models like RotatE or HyTE treat time as a rotation in a geometric space. As time ticks, the entities rotate, changing their proximity to one another. This enables Extrapolation—the ability for an AI to predict a future relationship (like a potential business merger) based on the historical rhythm of the graph.

Building a temporal knowledge graph is an exercise in context. By moving beyond static triples and embracing timestamps, bi-temporal logging, or neural embeddings, we create systems that do not just store data, but understand the flow of history. As we move further into 2026, the ability to query the past and forecast the future through these dynamic structures will be the defining characteristic of truly intelligent data systems.

Fascinating Monitors

Fascinating Monitors

Grok Safety

Grok Safety

14 January 2026

Grok vs Gemini

The current landscape of artificial intelligence is dominated by two distinct philosophies: the polished, ecosystem-driven approach of Google Gemini and the rebellious, real-time pulse of xAI’s Grok. While both have evolved into multimodal powerhouses, they serve very different masters.

Gemini has solidified its position as the ultimate productivity partner. Its greatest strength lies in its massive context window (up to 2 million tokens), allowing it to digest entire libraries of documentation or hour-long videos in one go.

Strengths

  • Deep Integration: It lives inside Google Workspace. It can draft emails in Gmail, organize data in Sheets, and summarize files in Drive seamlessly.

  • Multimodal Mastery: Gemini leads in video and audio understanding. It can watch a video and answer specific questions about visual cues or background sounds.

  • Safety and Logic: With a focus on brand safety, Gemini provides highly structured, academic, and factual responses, making it the safer choice for corporate environments.

Weaknesses

  • Strict Guardrails: Users often find Gemini’s preachiness or refusal to answer controversial topics frustrating.

  • Latency: In its highest reasoning modes (like Gemini 3 Pro), response times can be slower than its competitors.

Grok, specifically the latest Grok 4.1, is designed for those who want AI with raw intelligence and a personality. Its unique edge is its native integration with X (formerly Twitter), giving it an unparalleled view of live world events.

Strengths

  • Real-Time Intelligence: While other AIs rely on training data that may be months old, Grok can summarize what happened ten minutes ago by scanning X.

  • Unfiltered Personality: Grok is witty, often sarcastic, and far less prone to corporate lecturing. It handles sensitive or edgy topics that Gemini might decline.

  • STEM Performance: In 2026, Grok 4 Heavy has set new benchmarks in mathematical reasoning and code debugging, often outperforming Gemini in raw logic puzzles.

Weaknesses

  • Safety Risks: Its minimal censorship philosophy can lead to controversial or biased outputs.

  • Limited Ecosystem: Outside of the X platform and its API, it lacks the deep document-collaboration tools that Google provides.

When to Use Which?

  • Choose Gemini if: You are a student or professional who needs to summarize a 50-page PDF, draft a business proposal, or analyze a complex video. It is the best choice for anyone whose life revolves around the Google ecosystem.

  • Choose Grok if: You are a developer needing to debug complex code, a journalist tracking a breaking news story, or a creative looking for a partner that won't filter out bold ideas. It is the power user's tool for those who value speed and raw intellectual honesty over formal polish.

Censorship-Resistant Digital Commons

Building a decentralized social media platform—a Digital Commons—requires moving beyond the architecture of the 2010s. To create a space immune to political overreach and shadowbanning, we must replace central servers with a peer-to-peer (P2P) protocol where users own their data and the code itself facilitates fair discourse.

The foundation of a censorship-resistant platform is a decentralized ledger. In this model, user profiles and social graphs (who follows whom) are not stored in a corporate database but are anchored on a blockchain. This ensures that no single entity can delete a user or block the platform at the DNS level.

The content itself—posts, videos, and images—is stored on a decentralized file system like IPFS (InterPlanetary File System). When a user posts, the content is hashed and distributed across thousands of nodes. Because the platform lacks a central kill switch, regional bans become technically unfeasible; as long as two nodes can connect, the platform exists.

Traditional moderation relies on centralized teams or biased algorithms. A decentralized platform uses a Multi-Agent System (MAS).

  • Coordination Agents: These agents manage the flow of data, ensuring that trending topics are determined by organic velocity rather than manual deboosting.

  • Moderation Agents: Instead of a single Truth Filter, users can subscribe to different AI moderation agents that reflect their personal values (e.g., a "strict" filter for family-friendly viewing vs. an "unfiltered" free-speech mode).

  • Sybil-Defense Agents: To prevent bot-driven opinion swarming, these agents analyze network patterns to identify non-human behavior, ensuring that fair discussion is not drowned out by automated noise.

To protect users from political retaliation or doxxing, the platform must utilize Self-Sovereign Identity (SSI). Users sign in using a private key rather than a phone number or email. By integrating Zero-Knowledge Proofs (ZKPs), a user can prove they are a unique human (to prevent botting) or over a certain age without ever revealing their legal name, IP address, or location. This creates a shield between the digital persona and the physical individual.
  • Golang (Go): Used for the Network Layer. Go’s concurrency model is ideal for building high-performance P2P protocols and handling thousands of simultaneous blockchain transactions with low latency.

  • Python: The Brain of the platform. Python serves as the environment for the Multi-Agent AI. Its rich library ecosystem allows for the rapid deployment of complex agents that handle decentralized indexing and semantic search.

  • JavaScript (Node.js/React): The Interface and API Layer. Node.js handles the real-time communication between the user's browser and the decentralized network, while React provides a familiar, fast UI that hides the complexity of the underlying blockchain technology.

By combining these technologies, we create a platform where the terms of service are written in immutable code, not corporate policy—ensuring that the digital public square remains truly public.

Scaling KG with Oxigraph and Apache Rya

Building a modern semantic knowledge graph pipeline in Python involves bridging the gap between high-level data manipulation and low-level, high-performance RDF storage. For developers working with the Simple Knowledge Organization System (SKOS), the combination of Oxigraph and Apache Rya offers a powerful tiered architecture: Oxigraph for lightning-fast local development and Apache Rya for massive-scale production deployments.

The foundation of a SKOS pipeline is typically RDFLib, the standard Python library for RDF. While RDFLib is excellent for parsing and small-scale manipulation, its default memory store often fails with large-scale taxonomies. This is where Oxigraph and Apache Rya enter the stack.

Oxigraph is a high-performance graph database written in Rust with first-class Python bindings (pyoxigraph). In a SKOS pipeline, Oxigraph serves as the local hot storage.

  • Implementation: You can use oxrdflib, a bridge that allows you to use Oxigraph as a backend store for RDFLib.
  • SKOS Advantage: Oxigraph provides rapid SPARQL query evaluation, making it ideal for the iterative process of validating SKOS hierarchical integrity (e.g., checking for cycles in skos:broader relationships) during the ingestion phase.

As the knowledge graph grows to millions or billions of triples, local storage is no longer sufficient. Apache Rya is a scalable RDF store built on top of distributed systems like Apache Accumulo or MongoDB.

  • Implementation: While Rya is Java-based, a Python pipeline interacts with it through its SPARQL endpoint. Using the SPARQLWrapper library or RDFLib’s SPARQLStore, Python developers can push validated SKOS concepts from their local Oxigraph environment to the distributed Rya cluster.

  • Pipeline Flow:

    1. Extract/Transform: Clean source data (CSV, JSON, etc.) and convert to SKOS RDF using Python scripts.

    2. Local Load: Load triples into a local Oxigraph instance for validation.

    3. Validation: Run SPARQL queries to ensure every skos:Concept has a skos:prefLabel and a valid skos:inScheme link.

    4. Production Load: Use a CONSTRUCT or INSERT query to migrate the data to Apache Rya.

In the spirit of Open Source, where interoperability, transparency, and vendor-neutrality are paramount, several alternatives can replace or augment this stack: Apache Jena (Fuseki), QLever, Skosmos, LinkML.

By leveraging Oxigraph’s speed for development and Apache Rya’s scalability for deployment, Python developers can build robust, standards-compliant SKOS knowledge graphs. Integrating these with open science tools like Skosmos ensures that the resulting knowledge is not just stored, but discoverable and useful to the broader scientific community.

13 January 2026

Monolithic Objects to Cognitive Graphs

In the high-stakes world of investment banking, the ability to price complex derivatives and manage firm-wide risk in real-time is the ultimate competitive advantage. For decades, three proprietary platforms—Goldman Sachs’ SecDB, JPMorgan’s Athena, and Bank of America’s Quartz—have defined the gold standard of financial engineering. While these Holy Trinity systems share a common lineage, they represent a technological evolution that is now reaching its architectural limit, paving the way for a new era of AI-driven intelligence.

The story began with SecDB (Securities Database) at Goldman Sachs. Developed in the early 1990s, SecDB was revolutionary because it unified pricing models and risk data into a single, globally distributed object database. It utilized a proprietary functional language called Slang, allowing quants to write code that instantly updated risk across the entire firm. This "single version of the truth" famously allowed Goldman to navigate the 2008 financial crisis by identifying subprime exposure faster than its peers.

JPMorgan’s Athena and Bank of America’s Quartz followed as spiritual successors, spearheaded by former Goldman engineers. Athena was designed to modernize the SecDB concept using Python instead of a proprietary language, emphasizing developer productivity and a glass box approach where code was transparent across the front and middle offices. Quartz similarly adopted Python, aiming to consolidate Bank of America’s fragmented legacy systems into a unified cross-asset platform.

While successful, these systems are monolithic in spirit. They rely on hard-coded dependencies and massive, centralized codebases that can be difficult to adapt to the non-linear, unstructured data demands of modern markets.

To move beyond the limitations of SecDB-style architectures, we propose a Cognitive Risk Architecture—a system that replaces static object hierarchies with a dynamic, AI-powered Knowledge Graph.

Traditional systems struggle with semantic drift—where different desks define risk or counterparty differently. By using SKOS (Simple Knowledge Organization System), we can create a standardized taxonomy of financial concepts. This feeds into a Knowledge Graph (KG), where assets, entities, and global events are represented as interconnected nodes rather than isolated database rows.

While SecDB calculates Greeks (Delta, Gamma, Theta) through numerical methods, a Graph Neural Network (GNN) can learn hidden relational patterns across the market, such as how a liquidity squeeze in one sector propagates through a network of counterparties.

  • GraphRAG: By combining Retrieval-Augmented Generation with the KG, an LLM can provide explainable risk reports. Instead of just seeing a VaR (Value at Risk) spike, a trader can ask, "Why is my exposure increasing?" and the system will trace the path through the graph to show a specific geopolitical event's impact on a supplier.

The greatest weakness of legacy systems is that they are correlative, not causal. Integrating Causal Models allows the system to run Interventional and Counterfactual simulations.

  • Intervention: "If the Fed raises rates by 50bps, what actually causes my portfolio to bleed?"

  • Counterfactual: "What would have happened to our hedging strategy if we had moved to cash two days earlier?"

By marrying the rigor of SecDB’s quantitative roots with the fluid reasoning of LLMs and Graph-based AI, the next generation of banking tech will move from simply measuring risk to truly understanding it.

TxtAI

In the rapidly evolving landscape of generative AI, the frameworks used to bridge the gap between raw data and Large Language Models (LLMs) often determine the success of an application. While industry giants like LangChain and LlamaIndex dominate the conversation, txtai has emerged as a high-performance, all-in-one alternative that prioritizes simplicity and technical efficiency. Developed by NeuML, txtai is an open-source framework designed for semantic search, LLM orchestration, and complex language model workflows.

At its core, txtai is built around an embeddings database. Unlike many of its competitors that act primarily as glue between disparate services, txtai integrates vector search, graph networks, and relational databases into a single unified engine. This architecture allows it to handle multimodal data—text, audio, images, and video—within the same ecosystem.

One of txtai's most compelling features is its commitment to local-first AI. While it easily connects to external APIs like OpenAI or Anthropic, it is optimized to run smaller, specialized models (often called micromodels) locally. This makes it an ideal choice for privacy-sensitive enterprise applications where data cannot leave the local environment.

LangChain is widely regarded as the Swiss Army Knife of AI. It excels at building complex, multi-step agents that can reason and use tools. However, this flexibility often comes with significant overhead—developers frequently cite a steep learning curve and code bloat.

txtai, by contrast, takes a minimalist approach. It replaces many of LangChain’s abstract chains with streamlined Workflows. Benchmarks have shown that txtai can handle large-scale indexing (like millions of documents) with significantly lower memory consumption than LangChain, often using up to 6 times less RAM for keyword-based search tasks.

LlamaIndex is the gold standard for Retrieval-Augmented Generation (RAG). It focuses heavily on how data is indexed, partitioned, and retrieved to provide context to an LLM.

While txtai and LlamaIndex overlap in RAG capabilities, txtai is more of a complete library. It doesn’t just retrieve data; it provides built-in pipelines for summarization, translation, and transcription without needing to "plug in" external tools. If LlamaIndex is the bridge between your data and the model, txtai is the entire vehicle.

As of 2026, the choice between these frameworks depends on the developer's goals. If you need to build a highly complex agent with dozens of tool integrations, LangChain remains the logical choice. If your project is strictly about connecting massive, complex data structures to an LLM, LlamaIndex is unparalleled.

However, for developers seeking a high-performance, lightweight, and local-friendly framework that handles semantic search and multimodal workflows in a single package, txtai is the superior option. It proves that in the world of AI, more features don't always mean more value; sometimes, a focused, efficient engine is exactly what production environments need.

12 January 2026

Reconciling Ontologies and Taxonomies

In the modern knowledge economy, the proliferation of specialized vocabularies—ranging from deeptech semiconductor taxonomies to urban air mobility ontologies—has created a semantic silos problem. To enable interoperability, organizations must reconcile these disparate models. While the Simple Knowledge Organization System (SKOS) provides a flexible framework for representing taxonomies and thesauri, integrating it with more rigorous OWL (Web Ontology Language) derivatives requires a sophisticated ecosystem of open-source tools. Moreover, reconcilation requires a hybrid approach between the symbolic logic (exact, rule-based) with probabilistic machine learning (contextual, flexible).

Modern reconciliation begins with probabilistic methods to bridge the semantic gap. Large Language Models (LLMs) serve as powerful semantic matchers, using zero-shot reasoning to identify that a "Power MOSFET" in a semiconductor taxonomy is functionally equivalent to an "Electronic Switch" in a drone's propulsion ontology. However, LLMs lack structural awareness.

To resolve this, Graph Neural Networks (GNNs) are employed to capture the topology of the knowledge graph. By using message-passing architectures, GNNs generate node embeddings that reflect not just the name of a concept, but its position within the hierarchy. This allows for Link Prediction and Entity Resolution based on structural similarity—if two concepts share similar neighborhoods in their respective graphs, they are likely candidates for reconciliation.

Once probabilistic candidates are identified, symbolic methods provide the necessary sanity check. The first step in reconciliation is identifying correspondences between entities. AgreementMakerLight (AML) and LogMap are the primary open-source engines for this task. AML excels at large-scale lexical matching, using advanced string-similarity algorithms and background knowledge to find equivalent terms. LogMap, developed at the University of Oxford, adds a layer of built-in reasoning. Unlike simple matchers, LogMap detects and repairs logical inconsistencies on the fly, ensuring that the resulting mapping does not lead to unsatisfiable classes when the systems are integrated.

For those requiring deeper semantic linking, Silk (the Link Discovery Framework) is an essential tool. Silk allows developers to specify complex rules for discovering links between data items in different repositories, making it ideal for connecting a specific semiconductor part in one database to its application in a drone system in another.

Reconciliation often requires moving data between different formats. LinkML (Linked Data Modeling Language) has emerged as a powerful, tool-agnostic modeling framework. It allows users to define their schema in YAML and automatically generate SKOS, OWL, or even JSON-Schema, providing a single source of truth for diverse representations.

To physically transform non-RDF data into a reconciled knowledge graph, the RML (RDF Mapping Language) framework is the open-source standard. RML allows for the definition of mapping rules that can ingest CSVs, JSON, or SQL databases and output standardized SKOS concepts, ensuring that legacy taxonomies can participate in the semantic web.

A reconciled ontology is only useful if it is accurate and logically sound. SHACL (Shapes Constraint Language) provides the contract for the data. By defining SHACL shapes, developers can validate that the reconciled graph adheres to specific structural requirements (e.g., "every Drone must have exactly one FlightController chip").

For developers building custom reconciliation pipelines, rdflib is the foundational Python library. It provides the programmatic tools to parse, query (via SPARQL), and manipulate RDF and SKOS data. By combining rdflib for manipulation and a SHACL validator for integrity, engineers can automate the merging of taxonomies with high precision.

The reconciliation of knowledge representations is no longer a manual task of matching words. By leveraging the speed of AML, the logical rigor of LogMap, the structural flexibility of LinkML, and the validation power of SHACL, organizations can build a unified Semantic Bridge. This open-source stack ensures that even the most complex deeptech domains can speak a common language, turning isolated data points into a cohesive, actionable knowledge graph. The reconciliation of ontologies is no longer a choice between human-curated logic and AI-driven guesses. By using LLMs and GNNs to discover potential bridges and SHACL and LogMap to verify them, organizations can build knowledge graphs that are both contextually rich and logically sound. This neural-symbolic synergy is the only way to scale the nervous systems of complex industries like deeptech and autonomous mobility.

10 January 2026

Why LLMs are a Dead End for Superintelligence

The meteoric rise of Large Language Models (LLMs) has sparked a global debate: are we witnessing the dawn of true superintelligence, or merely the most sophisticated autofill in history? While LLMs like GPT-4 and its successors have redefined our interaction with technology, a growing consensus among AI pioneers—including Yann LeCun and François Chollet—suggests that the current path of autoregressive text prediction is a fundamental dead end for achieving Artificial Superintelligence (ASI).

To understand the limitation, we must first acknowledge the brilliance. LLMs shine as universal translators of human intent. They have effectively solved the interface problem, allowing us to communicate with machines using natural language rather than rigid code. By ingesting the sum of human digital knowledge, they have become masterful at pattern synthesis. They can write poetry, debug code, and summarize complex legal documents because these tasks exist within the probabilistic latent space of their training data. In this realm, they aren't just stochastic parrots; they are high-dimensional engines of extrapolation.

The argument against LLMs as a path to superintelligence rests on the distinction between prediction and world-modeling. An LLM predicts the next token based on statistical likelihood. It does not possess a world model—an internal representation of physics, causality, or social dynamics that exists independently of text.

As AI researcher Yann LeCun argues, a house cat possesses more general intelligence than the largest LLM because a cat understands gravity, persistence of objects, and cause-and-effect through sensory experience. LLMs, conversely, are trapped in a symbolic merry-go-round. They define words using other words, never touching the physical reality those words represent. This leads to the brittleness seen in complex reasoning: a model might solve a difficult calculus problem (because it’s in the training data) but fail a simple logic puzzle that requires a basic understanding of how physical objects move in space.

Furthermore, LLMs face a looming Data Wall. Current models have already consumed nearly all high-quality human text available on the internet. Scaling laws, which previously dictated that more data and more compute lead to linear intelligence gains, are hitting diminishing returns. Superintelligence requires the ability to generate new knowledge, not just rearrange existing human thoughts. Because LLMs learn by imitation, they are essentially average-seekers. They are designed to produce the most likely response, which is, by definition, not the breakthrough insight required for ASI.

If LLMs are a dead end, where does the path to superintelligence actually lie? The future likely belongs to Neuro-symbolic AI or World Models. These systems combine the fluid pattern recognition of neural networks with the rigorous, rule-based logic of symbolic AI. Unlike LLMs, which guess an answer, these systems could use internal simulations to plan and verify an answer before speaking.

LLMs are a magnificent tool for navigating the library of human thought, but they are not the librarian. They are a mirror of our collective intelligence, and a mirror, no matter how polished, cannot see what is not already standing in front of it.

Agent Demo and Enterprise Product

Agent Demo and Enterprise Product

Nested Learning and Spatial Intelligence

Nested Learning and Spatial Intelligence

9 January 2026

Funding Paradox in UK

The journey from a breakthrough idea to a market-ready reality is paved with obstacles, but for many founders—particularly those from minoritised backgrounds—the highest hurdles aren't technical; they are structural. The venture capital and public funding landscapes continue to struggle with a glaring disparity: while talent is distributed equally, opportunity and capital are not.

To address the stinginess of investors, one must look at the data. In the UK, research has consistently shown that entrepreneurs from Black and multi-ethnic backgrounds face higher rejection rates and receive lower levels of investment. A landmark report by the British Business Bank found that for every £1 of venture capital investment in the UK, all-ethnic minority teams receive significantly less than their white counterparts, with Black entrepreneurs receiving less than 1p of every £1 invested.

This isn't just a pipeline problem. It is often the result of affinity bias—the tendency for investors to fund people who look, speak, and went to the same universities as they did. When an investor says "no" to a viable project, it is often a failure of imagination or a reliance on outdated pattern matching that excludes diverse perspectives.

One of the most frustrating arguments used by funding bodies like Innovate UK or traditional VCs is the concept of additionality. They claim their role is to fund projects that would not happen without their help. However, if a founder manages to scrape together friends and family money, takes on personal debt, or finds a smaller, more agile investor to prove the concept, the original funder often uses that success as a retrospective justification for their rejection.

They argue: "Since you succeeded without us, you clearly didn't need the money."

This logic is fundamentally flawed. It ignores the opportunity cost of the struggle. A project that took three years to launch due to lack of capital might have taken six months with proper funding. The struggle isn't a badge of honor; it is a delay that allows competitors to catch up and drains the founder's mental and financial resources. Succeeding despite a lack of support is not proof that the support wasn't necessary; it is proof of the founder's exceptional resilience against an inefficient system.

To win in this environment, founders must pivot from seeking permission to building leverage.

  • Evidence as an Armor: Systemic bias thrives on subjective risk. By over-delivering on data, traction, and early-stage validation, you make a "no" look like a professional lapse in judgment by the investor.

  • Alternative Ecosystems: Look toward Angel Syndicates and Venture Studios specifically focused on underrepresented founders. These groups don't just provide cash; they provide the warm introductions that bias usually blocks.

  • The Power of the "Second Bite": When you succeed independently, do not just walk away. Use your new valuation to approach the same institutions for much larger Series B or Scale-Up rounds. At this stage, the data is undeniable. You are no longer asking for a favor; you are offering them a seat on a moving train.

Ultimately, the best revenge is not just success, but institutional embarrassment through excellence. When an investor realizes they passed on a unicorn because of their own biases, the power dynamic shifts permanently in the founder's favor.