Distributed systems are typically documented and operated through infrastructure-centric views that catalog servers, containers, and services. This post introduces the Holonic Flow Model (HFM), a functional approach that fundamentally reframes system understanding around capabilities, transformations, and value delivery.
By adopting the holon—a self-contained functional unit that’s both whole and part—this model provides a powerful framework for reasoning about, operating, and evolving complex systems. Its visual language and principles offer a practical alternative to traditional architectural documentation, improving incident response, ownership boundaries, and system evolution.
Managing Complex Systems: HFM operationalizes the principles of the OCO framework by organizing these essential management functions around self-contained functional units. Where OCO defines what capabilities any complex system needs, HFM demonstrates how to structure those capabilities through holons and their flows. Though HFM emerged from distributed systems challenges, its alignment with the universal OCO principles suggests broader applicability. Any system requiring effective observation of state, control of behavior, and practical interfaces for operation could benefit from holonic organization.
To facilitate steering of systems, HFM establishes a standardized architecture of holonic interaction grounded in inlets and outlets. Each holon, irrespective of its internal complexity or the capability it embodies, exposes these standard elements. An inlet serves as the boundary point where a holon assumes responsibility for a unit of work, while an outlet represents the point where it relinquishes custody. These boundaries are regarded as first-class operational entities, signifying the narrow points where traffic, decisions, and contracts converge.

HFM fundamentally serves as a model for comprehending and reasoning about a system, rather than providing a rigid prescription for a specific runtime implementation. This distinction is crucial for appreciating its true value. The power of HFM lies in its nature as a top-down abstraction. It offers a functional perspective of a system organized around business capabilities and the flow of value, which can subsequently be mapped onto any underlying infrastructure, whether that entails Kubernetes, service meshes, or API gateways.
The model’s fractal design, which enables zooming from a high-level business capability down to its constituent sub-flows, is a core feature that facilitates this flexibility. It allows for analysis at any level of granularity required, from a high-level executive dashboard to a detailed engineering view, without altering the fundamental language or concepts being employed.
The model facilitates the aggregation of numerous small, underlying interactions into a single, coherent flow at the boundary of the parent holon. This ensures consistent computation of signals like throughput, work-in-progress, and lead time, irrespective of whether you’re analyzing the entire system or a single subcomponent.
The capacity to establish a comprehensive, multi-layered perspective that preserves flow and is observable across diverse scales ultimately provides the profound situational awareness that distinguishes HFM from purely bottom-up approaches. The conventional incident management approach starts with alerts from infrastructure, requiring operators to mentally reconstruct business processes to understand the impact, like navigating an emergency city using building blueprints. HFM shifts this paradigm by assessing capability health and only delving into infrastructure when functional issues arise. Alerts are framed in business impact, clarifying priority and context. This transition reduces cognitive load and enables efficient, timely responses.
HFM isn’t meant to replace knowledge of the underlying infrastructure but to organize it into a more meaningful, purpose-driven hierarchy. It provides the “why” (business purpose) that gives context to the “what” (technical components), offering a unified conceptual framework for managing complexity. HFM takes the fundamental principle of flow, which is already the native language of the network, and elevates it into a universal model for the entire system. It argues that just as network traffic flows between devices, work flows between business capabilities. By applying this consistent, flow-based model at every level, HFM provides the comprehensive situational awareness that’s impossible to achieve when stitching together disparate, object-centric views.
The Ghost in the Machine
Arthur Koestler introduced holons in The Ghost in the Machine (1967) to describe entities that are simultaneously wholes and parts in a hierarchy. A classic example is a cell, which is a whole organism but also a part of a larger one. Koestler identified key attributes of a holon: self-contained, participatory in a larger system, hierarchical, and autonomous. HFM adopts this definition, treating holons as functional units of capability with clear boundaries and purposes, interacting with other holons to deliver higher-level outcomes.
- A holon acts as a whole that encapsulates a business capability end-to-end. It encompasses all the necessary logic, state, and subcomponents required to perform that function. A whole has its own purpose, boundaries, and internal coherence.
- A single holon can serve as a component within a larger flow or a higher-level holon. The model underscores that holons operate within larger holarchies.
- Holons possess autonomy, each maintaining its own health state and capable of independent adaptation or degradation, akin to a cell transitioning to emergency mode without waiting for the entire organism.
- Holons are hierarchical, allowing infinite nesting of holons, or sub-holons within holons. Each sub-holon utilizes the same interface or protocol as any holon, facilitating clean composition and substitutability.
Koestler’s concept of a holon also encompassed the notion of dual tendencies. A holon must assert its individuality (self-assertive tendency) while simultaneously integrating into the larger whole (integrative tendency). HFM effectively captures this duality by providing each holon with internal controls (for self-management) and external interfaces (for cooperation). For instance, a holon can independently handle contingencies without seeking higher authorities (autonomy) while simultaneously being subject to control from higher authorities (meaning it can be governed by system-wide policies or oversight). In HFM, this might translate to each holon managing its workload independently, but higher-level orchestration can redistribute work if necessary. This balance ensures the stability of the individual holon while maintaining system coherence.
The Holonic Flow Model
HFM is a proposed framework for system operations that shifts focus from physical infrastructure to functional capabilities and flows. It represents systems as networks of holons (self-contained functional units that are simultaneously wholes and parts) through which work flows to deliver value.
HFM explicitly draws upon Stafford Beer’s VSM, particularly the concept of recursive organizational structure. In VSM, every viable system comprises smaller viable systems, each with a degree of autonomy constrained by higher-level limitations. This phenomenon is referred to as cybernetic isomorphism, wherein an identical organizational pattern recurs at every nested level. The model mirrors this principle by organizing system functions hierarchically, ensuring that each holon (functional unit) is composed of sub-holons and is integral to a larger holon. This fractal, recursive design enables each level to operate autonomously while adhering to the policies of the overarching system, akin to Beer’s model.
HFM puts forward the view that holons engage in local interactions, such as passing work and applying backpressure, which collectively generate emergent global behavior and self-adaptation within the entire system. This aligns with the principles of Complex Adaptive Systems (CAS): the overall behavior of a complex system isn’t predetermined by any single component but emerges from the collective adaptation of numerous small components. For instance, the model’s emphasis on flow through holons and dynamic load-balancing mirrors CAS concepts of nonlinear interactions and feedback. When one holon slows down, leading to backlogs, upstream holons perceive backpressure and can adjust their flow rates. This form of local feedback contributes to the stabilization of the entire system, akin to homeostatic behavior.
The model draws an analogy to James G. Miller’s Living Systems Theory (LST) by proposing that holonic systems exhibit homeostasis, adaptation, and hierarchical regulation, similar to biological organisms. LST posits that living organisms (and by extension organizations) possess critical subsystems (such as respiratory, regulatory, etc.) and exist in nested levels (cell → organ → organism → society), each with regulatory feedback.
HFM attempts to introduce homeostatic regulation within each holon by defining health states and control mechanisms. For instance, each holon can be in one of seven states: Converging, Stable, Diverging, Erratic, Degraded, Defective, or Down, based on functional measures like throughput, success rates, and invariants. These states trigger local control actions, such as circuit breakers or scaling, akin to how an organism maintains internal stability. Hierarchical regulation is observed where higher-level holons oversee the collective behavior of sub-holons, similar to how organs regulate cells. The model’s control surfaces, including ingress/egress controls, internal holon controls, and more, function as feedback loops to ensure the system remains within its operational limits. This aligns conceptually with LST’s view that each level in a living hierarchy must have mechanisms for processing information, making decisions, and correcting behavior to serve the larger whole. In HFM, each holon’s control interface can be seen as fulfilling a “supervisory” role for that unit.
Flow as an Organizing Principle
Most outages resemble navigating a city using building blueprints. While we have pod names and dashboards, the customer journey remains obscured. This infrastructure-centric only view creates several problems:
- Cognitive overload: Operators must mentally reconstruct business flows from technical components
- Hidden dependencies: Functional relationships are obscured by deployment boundaries
- Brittle documentation: Infrastructure changes constantly, making documentation obsolete
- Poor incident response: Infrastructure alerts don’t directly map to business impact
- Unclear ownership: Technical boundaries rarely align with functional responsibilities
HFM restores the line of sight by organizing operations around holons (units of purpose), flow (the movement of work), stages (the maturation of work), and boundaries (the areas we can observe and control). In practice, we start with capability health and only then move on to infrastructure. Here, the risk to outcomes is prioritized.

Every holon possesses inlets (how work arrives) and outlets (how work leaves), serving as the points of entry and exit for incoming and outgoing information. Boundaries act as the natural control surfaces, delineating the narrow intersections where traffic, decisions, and contracts converge. The fundamental principle is to explicitly define and label these boundaries, treating them as primary operational entities.
A Visual Language for Flow
HFM complements its conceptual abstraction with a specific visual language designed to represent the system’s functional anatomy. This language isn’t intended merely as static documentation but as a dynamic, operational tool for system comprehension and diagnosis. The core elements of this visual grammar include:
- Holons: Depicted as nested containers representing the functional capabilities.
- Channels: Pathways for work entering and exiting the overall system or a parent holon.
- Inlets / Outlets: Boundary points through which work flows between holons within the system.
The hierarchical and fractal nature of the language ensures each sub-holon possesses the same standardized elements as its parent. This recursive structure facilitates consistent operations across various scales, enabling seamless transitions from a broad overview to a detailed subcomponent. It also allows for the clean composition of novel capabilities and the substitutability of holons without compromising the external interface.
Observing at the Boundary
The model’s operational framework is entirely dependent on a systematic approach to instrumentation. The strategy mandates that observability be implemented at specific, strategic locations within the architecture:
- Holon: A holon must emit signals about its own state and the business events it processes.
- Inlet/Outlet: The boundary between every holon must be instrumented to capture motion.
The specific data points required include:
- Request and response counts, and payload sizes.
- Error rates, categorized by type (e.g., transient vs. terminal).
- Queue depths, wait times, and processing times for asynchronous flows.
- Internal state transitions and invariant checks.
Most observability stacks aim to minimize complexity by employing sampling or collapsing detail into heatmaps. HFM, on the other hand, reduces complexity by selectively counting relevant data points. While calls may be intriguing, custody changes hold decisive significance. By measuring only at inlets and outlets, the model achieves conservation: arrivals minus departures equals the change in work-in-progress. From this singular invariant, the operator derives four crucial signals that govern a living system—throughput in, throughput out, work-in-progress, and lead time—computed at any level of aggregation that’s currently beneficial.
To provide deeper insight into how HFM operates, it’s essential to clarify that the model is built on counting boundary crossings rather than tracking individual work. This approach provides two powerful modes of analysis: paired flow for simple interactions and correlated flow for complex transformations.
In a synchronous RPC, the inlet and outlet are conceptually paired and often share the same name. A request increments the inlet counter, and the corresponding response increments the outlet counter. This one-to-one pairing allows direct measurement of classic flow metrics like throughput, work-in-progress, and lead time. The capability’s health can be understood by comparing arrival and departure rates over time.
The model’s true power emerges in more complex, asynchronous scenarios involving transformations like fan-out (one job splitting into many) or fan-in (many results being merged). In these cases, a simple 1:1 count between a single inlet and outlet is no longer meaningful for the parent holon. Here, the focus shifts from tracking individual transit time to analyzing the correlation between flow rates. The system is less concerned with “how long did job A take?” and more interested in questions like:
- Throughput: What’s the dynamic relationship between the rate of arrivals at one inlet to the rates at outlets.
- Correlation: Can we ascertain for every emission received at inlet X, there’ll be N emissions at outlet Y?
In this mode, HFM acts as a tool for understanding system dynamics. It measures the rates and ratios between interconnected boundaries to reveal the system’s operational patterns. This allows operators to move from simple counting to predictive analysis based on historical performance, identifying bottlenecks or deviations from expected behavior in complex, multi-stage processes. The model’s strength is its ability to quantify the relationship between a stimulus at one boundary and the expected response at one or more others.
Crucially, the model promotes natural aggregation. Initially, each holon has one inlet and one outlet, regardless of the actual number of endpoints or topics. This approach allows for immediate identification of work accumulation, drainage, or starvation. When a trend or anomaly emerges, the model determines the appropriate level of structure to reveal. This can involve dividing the boundary into multiple channels or creating a sub-holon to conceal contention. The model can collapse again once the question has been answered. By controlling the granularity, the model ensures that the mathematical calculations remain accurate, as only boundary crossings are counted.
HFM diverges from distributed tracing as an operational paradigm. While tracing elucidates the sequence of an invocation, HFM provides a more granular view of the actual execution of work. Traces exhibit fragility at critical points, such as asynchronous hops, retries, human intervention, and batch windows, precisely because call causality diminishes in these areas. Conversely, HFM signals remain unaffected. Sampling can’t obliterate a boundary crossing, and context propagation anomalies can’t disrupt conservation. Even in the absence of spans within a legacy domain, boundary measurements still yield decision-grade signals.
Flow Control
Due to the halon’s public face being its boundary, control should also reside there. Admission and rate limiting are no longer arbitrary throttles; they’re policy instruments linked to backlog age and drain time. When a class of work begins to age beyond its target, the holon can defer low-priority arrivals, shape traffic into batches, or allocate scarce resources on a time-sliced basis. When outflow lags, the holon can scale workers if capacity is available; otherwise, it can declare itself degraded and request upstream holons to reduce the influx of work. These aren’t vague SRE rituals; they’re direct manipulations of the same variables that are observed.
Inside the holon, adaptive control valves become surgical. Circuit breakers guard fragile dependencies and contain the blast radius. Fallbacks activate degraded modes that preserve flow against a partial loss of fidelity. Scaling policies follow flow instead of CPU graphs: they look at WIP and backlog age, not just host utilization. Sagas and compensations are first-class, with their own signals, because reversibility is part of keeping global flow coherent when local steps fail. Every actuation emits its own sign—what changed, when, why—so the control surface is observable too.
The control loop is straightforward and localized: measure at the boundary, assess the state (stable, converging, degraded), make a decision based on policy, take action, and signal the outcome. These localized states then aggregate holarchically to form a global posture.
Comparison to Infrastructure-Centric Models
HFM shifts from static inventories to a functional ecosystem view, prioritizing flow, capabilities, and value delivery over structure, components, and technical detail. This aligns with modern systems theory and business objectives, improving understanding, agility, and reliability. While traditional models are necessary for low-level work, they’re insufficient for grasping systemic behaviors and business impact quickly. The holonic model complements existing infrastructure maps, organizing knowledge of servers and microservices into a more meaningful hierarchy.
The following is comparative analysis covering both theoretical and practical aspects of holonic and traditional infrastructure models.
Level of Abstraction: Traditional diagrams (e.g., network topologies, deployment architecture) operate at the level of machines, services, and technical domains. HFM operates at the level of business capabilities and workflows. The theoretical advantage is cognitive: humans understand and reason better about systems described in terms of real-world functions and flows. The holonic approach scales with complexity. Traditional models lead to cognitive overload as operators mentally reconstruct business processes. Holonic models reduce this load by presenting the end-to-end flow in one view, making complex interactions explicit. This clarity improves on-call problem solving by allowing operators to see which holon (capability) is failing and its impact on upstream/downstream holons, instead of chasing low-level logs across dozens of services.
Handling of Dependencies: Service maps often obscure functional dependencies with physical ones. For instance, you might know Service A calls Service B, but you might not see this as part of fulfilling a “customer order” without business context. HFM groups relevant components and shows the flow of work, making dependencies and bottlenecks visible at the business level. In traditional views, dependencies can be hidden or misaligned, such as two microservices being part of the same transaction but owned by different teams, with no indication of this coupling in the infra diagram. This can lead to missed impacts or hidden dependencies that surprise operators. The holonic approach surfaces these by drawing boundaries around functionality that might cross technical boundaries, revealing inter-service relationships in terms of the work flowing between them.
Adaptability to Change: Infrastructure-centric documentation is often brittle because diagrams get outdated when infrastructure changes. Business capabilities, however, tend to be more stable, with system purposes changing less frequently than implementations. For example, a “Payment Processing” capability might remain even if you refactor from one payment service to three microservices. A holonic model is more future-proof and requires less churn in docs. It explicitly encourages drawing boundaries around long-lived capabilities rather than current implementations. Traditional models require continuous updates, losing relevance if they don’t. The holonic model acts as a stable backbone, with occasional updates to the mapping of infrastructure to holons, while the top-level picture remains consistent, easing maintenance of documentation.
Incident Response and Monitoring: The biggest practical difference lies in how outages and performance issues are understood. In a traditional setup, alerts come from infrastructure. Operators must manually connect these to business impact, which can slow response or lead to poor prioritization. With HFM, monitoring aligns to capabilities, such as “Checkout (Holon) degraded – success rate down to 80%” or “Payments (Holon) experiencing backpressure”. These alerts directly indicate which business functionality is hurting, allowing a faster and more appropriate response. Traditional models alert on technical symptoms, while holonic models alert on functional impact, improving operations and clarifying ownership.
Cultural and Organizational Alignment: Traditional infrastructure-centric views often lead to siloed teams (e.g., DBA, Frontend teams), each focusing on their component. The holonic approach aligns with cross-functional, product-aligned teams, which is more in tune with Agile/DevOps. While not purely technical, this affects deployment. Bounded Contexts in Domain-Driven Design correspond to business capabilities, ideally owned by one team. The holon is analogous to a bounded context, a cohesive unit of function. Every bounded context should align closely with a specific business function for coherence. The holonic model enforces this, unlike traditional models that break bounded contexts across diagrams and owners. Companies that transitioned from component teams to feature teams experienced better flow and fewer handoffs. Similarly, moving from infra-models to holonic models could reduce gaps between development and operations by providing a common, capability-focused language.
Complexity Management: Traditional architecture diagrams are either too high-level (simplistic boxes-and-arrows) or too low-level (sprawling network graphs). HFM offers a fractal visual language that zooms in on holons to see sub-flows. This approach allows for different granularities without switching notations, unlike traditional methods that use separate diagrams for different levels. The holonic model’s unified visuals facilitate consistent drilling down from symptoms to causes, addressing the integration of perspectives. It provides a single-pane-of-glass view where a VP can see high-level capability health and an SRE can zoom into specific sub-holon signals and status without context or tool switching.
Closing
HFM presents a compelling and theoretically grounded reframing of system operations. It adopts Beer’s VSM’s recursive viability, Complex Adaptive Systems’ emergent behavior, and Living Systems Theory’s hierarchical regulation principles. These principles inform the model’s design, such as holons with autonomy, local interactions, and feedback loops. Conceptually, it uses holons as part-whole systems, leveraging their power to resolve modern IT system complexity, much like in manufacturing, multi-agent systems, and organizational design. This lends the model a solid philosophical foundation and places it in lineage of holonic architectures that have demonstrated benefits like scalability, robustness, and control.
From a practical perspective, the fundamental concepts of HFM aren’t only valid but also becoming increasingly essential. In today’s complex real-world systems, traditional, infrastructure-centric approaches fall short. They lead to cognitive overload, missed insights, and sluggish responses. In contrast, focusing on the flow of work and capabilities aligns with the evolving practices of DevOps and SRE teams. These teams are adopting strategies such as monitoring business transactions, mapping user journeys, and aligning teams with features.
HFM’s minimal edge design requires only boundary counts (arrivals and completions) to produce decision-grade signals (inflow, outflow, WIP, lead/age). This makes adoption incremental and tool-agnostic. Organizations can start with a single holon (one inlet and one outlet) using existing gateway and queue metrics without changing service code or tracing. Granularity is optional: name and pair ports for synchronous paths, declare split/merge stages where counts aren’t 1:1, and add boundary policies (admission by class, shaping) tied to backlog age. In practice, this boundary-only entry point provides situational awareness that bottom-up stacks lack, while preserving distributed tracing as a complementary microscope inside a holon.
Appendix A: Holon Health States
Each holon classifies its state using functional signals (throughput, success rate, invariants):
- Converging – recovering toward target behavior
- Stable – within normal bounds
- Diverging – trending away from stability
- Erratic – high variance, unpredictable behavior
- Degraded – functioning with reduced fidelity/capacity
- Defective – failing to meet critical invariants
- Down – not processing work
These states drive boundary‑level controls (admission, rate, batching) and internal responses (circuit breakers, fallbacks, reversible compensations), with every actuation producing its own telemetry for closed‑loop governance.
Appendix B: What to Instrument
- Counts, latencies, and payload sizes at inlets/outlets
- Error rates by type (transient vs terminal)
- Queue depths, wait times, processing times (async)
- Contract validation success/failure at inlets
- Business‑event emissions and internal state transitions
- Explicit invariant checks that gate “healthy” work
Collecting these at boundaries gives you conservation, consistent roll‑ups, and a reliable basis for health and control—without exhaustively tracing every internal hop.
Appendix C: Example Boundary Points
- Service Calls
- Amazon CloudFront
- AWS API Gateway
- Application/Network Load Balancers
- Kubernetes Ingress
- Service Mesh Gateways & Sidecars
- Reverse Proxies / API Platforms
- Messaging
- Amazon EventBridge
- Amazon SQS
- Amazon SNS .
- Amazon MSK (Kafka)
- Amazon Kinesis
- Amazon MQ
- Orchestration
- AWS Lambda
- AWS Step Functions
- AWS Batch
- ECS/EKS Workers