Jan Bosch is a research center director, professor, consultant and angel investor in startups. You can contact him at jan@janbosch.com.

Opinion

Becoming an AI-first software-intensive company

Reading time: 5 minutes

For software-intensive systems companies, the opportunity isn’t to add intelligence to existing products and processes, but to reimagine the company itself as an intelligent system.

Over the past decades, software-intensive systems companies have gone through several major paradigm shifts. We moved from hardware-centric products to software-defined systems, from waterfall to Agile, from projects to products and now superset platforms, and from episodic releases to continuous deployment. When each of these shifts became relevant, it was initially resisted and, intentionally or unintentionally, misunderstood, becoming obvious in hindsight. Today, we’re at the beginning of another such shift: the transition to becoming an AI-first company.

This isn’t about using AI or adding AI features to existing products. Many companies already do that. AI-first means something more fundamental: every function, feature, process and interaction is designed from the assumption that intelligence is available and should be used. Just as “software-first” once meant that hardware design had to adapt to software realities, “AI-first” means that organizational structures, system architectures and decision-making models must adapt to intelligence being pervasive.

In my experience, most organizations I work and interact with today still treat AI as an add-on. A separate team explores use cases, pilots a few models and occasionally deploys something interesting as an extension to an existing product or as a standalone offering. The core of the organization, including development processes, deployment pipelines, customer interactions, operational decision-making and so on, remains largely unchanged.

An AI-first mindset turns this upside down. Instead of taking the perspective of “adding AI,” the question becomes more revolutionary and we focus on redesigning from scratch. This applies to product functionality, internal tools, testing strategies, deployment decisions, customer support, pricing and even governance. The key question is how we can automate activities using AI that we couldn’t with algorithmic approaches. However, we should really focus on a zero-based thinking approach where we return to first principles and start the design from there.

For example, why would a development activity rely on static rules or manual reviews if an AI system can continuously assess code quality, architectural compliance, security risks and technical-debt trends? Why would deployment policies be fixed if an AI system can dynamically balance risk, performance and customer impact in real-time? Why would customer interactions follow predefined scripts if systems can adapt their behavior based on context, history and intent? AI-first isn’t a technology choice; it’s a design principle.

The real promise of AI in software-intensive systems isn’t efficiency, but rather its ability to create new value. Intelligent systems can deliver fundamentally better results by learning from usage, context and outcomes. At the individual customer level, this enables deep customization. Systems can adapt behavior, interfaces, performance characteristics and even business logic to optimize for specific goals. Reinforcement learning approaches are particularly powerful here, allowing systems to continuously experiment, learn and improve based on real-world feedback rather than predefined assumptions.

At the same time, intelligence can operate at the fleet or ecosystem level. By learning across large populations of deployed systems, companies can optimize globally. Federated learning and related approaches allow models to improve using distributed data without centralizing it, enabling insights that no single system instance could achieve on its own. This allows companies to learn from a fleet of products without violating regulations such as the EU data and AI acts.

This combination of local adaptation and global optimization is a step change compared to traditional software. It allows companies to move from delivering features to delivering outcomes, from selling products to continuously improving customer value over time.

Of course, as AI allows for automating a variety of business processes, including decision processes, many decisions in AI-first companies will no longer be made by humans. In fact, these decisions shouldn’t be made by humans, because models will be able to do a better job.

In most organizations, decision-making is still heavily manual. Engineers decide when to release, operators decide how to respond to incidents, managers decide how to allocate resources and experts decide how systems should behave in edge cases. This isn’t because humans are particularly good at these decisions, but because historically there was no alternative.

AI changes this equation. Wherever decisions are frequent, data-rich and time-critical, machines will outperform humans, obviously provided they’re given clear intent and well-designed guardrails. The role of humans shifts from making individual decisions to defining objectives, constraints and acceptable risk boundaries.

This doesn’t reduce human responsibility; it changes its nature. Humans remain accountable for outcomes, but they express their intent at a higher level: what should be optimized, what trade-offs are acceptable and what values must never be violated. AI systems then operate continuously within these boundaries, making thousands or millions of micro-decisions that would be impossible for humans to manage.

This is a profound organizational shift. It challenges traditional notions of control, authority and expertise. But it’s also unavoidable if companies want to operate systems that are too complex, dynamic and interconnected for manual oversight.

Becoming AI-first isn’t something that can be delegated to an innovation lab or a single department. It affects architecture, data infrastructure, skills, governance and culture. Architecturally, systems must be designed to observe themselves, learn from outcomes and adapt safely. Data becomes a first-class product, not a by-product. Models must be treated as evolving system components, with lifecycle management, monitoring and continuous improvement.

Organizationally, roles and incentives need to change. Teams need to be comfortable with probabilistic behavior, continuous learning and systems that don’t behave exactly the same way twice. Leadership must accept that the illusion of predictability that many use is replaced by managed uncertainty and that progress comes from learning loops rather than upfront certainty. Most importantly, companies must be willing to rethink long-held assumptions about how value is created and controlled.

Just as with Agile and continuous delivery, becoming AI-first is currently a strategic choice. You can decide to experiment, learn and gradually transform, but you can also decide to ignore this for now. But history suggests that this window won’t remain open forever. As intelligent systems begin to consistently outperform traditional approaches, delivering more value, adapting faster and operating more efficiently, the question will shift from “Should we become AI-first?” to “How do we become AI-first?” And for the companies that start the shift late, it will be a struggle to change fast enough to stay competitive or even in business.

For software-intensive systems companies, the opportunity isn’t to add intelligence to existing products and processes, but to reimagine the company itself as an intelligent system – one that senses, learns, decides and evolves continuously. Those who make that shift deliberately will shape their future. Those who don’t will still experience the shift, just not on their own terms. To end with Norbert Wiener, “The world of the future will be an ever more demanding struggle against the limitations of our intelligence.”

Related content