iCentric Insights Insight

AI Company: What They Do & How to Choose One

What an AI company is, the types that exist, how they're structured, and how to pick the right AI partner for your business. A practical UK guide.

May 15, 2026
ai company
AI Company: What They Do & How to Choose One

Search "ai company" on Google and the first page is a wall of lists: Forbes ranking the world's most promising labs, Wikipedia cataloguing hundreds of vendors by region, US News flagging stocks to watch. Useful if you want a leaderboard. Less useful if you actually need to commission AI work, partner with a credible firm, or understand where your own organisation sits in the value chain.

This guide takes the opposite approach. Instead of ranking 50 names you'll forget by next week, it explains what an AI company really is, the different shapes they take, how they make money, what they deliver, and — most importantly — how to choose one that won't leave you stranded in pilot purgatory. It's written for buyers, founders and operators in the UK who want a working mental model, not marketing fluff.

What an AI company actually is

The phrase "AI company" gets used so loosely that it has almost lost meaning. A consumer app that wraps a single API call to GPT calls itself an AI company. So does Nvidia, whose chips power the data centres that train frontier models. So does a 200-year-old bank that has hired a head of AI. They are all, in some defensible sense, AI companies — but they are clearly not the same kind of business.

A workable definition is this: an AI company is an organisation whose core economic value depends on machine learning systems that it either builds, deploys, or productises. The emphasis is on "core". If you removed the AI, would the company still have a product? If yes, it's an AI-enabled business. If no, it's an AI-native one.

That distinction matters because it shapes everything downstream — how the company hires, how it sells, how it manages risk, and how much you should trust its claims. AI-native firms typically employ research engineers, run their own evaluation infrastructure, and treat model performance as a product KPI. AI-enabled firms treat AI as a feature, usually outsourced to a foundation model provider or a specialist agency.

It's also useful to think in three stacked layers:

  • Infrastructure layer. Chips, data centres, training clusters, the deep plumbing. Nvidia, AMD, the hyperscalers, specialist GPU clouds.
  • Model layer. The labs that train foundation models — OpenAI, Anthropic, Google DeepMind, Meta, Mistral, Cohere, Stability — and the increasingly important open-weight community.
  • Application layer. Everything that turns models into something a user or business actually pays for. Vertical SaaS, copilots, agentic systems, AI-powered services.

When someone says "AI company", they almost always mean the application layer, sometimes the model layer, rarely the infrastructure layer. The economics, talent profile and competitive dynamics differ sharply between them.

Finally, watch for AI-washing. Slapping a chatbot on a marketing site, or adding "AI" to a deck, doesn't reshape a business model. Signals that a company is genuinely AI-first include published evaluation methodologies, a measurable share of engineering effort on ML pipelines, named technical leadership with deep ML credentials, and a willingness to talk about failure modes — not just demos.

The different types of AI companies

Once you accept that "AI company" is a spectrum, it helps to categorise the field. These categories overlap, but they describe meaningfully different businesses with different procurement profiles.

1. Foundation model labs. These are the headline names: OpenAI, Anthropic, Google DeepMind, Meta AI, Mistral, Cohere, xAI, Stability, Black Forest Labs. They train large general-purpose models, license them via API, and increasingly sell direct enterprise products. Capital intensive, research-heavy, and concentrated in a handful of cities.

2. Infrastructure and chip companies. Nvidia is the obvious one, but it also includes AMD, Intel's Gaudi line, Cerebras, Groq, SambaNova, and the GPU specialist clouds like CoreWeave, Lambda, Crusoe and Nebius. The hyperscalers — AWS, Azure, Google Cloud, Oracle — also sit here, packaging compute with managed model services.

3. Vertical AI product companies. These pick a sector and build deep. Examples include Harvey and Robin AI in legal, Tempus and Hinge Health in healthtech, Glean and Sana in enterprise search and learning, Wayve in autonomous driving. They typically combine fine-tuned models with proprietary data and workflow software.

4. Horizontal AI tooling. The picks-and-shovels layer for everyone else building AI. Vector databases (Pinecone, Weaviate, Qdrant), orchestration frameworks (LangChain, LlamaIndex), evaluation platforms (Braintrust, LangSmith, Arize), inference platforms (Modal, Replicate, Together), guardrail vendors (Lakera, Protect AI). Buyers here are mostly developers.

5. AI consultancies and systems integrators. Big Four firms, Accenture, Capgemini, BCG X, McKinsey QuantumBlack, Slalom — plus specialist boutiques. They sell advisory, transformation and large delivery programmes. Strong on change management, weaker on cutting-edge engineering.

6. AI agencies. Smaller, design-led or engineering-led firms that build production systems end-to-end. They typically blend AI with web, product and brand work. iCentric Agency sits in this category. Faster than consultancies, more bespoke than SaaS, more accountable than freelancers.

7. In-house AI teams. Increasingly, large enterprises spin up internal "AI factories" or "intelligence platforms" with their own product names. JPMorgan's LLM Suite and Goldman Sachs' GS AI are examples. They aren't AI companies in the commercial sense, but they compete for the same talent and they shape buyer expectations.

A single procurement decision often touches several of these categories at once. You might license a foundation model from a lab, host it on a hyperscaler, deploy via an open-source orchestration framework, hire a boutique agency to build the application, and bring in a consultancy to handle change management. Understanding who does what stops you paying twice for the same capability.

How AI companies make money

The revenue model tells you more about an AI company than its homepage ever will. There are five dominant patterns, and most firms blend two or three.

Usage-based pricing. Tokens, API calls, minutes of audio, images generated. Frontier labs and inference platforms rely on this almost exclusively. It aligns cost with value, but it makes budgeting hard for buyers and exposes vendors to compute swings. Expect this to remain the default at the model and tooling layers.

Seat-based SaaS. Familiar to anyone who buys software. Copilots like Microsoft 365 Copilot, GitHub Copilot, Glean, Notion AI and ChatGPT Enterprise all charge per user per month. Predictable for buyers, but margins compress as inference costs swing.

Outcome-based contracts. Increasingly common in services. The AI company gets paid a share of measurable value created — collections recovered, hours saved, conversion lift, claims processed. Harder to structure, but it builds genuine alignment. Boutique agencies and some vertical AI vendors are leaning into this.

Licensing and enterprise agreements. Large, multi-year contracts that bundle model access, professional services, security guarantees and roadmap influence. The hyperscalers and frontier labs all sell this way to FTSE 100 and Fortune 500 buyers.

Services revenue. Discovery sprints, builds, MLOps, evaluation, training. Most consultancies and agencies make the majority of their money here. The healthier ones layer recurring managed services on top of one-off builds.

Margin profiles vary enormously. A frontier model lab spends billions on training before charging a single user. A vertical SaaS vendor can hit software-like gross margins once the model is stable. An agency lives on utilisation and project margin. If you're picking a partner, ask how they make money — and how they expect you to make money — before you ask about technology.

The global AI company landscape

The reason the SERP for "ai company" is dominated by lists is simple: nobody has settled on a single canonical taxonomy, so journalists keep ranking the field. Forbes' AI 50, CB Insights' AI 100, Wikipedia's exhaustive list, US News' investing angle — each cuts the market differently.

A few patterns are stable across all of them.

Geographic concentration. The US still dominates by capital and headline talent. San Francisco and the wider Bay Area host the largest cluster of labs and tooling vendors. New York leads in financial AI and media applications. Seattle anchors the hyperscaler ecosystem around AWS and Microsoft.

China's parallel stack. Baidu, Alibaba, Tencent, ByteDance, Huawei, Zhipu, Moonshot, DeepSeek and SenseTime form a separate ecosystem with its own models, regulators and supply chain. For UK buyers, the relevance is mostly indirect — supply chain exposure, geopolitical risk, and the open-weight releases (DeepSeek, Qwen) that have changed cost expectations across the industry.

Europe's specialists. Mistral in France, DeepL in Germany, Aleph Alpha for regulated workloads, Synthesia and ElevenLabs in media, Stability and Wayve in the UK. Europe has fewer foundation labs but a strong applied layer, often built around sectoral expertise.

Sovereign AI. Governments increasingly want compute, data and models hosted within their own borders. The UK's AI Safety Institute, France's commitment to Mistral, Germany's industrial AI investments, and the Gulf states' aggressive infrastructure spending all reshape vendor shortlists. Procurement teams should ask where models are trained, where they're hosted, and which jurisdiction owns the logs.

If you do need a starter list of names — for benchmarking, competitive intelligence or simply curiosity — the Forbes AI 50, the CB Insights AI 100, and the LinkedIn Top Startups list together cover most credible firms. Wikipedia's list is broader and useful for the long tail. Treat any single ranking as a starting point, not a shortlist.

The UK AI company ecosystem

The UK punches well above its weight in AI. London hosts Google DeepMind, the lab that effectively created the modern deep-learning playbook. Wayve has become the most credible end-to-end autonomous driving company outside Tesla and Waymo. Stability AI, ElevenLabs (Polish-British roots), Synthesia, Faculty, PolyAI, Builder.ai, Tractable, Cleo and Multiverse all have meaningful UK presence.

The ecosystem clusters around three poles:

  • London for capital, enterprise customers, and applied AI in financial services, legal, media and creative industries.
  • Cambridge for hardware, biotech AI, and deep research spinouts from the university and the Wellcome Trust ecosystem.
  • Edinburgh and the wider Scottish belt for NLP, robotics and language technology, building on long-standing university strengths.
  • Oxford for biomedical AI, robotics and a steady pipeline of research spinouts.

On the regulatory side, UK buyers benefit from a relatively pragmatic stance. The Information Commissioner's Office (ICO) has published clear guidance on automated decision-making, DPIAs and lawful basis for AI. The Competition and Markets Authority (CMA) has set out principles for foundation model markets. The AI Safety Institute, the first of its kind globally, runs evaluations on frontier models before public release. The forthcoming AI Bill is expected to be principles-based rather than rules-heavy, in deliberate contrast to the EU AI Act.

For procurement, the practical implications are:

  • Data residency. Insist on UK or EU hosting where personal or regulated data is involved, and confirm where training data and prompt logs live.
  • Onshore support. A UK-based AI company means shared timezone, contract law you understand, and accountable named individuals rather than offshore ticket queues.
  • Sector regulators. The FCA, MHRA, Ofcom and others have their own AI expectations. A credible UK AI partner should be fluent in the rules that apply to your sector, not just GDPR.
  • Standards. Look for Cyber Essentials Plus as a floor, ISO 27001 for serious data handling, and SOC 2 if you operate across the Atlantic. ISO 42001, the new AI management system standard, is rapidly becoming a useful differentiator.

Notable UK AI agencies and applied firms worth knowing alongside the bigger product companies include Faculty, Mind Foundry, ContactEngine (now NICE), Causaly, Quantexa, Peak, Featurespace, and a long tail of design-led studios like iCentric Agency that combine AI engineering with brand, web and product work.

Services a modern AI company delivers

If you strip away marketing language, almost every credible AI company sells some combination of the following services. Knowing the menu helps you write a proper brief.

AI strategy and opportunity mapping. A structured workshop programme that audits where AI can move the needle, prioritises by value and feasibility, and produces a sequenced roadmap. The best engagements deliver business cases for two or three concrete use cases, not 40 PowerPoint ideas.

Data readiness and governance. Most AI projects stall on data, not models. Good vendors will assess data quality, lineage, access controls, retention, and the legal basis for use. They will design pipelines that survive contact with production: schema management, drift detection, PII handling, and clean separation between training and serving data.

Custom model work. Increasingly rare to train a foundation model from scratch — and rightly so. But fine-tuning, distillation, retrieval-augmented generation, and prompt engineering are bread and butter. So is choosing between frontier closed models (GPT, Claude, Gemini) and open-weight options (Llama, Mistral, Qwen, DeepSeek, Gemma) based on cost, performance and data sensitivity.

Agentic system design. Multi-step systems that plan, call tools, and act. This is where most of the interesting recent progress sits — and where most of the failures happen. A good AI company treats agent design as a software engineering discipline with explicit state, observable steps, and bounded autonomy, not as prompt-stuffing.

MLOps and lifecycle management. Deployment, monitoring, retraining triggers, versioning, rollback, evaluation in production. The boring infrastructure that decides whether your AI product is reliable enough to bet on. Expect serious vendors to talk about evaluation pipelines, regression suites and canary rollouts the way mature software teams talk about CI/CD.

Front-end and product experience. AI is only as useful as the interface that delivers it. The best AI companies treat UI and UX as first-class — designing for trust, transparency, fallback and human-in-the-loop. Agencies like iCentric, which combine design, product and AI engineering under one roof, have a structural advantage here.

Change management, training and adoption. Even the best system fails if users don't trust or use it. Adoption work — playbooks, training, success metrics, internal champions — is often the difference between a celebrated POC and a sustained business impact.

The tech stack inside an AI company

You don't need to be technical to procure AI, but a baseline understanding of the stack saves money and protects you from vendor theatre. Most production systems combine the following layers.

Model layer. Frontier closed models — OpenAI's GPT family, Anthropic's Claude, Google's Gemini — dominate by raw capability. Open-weight models — Meta's Llama, Mistral, Qwen, DeepSeek, Gemma — close the gap fast on most enterprise tasks and offer dramatically more flexibility. Specialised models exist for embedding (OpenAI, Cohere, Voyage), reranking (Cohere, Jina), speech (Whisper, ElevenLabs, Deepgram), vision (CLIP, SAM, GPT-4o vision) and code (Claude, GPT, Codestral).

Orchestration. LangChain, LlamaIndex and Microsoft's Semantic Kernel are the most common frameworks. Many serious teams now build thin custom orchestration on top, because heavy frameworks can hide control flow that you'll later need to debug. Expect a credible vendor to explain why they chose what they did.

Retrieval and search. Vector databases — Pinecone, Weaviate, Qdrant, Milvus — and increasingly Postgres with pgvector for teams that don't want a new piece of infrastructure. Elastic, Vespa and Typesense remain strong for hybrid lexical-and-semantic search. Good RAG systems combine all three.

Evaluation. This is the single biggest differentiator between mature and immature AI companies. Tools like Ragas, LangSmith, Braintrust, Arize, Humanloop and Weights & Biases let teams measure performance on real tasks, not vibes. If a vendor can't show you their eval harness, walk away.

Inference and hosting. Hyperscaler managed services — AWS Bedrock, Azure AI Foundry, Google Vertex AI — for enterprise-grade compliance. Specialist inference platforms — Modal, Replicate, Together, Fireworks, Anyscale, Baseten — for cost-efficient open-weight serving. Self-hosting on GPUs remains common for sensitive data.

Observability and guardrails. Prompt and response logging, PII redaction, jailbreak detection, output validation. Lakera, Protect AI, NeMo Guardrails, Guardrails AI, plus general-purpose APM tools like Datadog and New Relic that have added LLM-specific features.

Data and feature platforms. Snowflake, Databricks, BigQuery for warehousing. dbt for transformation. Airbyte and Fivetran for ingestion. Feature stores like Tecton or Feast where classical ML still matters.

When you're evaluating an AI company, ask them to draw the stack they'd build for your problem on a whiteboard. The clarity and confidence of that answer tells you almost everything.

AI agencies vs consultancies vs product vendors

Procurement teams routinely confuse these three. They overlap, but they are fundamentally different businesses.

Tier-one consultancies — McKinsey, BCG, Bain, Deloitte, EY, KPMG, PwC, Accenture, Capgemini — sell strategy, transformation and large-scale delivery. They are excellent at navigating board politics, managing change, and running multi-hundred-person programmes. They are typically slower, more expensive, and reliant on partner ecosystems for cutting-edge engineering. Use them when the problem is organisational as much as technical.

Boutique AI agencies — iCentric Agency and a small number of peers — sell senior, opinionated delivery. They tend to have flat teams, no offshore handoffs, and the ability to take a problem from a one-line brief to a production system in weeks. They are best for organisations that have decided what they want to build and need it done well. They are not ideal for multi-thousand-headcount transformation programmes.

Product vendors — copilots, vertical SaaS, horizontal tooling — sell predictable software. They are the fastest route to value when the off-the-shelf product is a good fit. They are the worst choice when your differentiation depends on bespoke workflows or proprietary data, because you'll end up reshaping your business to fit their product.

A mature programme often uses all three. The consultancy frames the strategy and runs change. The agency builds the bespoke pieces. The product vendor handles commoditised workflows. Procurement teams that try to make one vendor do all three jobs usually pay more and get less.

How to choose the right AI company for your project

There is no universal best AI company. There is, however, a defensible process for shortlisting one for your project.

Start with the business problem. Not the technology. "We want to reduce average handling time in the contact centre by 20%" beats "we want to use generative AI" every time. Write the problem in one paragraph that a non-technical executive can sign off. If you can't, you aren't ready to procure.

Match vendor type to maturity. If you've never shipped an AI product, you almost certainly want an agency or specialist consultancy, not a frontier lab and a green-field build. If you're a sophisticated buyer with an internal platform team, you may only need point vendors and tooling. Honest self-assessment saves months.

Ask about evaluation before case studies. Anyone can show you a successful deployment. Far fewer vendors can show you the eval harness, the regression test suite, the failure mode taxonomy, and the production incident reviews that sit underneath it. Those artefacts are the truest sign of competence.

Insist on production references. A pilot is not a deployment. Ask for references where the system is handling real users at real volumes, ideally with named contacts you can speak to. Ask what broke and how it was fixed. The vendors who answer cheerfully are the ones to keep.

Stress-test the commercial model. Build a usage forecast that goes 10x, 50x, 100x. Plug it into the vendor's pricing. Many AI engagements look attractive at pilot scale and become unworkable at production scale. The best vendors will run this exercise with you proactively.

Consider lock-in. Who owns the prompts, the fine-tuned weights, the evaluation data, the orchestration code? If the answer is anything other than "you do, with full portability", price that risk in. Good agencies hand over a complete, documented codebase at the end of every engagement.

An evaluation checklist for shortlisting AI vendors

When you reach the RFP or capability deck stage, the following checklist separates the credible from the cosmetic.

  • Named delivery team. Who actually does the work, with bios? Beware the bait-and-switch where senior partners pitch and juniors deliver.
  • Technical credentials. Real engineering CVs, published work, open-source contributions, conference talks. "AI expert" on LinkedIn is not a credential.
  • Evaluation methodology. Documented, repeatable, with examples of how it surfaced problems before launch.
  • Security posture. SOC 2 Type II, ISO 27001, Cyber Essentials Plus as a UK baseline, ISO 42001 as a forward signal. Penetration test reports available under NDA.
  • Data handling. Where is data stored? For how long? Is it used to train models? Can it be deleted on demand? Are prompts and outputs logged, and if so, where?
  • IP terms. Customer owns prompts, fine-tunes, evaluation data and bespoke code. Vendor retains generic frameworks. Anything less is a red flag.
  • Exit plan. Documented handover, code in your repos, secrets in your vaults, observability dashboards transferable. If you can't leave cleanly, you shouldn't sign.
  • Roadmap alignment. Does the vendor's plan for the next 12-18 months match where your problem is heading?
  • Cultural fit. AI delivery is collaborative and iterative. If you can't imagine working with these people every day, the project will be miserable regardless of skill.

Engagement models and how AI work is typically structured

The shape of the contract matters as much as the vendor inside it. Most credible AI companies will offer some combination of the following.

Discovery sprints. A short, fixed-scope engagement — typically two to six weeks — that audits the opportunity, defines success metrics, prototypes the highest-value use case, and produces a build plan. The output is a go/no-go decision with evidence. Good for organisations that are unsure where to start.

Time-and-materials build phases. The default for bespoke product engineering. The vendor commits a team for a defined period against an agreed backlog. Scope flexes, hours are tracked, deliverables are demoed every fortnight. Works well when the problem is well-understood and the unknowns are technical.

Fixed-price builds. Possible but rarer in AI than in classical software, because the unknowns are larger. Where they work, they're usually layered on top of a discovery that has de-risked the architecture.

Outcome-based contracts. The vendor's compensation is tied to a measurable business KPI — calls deflected, hours saved, revenue uplift. Aligns incentives beautifully when the KPI is clean and attributable. Requires a high-trust relationship and unambiguous baselines.

Managed services. Ongoing operations: model monitoring, retraining, evaluation, incident response, gradual capability expansion. The unglamorous work that keeps AI products alive and improving.

Embedded teams and staff augmentation. Senior AI engineers integrated into your team for a fixed period. Works well when you have product leadership but lack ML depth. Less suitable when you need a turnkey delivery.

The healthiest engagement patterns combine a discovery sprint, a build phase, and a managed service tail — with explicit gates between each. Watch for vendors that push straight into multi-year, multi-million-pound transformations without staged commitments.

Industry applications and where AI companies add the most value

Different sectors are at radically different points in the AI adoption curve. Some snapshots of where serious value is being captured.

Financial services. Document-heavy workflows — KYC, AML, credit memo drafting, research summarisation — are being transformed by retrieval-augmented systems. Advisor and analyst copilots are now mainstream at the largest banks. Fraud and AML detection models, decades old, are being supplemented with graph-based and LLM-assisted approaches. Regulatory expectations (FCA, PRA) make evaluation discipline non-negotiable.

Healthcare and life sciences. Medical imaging triage, clinical note summarisation, ambient scribing, trial recruitment, drug discovery target identification, protein structure prediction. The UK has notable strengths via DeepMind's AlphaFold lineage, BenevolentAI, Exscientia and a healthy NHS-aligned ecosystem. MHRA and the NHS AI Lab set the regulatory floor.

Legal. Contract review, e-discovery, drafting assistance, knowledge management. Harvey, Robin AI, Luminance and Spellbook lead a crowded application layer. The biggest gains come from combining LLM drafting with firm-specific precedent retrieval — generic ChatGPT is a non-starter for most firms because of confidentiality and hallucination risk.

Retail and e-commerce. Personalised search, dynamic merchandising, agentic shopping assistants, product content generation, returns prediction, supply chain optimisation. Conversion lift of even a few percentage points pays back fast at scale.

Manufacturing, energy and utilities. Predictive maintenance, vision-based quality control, demand forecasting, grid optimisation, anomaly detection on sensor data. Classical ML still dominates here; generative AI is mostly used internally for knowledge management and engineering productivity.

Media and creative industries. Generative tooling — image, video, audio, copy — is reshaping production pipelines. Synthesia, Runway, ElevenLabs and Adobe's Firefly ecosystem are the headline names. The interesting commercial questions are about rights, attribution and brand control rather than capability.

Public sector. Case management, citizen services, document processing, fraud detection. The UK Government has set up a central AI capability (the Incubator for AI, i.AI) and is rolling out cross-departmental copilots. Procurement is via G-Cloud, CCS frameworks, and Crown Commercial routes; AI vendors that aren't on those frameworks effectively can't sell to the centre.

The pattern across all sectors is consistent: highest ROI sits where there is high-volume, semi-structured work currently done by humans, where errors are tolerable or detectable, and where the data needed already exists in machine-readable form.

Governance, safety and compliance

A serious AI company treats governance as a first-class engineering discipline, not a legal afterthought. The frameworks worth knowing:

EU AI Act. Risk-tiered: unacceptable, high, limited, minimal. High-risk systems (employment, credit, education, critical infrastructure, law enforcement) carry significant compliance obligations including documentation, human oversight, data governance and post-market monitoring. UK companies selling into the EU are in scope even if they aren't EU-established.

UK regime. Principles-based, sector-led, with the ICO, FCA, MHRA, Ofcom, CMA and others applying their own rules. The forthcoming AI Bill is expected to formalise this. The AI Safety Institute provides independent evaluation of frontier models.

ICO guidance. Practical, accessible, and the de facto floor for UK AI deployments. Covers lawful basis for processing, automated decision-making rights under UK GDPR, DPIAs, and bias.

NIST AI Risk Management Framework. US-origin but widely used as a structured way to identify and mitigate AI risks across the lifecycle. Useful as an internal scaffolding even outside the US.

ISO 42001. The first international management system standard for AI. Equivalent in spirit to ISO 27001 for security. Expect this to become a procurement expectation across the enterprise market.

Practical artefacts a credible AI company should produce:

  • Model cards and system cards documenting capabilities, limitations and intended use.
  • Data sheets for training and evaluation datasets.
  • Evaluation reports including bias testing across protected characteristics.
  • Red-team reports showing adversarial testing results.
  • Incident response playbooks with named owners and escalation paths.
  • Audit trails of prompts, outputs and decisions for systems that affect customers.

Governance is not a cost. It is the thing that lets you scale AI in regulated environments without spending the next decade in remediation.

Risks, red flags and common failure modes

The public success stories are loud. The failures, much more numerous, are quiet. The most common patterns to avoid:

Pilot purgatory. Industry surveys consistently put the share of AI proofs-of-concept that fail to reach production at 70-85%. Causes are usually organisational, not technical: no clear owner, no integration plan, no budget for the unglamorous middle 80% of the work.

Demo-driven delivery. A demo is a controlled environment with cherry-picked inputs. Production is the opposite. If a vendor's pitch leans heavily on a single shiny demo and lightly on evaluation methodology, recalibrate.

Single-model dependence. Architectures that hard-wire one foundation model — usually GPT — into every layer become fragile when prices change, terms change, or a better model arrives. Good systems abstract the model behind an interface and re-evaluate quarterly.

Hallucination without grounding. Generative models will confidently invent answers. Production systems need retrieval, citations, validation, and human-in-the-loop where the cost of error is meaningful. Vendors who shrug at hallucinations are not ready for regulated work.

Data leakage. Prompts containing customer data sent to public APIs and logged indefinitely. Fine-tuning datasets that bleed PII into model weights. Test data contaminated with production secrets. All preventable, none rare.

Inference cost surprises. A workflow that costs pennies per task at pilot scale can cost meaningful sums per task at production scale, and large amounts in aggregate. Forecast usage curves before signing.

Hidden human cost. Many "AI" products quietly depend on human reviewers behind the scenes. Sometimes that's fine and explicit. Sometimes it's a liability waiting to happen. Ask.

Vendor concentration. Putting strategy, build, hosting and managed service with a single vendor maximises convenience and lock-in. Spread the layers if you value optionality.

Build, buy or partner: which path fits your organisation

The build-buy-partner decision is older than AI, but AI sharpens it. A workable decision matrix:

Buy when the use case is generic (meeting transcription, internal search, code completion, marketing copy), several mature products exist, your data isn't a meaningful differentiator, and switching costs are modest. Most organisations should buy more and build less than they currently plan to.

Build when the workflow is specific to your business, your data is genuinely proprietary, the off-the-shelf options force you to compromise on differentiation, and you have — or can hire — the engineering depth to maintain the system. Build assumes ongoing investment, not a one-off project.

Partner when you have a clear use case but no internal team yet, or when you want to accelerate the first two or three production systems while a permanent team is hired and grown. The partner does the heavy lifting, transfers knowledge, and exits cleanly. This is the sweet spot for most mid-market organisations.

Hybrid patterns are increasingly common: buy the platform, partner on the workflows, build the differentiated layer. A retailer might buy a generic conversational AI platform, partner with an agency to build merchandising-specific agents on top, and build proprietary models for personalisation in-house.

The worst decision is usually the absent decision — drifting into builds because buying felt embarrassing, or buying expensive platforms that nobody integrates. Make the call deliberately and revisit it annually.

Trends shaping AI companies

A handful of forces are reshaping the AI company landscape, and they will affect any procurement decision made in the next several years.

Agentic systems moving into production. The shift from single-turn chatbots to multi-step agents that plan, call tools and act on behalf of users is well underway. The early production wins are narrow and well-bounded — research, scheduling, code modification, customer operations — rather than the autonomous everything-agents that dominate Twitter. Expect this to keep accelerating.

Smaller, fine-tuned models challenging frontier giants. Open-weight models from Meta, Mistral, Alibaba and DeepSeek now match frontier performance on many enterprise tasks at a fraction of the inference cost. Sophisticated buyers increasingly run a portfolio of models, routing tasks by complexity. Frontier labs will retain the lead on the hardest tasks; everything else commoditises.

Multimodal as default. Voice, image, video and document inputs are converging into the same models. Interfaces are following: AI copilots that see your screen, hear your meetings, read your documents and respond in any modality. This raises the bar for design and privacy alike.

Vertical copilots replacing horizontal chatbots. The first wave of generative AI was general-purpose. The second is sector-specific: legal, medical, financial, engineering copilots that combine fine-tuned models with sector data and workflows. Generic chatbots are losing ground inside the enterprise.

Sovereign and onshore deployments. Regulated buyers, especially in the UK and EU, are demanding hosting, training data and operational support within their jurisdiction. AI companies that can offer credible onshore deployments are winning the procurement battles, even when their headline benchmarks lag.

Tooling consolidation. The Cambrian explosion of LLM tooling — orchestration, eval, observability, vector databases — is consolidating. Expect the strongest players in each category to acquire or out-distribute the rest. Buyers should bias toward portable standards (OpenTelemetry, OpenAI-compatible APIs, vendor-neutral eval frameworks) over proprietary lock-in.

Workforce reshaping. AI is changing how AI companies themselves work. Engineering teams ship faster with AI assistance. Design and content production cycles compress. The economics of running a small, senior agency improve materially — which is one reason boutique AI firms are punching above their weight against tier-one consultancies.

Working with iCentric Agency as your AI company

iCentric Agency is a UK-based AI and digital agency that helps ambitious organisations turn AI from a slide into a working system. We sit deliberately in the boutique agency category described above — senior, opinionated, accountable, and small enough to care about outcomes.

What that means in practice:

  • Senior delivery, no offshore handoffs. The people who scope your project are the people who build it. Every engagement is led by a named technical principal.
  • Pragmatic stack. We use frontier closed models where capability matters, open-weight models where cost and control matter, and we tell you which is which. We are not affiliated with any single foundation model provider.
  • Evaluation-first methodology. Every project starts with the metric that defines success and the harness that measures it. We publish results to you, not vibes.
  • Integrated capability. Because we also run web and product engagements, the AI we build sits inside experiences that look and feel like they belong. This matters more than most procurement decks acknowledge.
  • Right-sized engagements. From week-long opportunity sprints through to multi-quarter builds and ongoing managed AI. We will tell you when the right answer is to buy a SaaS product rather than commission us.
  • UK-native. Cyber Essentials Plus, UK data residency by default, contracts under English law, and a team you can meet in person.

If you want to explore whether iCentric is the right AI company for a specific project, the fastest route is to book an AI opportunity assessment. You'll come out with a prioritised list of opportunities, a recommended approach for the top two, and a clear view of where we add value and where we don't.

Frequently asked questions

What counts as an AI company today? The most useful definition is an organisation whose core economic value depends on machine learning systems that it builds, deploys or productises. That includes foundation model labs, infrastructure providers, vertical AI products, horizontal tooling, and applied AI agencies and consultancies. It excludes companies that have added a chatbot to a marketing site and rebranded.

How is an AI company different from a software company? Most modern AI companies are software companies, but with a distinct economic profile: heavier compute costs, more variable margins, deeper dependence on data quality, and a stronger evaluation culture. The differentiator is whether the product would still function with the AI removed.

Do I need a UK-based AI company specifically? Often, yes. Data residency, contract law, shared timezone, regulatory familiarity (ICO, FCA, MHRA), and proximity for senior workshops all matter for serious deployments. UK buyers in regulated sectors should bias heavily toward UK or EU-based partners.

How long does a typical AI engagement take? A discovery sprint runs two to six weeks. A first production build is usually three to six months. Ongoing managed AI is a multi-year relationship. Transformation programmes can span years, but should be broken into shippable phases of no more than two quarters each.

How do I avoid AI vendor lock-in? Demand portability: own your prompts, your fine-tunes, your evaluation data, your orchestration code. Use vendor-neutral standards where possible. Architect with an abstraction layer between application code and the underlying model. Re-evaluate model choices on a regular cadence rather than treating them as permanent.

What's the difference between an AI agency and an AI consultancy? Consultancies primarily sell advice, strategy and large-scale change. Agencies primarily sell delivery — they build the product. The best engagements often combine both: consultancy-led framing and agency-led build. Confusing the two leads to slides without systems, or systems without organisational adoption.

Can a small business afford to work with an AI company? Yes. The economics have improved sharply as open-weight models, managed inference and agile boutique agencies have matured. Most small businesses should start with a single, well-scoped use case delivered in weeks rather than a sweeping transformation. The payback horizon for the right use case is typically months, not years.

What questions should I ask in a first meeting with an AI company? What is your evaluation methodology? Show me a recent production system you've shipped. Who specifically will lead my engagement? How do you handle hallucinations and data leakage? What does your handover look like at the end of an engagement? Where do you say no? The answers to those six questions will tell you more than any capability deck.

What counts as an AI company today?

An AI company is an organisation whose core economic value depends on machine learning systems that it builds, deploys or productises. That includes foundation model labs, infrastructure providers, vertical AI product firms, horizontal AI tooling vendors, and applied AI agencies and consultancies. It excludes companies that have simply added a chatbot to a marketing site and rebranded themselves accordingly.

How is an AI company different from a software company?

Most modern AI companies are software companies, but with a distinct economic profile: heavier compute costs, more variable margins, deeper dependence on data quality, and a stronger evaluation culture. The clearest test is whether the product would still function if the AI were removed. If the answer is no, you are looking at an AI-native business rather than an AI-enabled one.

Do I need a UK-based AI company specifically?

For serious deployments, often yes. Data residency, English contract law, a shared timezone, and familiarity with UK regulators such as the ICO, FCA and MHRA all matter. UK buyers in regulated sectors should bias heavily toward UK or EU-based partners who can offer onshore hosting and named, accountable individuals rather than offshore ticket queues.

How long does a typical AI engagement take?

A discovery sprint usually runs two to six weeks. A first production build is typically three to six months from kickoff to live deployment. Ongoing managed AI services are multi-year relationships. Larger transformation programmes can span longer, but they should always be broken into shippable phases of no more than two quarters each so value is delivered continuously.

How do I avoid AI vendor lock-in?

Demand portability of every artefact: prompts, fine-tuned weights, evaluation datasets, orchestration code and observability dashboards should all be yours. Use vendor-neutral standards where possible and architect with a clean abstraction layer between application code and the underlying model. Re-evaluate model choices on a regular cadence rather than treating any single provider as a permanent dependency.

What is the difference between an AI agency and an AI consultancy?

Consultancies primarily sell advice, strategy and large-scale change management. Agencies primarily sell delivery: they build the actual product or system. The best programmes often combine both, with consultancy-led framing and agency-led build. Confusing the two leads either to slide decks without working systems, or to systems that nobody in the organisation adopts.

What questions should I ask in a first meeting with an AI company?

Ask about their evaluation methodology, a recent production system they have shipped, who specifically will lead your engagement, how they handle hallucinations and data leakage, and what handover looks like at the end of the engagement. Also ask where they say no. The answers to those questions will reveal far more about competence and cultural fit than any capability deck.

ai company

Get in touch today

Book a call at a time to suit you, or fill out our enquiry form or get in touch using the contact details below

iCentric
May 2026
MONTUEWEDTHUFRISATSUN

How long do you need?

What time works best?

Showing times for 18 May 2026

No slots available for this date