Can Ai Scale?
Updated: Jun 19, 2022
This one should be easy, you'd think. Isn't AI (or rather, Machine Learning) built for scale?
SCALE means different things to different people. This post specifically deals with implications for GROWTH.
As investors look into Ai technology companies, they are interested in how successful they can be, and to that end, a company's ability to scale sales and operation is critical. It is true that Ai applications rely on hoards of data to train. It is also true that Ai applications are justifiable in case there are too much data, and noisy at that, to retrieve meaningful insights. But even if your software churns terabytes of information, it does not mean your business is scalable.
In order for your business to scale, you need an accelerating number of customers, resulting in accelerating revenue.
And how would you do that? The book says: Find a large problem, with as many potential customers vying to solve it.
The Challenges of Scaling Up Ai
But is Ai, as a technology, scalable? The book says: "It depends", so let us analyze what is at play.
When we talk about Ai (Artificial intelligence) we usually mean application of Machine Learning (ML), which Tom Mitchell defined as "the study of computer algorithms that can improve automatically through experience and by the use of data".
Without going into technical details, ML implementations start with a training phase, during which data is processed and results are rated against known results. A simplistic example is scanning thousands of images to decide whether they depict cats or dogs, with humans assessing the results and grading the success or failure of each identification by the machine.
After a while the system, by feeding the successes and failures back, learns to discern between dogs and cats in ever increasing precision. At some point, the error rate equates or falls below human threshold, and this is usually where the training stops, or deemed "good enough".
At this stage, production begins, which is where real data is processed to reveal who's the dog here, and who's the cat.
This strategy is useful when there is a large enough available data to train on, and when the data keeps coming from similar sources, while the research question remains quite similar.
But what happens if there is no training data, when that data is customer specific, or when it belongs to parties that would not share it freely?
Would you be able to train a model, then implement your findings over a large market?
Here stand the challenges for Ai solution providers: Have they found a large and reasonably homogenous market? Do they have an adequate data set to train their system on?
Is this barrier to scale surmountable?
Go To Market Dilemma: Product or Service?
Consider an Ai vendor penetrating a new market, based on their learning in previous successful implementation. Said vendor would have to train their model on available data, and only then they will be able to reach out to new customers with a viable offering.
A case in point is machine translation in particular, and language applications in general. By and large, machine translation is a high volume business in most living languages, with both large datasets to train on and large target audience and expected usage.
The result is a rich set of language applications, ever evolving to domains from Machine translation to Speech To Text, to Text To Speech, Summarization, and beyond.
Is it possible that the universal nature of language processing has a positive impact on the availability of general use language applications? After all, industry specific jargons are easily contained within a language syntax, and mastering new domains is easier when there are available data.
In contrast, we may consider machine vision problems, which tend to be fiendishly specific. It is not trivial to train a system to discern between cats and dogs, when the data fed to it for training does not contain enough of those, or when the input images diverge from certain levels.
Is it not fair to say, then, that machine vision requires more domain specific training?
Scale and Modality
Let's analyze this topic with eye on scalability and growth, as impacted by the go-to-market modality:
Suppose you were to develop a solution to a problem - Will you cater to your customers in professional service (project) mode, building a team, employing your knowhow to deliver a working system in place? Or will you provide a tool (product) which the customers could operate and deploy themselves?
This dilemma has wide repercussions on your business model, on the fees you must charge to recuperate your costs. With professional services, you must deploy highly trained people to solve a single customer's needs. It will invariably reflect on your cost structure, on your pricing, and as a result, on the number of prospective customers.
The professional service model also impacts your throughput in terms of customers / per period, given constraints of your team size, and while your team builds solutions for newly acquired customers, it cannot process new ones.
Compare a hypothetical Growth patterns for professional services (blue), Enterprise software (red), and SaaS (Green):
Each modality has different dynamics, depending on sale cycle, integration period, and capacity constraints.
As the solution matures (on the Project➛Product axis) , its implementation becomes ubiquitous and can be relegated to customers, resulting in higher scale.
Do note, however, that at a given level of internal resources, Professional Services do not scale.
The Tragedy of the Commons and Ai's barriers to Scale
We previously discussed the difference between general purpose problems (language applications) and specialty problems (machine vision applications). This comparison isn't very fair though, and why?