Home News The Case for Decentralizing Your AI Tech Stack

The Case for Decentralizing Your AI Tech Stack

0
The Case for Decentralizing Your AI Tech Stack

A lot of the conversation on AI development has turn out to be dominated by a futuristic and philosophical debate – should we approach general artificial intelligence, where AI will turn out to be advanced enough to perform any task the way in which a human could? Is that even possible?

While the acceleration versus deceleration discussion is essential and timely with advancements just like the Q-star model, other points matter, too. Mainly, the importance of decentralizing your technology stack, and easy methods to try this without making it an excessive amount of of a value burden. These two challenges can feel at odds: constructing and deploying models is incredibly expensive, but over-relying on one model might be detrimental in the long term. I do know this challenge personally as an AI founder.

To construct intelligence, you wish talent, data, and scalable compute. To speed up time to market and do more with less, many corporations will select to construct on top of existing models, somewhat than construct from the bottom up. And the approach is sensible when what you’re constructing is so resource-intensive. Compounding this challenge is that, unlike software, a lot of the gains to date in AI have been made by adding more scale, which requires more computing power and subsequently cost.

But what happens when the corporate by which you’ve built your solution experiences a governance failure or a product outage? From a practical standpoint, counting on a single model to construct your product signifies that you at the moment are a part of a negative ripple effect for anything that happens.

We also must remember the risks of working with systems which might be probabilistic. We aren’t used to this and the world we live in to date has been engineered and designed to operate with a definitive answer. Models are fluid by way of output, and corporations continually tweak the models as well, which implies the code you may have written to support these and the outcomes your customers are counting on can change without your knowledge or control.

Centralization also creates safety concerns since it introduces a single point of failure. Every company is working in one of the best interest of itself. If there’s a security or risk concern with a model, you may have much less control over fixing that issue or less access to alternatives.

Where does that leave us?

AI is indisputably going to enhance how we live. There may be a lot that it’s able to achieving and fixing, from how we gather information to how we understand vast amounts of information. But with that chance also comes risk. If we over-rely on a single model, all corporations are opening themselves as much as each safety and product challenges.

To repair this, we want to bring the inference costs down and make it easier for corporations to have a multi-model approach. And naturally, every part involves data. Data and data ownership will matter. The more unique, top quality, and available the information, the more useful it’s going to be.

For a lot of problems, you possibly can optimize models for a particular application. The last mile of AI is corporations constructing routing logic, evaluations, and orchestration layers on top of those different models, specializing them for various verticals.

There have been multiple substantial investments on this space which might be getting us closer to this goal. Mistal’s recent (and impressive) funding round is a promising development towards an OpenAI alternative. There are also corporations helping other AI providers make cross-model multiplexing a reality and reducing inference costs via specialized hardware, software, and model distillation, as a couple of examples.

We’re also going to see open-source take off, and government bodies must enable open source to stay open. With open-source models, it’s easier to have more control. Nevertheless, the performance gaps are still there.

I presume we are going to find yourself in a world where you’ll have junior models optimized to perform less complex tasks at scale while larger super-intelligent models will act as oracles for updates and can increasingly spend compute on solving more complex problems. You won’t need a trillion-parameter model to answer a customer support request. I liken it to not having a senior executive manage a task that an intern can handle. Very similar to we have now multiple roles for human counterparts, most corporations will even depend on a group of models with various levels of sophistication.

To realize this balance, you wish a transparent task breakdown and benchmarking, considering the time, computational complexity, cost, and required scale. Depending on the use case, you possibly can prioritize accordingly. Determine a ground truth, a perfect end result for comparison, and an example input and output data, so you possibly can run various prompts to optimize and get the closest end result to the bottom truth.

If AI corporations can successfully decentralize their tech stack and construct on multiple models, we are able to improve the security and reliability of those tools and thereby maximize the positive impact of AI. We are not any longer in a spot for theoretical debates – it’s time to concentrate on easy methods to put AI to work to make these technologies more practical and resilient.

LEAVE A REPLY

Please enter your comment!
Please enter your name here