内容简介:“What’s the accuracy of this machine learning (ML) model?”“How long is the training time?”“How much training data do you need?”
And Three Things You Can do to Avoid it Happening to you.
Mar 16 ·4min read
“What’s the accuracy of this machine learning (ML) model?”
“How long is the training time?”
“How much training data do you need?”
Working for a company that builds machine learning software for robotics, I hear these questions every day. Machine learning has become a shiny object that everyone wants to pursue. Over 80% of the companies are looking into at least one AI project.
Users generally want to know how long it would take to onboard a new item and how well the models perform or generalize. They want a way to measure the overall cost against performance. However, answers to the above questions don’t give you a full picture. Even worse, they are misleading.
Model training is only the tip of the iceberg. What most users and AI/ML companies overlook is the massive hidden cost of acquiring appropriate datasets, cleaning, storing, aggregating, labeling, and building reliable data flow as well as infrastructure pipeline.
Based on recent research, companies spend more than 80% of the time in AI/ML projects on data preparation and engineering tasks. In other words, if you pay most of your attention to building and training models, your total engineering effort and cost could be five times more than your initial estimation.
Furthermore, machine learning blurs the line between user and software providers.
We have begun to see the emergence of AIaaS or MLaaS. Models in the cloud continue to improve with more data. This is the biggest benefit of machine learning but also what makes MLaaS a more challenging business than SaaS.
Models learn from training data. Without quality data, models won’t perform well. Users don’t always know the best practice of generating or annotating appropriate datasets. In fact, most of the time, they don’t.
When systems underperform, users tend to blame models. Therefore, AI/ML companies generally spend significant time and resources training and working with users to ensure data quality, which becomes a shared responsibility between AI companies and their customers.
For example, to train a model for defect inspection on the production line, a computer vision company needs to work with its client to install cameras with the right angle and position, check the resolution and frame rate, make sure that there are enough positive and negative training samples for every scenario.
Data collection is even more time consuming and costly with robotics or self-driving car applications because someone needs to control robots or vehicles to perform certain actions.
Even with all the training sessions, user manuals, and guidelines, you still do not have full control over user-generated data. A machine vision camera company told me that they have engineers manually verify 100% of the incoming data to ensure data integrity.
All these additional and often overlooked training, manual checking, data cleaning, and labeling tasks lead to significant overheads for AI companies, making MLaaS a more challenging business with a lower gross margin.
To build a more scalable AI/ML business, there are three things we can do:
Scalability is King. Identify the right use cases that a significant number of customers would pay you for and you can solve with the same model architecture. The last thing you want to do is to build and train different models for different companies without a standard product offering.
Make as many things self-serve as possible. Automate your training and data pipeline as much as possible to increase operational efficiency and reduce dependency on manual labor. Companies tend to prioritize customer visible features over internal tools or automation. But the latter pay off quickly. Make sure that you allocate enough resources for internal process automation.
Lastly, identify and track your costs, especially the hidden ones.How much time do your engineers spend cleaning, filtering, or aggregating data? How much time to they spend ensuring annotation is done correctly by a third party? How often do they need to help customers set up the environments and correctly gather data? How much of this can be automated or outsourced?
以上所述就是小编给大家介绍的《The Single Biggest Reason Why AI/ML Companies Fail to Scale?》,希望对大家有所帮助,如果大家有任何疑问请给我留言,小编会及时回复大家的。在此也非常感谢大家对 码农网 的支持!
猜你喜欢:本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们。
松本行弘的程序世界
松本行弘 / 柳德燕、李黎明、夏倩、张文旭 / 人民邮电出版社 / 2011-8 / 75.00元
《松本行弘的程序世界》是探索程序设计思想和方法的经典之作。作者从全局的角度,利用大量的程序示例及图表,深刻阐述了Ruby编程语言的设计理念,并以独特的视角考察了与编程相关的各种技术。阅读《松本行弘的程序世界》不仅可以深入了解编程领域各个要素之间的关系,而且能够学到大师的思考方法。 《松本行弘的程序世界》面向各层次程序设计人员和编程爱好者,也可以供相关技术人员参考。一起来看看 《松本行弘的程序世界》 这本书的介绍吧!