Design Thinking for Artificial Intelligence Projects

栏目: IT技术 · 发布时间: 4年前

内容简介:IBM uses an interesting workflow for AI projects that is based on design thinking principles, shown in detail in itsDesign thinkingis a type of thinking used when developing design concepts (for example when designing a new building or tech product), which

Design Thinking for Artificial Intelligence Projects

How IBM adapted design thinking principles to build a workflow for AI projects

Why should you read this article?

IBM uses an interesting workflow for AI projects that is based on design thinking principles, shown in detail in its specialisation on Coursera . It’s a long MOOC, that goes into thorough detail on implementing AI apps, and its worth taking, if you have the time and technical knowledge. If you are only interested in learning the workflow and how to implement it on AI projects, this article can help you.

What is design thinking?

Design thinkingis a type of thinking used when developing design concepts (for example when designing a new building or tech product), which seeks to understand the product from the end user’s point of view. This is done not only by empathising with the user, but also by generating multiple ideas in brainstorming sessions, prototyping , and then testing those ideas. It is, thus, an iterative and hands-on approach by design. The main phases of the process are: empathise , define , ideate , prototype and test . I’ll not get into the details of design thinking for product creation, but will focus instead on how this process can be adapted to AI projects.

The workflow

The workflow presented by IBM for AI projects is the following, with the corresponding name in design thinking:

  • Data collection ( empathise )
  • Exploratory analysis ( define )
  • Transformation ( ideate )
  • Modelling ( prototype )
  • Testing ( test )

Let’s now get into further detail of each of these steps.

Data collection

This is where the data scientists talk to the people closest to the data to articulate the business opportunity and translate it into testable hypothesis or hypotheses. It includes defining a timeline, cost, feasibility, etc. Finally you proceed to gather data for the project.

  1. Get as close to the source of data as possible usually by interviewing the people involved
  2. Identify the business problem
  3. To articulate the business question, enumerate possible questions, and prioritise according to domain knowledge, feasibility and impact (impact can be measured using a back-of-the-napkin ROI calculation).
  4. Obtain all of the relevant the data
  5. Translate the business problem into a testable hypothesis or hypotheses

Exploratory data analysis

At this point, you try to visualise your data, check for missing values (and decide how to deal with each of them) and potentially test hypotheses.This part is composed by data visualisation and hypothesis tests .

Start your analysis with simple csv files, to make sure your model will have value, before building a full data ingestion pipeline.

Data visualization

Check for missing data, understand what features and observations have missing values, and why (go back to “empathise” and talk to the client)

Look at how the missing observations behave, compared to the rest:

  • MCAR (Missing Completely At Random) : missing cases are, on average, identical to non-missing cases, with respect to the feature matrix. Complete case analysis (removing those observations from the dataset) will reduce the power of the analysis, but will not affect bias
  • MAR (Missing At Random) : missing data often have some dependence on measured values, and models can be used to help impute what the likely data would be. For example, in an MLB survey, there may be a gender bias when it comes to completing all of the questions
  • MNAR (Missing Not At Random) : missing data depend on unmeasured or unknown variables. There is no information available to account for the missingness.

If needed, try a few different imputation methods, but make sure to come back and re-assess them during the modelling phase, knowing by then which ones yielded the best results. The process of trying different imputation methods is called multiple imputation.

  • Univariate imputation : mean or median of the missing feature
  • Multivariate imputation : use other variables to predict the missing feature

Create visual summaries detailing data, including missing values, outliers, classing imbalance issues, and try to identify factors that can be useful for your specific business problem and start formulating hypothesis.

Use plots and tables to create a first presentation or report telling a story related to your business problem (make sure all of them have at least one or two lines explaining the main point/conclusion). Finish with conclusions and suggestions on next steps.

Transformation

where the goal is to transform your data so that it becomes consumable by models. This is where all the feature engineering magic happens.

Modelling

At this point you define an evaluation metric and compare different models for your problem. Start with simple models, and build up from there.

Testing

Here the goal is not only to perform unit tests on your code, but also criticise the choices you have previously made and go back if needed (it usually is). It culminates with the deployment of the chosen solution, which is also subject to testing: even running models are subject to scrutiny and ongoing performance testing.


以上所述就是小编给大家介绍的《Design Thinking for Artificial Intelligence Projects》,希望对大家有所帮助,如果大家有任何疑问请给我留言,小编会及时回复大家的。在此也非常感谢大家对 码农网 的支持!

查看所有标签

猜你喜欢:

本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们

第三次浪潮

第三次浪潮

托夫勒 / 黄明坚 / 中信出版社 / 2006-6 / 38.00元

《第三次浪潮》作者托夫勒在20多年前预见的未来是:跨国企业将盛行;电脑发明使SOHO(在家工作)成为可能;人们将摆脱朝九晚五工作的桎梏;核心家庭的瓦解;DIY(自己动手做)运动的兴起……时过境迁,如今我们才发现托夫勒的预言竟大多已成为了现实。   20年前的《第三次浪潮》在打开国门之初给人们心灵造成的冲击,其影响至今仍然连绵不绝。托夫勒在这本书中将人类社会划分为三个阶段:第一次浪潮为农业阶段......一起来看看 《第三次浪潮》 这本书的介绍吧!

HTML 编码/解码
HTML 编码/解码

HTML 编码/解码

正则表达式在线测试
正则表达式在线测试

正则表达式在线测试

HEX HSV 转换工具
HEX HSV 转换工具

HEX HSV 互换工具