内容简介:Bahador Khaleghi is a Customer Data Scientist and Solution Engineer at H2O.ai. His unique technical background, which he has gained over the last thirteen years, is quite diverse entailing a wide range of disciplines including machine learning, statistical
About the speaker
Bahador Khaleghi is a Customer Data Scientist and Solution Engineer at H2O.ai. His unique technical background, which he has gained over the last thirteen years, is quite diverse entailing a wide range of disciplines including machine learning, statistical information fusion, and signal processing. Bahador obtained his PhD from CPAMI at the University of Waterloo. Over the last six years, he has actively contributed to industrial R&D projects in various domains including Telematics, mobile health, predictive maintenance, and customer analytics. Acting as the (former) technical lead of the explainability team at Element AI, he is currently focused on developing novel methodologies that enhance transparency, trustability, and accessibility of AI solutions.”
About the talk
“Modern AI systems are increasingly capable of tackling real-world problems. Yet the black box nature of some AI systems, giving results without a reason, is hindering the mass adoption of AI. According to an annual survey by PwC, the vast majority (82%) of CEOs agree that for AI-based decisions to be trusted, they must be explainable. As AI becomes an ever more integral part of our modern world, we need to understand why and how it makes predictions and decisions. These questions of why and how are the subject of the field of Explainable AI, or XAI. Like AI itself, XAI isn’t a new domain of research, and recent advances in the theory and applications of AI have put new urgency behind efforts to explain it. In this talk we will present a technical overview of XAI. The presentation will cover the there key questions of XAI: “What is it?”, “Why is it important?”, and “How can it be achieved?”. The what of XAI part takes a deep dive into what it really means to explain AI models in terms of existing definitions, the importance of explanation users’ roles and given application, possible tradeoffs, and explanation studies beyond the AI community. In the why of XAI part, we explore some of the most important drivers of XAI research such as establishing trust, regulatory compliance, detecting bias, AI model generalization and debug. Finally, in the how of XAI part we discuss how explainability principles can be applied before, during, and after the modelling stage of AI solution development. In particular, we introduce a novel taxonomy of post-modelling explainability methods, which we then leverage to explore the vast XAI literature work.”
以上所述就是小编给大家介绍的《An Explanation of What, Why, and how of Explainable AI (XAI)》,希望对大家有所帮助,如果大家有任何疑问请给我留言,小编会及时回复大家的。在此也非常感谢大家对 码农网 的支持!
猜你喜欢:本站部分资源来源于网络,本站转载出于传递更多信息之目的,版权归原作者或者来源机构所有,如转载稿涉及版权问题,请联系我们。
马化腾自述-我的互联网思维
赵黎 / 石油工业出版社 / 2014-8-1 / 35
马化腾自述:我的互联网思维》讲述了些人说移动互联网就是加了“移动”两个字,互联网十几年了,移动互联网应该是个延伸。我的感受是,移动互联网远远不只是一个延伸,甚至是一个颠覆。互联网是一个开放交融、瞬息万变的大生态,企业作为互联网生态里面的物种,需要像自然界的生物一样,各个方面都具有与生态系统汇接、和谐、共生的特性。开放和分享并不是一个宣传口号,也不是一个简单的概念。开放很多时候被看作一种姿态,但是我......一起来看看 《马化腾自述-我的互联网思维》 这本书的介绍吧!