Sign In Start Free Trial
Account

Add to playlist

Create a Playlist

Modal Close icon
You need to login to use this feature.
  • Book Overview & Buying Hands-On Explainable AI (XAI) with Python
  • Table Of Contents Toc
Hands-On Explainable AI (XAI) with Python

Hands-On Explainable AI (XAI) with Python

By : Denis Rothman
4.4 (12)
close
close
Hands-On Explainable AI (XAI) with Python

Hands-On Explainable AI (XAI) with Python

4.4 (12)
By: Denis Rothman

Overview of this book

Effectively translating AI insights to business stakeholders requires careful planning, design, and visualization choices. Describing the problem, the model, and the relationships among variables and their findings are often subtle, surprising, and technically complex. Hands-On Explainable AI (XAI) with Python will see you work with specific hands-on machine learning Python projects that are strategically arranged to enhance your grasp on AI results analysis. You will be building models, interpreting results with visualizations, and integrating XAI reporting tools and different applications. You will build XAI solutions in Python, TensorFlow 2, Google Cloud’s XAI platform, Google Colaboratory, and other frameworks to open up the black box of machine learning models. The book will introduce you to several open-source XAI tools for Python that can be used throughout the machine learning project life cycle. You will learn how to explore machine learning model results, review key influencing variables and variable relationships, detect and handle bias and ethics issues, and integrate predictions using Python along with supporting the visualization of machine learning models into user explainable interfaces. By the end of this AI book, you will possess an in-depth understanding of the core concepts of XAI.
Table of Contents (16 chapters)
close
close
14
Other Books You May Enjoy
15
Index

Chapter 1, Explaining Artificial Intelligence with Python

  1. Understanding the theory of an ML algorithm is enough for XAI. (True|False)

    False. Implementing an ML program requires more than theoretical knowledge.

    True. If a user just wants an intuitive explanation of an ML algorithm.

  2. Explaining the origin of datasets is not necessary for XAI. (True|False)

    True. If a third party has certified the dataset.

    False. If you are building the dataset, you need to make sure you are respecting privacy laws.

  3. Explaining the results of an ML algorithm is sufficient. (True|False)

    True. If a user is not interested in anything else but the result.

    False. If it is required to explain how a result was reached.

  4. It is not necessary for an end user to know what a KNN is. (True|False)

    True. If the end user is satisfied with results.

    False. If the end user is a developer that is deploying the program...

CONTINUE READING
83
Tech Concepts
36
Programming languages
73
Tech Tools
Icon Unlimited access to the largest independent learning library in tech of over 8,000 expert-authored tech books and videos.
Icon Innovative learning tools, including AI book assistants, code context explainers, and text-to-speech.
Icon 50+ new titles added per month and exclusive early access to books as they are being written.
Hands-On Explainable AI (XAI) with Python
notes
bookmark Notes and Bookmarks search Search in title playlist Add to playlist download Download options font-size Font size

Change the font size

margin-width Margin width

Change margin width

day-mode Day/Sepia/Night Modes

Change background colour

Close icon Search
Country selected

Close icon Your notes and bookmarks

Confirmation

Modal Close icon
claim successful

Buy this book with your credits?

Modal Close icon
Are you sure you want to buy this book with one of your credits?
Close
YES, BUY

Submit Your Feedback

Modal Close icon
Modal Close icon
Modal Close icon