Use coupon code “SUMMER20” for a 20% discount on all items! Valid until 2024-08-31

Site Logo
Search Suggestions

      Royal Mail  express delivery to UK destinations

      Regular sales and promotions

      Stock updates every 20 minutes!

      Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

      2 in stock

      Firm sale: non returnable item
      SKU 9783030289539 Categories ,
      The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” ...

      £79.99

      Buy new:

      Delivery: UK delivery Only. Usually dispatched in 1-2 working days.

      Shipping costs: All shipping costs calculated in the cart or during the checkout process.

      Standard service (normally 2-3 working days): 48hr Tracked service.

      Premium service (next working day): 24hr Tracked service – signature service included.

      Royal mail: 24 & 48hr Tracked: Trackable items weighing up to 20kg are tracked to door and are inclusive of text and email with ‘Leave in Safe Place’ options, but are non-signature services. Examples of service expected: Standard 48hr service – if ordered before 3pm on Thursday then expected delivery would be on Saturday. If Premium 24hr service used, then expected delivery would be Friday.

      Signature Service: This service is only available for tracked items.

      Leave in Safe Place: This option is available at no additional charge for tracked services.

      Description

      Product ID:9783030289539
      Product Form:Paperback / softback
      Country of Manufacture:GB
      Series:Lecture Notes in Artificial Intelligence
      Title:Explainable AI: Interpreting, Explaining and Visualizing Deep Learning
      Authors:Author: Andrea Vedaldi, Wojciech Samek, Lars Kai Hansen, Klaus-Robert Muller, Gregoire Montavon
      Page Count:439
      Subjects:Information technology: general topics, Information technology: general issues, Computer hardware, Computer security, Network security, Artificial intelligence, Computer vision, Computer hardware, Computer security, Network security, Artificial intelligence, Computer vision
      Description:The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner. The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.

      The development of "intelligent" systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to "intelligent" machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner.

      The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.


      Imprint Name:Springer Nature Switzerland AG
      Publisher Name:Springer Nature Switzerland AG
      Country of Publication:GB
      Publishing Date:2019-08-30

      Additional information

      Weight698 g
      Dimensions158 × 234 × 30 mm