content.dataiku.com
Open in
urlscan Pro
35.174.108.193
Public Scan
Submitted URL: https://pages.dataiku.com/e3t/Ctc/GA%20113/cfvmy04/VWskLb548KMDW4tZmHK1s_rFJW4gSXWJ4TbJcfN3DlwGV3q3pBV1-WJV7CgQmJVl5YlQ6YK...
Effective URL: https://content.dataiku.com/c/oreilly-responsible-ai?x=y8gWw0&l&utm_campaign=GLO%20Content%20O%27Reilly%20Responsible%20AI%2...
Submission: On December 06 via api from US — Scanned from DE
Effective URL: https://content.dataiku.com/c/oreilly-responsible-ai?x=y8gWw0&l&utm_campaign=GLO%20Content%20O%27Reilly%20Responsible%20AI%2...
Submission: On December 06 via api from US — Scanned from DE
Form analysis
0 forms found in the DOMText Content
Thumbnails Document Outline Attachments Layers Current Outline Item Cover Dataiku Copyright Table of Contents Preface Who Should Read This Book What Readers Will Learn Preliminary Book Outline Part 1 Part 2 Stop Going Fast and Breaking Things Conventions Used in This Book Using Code Examples O’Reilly Online Learning How to Contact Us Acknowledgments Chapter 1. Contemporary Model Governance Basic Legal Obligations AI Incidents Organizational and Cultural Competencies for Responsible AI Accountability Drinking Your Own Champagne Diverse and Experienced Teams “Going Fast and Breaking Things” Organizational Processes for Responsible AI Forecasting Failure Modes Model Risk Management Beyond Model Risk Management Case Study: The Rise and Fall of Zillow’s iBuying Fallout Lessons Learned Chapter 2. Interpretable and Explainable Machine Learning Important Ideas for Interpretability and Explainability Interpretable Models Additive Models Decision Trees An Ecosystem of Interpretable Machine Learning Models Post-hoc Explanation Feature Importance Surrogate Models Plots of Model Performance Cluster profiling Stubborn Difficulties of Post-hoc Explanation in Practice Case: Graded by Algorithm Resources Chapter 3. Debugging Machine Learning Systems for Safety and Performance Training Reproducibility Data Quality and Feature Engineering Model Specification Model Debugging Software Testing Traditional Model Assessment Residual Analysis for Machine Learning Sensitivity Analysis Benchmark Models Machine Learning Bugs Remediation: Fixing Bugs Deployment Domain Safety Model Monitoring Case Study: Death by Autonomous Vehicle Fallout An Unprepared Legal System Lessons Learned Resources Chapter 4. Managing Bias in Machine Learning ISO and NIST Definitions for Bias Systemic Bias Statistical Bias Human Biases and Data Science Culture United States Legal Notions of ML Bias Who Tends to Experience Bias from ML Systems Harms That People Experience Testing for Bias Testing Data Traditional Approaches: Testing for Equivalent Outcomes A New Mindest: Testing for Equivalent Performance Quality On the Horizon: Tests for the Broader ML Ecosystem Summary Test Plan Mitigating Bias Technical Factors in Mitigating Bias The Scientific Method and Experimental Design Bias Mitigation Approaches Human Factors in Mitigating Bias Case: The Bias Bug Bounty Resources Chapter 5. Security for Machine Learning Security Basics The Adversarial Mindset CIA Triad Best Practices for Data Scientists Machine Learning Attacks Integrity Attacks: Manipulated Machine Learning Outputs Confidentiality Attacks: Extracted Information General AI Security Concerns Counter-measures Model Debugging for Security Model Monitoring For Security Privacy-enhancing Technologies Robust Machine Learning General Countermeasures Case Study: Real-world Evasion Attacks Lessons Learned Resources Chapter 6. Explainable Boosting Machines and Explaining XGBoost Concept Refresher Additivity vs. Interactions Steps Toward Causality with Constraints Partial Dependence and Individual Conditional Explanation Shapley Values Model Documentation The GAM Family of Interpretable Models Elastic Net Penalized GLM w/ Alpha and Lambda Search Generalized Additive Models GA2M and Explainable Boosting Machines XGBoost with Constraints and Explainable Artificial Intelligence Constrained and Unconstrained XGBoost Explaining Model Behavior with Partial Dependence and ICE Decision Tree Surrogate Models as an Explanation Technique Shapley Value Explanations Resources Chapter 7. Red-teaming XGBoost Concept Refresher CIA Triad Attacks Countermeasures Model Training Attacks for Red-teaming Conclusion Resources: About the Authors Previous Next Highlight All Match Case Match Diacritics Whole Words Color Size Color Thickness Opacity Presentation Mode Open Print Download Current View Go to First Page Go to Last Page Rotate Clockwise Rotate Counterclockwise Text Selection Tool Hand Tool Page Scrolling Vertical Scrolling Horizontal Scrolling Wrapped Scrolling No Spreads Odd Spreads Even Spreads Document Properties… Toggle Sidebar Find Previous Next (1 of 228) Presentation Mode Open Print Download Current View FreeText Annotation Ink Annotation Tools Zoom Out Zoom In Automatic Zoom Actual Size Page Fit Page Width 50% 75% 100% 125% 150% 200% 300% 400% AI Governance With Dataiku The Platform for Everyday AI Leveraging one central solution for AI means more transparent, repeatable, and scalable AI. Dataiku gives people (whether technical and working in code or on the business side and low- or no-code) the ability to make better day-to-day decisions with data. Keep critical AI initiatives up and running with Dataiku, which offers model monitoring, drift detection, automatic model retraining, easy production project model updates, automated CI/CD, and more. Dataiku provides critical capabilities for explainable and responsible AI, including reports on feature importance, partial dependence plots, subpopulation analysis, model fairness, and individual prediction explanations. Advanced AI Governance including customizable governance plans, production sign-off, risk and value analysis together with permissions management, user directory integration, audit trails, and secure API access make Dataiku the perfect choice for streamlined AI Governance across the organization. Deploy, Monitor, & Manage Machine Learning Projects in Production Understand Outputs, Increase Trust, & Identify Potential Bias Manage Risk & Ensure Compliance at Scale ©2022 dataiku | dataiku.com Patrick Hall, James Curtis, and Parul Pandey Machine Learning for High-Risk Applications Techniques for Responsible AI FIRST EDITION Boston Farnham Sebastopol TokyoBeijing Boston Farnham Sebastopol TokyoBeijing More Information Less Information Close Enter the password to open this PDF file. Cancel OK File name: - File size: - Title: - Author: - Subject: - Keywords: - Creation Date: - Modification Date: - Creator: - PDF Producer: - PDF Version: - Page Count: - Page Size: - Fast Web View: - Close Preparing document for printing… 0% Cancel Next Next Deploy and Maintain More Models in Production With Dataiku 10 Dataiku 10 builds on core strengths related to model deployment and brings additional tools to ML operators maintaining live models in production. LinkedIn LinkTwitter LinkFacebook LinkEmail LinkLike Button GET DATAIKU NEWS GET MORE CONTENT FROM DATAIKU! You seem new here, sign up to our newsletter for updates on new content and videos O'Reilly Machine Learning for High-Risk Applications: Techniques for Responsible AIpdf Deploy and Maintain More Models in Production With Dataiku 10webpage [VIDEO] Who's Responsible for Responsible AI?video Navigating Targeted Ad Bias With Responsible AIwebpage [VIDEO] Responsible AI Is About More Than Modelsvideo So You've Built a Fair Model, Now What?webpage [DEMO] Dataiku for Coders (Data Scientists, Engineers, & More)video GET MORE CONTENT FROM DATAIKU! You seem new here, sign up to our newsletter for updates on new content and videos