• Aucun résultat trouvé

Contents

N/A
N/A
Protected

Academic year: 2022

Partager "Contents"

Copied!
2
0
0

Texte intégral

(1)

SafeAI 2020 Table of Contents

Table of Contents

Session 1: Adversarial Machine Learning

Bio-Inspired Adversarial Attack Against Deep Neural Networks . . . 1 Bowei Xi, Yujie Chen, Fei Fan, Zhan Tu and Xinyan Deng

Adversarial Image Translation: Unrestricted Adversarial Examples in Face Recognition Systems . . . 6

Kazuya Kakizaki and Kosuke Yoshida

Session 2: Assurance Cases for AI-based Systems

Hazard Contribution Modes of Machine Learning Components . . . 14 Ewen Denney, Ganesh Pai and Colin Smith

Assurance Argument Patterns and Processes for Machine Learning in Safety-Related

Systems. . . 23 Chiara Picardi, Colin Paterson, Richard Hawkins, Radu Calinescu and Ibrahim Habli Session 3: Considerations for the AI Safety Landscape

Founding The Domain of AI Forensics . . . 31 Vahid Behzadan and Ibrahim Baggili

Exploring AI Safety in Degrees: Generality, Capability and Control. . . 36 John Burden and Jos´e Hern´andez-Orallo

Session 4: Fairness and Bias

Fair Enough: Improving Fairness in Budget-Constrained Decision Making Using

Confidence Thresholds . . . 41 Michiel Bakker, Humberto Riveron Valdes, Duy Patrick Tu, Krishna Gummadi, Kush Varshney, Adrian Weller and Alex Pentland

A Study on Multimodal and Interactive Explanations for Visual Question Answering . . . 54 Kamran Alipour, Jurgen P. Schulze, Yi Yao, Avi Ziskind and Giedrius Burachas

You Shouldn’t Trust Me: Learning Models Which Conceal Unfairness From Multiple

Explanation Methods . . . 63 Botty Dimanov, Umang Bhatt, Mateja Jamnik and Adrian Weller

Session 5: Uncertainty and Safe AI

A High Probability Safety Guarantee with Shifted Neural Network Surrogates . . . 74 M´elanie Ducoffe, Jayant Sen Gupta and Sebastien Gerchinovitz

Benchmarking Uncertainty Estimation Methods for Deep Learning With Safety-Related Metrics . . . 83

Maximilian Henne, Adrian Schwaiger, Karsten Roscher and Gereon Weiss

1

(2)

SafeAI 2020 Table of Contents

PURSS: Towards Perceptual Uncertainty Aware Responsibility Sensitive Safety with ML . 91 Rick Salay, Krzysztof Czarnecki, Maria Elli, Igancio Alvarez, Sean Sedwards and Jack Weast

Poster Papers

Simple Continual Learning Strategies for Safer Classifers . . . 96 Ashish Gaurav, Sachin Vernekar, Jaeyoung Lee, Vahdat Abdelzad, Krzysztof

Czarnecki and Sean Sedwards

Fair Representation for Safe Artificial Intelligence via Adversarial Learning of Unbiased Information Bottleneck . . . 105

Jin-Young Kim and Sung-Bae Cho

Out-of-Distribution Detection with Likelihoods Assigned by Deep Generative Models

Using Multimodal Prior Distributions . . . 113 Ryo Kamoi and Kei Kobayashi

SafeLife 1.0: Exploring Side Effects in Complex Environments . . . 117 Carroll Wainwright and Peter Eckersley

(When) Is Truth-telling Favored in AI Debate? . . . 128 Vojtech Kovarik and Ryan Carey

NewsBag: A Benchmark Multimodal Dataset for Fake News Detection . . . 138 Sarthak Jindal, Raghav Sood, Richa Singh, Mayank Vatsa and Tanmoy Chakraborty

Algorithmic Discrimination: Formulation and Exploration in Deep Learning-based Face Biometrics . . . 146

Ignacio Serna, Aythami Morales, Julian Fierrez, Manuel Cebrian, Nick Obradovich and Iyad Rahwan

Guiding Safe Reinforcement Learning Policies Using Structured Language Constraints . . . . 153 Bharat Prakash, Nicholas Waytowich, Ashwinkumar Ganesan, Tim Oates and

Tinoosh Mohsenin

Practical Solutions for Machine Learning Safety in Autonomous Vehicles . . . 162 Sina Mohseni, Mandar Pitale, Vasu Singh and Zhangyang Wang

Continuous Safe Learning Based on First Principles and Constraints for Autonomous

Driving . . . 170 Lifeng Liu, Yingxuan Zhu, Tim Yuan and Jian Li

Recurrent Neural Network Properties and their Verification with Monte Carlo Techniques 178 Dmitry Vengertsev and Elena Sherman

Toward Operational Safety Verification Via Hybrid Automata Mining Using I/O Traces of AI-Enabled CPS . . . 186

Imane Lamrani, Ayan Banerjee and Sandeep Gupta

2

Références

Documents relatifs

I will describe three efforts in the University of Arizona School of Information to model aspects of learning statistically: Teaching robots the meanings of

There are two types of arguments: positive arguments are used to explain (or argue) why a certain learning example is in the class as given, and negative ones give reasons against

In contrast to that we consistently observed, that the LC method is working considerably better with VGG16 for its confidence predictions, which is why we selected it for

Section 2 briefly reviews the two main automotive industry safety standards and identifies their five fundamental gaps with machine learning algorithms. Section 3 describes prac-

This paper investigates the applicability of machine learning approaches for predicting tool wear in indus- trial milling processes based on real-world sensor data on exerted

To support this, based on approaches for re- quirements assurance in traditional safety-related software, we develop nine specific areas where confidence is required

We have used such counterfactual explanations with pre- dictive AI systems trained on two data sets: UCI German Credit 1 – assessing credit risks based on applicant’s personal

The goal of this paper is to show through several practi- cal examples that the application of machine learning tech- niques on manufacturing problems can lead to increased