X Close

UCL Department of Science, Technology, Engineering and Public Policy

Home

Applied in Focus. Global in Reach

Menu

Archive for the 'Master’s of Public Adminstration' Category

Adversarial Attacks, Robustness and Generalization in Deep Reinforcement Learning

By Ezgi Korkmaz, on 20 December 2023

Reinforcement learning has achieved substantial progress on successfully completing tasks, from solving complex games to large language models (i.e. GPT-4) including many different fields from medical applications to self-driving vehicles and finance, by learning from raw high-dimensional data with the utilization of deep neural networks as function approximators.

The vulnerabilities of deep reinforcement learning policies against adversarial attacks have been demonstrated in prior studies [1,2,3,4]. However, a recent study takes these vulnerabilities one step further and introduces natural attacks (i.e. natural changes to the environment given that these changes are imperceptible) while providing a contradistinction between adversarial attacks and natural attacks. The instances of such changes include, but are not limited to creating a blur, introduction of compression artifacts, or perspective projection of the state observations at a level that humans cannot perceive the change.

Intriguingly, the results reported demonstrate that these natural attacks are at least equally, and often more imperceptible compared to adversarial attacks, while causing larger drop in policy performance. While these results carry significant concerns regarding artificial intelligence safety [5,6,7], they further raise questions on the model’s security. Note that the prior studies on adversarial attacks on deep reinforcement learning rely on the strong adversary assumption, in which the adversary has access to the policy’s perception system, training details of the policy (e.g. algorithm, neural network architecture, training dataset), and the ability to alter observations in real time with simultaneous modifications to the observation system of the policy with computationally demanding adversarial formulations. Thus, the fact that natural attacks described in [8] are black-box adversarial attacks, i.e. the adversary does not have access to the training details of the policy and the policy’s perception system to compute the adversarial perturbations, raises further questions on machine learning safety and responsible artificial intelligence.

Furthermore, the second part of the paper investigates the robustness of adversarially trained deep reinforcement learning policies (i.e. robust reinforcement learning) under natural attacks, and demonstrates that vanilla trained deep reinforcement learning policies are more robust than adversarially trained policies. While these results reveal further security concerns regarding the robust reinforcement learning algorithms, they further demonstrate that adversarially trained deep reinforcement learning policies cannot generalize at the same level as straightforward vanilla trained deep reinforcement learning algorithms.

This study overall, while providing a contradistinction between adversarial attacks and natural black-box attacks, further reveals the connection between generalization in reinforcement learning and the adversarial perspective.

Author’s Note: This blog post is based on the paper ‘Adversarial Robust Deep Reinforcement Learning Requires Redefining Robustness’ published in AAAI 2023.
References:
[1] Adversarial Attacks on Neural Network Policies, ICLR 2017.
[2] Investigating Vulnerabilities of Deep Neural Policies. Conference on Uncertainty in Artificial Intelligence (UAI).
[3] Deep Reinforcement Learning Policies Learn Shared Adversarial Features Across MDPs. AAAI Conference on Artificial Intelligence, 2022. [Paper Link]
[4] Detecting Adversarial Directions in Deep Reinforcement Learning to Make Robust Decisions. International Conference on Machine Learning, ICML 2023. [Paper Link]
[5] New York Times. Global Leaders Warn A.I. Could Cause ‘Catastrophic’ Harm, November 2023.
[6] The Washington Post. 17 fatalities, 736 crashes: The shocking toll of Tesla’s Autopilot, June 2023.
[7] The Guardian. UK, US, EU and China sign declaration of AI’s ‘catastrophic’ danger, November 2023.
[8] Adversarial Robust Deep Reinforcement Learning Requires Redefining Robustness, AAAI Conference on Artificial Intelligence, 2023. [Paper Link]

Is a total ban of plastic bags good and inclusive? Lessons from Rwanda

By katerynatsybenko, on 24 September 2021

In June 2021, Ukraine adopted a law to ban plastic bags. The ban will be implemented in stages: in December 2021, bags up to 50 microns thick will be banned; on March 2022, bags 15 microns thick will be banned. Only very small thin bags for transporting fish, meat, ice will be allowed but for a limited period of time. Starting from January 1 2023, only biodegradable bags will be allowed. Similar bans have been imposed in other countries, such as Rwanda and the UK, and in the EU. Radical policies to ban plastic bags may improve environmental sustainability, but there can be unintended consequences. They should be anticipated and carefully planned for.

plastic bags in different colours at lanfill

Photo by Brian Yurasits on Unsplash

The new Ukrainian law stipulates fines for using plastic bags: 1700-8500 UAH (45-215GBP, while 150GBP is a minimum salary) from December 2021, and 8500-34000 UAH (215-850 GBP) from March 2022.

(more…)

Walking, cycling and using public transport: how the UK government offers to develop urban mobility

By katerynatsybenko, on 16 December 2020

Kateryna Tsybenko is an MPA candidate in Science, Engineering and Public Policy

Recently, the Ukrainian Minister of Infrastructure Vladyslav Krykliy announced that Ukraine plans to replace all public transport with electric transport by 2030. It will take place within the framework of the implementation of the National Transport Strategy.

I currently study in the UK, and I researched the UK’s urban mobility strategy; and in this blog, I want to share urban mobility trends in this country. The key in it is the emphasis on inclusiveness, encouraging citizens to use public transport instead of individual transport, ride-sharing, walking and cycling, and broad support for all these means of transportation, including through open data.

Image of trams and bus

Anne Burgess / Integrated Public Transport

(more…)

The Infinite Game of Disinformation

By Alex Shepherd, on 15 October 2020

Alex Shepherd (@palexshepherd) is a nationally recognised subject matter expert on disinformation. He has delivered talks on the subject at the University of Oxford and the University of Cambridge, and has actively engaged with representatives from the UK government’s Sub-Committee on Disinformation. He is currently a senior AI researcher at Oxford Brookes University and a Digital Technologies and Policy MPA candidate at UCL STEaPP. 

Disinformation is one of the most important issues we face today, not only due to the massive social impact and disruption it creates globally, but also due to its exceptionally robust nature. This blog post, inspired by the tweetstorm “Some thoughts on disinformation”, attempts to explain disinformation’s robustness through the lens of game theory and analysis of technology trends.

Man using tablet to view fake news website

The concept of infinite games and finite games was popularised by Simon Sinek in his book, The Infinite Game, and at a keynote speech he delivered at a New York Times event. The book was influenced, in part, by James P. Carse’s book Finite and Infinite Games, which in turn was influenced by basic game theory.

(more…)

COVID-19: IoT and Cybersecurity

By fredrikskippervold, on 27 August 2020

Fredrik Johan Skippervold is a UCL MPA Graduate within Digital Technologies and Policy 18/19. He holds a Bachelor of Law with Spanish and is currently a researcher in the PETRAS National Centre of Excellence for IoT Systems Cybersecurity.

Introduction

Over the past four months (April – July) my colleague Dr Catherine Wheller and I have been following the impacts of COVID-19 on cybersecurity and the Internet of Things (IoT) within the UK and beyond. The pandemic has inspired a range of IoT innovations to help stop the spread of the virus. We have written weekly landscape briefings (LB) that provide up to date information on the latest developments in this area. In this blog I will talk about how we set about collecting information and how we put together these reports, as well as highlight some of the major developments which include discussions surrounding privacy and ethics. To note, a final summary briefing will be posted alongside this blogpost. The summary, which can be found here, includes a detailed timeline of events, provides an overview of how IoT devices are helping to stop the spread of the virus (UK and globally) and presents discussions around so-called ‘immunity passports’.

Cybersecurity

(more…)