English
 
Help Privacy Policy Disclaimer
  Advanced SearchBrowse

Item

ITEM ACTIONSEXPORT
 
 
DownloadE-Mail
  Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time Attacks

Mohammadi, M., Nöther, J., Mandal, D., Singla, A., & Radanovic, G. (2023). Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time Attacks. Retrieved from https://arxiv.org/abs/2302.13851.

Item is

Files

show Files
hide Files
:
arXiv:2302.13851.pdf (Preprint), 2MB
 
File Permalink:
-
Name:
arXiv:2302.13851.pdf
Description:
File downloaded from arXiv at 2023-03-06 10:04
OA-Status:
Visibility:
Private
MIME-Type / Checksum:
application/pdf
Technical Metadata:
Copyright Date:
-
Copyright Info:
-

Locators

show

Creators

show
hide
 Creators:
Mohammadi, Mohammad1, Author           
Nöther, Jonathan2, Author
Mandal, Debmalya1, Author           
Singla, Adish3, Author           
Radanovic, Goran1, Author           
Affiliations:
1Group K. Gummadi, Max Planck Institute for Software Systems, Max Planck Society, ou_2105291              
2External Organizations, ou_persistent22              
3Group A. Singla, Max Planck Institute for Software Systems, Max Planck Society, ou_2541698              

Content

show
hide
Free keywords: Computer Science, Learning, cs.LG,Computer Science, Artificial Intelligence, cs.AI,Computer Science, Cryptography and Security, cs.CR,Computer Science, Multiagent Systems, cs.MA
 Abstract: In targeted poisoning attacks, an attacker manipulates an agent-environment
interaction to force the agent into adopting a policy of interest, called
target policy. Prior work has primarily focused on attacks that modify standard
MDP primitives, such as rewards or transitions. In this paper, we study
targeted poisoning attacks in a two-agent setting where an attacker implicitly
poisons the effective environment of one of the agents by modifying the policy
of its peer. We develop an optimization framework for designing optimal
attacks, where the cost of the attack measures how much the solution deviates
from the assumed default policy of the peer agent. We further study the
computational properties of this optimization framework. Focusing on a tabular
setting, we show that in contrast to poisoning attacks based on MDP primitives
(transitions and (unbounded) rewards), which are always feasible, it is NP-hard
to determine the feasibility of implicit poisoning attacks. We provide
characterization results that establish sufficient conditions for the
feasibility of the attack problem, as well as an upper and a lower bound on the
optimal cost of the attack. We propose two algorithmic approaches for finding
an optimal adversarial policy: a model-based approach with tabular policies and
a model-free approach with parametric/neural policies. We showcase the efficacy
of the proposed algorithms through experiments.

Details

show
hide
Language(s): eng - English
 Dates: 2023-02-272023
 Publication Status: Published online
 Pages: 27 p.
 Publishing info: -
 Table of Contents: -
 Rev. Type: -
 Identifiers: arXiv: 2302.13851
URI: https://arxiv.org/abs/2302.13851
BibTex Citekey: Mohammadi2302.13851
 Degree: -

Event

show

Legal Case

show

Project information

show

Source

show