[en] We address dynamics in abstract argumentation using a logical theory where an agent’s belief state consists of an argumentation framework (AF, for short) and a constraint that encodes the outcome the agent believes the AF should have. Dynamics enters in two ways: (1) the constraint is strengthened upon learning that the AF should have a certain outcome and (2) the AF is expanded upon learning about new arguments/attacks. A problem faced in this setting is that a constraint may be inconsistent with the AF’s outcome. We discuss two ways to address this problem: First, it is still possible to form consistent fallback beliefs, i.e., beliefs that are most plausible given the agent’s AF and constraint. Second, we show that it is always possible to find AF expansions to restore consistency. Our work combines various individual approaches in the literature on argumentation dynamics in a general setting.