• Graduate Programs
    • Tinbergen Institute Research Master in Economics
      • Why Tinbergen Institute?
      • Research Master
      • Admissions
      • All Placement Records
      • PhD Vacancies
    • Facilities
    • Research Master Business Data Science
    • Education for external participants
    • Summer School
    • Tinbergen Institute Lectures
    • PhD Vacancies
  • Research
  • Browse our Courses
  • Events
    • Summer School
      • Applied Public Policy Evaluation
      • Deep Learning
      • Development Economics
      • Economics of Blockchain and Digital Currencies
      • Economics of Climate Change
      • The Economics of Crime
      • Foundations of Machine Learning with Applications in Python
      • From Preference to Choice: The Economic Theory of Decision-Making
      • Inequalities in Health and Healthcare
      • Marketing Research with Purpose
      • Markets with Frictions
      • Modern Toolbox for Spatial and Functional Data
      • Sustainable Finance
      • Tuition Fees and Payment
      • Business Data Science Summer School Program
    • Events Calendar
    • Events Archive
    • Tinbergen Institute Lectures
    • 2026 Tinbergen Institute Opening Conference
    • Annual Tinbergen Institute Conference
  • News
  • Summer School
  • Alumni
    • PhD Theses
    • Master Theses
    • Selected PhD Placements
    • Key alumni publications
    • Alumni Community
Home | Events Archive | Establishing error bounds for implicit and explicit score-driven filters
Research Master Pre-Defense

Establishing error bounds for implicit and explicit score-driven filters


  • Location
    Tinbergen Institute Amsterdam, room 1.60
    Amsterdam
  • Date and time

    August 22, 2023
    13:30 - 15:00

We present a novel error bound analysis for explicit score-driven (ESD) filters, better known as generalized autoregressive score (GAS) filters, and their implicit counterparts, referred to as implicit score-driven (ISD) or proximal-parameter (ProPar) filters. This analysis considers potential misspecification with respect to the data generating process (DGP) and the learning rate. Specifically, we establish upper bounds on the asymptotic root mean squared filtering errors (RMSEs) of ISD and ESD iterates. We derive these asymptotic error bounds under more general conditions regarding the DGP compared to conventional error bound analyses. In particular, we find that ISD filtering error iterates remain bounded asymptotically for any learning rate that is a scalar multiple of the identity matrix. In contrast, ESD requires the learning rate to be sufficiently small. These findings are confirmed in a Monte Carlo study examining nine univariate DGPs. Lastly, our theoretical analysis shows that even when the DGP is non-stationary, we can still guarantee asymptotic error bounds when using an identity prediction step in the filter.