Papers
Topics
Authors
Recent
Search
2000 character limit reached

Learning a Generator Model from Terminal Bus Data

Published 3 Jan 2019 in stat.ML, cs.LG, and math.OC | (1901.00781v1)

Abstract: In this work we investigate approaches to reconstruct generator models from measurements available at the generator terminal bus using ML techniques. The goal is to develop an emulator which is trained online and is capable of fast predictive computations. The training is illustrated on synthetic data generated based on available open-source dynamical generator model. Two ML techniques were developed and tested: (a) standard vector auto-regressive (VAR) model; and (b) novel customized long short-term memory (LSTM) deep learning model. Trade-offs in reconstruction ability between computationally light but linear AR model and powerful but computationally demanding LSTM model are established and analyzed.

Citations (4)

Summary

  • The paper demonstrates a generator-agnostic approach to reconstructing generator models from terminal bus data using machine learning techniques.
  • The paper compares a VAR model, known for rapid and efficient prediction in linear dynamics, with a custom LSTM model that excels in handling non-stationary data.
  • The paper shows that the LSTM-based emulator significantly reduces NRMSE under varying noise conditions, indicating its promise for real-world power grid applications.

Learning a Generator Model from Terminal Bus Data

Introduction

The paper "Learning a Generator Model from Terminal Bus Data" (1901.00781) explores the use of machine learning techniques to reconstruct generator models based solely on data available at the terminal bus of the generator. The research is centered around developing online trainable emulators capable of rapid predictive computations. Two key models are investigated: the vector auto-regressive (VAR) model, which offers simplicity and computational efficiency, and a custom long short-term memory (LSTM) model, which provides superior reconstruction capabilities at a higher computational cost. The study leverages synthetically generated data to demonstrate the efficacy of these models, emphasizing a generator-agnostic approach.

Technical Background

Vector Auto-Regressive Model

The VAR model is utilized for its proven effectiveness in handling stationary time series. In this context, it is extended to manage multivariate data through equations that incorporate both endogenous and exogenous variables. The paper outlines how conditional dependence within time-series data informs the selection of the model order pp, where pp is critical for accurate prediction of generator dynamics. Figure 1

Figure 1: Colormap interpretation of inverted correlation matrices for each pair of variables. Upper triangular part of each plot corresponds to degree of conditional dependence of first variable's present and second variable's past. Largest sequence of non-zero elements ending with zeros constitutes a measure for model order pp.

Long Short-Term Memory Network

The LSTM model, noted for its efficacy in modeling long sequences, is enhanced using weight-drop techniques to mitigate overfitting. The customized model addresses the challenges of non-stationary data, directly applying to input variables without requiring transformations for statistical stationarity. This model leverages the architecture's ability to learn complex temporal dependencies and manage large fluctuations in data effectively.

Experimental Framework and Data

The research employs synthetic data generated via the OpenModelica tool and OpenIPSL package, simulating a typical power generator scenario. This data includes stochastic perturbations to model uncertainty, following dynamics outlined by a telegraph process. Such synthetic data provides a controlled environment to evaluate the robustness and scalability of the ML models. Figure 2

Figure 2

Figure 2: Telegraph process.

Results and Discussion

The study demonstrates the VAR model's capability to rapidly adapt and predict generator states under varied noise conditions, albeit with limitations when dealing with strong non-linear dynamics. In contrast, the WD-LSTM model excels in handling both typical and adversarial data conditions, showcasing a significant reduction in normalized root mean squared error (NRMSE) across scenarios. Notably, the LSTM-based model exhibits resilience to high-order noise and parameter randomization, indicating its robustness and potential applicability in real-world scenarios. Figure 3

Figure 3: WD-LSTM: Estimated density of the NRMSE distribution.

Figure 4

Figure 4: WD-LSTM: predicted values vs True values.

Conclusion

This work highlights the potential of adaptive machine learning frameworks in generator model reconstruction, showcasing a VAR model suitable for linear dynamics and an LSTM framework adept at capturing complex, nonlinear temporal dependencies. While the former provides a lightweight implementation, the latter affords comprehensive modeling power, meriting consideration for integration into current power grid predictive practices.

The paper concludes with a roadmap for future work encompassing physical model integration, application to real-world data, and expansion to other energy system components. These endeavors aim to refine the predictive toolkit available for power system management, enhancing operational reliability and efficiency.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

Knowledge Gaps

Unresolved knowledge gaps, limitations, and open questions

The following list enumerates what remains missing, uncertain, or unexplored in the paper, phrased to be concrete and actionable for future work.

  • Validate the proposed VAR and WD-LSTM models on real PMU/SCADA datasets from diverse generators and grid conditions, including measurement noise, time synchronization errors, and missing data.
  • Quantify the end-to-end computational cost (training and inference), memory footprint, and latency of the WD-LSTM and VAR models at typical PMU sampling rates (30–60 Hz), and assess real-time feasibility.
  • Resolve the mapping-direction inconsistency: the paper states learning a mapping from (P, Q) to (V, φ), but the LSTM is trained to map (V, φ) to (P, Q). Clarify intended use (emulation vs estimation) and evaluate both directions.
  • Establish formal stationarity diagnostics (e.g., ADF/KPSS tests) for input variables and their increments; quantify how differencing affects model bias, variance, and predictive fidelity.
  • Provide a principled, rigorously validated method for selecting VAR order under non-Gaussian, nonstationary noise (telegraph and fault processes), and compare to AIC/BIC/HQIC across regimes.
  • Analyze identifiability of a generator model from terminal bus measurements in multi-machine systems; determine conditions under which mapping between (P, Q) and (V, φ) is unique or confounded by network interactions.
  • Examine sensitivity of the models to sampling rate changes and downsampling (e.g., from solver’s 100 kHz to 10 Hz vs 30–60 Hz), including aliasing and information loss effects.
  • Characterize performance under realistic measurement noise models (additive, multiplicative, colored noise), sensor bias/drift, and data outages; test robustness and imputation strategies.
  • Quantify how prediction error grows with horizon for both VAR and LSTM; implement and compare stabilization strategies (e.g., closed-loop corrections, Kalman filtering, scheduled re-initialization).
  • Investigate why reactive power (Q) estimation degrades more than active power (P), including feature importance, observability limits, and controller interactions (AVR/PSS/governor).
  • Report detailed LSTM training protocol (sequence length distributions, dropout rates, weight decay, learning rates, batch size, hidden sizes) and perform ablations to identify which components drive performance.
  • Provide a sample complexity analysis: how many time-series samples and what sequence lengths are required to achieve target accuracy across generator types and noise regimes.
  • Quantify fine-tuning requirements (data volume, number of epochs, runtime) for domain shifts (randomized generator parameters, high-order noise) and define triggering criteria for online adaptation.
  • Demonstrate true online learning: incremental parameter updates, adaptation speed to changing generator/grid states, and mechanisms to avoid catastrophic forgetting.
  • Compare against additional baselines beyond VAR (e.g., NARX, state-space models with observers, temporal convolutional networks, transformers, Gaussian processes), and hybrid physics-informed ML.
  • Incorporate and evaluate physics constraints (e.g., power balance, voltage/current limits, rate-of-change limits), using constrained learning or regularization to prevent physically implausible outputs.
  • Provide uncertainty quantification (prediction intervals, calibrated probabilistic outputs) and reliability metrics, especially for rare events or near-instability conditions.
  • Test performance near operational limits (e.g., low inertia, large faults close to protection thresholds, saturation of limiters, OEL/V/Hz) and during protective actions, tripping, or islanding.
  • Assess generalization across generator technologies (salient-pole vs round-rotor machines), controllers (AVR, PSS, governors), and parameter variations beyond those in the Kundur 13.2 example.
  • Evaluate multi-generator settings and network-wide deployment: scalability, cross-coupling effects, and whether single-generator emulators remain accurate under strong inter-machine interactions.
  • Clarify the supervised learning target: one-step vs multi-step prediction; specify loss aggregation across time and outputs, and explore multi-objective losses that weight variables differently (e.g., prioritize φ or V fidelity).
  • Analyze the impact of using increments for VAR vs raw signals for LSTM; test alternative preprocessing (de-trending, filtering, normalization schemes) and their effect on performance and stability.
  • Provide a rigorous explanation for the counterintuitive increase in optimal VAR order when noise magnitude decreases; determine whether this is overfitting, model misspecification, or a data-generation artifact.
  • Evaluate the representativeness of the telegraph-process fault model relative to real-world fault statistics (spatial/temporal correlations, weather-induced patterns); explore more realistic stochastic processes.
  • Specify dataset composition (number of time steps per sample, total duration, train/validation/test splits, fault parameter distributions) and release code/data to enable reproducibility.
  • Investigate causality and lag structure: confirm that models are strictly causal (no leakage of future information), and reconcile the reported long lag dependence (up to 32 steps) with near-memoryless fault dynamics.
  • Explore safety-aware deployment: define guardrails, anomaly detection, and fallback strategies when predictions deviate or the model encounters out-of-distribution conditions.
  • Assess how models integrate with control and operations (e.g., AGC, dispatch tools): potential feedback-loop effects, stability implications, and requirements for certification in grid practice.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.