Transformer Fault Prognosis Using Deep Recurrent Neural Network over Vibration Signals

Amin Zollanvari, Kassymzhomart Kunanbayev, Saeid Akhavan Bitaghsir, Mehdi Bagheri

Research output: Contribution to journalArticlepeer-review

Abstract

Vibration analysis is considered as a cost-efficient and nondestructive technique to monitor the transformer operating conditions and evaluate the transformer mechanical integrity. This method enables transformer fault prognosis before insulation catastrophic failure. In this work, the possibility of using deep neural networks in capturing the hidden patterns of vibration time series to predict the transformer underexcitation and overexcitation and interturn fault progress prediction in early stages is examined. This focus is warranted because deep learning techniques lend themselves to integrating the feature extraction into the predictive model construction stage. In this regard, deep recurrent neural network (RNN) architecture, including unidirectional and bidirectional gated recurrent units (GRUs) and long short-term memory (LSTM) models, are adopted. The constructed RNN for predicting excitation voltage exhibits a remarkable performance with a relative absolute error (RAE) of 0.56%. Predicting the interturn fault proved to be a more challenging problem and the constructed RNN for this purpose showed an RAE of 17.58%. The source code for implementing all constructed models is available at https://github.com/azollanvari/FaultPrognosisDL.

Original languageEnglish
Article number9205841
JournalIEEE Transactions on Instrumentation and Measurement
Volume70
DOIs
Publication statusPublished - 2021

Keywords

  • Deep learning
  • recurrent neural network (RNN)
  • time series
  • transformer fault
  • vibration analysis

ASJC Scopus subject areas

  • Instrumentation
  • Electrical and Electronic Engineering

Fingerprint Dive into the research topics of 'Transformer Fault Prognosis Using Deep Recurrent Neural Network over Vibration Signals'. Together they form a unique fingerprint.

Cite this