Transforming Electric Utility's Performance with Data Analytics
Data Analytics and Utilities
People disagree on many topics in today's world, but most would admit that data analytics, as a powerful tool, is transforming many businesses in various ways. From marketing to sales, from operations to R&D, data analytics and its derivatives play an increasingly important role in almost every aspect of a business's ecosystem, with no exception to the power utility industry. The good news is, the utility, or power industry in general, is never short of data. On the transmission side, traditional data sources such as SCADA record basic electric parameters such as voltage, frequency, and power flow every 2 or 4 seconds for just about every bus and line in the grid. The phasor measurement unit, or PMU, measures the same quantities with up to 120 samples per second to capture the transient events in the power grid. On the distribution side, smart meters and sensors record data at the feeder or individual household level every few seconds. For deregulated states in the U.S., the Independent System Operators (ISOs) calculate and publish prices, generator dispatches, and load and renewable forecasts on a 5 or 15 minutes basis. The list goes on...
While utilities are sitting on ample opportunities to benefit from data analytics, the reality is that few have appreciated the data enough nor utilized them to the full potential to guide the decision-making process in daily operations compared to other industries such as finance and marketing. So how will the data help utilities transform their business? The answer is in many ways, and the following are just a few of many possibilities:
Type of Data Analytics
The famous Gartner analytic ascendancy model below categorizes analytics into four types: descriptive, diagnostic, predictive, and prescriptive. Descriptive and diagnostic analytics analyze historical data and try to answer what happened and why it happened, while predictive and prescriptive analytics focus on the future and predict what will happen and what actions to take. The analysis will arguably become more complex as moving up the arrow but will also add more value and generates greater insights. This is not to say that one analytics is clearly better than the other. In fact, the need for the type of analytics varies at different stages of a company. In recent years, however, the trend is moving towards predictive and prescriptive analytics types 2.
Data Analytics at SMUD Grid Operations
Shortly after SMUD joined the CAISO Energy Imbalance Market (EIM) in April 2019, we decided to pursue a data-driven approach and initiated an analytics project to facilitate the after-the-fact EIM analysis, answering what happened and why it happened. Later, we added a predictive component to provide operators and traders the situational awareness of daily reliability and market operations. The project involves four steps: pulling data from the California ISO's Open Access Same-time Information System (OASIS) and Customer Market Results Interface (CMRI) portals and stored in the local database, cleaning and reconstructing the data, perform statistical analysis, and display to the users as shown below. The analytics part of the project was developed in R using various packages such as flexdashboard and shiny. The final product - an interactive dashboard - is hosted in Docker.
As illustrated below, the dashboard has five tabs, each displaying daily key EIM performance metrics such as real-time prices, congestion, intertie power flows, and generation dispatches. The user can pick a day from the date panel on the left and navigate through the tabs to review the metrics for that day. These displays focus on answering what happened and why it happened, which are critical in the after-the-fact analysis to learn from the mistakes and work toward operational excellence.
As demonstrated below, SMUD's EIM performance is also summarized on a rolling 7-day and 30-day (not shown) basis for engineers and management to track and take appropriate actions as needed without delay.
Recently, a predictive component is added to quantify risks and help engineers and traders understand how prices and flex requirements vary across different hours of a day. Based on statistical analysis using historical data, users will be able to predict the next day's price or flex requirement with a level of confidence and risk allowance. With this information in mind, schedules can be developed more accurately for real-time operations. Are these the most advanced data analytics techniques? No. The techniques used here are no more than data reconstruction and basic regressions. However, don't overlook simple techniques like these because the cost to implement something like this is usually relatively low, yet the benefit, which is what you should focus on, could be significant.
We also have explored more advanced machine learning techniques to predict flows on a transmission path (see below) and the reserve requirements. Based on two weeks of testing data, the Long Short Term Memory (LSTM) model predicted the flow 25% more accurately than the current method on the power flow, and 15% more accurately on the reserve requirement than the current method. However, there is still a long way to implement this type of method in production, mainly due to risk and compliance concerns.
Conclusions
The utility industry is undergoing a paradigm shift. To be successful in this transition, it is necessary for utilities to fully utilize the data they own to transform the business in many ways. The following are some suggestions to be considered:
Copyright by Hui Zhang, 2020, modified in 2022.
Sr Director @ ContourGlobal, a KKR company
3yEdited with the latest updates in 2022.