A Reason to Invest in IT by Chris Trayhorn, Publisher of mThink Blue Book, March 11, 2004 Apart from demonstrating the vulnerability of the electric grid, the American and European blackouts of 2003 put enormous pressure on energy companies to prove they’re serious about raising reliability by investing in infrastructure. But how? Physical infrastructure requires a huge capital investment – $10 billion a year for the next decade, at the very least. However, the first, quick steps toward greatly enhanced reliability need not be so costly. They involve processes and technology, not vast changes in physical infrastructure. Consider this: More than four hours before the August blackout in the northeastern United States, there were clear indicators of grid instability. Both the New England independent system operators (ISO) and the Pennsylvania-New Jersey-Maryland (PJM) ISO managed to avoid substantial involvement because they have the technology to provide early visibility to grid conditions and clear lines of authority in place that allowed operators to either isolate from the grid or bring on generation necessary to meet demands on reactive power. In contrast, the control area at the Midwest ISO is highly fragmented, with no overall visibility to conditions. So, despite the fact that practically every technology vendor in the space boasts of having contracts in the ISO, the blackout spread quickly. Simply said, the technology exists to prevent massive blackout conditions, but visibility and authority must be in place at larger regional transmission organizations (RTOs). The Key Objective Reliability has always been critical for the transmission control operators, but never more so than today. Although North American Electric Reliability Council (NERC) standards are voluntary – there are no penalties – that is expected to change quickly. Energy companies also look at reliability from the perspective of generation plant up-time and availability, as well as outage frequency and duration on the distribution grid. Most energy companies list enhanced reliability as a key corporate objective for the year ahead. Customers expect greater levels of reliability, and the regulators demand it. At the same time, the industry is suffering from extreme pressure on the balance sheet, forcing companies to limit capital expenditures. Faced with this conundrum, the industry needs to look at reliability from three perspectives: the grid at the control level; reliability of equipment; and planning to expand capacity. We’ll examine each of these. The Control Level Although the August blackout pointed out deficiencies in grid control, the move toward more interconnection in recent years didn’t cause this problem. In fact, greater availability of power from a greater number of sources has prevented more frequent localized outages. The future of regional markets may be in doubt; witness the furious debate over this issue with regards to the Energy Bill. However, there were interconnections between control areas before wholesale markets and there will be interconnections in the future. The energy industry must accommodate that reality. The real need is to model and simulate the grid on an ongoing basis in order to understand how it responds. Knowledge gained from simulations through tools such as state estimation and contingency analysis allows protections to be built into the system at all levels – the local transmission grid, the control areas (such as ISOs, power pools, or RTOs), and even between control regions and countries. Operators can also use this knowledge to recognize patterns leading up to potential failure and take corrective action long before the situation becomes a crisis. This same type of approach proved necessary in very complex manufacturing environments, such as semiconductor plants. State estimation combined with contingency analysis to support automated decision rules or human intervention is the most practical approach to addressing future grid vulnerability. There are three steps to follow: characterize the risks; install or identify monitoring equipment for incipient conditions; and establish operating rules to make the grid more resilient. The requirements to support this approach are: Access to historical control data for analysis. Time series control data is already being collected through supervisory control and data acquisition (SCADA) for the grid and energy management systems (EMS) for generation. Companies with data delivery platforms, such as OSIsoft and their experience with the California ISO, have a long history of working in the power industry to provide access to control system data. IT provides access, but the data must be shared across interconnected control areas. Analytical engines to predict potential hot spots and simulate possible mitigation strategies. The idea is to recreate the demand conditions, understand what can happen, and then change the rules at local nodes to determine how the system reacts. Simulating the grid is no easy task. Each interconnection node and generator tie-in must be included in the analysis. There are tens of thousands of nodes to be considered. While calculation capacity needs to be robust and must accommodate large amounts of data, it does not need to occur at super speeds. In addition to the traditional providers of state estimators and contingency analysis, Emerson, AspenTech, and Pavilion Technologies have developed optimization capabilities in the power industry that go beyond reliability to profitability and environmental compliance – AspenTech with generation unit dispatch within the context of location on transmission grid, and Pavilion with power plant emissions with emission credits. Protections built into control systems to prevent potential disasters. Control systems are good at reacting to local conditions. Based on what is learned from scenario testing, logic is added to controls systems at points of vulnerability. Companies such as ABB, Alstom ESCA, GE Harris, and Siemens provide control systems. However, there is often little communication between SCADA and EMS. SCADA needs to be better integrated with EMS systems at local, control area, and regional areas so that these systems can respond more rapidly to potentially disastrous conditions. Sharing of real-time information between interconnected regions to act as a final failsafe. The US-Canada Task Force cited both FirstEnergy and the Midwest Independent Systems Operator (MISO) for communication failures. Regardless of how the debate on regional electricity markets resolves, there will still be a need for visibility of voltage status, not ubiquity of control, across interconnected control regions in real time. A few years ago, security was important to prevent access to information that would lead to gaming the system. Now, the concern is protection against terrorist attack. Secure portals allow for implementation of security measures at each level. Reliability of Equipment As we stated earlier, commercially available and proven technology already exists to support equipment reliability. An early warning system can alert energy companies to potential failures so that there is time to react, as long as those systems are turned on and functioning. If the time window is wide enough and other conditions are right, the turbine, transformer, or section of network can be taken offline and repaired or replaced, averting failure. If there is not enough time, energy can be diverted around the problem. Essentially, this involves a focus on the assets in the energy supply and delivery network (see Figure 1). Of course, it is important that reliability is maintained, but within cost constraints. At the same time, an energy company cannot lose sight of top-line opportunities related to the asset. However, while the technology for optimizing performance of assets exists, reliability-centered maintenance (RCM), RCM2, and asset lifecycle management have yet to be accepted as a corporate mandate. There are some basic steps that do not require a large new investment in information technology, but may require installation of additional monitoring equipment: Enhance existing enterprise asset management (EAM) and work management (WM) applications with plug-ins for scheduling and reliability-centered maintenance. Scheduling modules speed work turnaround, while RCM provides the analytical tools to determine the optimal maintenance schedule. EAM vendors with experience in energy, such as Datastream, IFS, Indus, Intentia, JD Edwards, MRO Software, and Mincom, have experience working with each. The advanced application architecture of the latest EAM applications makes integration with new modules by the same or other vendors easier. Enable access to control data for fault identification. There is already an active control system in place – SCADA and EMS. Data historians make control data easily accessible for analysis. OSISoft has the largest presence in transmission and distribution. Its experiences with Idaho Power and the California ISO are particularly notable. InStep’s eDNA is used by Southern California Edison to facilitate interfaces with five separate GE Harris SCADA systems widely distributed throughout the territory. Most utilities already use data historians in their engineering departments; IT should seek out these tools and support engineering in applying them. Enable access to monitoring equipment in the field to identify conditions that indicate potential failure. SmartSignal has applied its experience in the airline industry to the statistical analysis of condition data for generators to identify potential generator failures at companies such as Entergy and Dynegy. Look also at new approaches to monitoring transmission line sag. Work with engineering to determine whether there are enough of the right monitors in place to perform a robust analysis. Develop a business case to determine whether it is worth the investment by operations to add more. Put in place the connections to initiate a work process to address a problem asset. EAM offered by SAP is linked and in production with data historians like PI to automate the creation of work orders to investigate, repair, or replace an asset. Similar links are developed between EAM applications by Indus, MRO Software and JD Edwards. Keep tabs on how these links are working as they come into production; there are still lessons to be learned. Portals provide the right information to engineering supervisors. Condition-based monitoring is best paired with repair histories for determining the best approach to repairing or replacing a failing piece of equipment. TransAlta has a notable portal for engineering supervisors that serves up PI data, asset content from NRX, repair history on the asset from SAP, and other outside data services through SAP’s xApps architecture. But just having access to condition data is not enough. Energy companies need the right analytics to quickly determine whether there is a problem and to initiate the proper maintenance approach. There will be times when engineering services are required. Complex problems may not easily lend themselves to automated solutions. Vendors of EAM applications also offer add-on engineering services to be used in conjunction with software to determine the right maintenance or repair approach for a piece of highly engineered equipment or process. Companies such as Data Systems and Solutions and Invensys are currently offering these services. Take stock of your resident expertise and their capabilities; in-house resources may be able to use the analytical tools to construct specialized approaches at a lower cost. Planning to Expand Capacity Investment now in information technology will reduce the need for massive new physical infrastructure. Energy companies have more data available for planning than any other industry. The operation of the grid depends on information technology. Ultimately, there is a need for new infrastructure investment to replace aging pipes and wires, or to add new substations, transformers, and other infrastructure, especially in areas of high load growth. But there are alternatives to massive investment in infrastructure, through demand management and better forecasting. There are some promising new concepts and existing applications that can help energy companies use that data: Meter data at the substation level and below can be used to assess the need for new distribution infrastructure. ONEOK and Xcel Energy are at the beginning stages using Silicon Energy technology recently acquired by Itron. CES International has also done exploratory work in this area. Systems integrator SAIC has done extensive work in this area. This approach is still at the experimental level; companies with a business model built on T&D will want to invest in pilots. Start to investigate planning and design tools for new transmission. CGI, through its acquisition of Cognicase, Itron, through its acquisition of Linesoft, and niche vendors PLSCADD, Cook/Hurlbert, and Enghouse have transmission and distribution design functionality. ABB has tools that will aid in transmission citing. There are likely to be barriers against location of new infrastructure, so look for design tools that can aid in taking advantage of existing rights of way. Ultimately, using load response and market pricing signals – an approach that EPRI also recommends – is a better way to prevent grid failure than any failsafe device (see the AMR Research report “Economic Demand Response: Increasing Margins on Commercial and Industrial Customers”). A market mechanism that provides a price signal (power becomes more expensive when supply is low relative to demand, or transmission is constrained) provides a way to relieve pressures on the grid to perform. Energy companies, grid operators, and control vendors should spend 2004 shoring up their existing information technology infrastructure supporting reliability. Filed under: White Papers Tagged under: Utilities About the Author Chris Trayhorn, Publisher of mThink Blue Book Chris Trayhorn is the Chairman of the Performance Marketing Industry Blue Ribbon Panel and the CEO of mThink.com, a leading online and content marketing agency. He has founded four successful marketing companies in London and San Francisco in the last 15 years, and is currently the founder and publisher of Revenue+Performance magazine, the magazine of the performance marketing industry since 2002.