In this information age, data is everywhere. How can we improve efficiencies and manage this information into useable nuggets of information? Plant and operations managers receive vast amounts of both structured and unstructured data every day. How can information be accessed quickly and affordably to improve performance?
Historians serve as a repository for data from many systems, making them a good source for advanced analytics. However, process historian tools are not ideal for automating the analysis of the data or search queries. They are ‘write’ optimized and not ‘read/analytics’ optimized. Finding the relevant historical event and building the process context is usually a time consuming and laborious task.
A level of operational intelligence and understanding of data are required to improve process performance and overall efficiency. Process engineers and other personnel must be able to search time series data over a specific timeline and visualize all related plant events quickly and efficiently. Part of this is the time series data generated by the process control and automation systems, lab systems and other plant systems as well as the usual annotations and observations made by operators and engineers.
PREDICTING PROCESS PERFORMANCE TODAY
In order to run a plant smoothly, process engineers and operators need to be able to accurately predict process performance or the outcome of a batch process, while eliminating false positives. Accurately predicting process events that will likely happen in a plant or facility requires accurate process historian or time-series search tools and the ability to apply meaning to the patterns identified within the process data.
While there exists a variety of process analytics solutions in the industrial software market, these largely historian-based software tools often require a great deal of interpretation and manipulation and are less than automated. They perform rear-looking trends or export raw data in Microsoft Excel. The tools used to visualize and interpret process data are typically trending applications, reports and dashboards. These can be helpful, but are not particularly good at predicting outcomes.
Predictive analytics, a relatively new dimension to analytics tools, can provide valuable insights about what will happen in the future based on historical data, both structured and unstructured. Many predictive analytics tools start by using a more enterprise approach and require more sophisticated distributed computing platforms such as Hadoop or SAP Hana. These are powerful and useful for many analytics applications, but represent a more complex approach to managing both plant and enterprise data. Companies that use this enterprise data management approach often must employ specialized data scientists to help organize and cleanse the data. In addition, data scientists are not intimately familiar with the process like engineers and operators, which limits their ability to achieve the best results.
Furthermore, many of these advanced tools are perceived as engineering-intensive “black boxes” in which the user only knows the inputs and expected outcome, without any insight into how the result was determined. Understandably, for many operational and asset related issues, this approach is too expensive and time consuming and require highly skilled data scientist. This why a lot of the vendors target only the 1 percent of critical assets, ignoring many other opportunities for process improvement.
HOW TO MANAGE BIG DATA WITHOUT A DATA SCIENTIST
There are just a handful of solution suppliers that are taking a different approach to providing industrial process data analytics and also leveraging unique multi-dimensional search capabilities for stakeholders. This approach combines the ability to visualize process historian time series data, overlay similar matched historical patterns and provide context from data captured by engineers and operators.
The ideal pattern recognition solution provides on-premise, packaged virtual server deployment that easily integrates to the local copy of the plant historian database archives and evolves over time towards scalable architecture to communicate with available enterprise distributed computing platforms.
This newer technology uses “pattern search-based discovery and predictive-style process analytics” targeting the average user. It is typically easily deployed in less than two hours, delivering immediate value, with no data modeling solution or data scientist required. Often called “self-service analytics” this software puts the power of extensive search and analytics into the hands of the process experts, engineers and operators, who can best identify areas for improvement.
Another problem typically presented by historian time series data is the lack of a robust search mechanism along with the ability to annotate effectively. By combining both the search capabilities on structured time series process data and data captured by operators and other subject matter experts, users can predict more precisely what is occurring or what likely will occur within their continuous and batch industrial processes.