Loading...

Asset health: doctor blade

Case Study

Context

Doctor blades are used in printing to remove the excess ink from a cylinder to create a uniform layer of ink.

If the blade is not sharpened correctly, wrongly adjusted, or is damaged, this can result in millions of dollars of asset damage.

In a nutshell, what Quasar delivered

  • Greatly increased accuracy of prediction and models by capturing the full waveform data and performing spectral analysis.
  • A shorter model to production loop that empowers data analystis and reliability engineers
  • 1,000X faster queries compared to a datawarehousing solution while using less cloud resources
  • Storage footprint divided by 10
  • Future proof solution thanks to built-in scalability

Challenge

The blade is changed or adjusted based on a technician’s hunch and experience or current practice.

However, collecting the vibration data coming out of the sensors makes it possible to identify common patterns present before problems happen.

The goal is to have a dashboard with precise information regarding the imminence of a problem and its type. In addition, it is possible with this approach to find the optimal time to change the doctor blade, preventing asset damage and optimizing parts inventory.

However, having this level of accuracy requires working on the entire waveform of the vibration data and performing a frequency analysis. Complete waveform data is usually several thousands of points per second per sensor, resulting in massive data sets.

The other challenge is that sometimes problems are identified just minutes before they happen, which means it’s imperative to have an end-to-end data latency update below the minute.

That update latency and write volume disqualify data warehousing technology immediately.

An approach that solves the write volume problem is to store the data in files and then have ad-hoc programs perform the feature extraction. This, however, means that each analysis may have a high CPU time cost, and it requires writing new programs if the feature extraction needs to change.

Finally, these programs must be highly efficient to not add to the data latency. If it takes 15 minutes to extract the feature, the information may come too late to take action.

The solution

Using Quasar, firms can capture the complete waveform data using dedicated connectors. The capture can happen at the edge or via MQTT.

Quasar will use its unique algorithm to compress the data faster than it arrives and use its micro-indexes to reduce future lookup and feature extraction time. This results in significantly reduced disk usage without impacting the queryability of the data.

Each sensor is stored in a dedicated table storing timeseries data. Tables can be tagged at will, enabling a flexing querying mechanism based on those tags. For example, obtaining all the sensors of an asset or all the sensors of a mill.

Quasar has built-in feature extraction capabilities that leverage the capabilities of the SIMD capabilities of the processor. It can, in real-time, perform complex statistical analysis, pivot the data, do FFT, and much more.

This means that feature extraction does not require ad-hoc programs and can be tailored by writing a simple SQL query.

Quasar can keep a deep history, meaning that the same system is used for operational reporting and building the strategy. This significantly reduces the time between devising a new approach and deploying it in production.

Benefits

  • Firms can perform advanced data analysis to protect their asset health and optimize their parts inventory, increasing competitiveness
  • Capture the entire waveform without any limitation in history depth, increasing accuracy and thus potential gains
  • Devising and deploying new strategies is extremely fast since Quasar contains both the history and the live data, increasing productivity of rare data scientists resources
  • Extracting feature does not require writing custom programs, increasing flexibility
  • Outstanding TCO: extracting features does not require writing custom programs and uses less CPU. Up to 20X less storage space is needed compared to a typical data

warehousing solution and up to 1,000X faster feature extraction compared to data lakes


Related Tags:
Top