Metrology has a key position in networked, adaptive production, with the task of a holistic and valid assessment of the state of various production scenarios. With the diminishing focus on a device-specific development towards an adaptive production network, which is less hierarchical in the sense of the “Internet of production”, and with the focus on the properties of cyber-physical systems (CPSs), new opportunities for the strengthening of metrology arise. Characteristic of these CPSs are sensors for multi-modal data acquisition, actuators for interaction with the environment, distributed computing power and the ability to spontaneously or permanently network itself. They form the basis for the creation of a “digital shadow” and thus are essential components of a model for process control. Current trends and challenges for metrology in networked production, such as multi-sensor systems, model-based measurements, virtual measurement processes or the integration into adaptable production systems, broaden the boundaries of future requirements of metrology, in particular with regard to its flexibility, speed and compatibility. A prerequisite is a scalable, specifiable information fusion. A solution to this is the service-based provision of sensor information, measurement data and decisions, which can be flexibly adapted to task-specific requirements. For this concept of “sensor information as a service”, development stages and prerequisites for its implementation as well as affected areas are discussed.
Infrastructure of the Internet of production (Brecher et al., 2017).
The core of the concept of Industry 4.0 is the connection of digitized products and production resources based on cyber-physical systems in real time (Bauernhansl et al., 2017). With the development of products and production resources towards intelligent systems and their connection to the “Internet of things” (IoT), services, methods and processes coming from the information technology sector are becoming increasingly important in the production environment (Bauernhansl et al., 2017; Brecher et al., 2017, 2014; Monostori, 2014). This trend goes hand in hand with the so-called “Internet of production” with its paradigm shift towards the focus on data-centric business models to secure competitive advantages (Brecher et al., 2017). This paradigm shift is based on several factors: on the one hand on cost-effective high-performance sensors and thus on a strongly increasing number of recorded production data. On the other hand, it is based on steadily decreasing costs for storage space and thus an increased number of available historical production data. In addition, the available computing power is exponentially increasing, enabling simple and fast data processing, data analysis and feedback of the results into production. The Internet of production focuses on real-time reliable and secure availability of information at any time and any place to enable systematic learning from the data collected for decision support. A schematic infrastructure of the Internet of production is shown in Fig. 1. The main driver of these three developments is the market of consumer electronics (Bauernhansl et al., 2017; World Economic Forum, 2017).
When comparing the digitization of different sectors, recurring trends and development stages can be identified, which can be represented, for example, by the so-called “digital maturity level” (O'Leary, 2009; Fenn and Linden, 2017; Schuh et al., 2017). This maturity assessment covers all areas of the company: from the degree of digitization of business processes to the management of IT processes, IT resources and IT technologies, to the assessment of a digital strategy and market development. A development that has just entered the field of production technology is “cloud manufacturing” (Wu et al., 2013; Xu, 2012; Soyata, 2015). Wu et al. (2013) define the concept of cloud manufacturing as a customer-centric production model that uses demand-based access to a shared network of different and distributed production tools to create temporary and reconfigurable production lines. Part of this development, which results from the low cost of data transfer, is the provision of customized services managed in the cloud and sold as a service. It is called “everything as a service” (XaaS) (Yang and Tate, 2012; Duan, 2012; Duan et al., 2015). A typical example is the provision of software without the user having to own or install the software, but only needing to access the software via the Internet (“software as a service”, SaaS). In the case of cloud-based services, this development is continuing, with the result that not only the software is offered as a service, but also, for example, the required infrastructure, hardware or entire platforms (“infrastructure as a service”, “platform as a service”). An overview of the current manifestations of the concept of everything as a service can be found in Duan et al. (2015).
The central questions of this article therefore are as follows. To what extent can the concept of everything as a service be transferred to metrology in a production environment? Which potentials result from this, and what could a possible design look like?
To answer these questions, it is worthwhile looking at current challenges and trends in metrology in networked production and related developments (Sect. 2). It is then explained what is meant by the term “sensor information as a service” and what these services can look like (Sect. 3). Based on this, necessary and expected developments are discussed, which are to be pursued for the implementation of the service concepts in metrology (Sect. 4).
The shift towards a digitized production influences the role of metrology within production. Current trends in production technology demand the further development of metrology, a process that can be described quite well with the terms “measuring in the model”, “model-based measurement processes”, “model-based process control” and “cyber-physical production systems” (CPPSs). The introduction of the “service” concept to metrology can be an answer to these challenges and can be seen as a complement to the presented trends.
An overview of the challenges and trends in metrology has already been
provided by Imkamp et al. (2016, 2012), as well as by Schütze and
Helwig (2017), and serves as a basis for the following overview. The key
challenges for measurement technology can be described as faster, more accurate, more secure, more
flexible and more holistic, faster meaning a requirement for higher process
cycle times which can be achieved by increased integration, a reduction of
measurement times or automated data processing. The demand for more accurate
metrology results from the trend towards smaller tolerances and higher
manufacturing precision in the production and is decisively reflected in the
requirements for measurement deviations and measurement uncertainty, meaning
the systematic and random deviations of measurement results as described in
the
Characteristics of a cyber-physical production system (CPPS) using the example of a free-chain assembly scenario. This figure is translated from Schmitt and Voigtmann (2017).
The use of cyber-physical systems (Monostori, 2014; Brecher et al., 2014;
Rajkumar et al., 2010; Broy, 2010), their interconnection and the
modularization of services as well as the encapsulation of subprocesses will
make production even more flexible. The future of production as a CPPS is
designed for individual production as well as a flexible adaptation of
production to the requirements of the customer (Monostori, 2014; Monostori et
al., 2016). The basic characteristics of a CPPS are given according to the
working group Industry 4.0 (Brecher et al., 2017, 2014):
data acquisition via production-integrated sensors and measurement systems
in real time, storage and evaluation of data for the purpose of modeling, interaction of the physical, human and digital world (multimodal human–machine
interfaces) by actuators and connection via digital communication devices among themselves as well as
with the Internet of things and services.
This concept, which is based on multi-modal metrology, allows parts, hardware and sensors to be monitored and
controlled within the entire production area; see Fig. 2. Production engineering concepts,
which combine the indisputable advantages of a
synchronous production flow with the free-networked island production (which
itself focuses on volatility and small batch production) in a “Beyond Toyota” approach, are based on a
mobile and modular design of all components. One example is the concept of a
free float flawless assembly (Huettemann et al., 2016). The goal is to make
products, production resources and technologies available as required in the
entire workspace (Huettemann et al., 2016; Brecher et al., 2017). For
example, in the case of highly precise production steps, a plurality of
sensors can therefore be used together or additional sensors can be driven to
the workplace in order to ensure more reliable monitoring and control of
the production step (Brecher et al., 2017).
The goal is an automated and autonomous process control, which collects the required information as needed. As a result of the continuous monitoring of the products and production processes, an expandable digital image of the product (“digital twin”), which is designed in product development and various application scenarios, and an application-specific image of the production state (digital shadow) are created across all production steps (Bauernhansl et al., 2017; Brecher et al., 2014; Spencer et al., 2004). In the future, this will enable the inspection of the product using the digital representation of the product. As a long-term outlook, it is to be assumed that separate inspection processes and corresponding procedures for the final inspection will be drastically reduced.
As a result of this trend towards distributed and connected measurement and production systems, further challenges arise for metrology. In order to enable the connection and communication between measurement and production equipment, different scales have to be homogenized, various hardware (sensors, machines, IT infrastructures) have to be combined and manufacturer-specific controls and visualizations have to be standardized. This is the only way to ensure a uniform data exchange and uniform analysis and interpretation of the data. The central challenge here is to define these uniform data formats and communication interfaces independent of the manufacturer. In particular the provision of metadata is of greater importance in order to be able to describe the data as specifically as possible and to be able to use the data universally. On the one hand data formats and interfaces must therefore be standardized; on the other hand they should be able to record data and associated metadata from all kinds of different measurement and production processes.
There are some industry and application-specific approaches to solve these challenges, but no internationally established standards. Current developments are the STEP data format for product data within the ISO 10303 series (ISO, 2018), the RAMI model for the architecture description of Industry 4.0 components (VDI, 2017; DIN, 2016; Löwen et al., 2017, 2016) and OPC-UA as a communication protocol for machine data (Candido et al., 2010; Hannelius et al., 2008; Montavon et al., 2017). Further approaches can be found in context-based data formatting (Perera et al., 2014a), self-describing data models (NetCDF 2017), a universal description of metadata (Guenther and Radebaugh, 2004; Koza, 2003) and especially adapted descriptions for sensor networks in Li et al. (2013), Qingping et al. (2014), Huang and Javed (2008), Goos et al. (2003), Compton et al. (2009), Barnaghi et al. (2009) and Kim et al. (2008).
Self-optimizing assembly of aircraft structures. This figure is translated from Schmitt and Voigtmann (2017).
If one follows this argument, it is to be expected that models for the feedback of the data with the aim of process control become more exact but at the same time more complex (Maropoulos et al., 2008, 2011; Pabst et al., 2010; Klocke et al., 2013; Schmitt et al., 2012; Möhring et al., 2010), since more and more process parameters and disturbance variables are considered in the process models for a more precise description and prognosis of the production processes. These parameters must be acquired and made available during operation. In addition, the quality of the data and their interpretation determines the quality and efficiency of the control. Hence, not only is it necessary to provide more data, but the data quality must also be ensured, meaning the completeness, consistency, timeliness and accuracy of the data which is required for its use. In combination with an increasing extent of monitoring and control of production processes, this leads to a high demand for valid, holistic process and status data. This requirement, however, can only be met by metrology. This is also known as information-rich metrology (Leach, 2016). Hence, in the future metrology will no longer be only necessary – in the metrological sense – for measurement and inspection tasks, but rather will play a dominant role in the production process with the provision of customer-tailored digital information. The current development of model-based process control will be illustrated by three examples. These examples have the fact that they use classical criteria of the inspection process capability as optimization parameters in a production-technological context in common.
An example of the model-based, metrology-assisted process control of production and assembly processes is the self-optimizing assembly of aircraft structures; see Fig. 3 (Permin et al., 2016; Schmitt et al., 2013a). The fulfillment of the measurement task is not the primary goal in this application, rather the fulfillment of the assembly task is. To optimize the assembly process, a model is created for the individual component-dependent geometric errors. The model allows compensation of geometrical errors during the assembly process. The compensation routine uses a large number of heterogeneous (sensor) data for this purpose. For this purpose, the pose of the robots and the part is recorded by means of iGPS and laser trackers, and the forces acting on the part, device and robots are recorded by means of strain gauges and load cells. The measurement technology serves as a link between the real and the virtual world. In addition, the CAD data and the robot control data are included. In this example, the challenge consists, on the one hand, of the modeling of the geometrical errors that has to be as realistic as possible, and, on the other hand, the homogenization of a wide range of sensor, plan and control data. A uniform definition of interfaces is essential for the modularization via services, in particular for the provision of the described sensor data as model input.
Another example of the use of complex process models in connection with metrology is the control of machine tools. By developing more comprehensive process models, the position and orientation of the tool center point can be controlled more precisely using external laser trackers than by the systems themselves, which are subject to load-dependent deformation. The model for the compensation of the kinematic errors uses the more precise position information of the laser trackers as well as the ambient temperature, the internal sensor system and the information on the axis errors (Schmitt et al., 2013b; Blunt, 2013; Altintas et al., 2005; Schwenke et al., 2005).
An example of the model-based and requirement-oriented usage of the measurement systems themselves has already been implemented for a distributed measurement system of infrared cameras (Franceschini et al., 2016, 2012; Galetto et al., 2015, 2010). The goal is to optimize the number of sensors needed in a sensor network. The measurement uncertainty depends on the number and the position of the sensors used and can be calculated for each configuration using the developed model. If a limit value for the measurement uncertainty exists, the service for determining the measured value can select and request an economically optimal number of sensors. The measurement uncertainty becomes the manipulated variable in the production control. In this way, the measurement uncertainty can automatically be optimized as a target value in the production scenario. It is thus possible, for example, to ensure that only the number of sensors is used that is actually required and thus the efficiency of the measurement is increased. Unnecessary sensors can be used for other measurement tasks. In contrast to the models for virtual measurement systems (see Sect. 3.3), this is an external model and not a measurement process-integrated model.
As described in the sections before, the focus of development in metrology will shift from device engineering to the provision and model-based linking of data. The deployment, inter-connection and flexibility of production in the vision of cyber-physical production systems is a central field of development and opportunity for metrology. The key to the implementation of model-based process control is sensor systems, which allow for a holistic and valid recording of the current state of the production scenarios. The central requirement for measurement technology is therefore a flexible and demand-oriented provision of homogenized measurement data from distributed measurement systems. One possible solution to meet this requirement is the modularized, standardized and model-based provision of sensor information using the concept of sensor information as a service, which allows an application-oriented high specificity. This concept of sensor information as a service as well as its stages of development and application scenarios are presented in the following chapters.
Connection of sensors with virtual services. This figure is translated from Schmitt and Voigtmann (2017).
The concept of sensor information as a service (S2aaS, see Kantarci and Mouftah, 2015; Perera et al., 2014b; Sheng et al., 2013; Zaslavsky et al., 2012) is based on a service-oriented architecture in computer science, which is increasingly applied to business processes with a high demand of flexibility and compatibility in distributed processes (Candido et al., 2011; Bell, 2012; Rosen, 2008). The core aspect is to divide business processes into individual activities and to encapsulate them with clearly defined tasks. The modules offer their functionalities in the form of services. The service consists of a retrievable service that is fixedly defined towards external users, without the implementation or execution of the service being visible. Communication is done via standardized interfaces. Due to the division into hierarchical levels and the access of higher services to services of lower hierarchical levels, complex services can be processed in a modularized manner. The two basic principles of a service-oriented architecture are the separation of responsibilities according to the provided services as well as the encapsulation of the technical details of the implementation. The goal is to maximize the reusability of the services and to maximize a flexible usage of the output. Sensor information as a service is therefore understood as the connection of physical sensors with (software) services, which can be accessed in a standardized manner using the Internet or local networks. The service provides one or more measurement or (relevant) status values as the primary output. However, for the user the generation and acquisition of the values stays hidden. Similar to the concept of everything as a service, the physical implementation and processing is separated from the provision of sensor data (Candido et al., 2010; Krafzig et al., 2005; Rosen, 2008). For the concept of sensor information as a service, it is therefore necessary in the first step to link physical sensors with services; see Fig. 4. Therefore the encapsulation levels for sensor systems must be expanded. Previously, sensors were designed without interfaces, but sensor information as a service requires interfaces to the Internet or local networks. By combining services of low abstraction levels (for example 1-D position information), services of higher abstraction levels (for example, 6-D position information) can be created. The scope of the services ranges from providing the hardware and the recording of the sensor data via the filtering and pre-processing of the data to the combination of several individual sensors.
Advantages of the service concept are generally a high degree of standardization, high flexibility, more availability and lower investment costs. With regard to metrology, this means that the two requirements of being more flexible and faster can be met by the sensor information as a service. The increase in flexibility is achieved by the fact that, if necessary, a large number of measured values are universally available and are initially offered without a direct application. In many cases, existing sensors can be used for tasks for which they were not intended in the first place. Therefore, services can be designed specifically for 1 : 1, n : 1, 1 : n or m : n links, but services can also access output information from other services. Their heterogeneity, combined with their specificity, make measurement data and feature values available more quickly by using the standardized and universal connectivity of the sensors as well as by using an automated analysis and combination of the sensor data. However, an increase in the accuracy or reduction of the measurement uncertainty in distributed systems can also be achieved by the service concept. If, for example, the measurement uncertainty is too high for the application scenario and a service supports the exchange or the addition of measurement systems, further sensors, possibly with another modality, can be addressed, which then in combination have a lower measurement uncertainty.
For the implementation, a service-oriented architecture of sensor data acquisition and processing is necessary. The first applicable approaches are described by Kantarci and Mouftah (2015), Perera et al. (2014b, c), Sheng et al. (2013) and Ibbotson et al. (2010). Zaslavsky et al. (2012), Broring et al. (2011) and Yuriyama and Kushida (2010) describe the connection of the sensors to the Internet and the cloud. On the one hand, a homogenization of the recorded data and of the communication interfaces is necessary for the concept of the sensor information as a service and, on the other hand, methods are necessary to validate the functionality of the services and to calibrate measurements carried out by the services. The homogenization of the interfaces is necessary, especially for the connection of arbitrary sensors and the establishment of service hierarchies as well as the universal use of these services for data analysis and within feedback loops. Homogenization involves the standardization of protocols, request structures, interfaces and data formats. In the future, the question will be how distributed data sets can be stored and described in the individualized production and how to deal with heterogeneous data. The goal is a flexible and product-centered collection of production process data, which is organized using partial homogenization and is provided ad hoc to users.
In order to acquire valid and comparable measurement data with a service, the systems for the generation and determination of the measurement and sensor information have to be calibrated. However, this requires, in addition to the calibration of the physical measurements, a calibration of the evaluation algorithms used in the service as well as a calibration of the underlying models – a change from the previous approaches. Although for a long time models have been used for the calculation of measurement uncertainties, e.g., according to the “Guide to the Expression of Uncertainty in Measurement” (GUM; the so-called measurement model), these are usually only validated experimentally. However, the models described above require a structural validation. Therefore, calibration methods are more and more required for the validity of the underlying logic and its implementation. These virtual standards serve as a reference for the calibration of the services. For software there are already approaches to solve this challenge using reference data sets, which are provided, for example, by central authorities such as the Physikalisch-Technische Bundesanstalt (PTB) or the International Organization for Standardization (ISO). Examples are the TraCIM project (PTB, 2017b), reference data sets for roughness measurements (PTB, 2017a) or reference data sets for the implementation of statistical process control (ISO, 2016). The additional software-specific traceability can also be used in the future in the case of more complex measurement systems, for example for computed tomography, in order to improve the traceability of the physical measurement in that the error of the classical calibration is already compensated beforehand by the error of the evaluation. In other words, the knowledge that is reflected in the application-specific digital shadow of the service modifies the underlying model itself.
The increasing number of sensors in use makes a regular and economically acceptable calibration using standards and references and the associated classic traceability more difficult. In order to be able to use these sensors for measurement tasks in the future and to ensure the validity of the data without regular calibration, current work focusses on fundamental self-calibration strategies and procedures to which the metrological infrastructure must then be adapted. Self-calibration has long been relevant to industries in which long-lasting sensors are used and which are not accessible to recalibration, e.g., in space travel or meteorology. Within these application scenarios, first methods for self-calibration can be found.
The methods themselves are based on different approaches. A common method is the use of substitute references. For example, the use of known quantities in the acquired image area for the calibration of optical systems (Ling and Strohmer, 2015; Lu et al., 2013), the calibration of inertial sensors via built-in microchips, which produce reference Coriolis forces (Aktakka and Najafi, 2016) or the usage of quantum-based electrical standards as developed by NIST; see for example the “NIST-on-a-Chip” program. (Flowers-Jacobs et al., 2016). A method specific to the area of multi-sensor systems is checking the consistency with other identical sensors and matching their calibration status (Goos et al., 2003; Kelly and Sukhatme, 2011; Dai et al., 2016). Other methods are based on internal consistency checks and mathematical methods for verifying validity, and can also be used, for example, for the calibration of coordinate measurement devices (Dang et al., 2006; van der Horn and Huijsing, 2011). As described in Sect. 2, models for controlling real production processes will be modular in the future. The data required in the models for control are then requested from services instead of directly from sensors and are provided with all necessary additional information (see Fig. 4). The development effort for new models is divided into the development of service-internal models and the modeling of the production processes. The development of the models for the services includes, on the one hand, models of the sensors and, on the other, models for combining sensors with higher order services. For example, a service that provides a pose, but which must combine this pose from several different sub-services using different measurement systems (laser tracker, iGPS, robot control), requires a model for the acquisition, combination and validation of the individual measurement systems. Generally speaking, with the use of services that use heuristics and adaptive learning methods, sensor data transform into smart data, which enables manual or automated event-driven decisions and adaptive behavior.
Development stages of the service concept in metrology, translated from Schmitt and Voigtmann (2017).
Another focus of development will be on the methods for a measurement system analysis and process capability studies as well as on the methods for process monitoring as these methods have to be adapted for the use on and with services. In particular, the process monitoring will shift from the monitoring of individual sensors to the monitoring of services and virtual sensor groups.
Analogous to the development of the concept of software as a service to the concept of Platform as a Service, the development of the sensor information as a service can be taken further, so that the services will take on additional tasks. Figure 5 shows, based on the development stages of the concept software as a service, different levels of services for metrology and different stages of outsourcing of individual tasks as a service. The responsibility and the management of the tasks shift from left to right more and more to the service provider. The first step is that without the provision of a measurement uncertainty, the value of a characteristic provided by a service is also worthless. For the determination of the measurement uncertainty, a large amount of information about the environmental conditions, as well as models for linking this information to the measurand, is necessary. Therefore, for the concept of “characteristic as a service” in addition to the value of the characteristic, all environmental conditions of the process and all uncertainties must be determined and combined. This information itself can, in turn, be collected using services or must be provided by the services for the sensor information in addition to the sensor measurement value. This means, that the provision of meta-information (temperatures, ambient conditions, calibration state, time and time dependency, location, etc.) must be included in the services, and mere provision of the sensor and measured values is not sufficient. Only with knowledge of the meta-information can a complete measurement uncertainty be derived via previously created models (measurement model). The starting point for the second step is the consideration that the relevant added value when using measurement data does not lie in the measurement data themselves or their analysis, but only in the subsequent decision, e.g., the decision to intervene to achieve process optimization or the decision for or against product conformity. From a user perspective, only this decision must be made available within an approved error probability. It is not important which sensors are used, how characteristics or process data are extracted or how large the tolerances for process control or product conformity are – for the user it is only important whether the value is within the tolerance limits and with what error probability this decision can be made. The uncertainty of the decision is determined using the ratio of the uncertainty of the determination of the measured values in the application and the distance to the tolerance limit. The tolerance itself can also be provided using a service, which may perform adaptive and dynamic tolerance adjustments. “Decision as a service” thus forms the highest hierarchy level and uses subordinate services, such as the provision of tolerance or the provision of the characteristic value.
Consequently, the provision of the physical sensor itself can also be offered as a service. In the cyber-physical production systems, described in Sect. 2.1, in which measurement systems are also mobile (e.g., employees' smartphones, laser trackers mounted on drones, CT devices on mobile platforms), the demand-dependent (e.g., measurement uncertainty-dependent or frequency-dependent) measurement systems can be automatically transported to the required location using services.
A noticeable trend in metrology is the time shift between the direct recording of the characteristic on the product to a measurement of the characteristic on a model of the product. This means that at first a digital model of the product is acquired independently of the measurement task and stored digitally. Individual characteristics from this model can be retrospectively extracted, measured and evaluated. The product models consist not only of point clouds representing the 3-D geometry, but they also consist of models that include the environmental conditions, e.g., to compensate for deformation of products under the effect of force (Forbes, 2015; Maropoulos et al., 2008) or to compensate for thermal effects (Schmitt and Peterek, 2015). The digital product model therefore is more than just the digital representation of the component. By taking the ambient conditions into account, the model-based measurements can be used more precisely and more flexibly than the direct measurement of the characteristic. In semiconductor manufacturing, for example, virtual measurement processes are already used because of the large number of pieces and the impossibility of direct measurements (Khan et al., 2007). All available sensor information is used and combined in the digital shadow of the product to precisely represent the product geometry and the behavior of the product. The complete 3-D digitization of products using photogrammetry is a classic example of a purely digital shadow. An example of a model-based measurement method is the optimization of laser interferometry measurements. The additional measurement of air temperature, atmospheric humidity and atmospheric pressure in several locations can be used to calculate the refractive index in the entire working area with the help of spatial statistics, thus improving the conversion of optical lengths to geometric lengths and thus the interferometry measurements. The model is flexible in such a way that acoustic tomography can also be used for the indication of the ambient temperature referenced by location and time (Schmitt et al., 2016).
Example of a virtual measurement process: compensation of geometry errors via temperature models of the component within the production environment. This figure is translated from Schmitt and Voigtmann (2017).
Another trend to reduce the uncertainty of the measurements and to increase the flexibility is virtual measurement processes. For the automated determination of value of a characteristic and its measurement uncertainty, it is necessary to automatically determine the measurement uncertainty under application conditions. For this purpose, application-specific models of the measurement processes for combining and simulating the influences on the measurement and the associated uncertainties are necessary. This is called a virtual measurement process. Typical examples are the virtual coordinate measurement machine or the virtual laser tracker. In both cases, the measurement uncertainty of a real-world measurement is estimated by simulation using environmental data as well as a model of the influence of the environmental conditions on the measurement uncertainty (Schmitt et al., 2016; Huo et al., 2009, 2010; Trapet, 1999). Virtual measurement processes form the basis for a future a priori assessment of the measurement uncertainty and thus the basis for an optimization of the measurement planning. With virtual measurement processes, there is also a chance that air-conditioned measurement rooms are no longer necessary to measure at reference temperatures but that they can be replaced by a sensor-assisted temperature recording and models for the temperature behavior of the measurement processes and the products within the production environment. A contribution to the compensation of temperature-related geometry errors in the production can be found in Ohlenforst et al. (2016, see Fig. 6). The integration of miniaturized temperature sensors into the product and the exact measurement of their position using laser trackers as well as a complete model of the product enable the compensation of temperature-related geometry errors. If these models are extended from the temperature to all ambient influences, which are kept constant in measurement rooms, it is foreseeable that, especially in production processes of technologically or dimensionally complex products, classical production approaches with a dedicated measurement space will be replaced or supplemented by virtual measurement processes in combination with the recording of the environmental data in the production. If the structure of the virtual measurement processes and the inputted production data are developed as services, they can react flexibly to different products and production conditions. A virtual metrology room is created according to current needs, which, if necessary, can be reconfigured locally and temporally in the production.
The service administrators and cloud service providers will be affected by this development on the one hand. For them, the effort to implement and maintain the services will increase. On the other hand, the users of the services will be able to simplify their acquisition of measurement data. A strong separation is taking place between experts of metrology and users of metrology. For users, the complexity of the information generation is reduced by the provision of simplified operating and adjustment concepts in favor of their competence to act as smart experts in their respective domains, e.g., the production. For sensor manufacturers, this means, in particular, a stronger focus on intelligent sensors, which are also referred to as smart sensors or sensor 4.0 (Schütze and Helwig, 2017; Rice, 2009; Meijer, 2008; Spencer et al., 2004), and their communication interfaces as well as their compatibility to standard data formats. However, it is unclear who will be the driving force behind this development and provision of these services. Whether this development is more likely to originate from sensor manufacturers, production-related software producers or non-production-related software providers (e.g., providers of cloud services or big data analysis) remains to be seen. It is possible that the large providers of cloud services and search engines, with their already existing know-how in the field of service-oriented concepts, will expand their business models.
The conversion to a service-based application of metrology also makes it possible to re-evaluate the cost-effectiveness of metrology. One of the main problems of metrology results from accounting: metrology components are often large investments, and rooms and personnel are billed as overhead costs; an estimate of their benefit in the sense of a contribution to the efficiency of the production is often impossible. Typically, services are billed per usage or for a period of use. Eventually, when services are offered not only internally, but also between companies, the value of the services, and thus the value of metrology, must be stated per usage or for a period of use. Depending on the scope of the service, this requires the determination of how much the value of sensor information, characteristic values, measurement uncertainties or decisions is per indication or per usage and how this is quantified. If this is the case, the value of the metrology or the value of the data provided and thus the usefulness of metrology can be presented more obviously. The central questions for the evaluation of sensor information as a service are as follows. What does sensor information cost? What does an inspection cost? What is the benefit for metrology per application? Sensor-related services here also allow for a better integration of metrology into business processes.
The digitization, networking and improvement in flexibility of production means, which lead to cyber-physical production systems and which can flexibly adapt to customer requirements, are central developments in production and an opportunity for metrology. In networked production, metrology forms the basis for a holistic, valid assessment of production scenarios and is the key for the implementation of a model-based process control. As a result, the development focus of metrology will shift from device engineering to the provision and linking of data. The main challenges for networking are the homogenization of data formats and interfaces. With the concept of sensor information as a service, a high degree of standardization and modularization can be achieved, which makes it possible to use sensors and measurement systems more flexibly and in a more appropriate manner. Therefore, it can be used as a central component for the creation of future cyber-physical production systems. Approaches and initial solutions for a service-oriented architecture of sensor systems and the connection to cloud solutions are already available, but further research and development is necessary for a comprehensive integration into production.
No data sets were used in this article.
The authors declare that they have no conflict of interest.
This article is part of the special issue “Evaluating measurement data and uncertainty”. It does not belong to a conference.
This is an updated and revised version of a paper written in German language and published in “Technisches Messen 84 (2017) 2” in 2017. It is published in JSSS with the kind permission of the publisher De Gruyter. Edited by: Klaus-Dieter Sommer Reviewed by:three anonymous referees