D6.13 Assessment planning - Manutelligence€¦ · 4.1.2 ECOGRAI ..... 15 4.1.3 Business benefit...
Transcript of D6.13 Assessment planning - Manutelligence€¦ · 4.1.2 ECOGRAI ..... 15 4.1.3 Business benefit...
Copyright Manutelligence Consortium 2015-2018 Manutelligence N°636951
Horizon 2020
Acronym: Manutelligence
Project No: 636951
Call: H2020-FoF-2014
Topic: FoF-05- Innovative product-service design using manufacturing intelligence
Type of action: RIA
Duration: 01.02.2015 - 31.01.2018
D6.13 – Assessment planning
Type Deliverable
Document ID: D6.13 – Assessment planning
Workpackage: WP6
Leading partner: VTT
Author(s): Iris Karvonen, Kim Jansson, Tapani Ryynänen –
VTT + all use cases & supporting partners
Dissemination level: Public
Status: Released
Date: 24.1.2017
Version: 2.0
Public
Copyright Manutelligence Consortium 2015-2018 Page 2 / 40
Versioning and contribution history
Version Description Contributors
0.1 28.11.16 Table of Content, draft section 4. IK, KJ
0.2 15.12.16 Results of Lugano workshop added + other
contributions
IK, all use cases and supporting
partners
0.3 22.12.16 new chapters added TR, KJ, IK
1.0 12.1.17 completed and modified version for peer review IK, KJ
2.0 Final version; peer review comments handled IK, KJ
Reviewers
Name Affiliation
Laura Cattaneo, Daniele Cerri Polimi
Deliverable Peer Review Summary
ID Comments Addressed ()
Answered (A)
1 Some minor comments
addressed
The comment about deliverable number: The number
D6.13 is correct according to Research Participant
Portal.
2
3
4
5
Public
Copyright Manutelligence Consortium 2015-2018 Page 3 / 40
Table of Contents
Executive summary ........................................................................................................... 5
Abbreviations ................................................................................................................... 6
1 Introduction and scope of this deliverable ................................................................. 7
1.1 Objectives .................................................................................................................... 7
1.2 Relation to other work packages ................................................................................... 8
2 Use case Assessment scope in Manutelligence ........................................................... 9
3 Requirement validation ........................................................................................... 10
3.1 Scope of Requirement validation in Task 6.5 ............................................................... 10
3.2 Validation workshop planning .................................................................................... 10
4 Assessment of Use case Business Benefits ............................................................... 13
4.1 Experience on business benefit assessment in previous projects .................................. 13
4.1.1 Balanced Scorecard ......................................................................................................... 14
4.1.2 ECOGRAI .......................................................................................................................... 15
4.1.3 Business benefit measurement based on process improvements .................................. 17
4.2 Simplified approach for indicator definition in Manutelligence .................................... 20
4.3 Lugano workshop for BPI definition ............................................................................ 22
4.4 First Definition of Business Benefit Indicators ............................................................. 22
4.4.1 Automotive use case ....................................................................................................... 22
4.4.2 Ship use case .................................................................................................................... 23
4.4.3 Fablab use case ................................................................................................................ 23
4.4.4 Smart house use case ...................................................................................................... 24
4.4.5 Summary of indicators ..................................................................................................... 24
4.4.6 Measurement method & process .................................................................................... 25
5 Use case experience – lessons learnt ....................................................................... 26
5.1 Objectives .................................................................................................................. 26
Public
Copyright Manutelligence Consortium 2015-2018 Page 4 / 40
5.2 Experience collection .................................................................................................. 26
6 Defining success of use cases ................................................................................... 27
7 Next steps: Plan for use case assessment activities .................................................. 28
References ...................................................................................................................... 29
Appendix 1: Draft Business Performance Indicators for Use cases .................................... 30
Automotive Case (Ferrari): ...................................................................................................... 30
Ship use case (Meyer) ............................................................................................................. 32
Fablab case (FCIM) ................................................................................................................. 34
Smart House case (Lindbäcks) ................................................................................................. 37
Public
Copyright Manutelligence Consortium 2015-2018 Page 5 / 40
Executive summary
This document describes the plan for the assessment of the 4 use case trials in Manutelligence
project (automotive, ship, fablab, smart house). The deliverable is related to Task T6.5 –
Assessment. The results of the assessment will be later reported in D6.14 Manutelligence
Assessment reporting (M36).
The use case trial assessment includes the following aspects:
- Requirement validation: the solution implementations, the pilots, are checked
against the aggregated requirements formulated in WP1. This validation is different
from that performed in WP1 where the aggregated requirements were validated
against the trial scenarios and original unstructured requirements. In the trial
assessment two dimensions will be used: matching of requirements and assessment
of their importance. The validation will use a workshop approach.
- Assessing business benefits of the use case trials: This is performed using Business
Performance Indicators. As the different trials have different objectives they all need
to define their own indicators. The document represents some approaches for the
business benefit indicator definition used in previous R&D projects. A simplified
version based on trial business objectives is selected to be used in Manutelligence.
The preliminary indicators for each trial are reported, based on a definition in a
specific workshop.
- Accumulation of use case experience, lessons learned and best practice. This will be
collected through questionnaires and/or workshops in the final phase of the project.
The success of the use case trials is assessed as a combination of these three aspects. Finally,
the schedule for the assessment tasks is presented.
Public
Copyright Manutelligence Consortium 2015-2018 Page 6 / 40
Abbreviations
AV Action Variable
BPI Business Performance Indicator
DoA Description of Action
DV Decision Variable
EU FP7 European Commission Framework Program 7
EU H2020 European Commission Horizont 2020-program
LCA Life Cycle Analysis
LCC Life Cycle Cost
PLM Product Lifecycle Management
SLM Service Lifecycle Management
P-S Product-Service
VRM Value Reference Model
WP Work package
Public
Copyright Manutelligence Consortium 2015-2018 Page 7 / 40
1 Introduction and scope of this deliverable
1.1 Objectives
According to the DoA (Manutelligence Description of Actions) the objectives of Task T6.5 –
Assessment are the evaluation and assessment of the Use Case Trials, regarding the following
aspects:
- Harmonising the experimentations, share experiences.
- Assessment - To ensure that the solutions actually meet the user’s needs, the
requirements and specifications were correct in the first place and that the solution
fulfils its intended use when placed in its intended business environment.
- Usage of Business Performance Indicators for assessing business benefits.
- Harmonized accumulation of use case experience, lessons learned and best practice.
In the first phase, the assessment methodology is developed and finally it is used to assess the
success of the use cases.
The first deliverable from this task is the public report D6.13 Manutelligence Assessment
planning, due at month 24 (this report). This deliverable presents the harmonized assessment
and performance measuring system for the use cases and a plan for its implementation in the
next phase.
The second and final deliverable from this task is the confidential report D6.14
Manutelligence Assessment reporting, due at the end of the project, at month 36. The
deliverable D6.14 reports on achievements in the use cases based on the requirement
validation, business benefit indicators and accumulation of use case experience and lessons
learned.
The task 6.5 is part of WP6 Use cases and thus the assessment is focused on the use cases and
how they see the solutions. The first objective of the assessment is to ensure that the solutions
actually meet the users’ needs. This means that the use case pilot implementations are
compared with the requirements identified and processed in WP1 Requirements. The second
objective is to assess the business benefits of the use cases using indicators. The third
objective is to cumulate experience and lessons learnt. The assessment activities should be
harmonized among the use cases, that is: the assessment should be performed with common
processes. This does not, however, mean that all the use cases have the same indicators.
The first results of this task, reported here in this deliverable D6.13, are;
- Methods for validation of requirements in use case implementation (chapter 3).
- Method for the assessment of use case business benefits, using business
performance indicators and first definition of business performance indicators for
each trial (chapter 4).
Public
Copyright Manutelligence Consortium 2015-2018 Page 8 / 40
- Method for use case experience accumulation (chapter 5).
- Next steps and planning till the end of the Task 6.5 (chapter 6).
1.2 Relation to other work packages
The task is highly dependent on the previous deliverables within the same work package WP6
(D6.1, D6.4, D6.7 & D6.10 - Use Case Scenario + Pilot stories and D6.2, D6.5, D6.8 & D6.11
Use case Pilots). The previous reports form the main input for the assessment work. The final
deliverable from WP6 is the D6.14 Manutelligence Assessment, reporting the assessment
results.
The Manutelligence Work Package WP1 “Engineering and business requirements for holistic
P-S development processes in the targeted application sectors”, performed requirement
elicitation based on the use case needs. As a result of this more than 200 unstructured and
heterogeneous requirements were identified. In the next phase the requirements were
processed through structuring and analysis into 20 aggregated requirements. As the final task
of WP1 the aggregated requirements were validated by the use cases against the pilot stories
and use case scenarios. This ended up to 21 aggregated requirements.
The requirements have been used in the technical work packages developing the platform
(WP4), simulation of PLM and SLM interaction (WP2), cross-section collaborative design
(WP3) and decision support for LCA & LCC issues (WP5). As described in D1.1, the
approach of requirement engineering in Manutelligence follows an iterative approach, and
thus the requirements are subject to modification during the development phase. The use case
scenario definition and the requirement elicitation were performed with a wide focus: it was
not the idea to implement all the defined use scenarios but make the selection later in the
project, and also to give space for new ideas coming during the project. In some use cases
new scenarios have emerged during the development.
Public
Copyright Manutelligence Consortium 2015-2018 Page 9 / 40
2 Use case Assessment scope in Manutelligence
As described in DoA the objective of Task 6.5 has three main tasks:
Requirement validation: To ensure that the solutions actually meet the user’s needs,
the requirements and specifications were correct in the first place and that the solution
fulfils its intended use when placed in its intended business environment.
Usage of Business Performance Indicators for assessing business benefits.
Harmonized accumulation of use case experience, lessons learned and best practice.
Requirement validation here is the process of checking that the use case solutions and
Manutelligence platform implementation fulfill the needs identified in WP1 Requirements.
WP1 also included a validation task but in WP1 the aggregated requirements were compared
against the pilot stories and use case scenarios. Even if the task is thus different here, the
main actors performing this validation are again the use case owners. The validation method
is described in Chapter 3.
Manutelligence platform and the pilot solutions are expected to offer the use case companies
business benefits. To assess these, Business Performance Indicators (BPI) will be defined.
The benefits may be linked to increased efficiency in Product-Service (PS) design, bringing
along savings in costs and time. They may mean increased quality and customer satisfaction,
new services and new business income. However, the indicators should be defined at a level
which is closely related to the pilot scenario and not too high at the company level. Probably
the indicator values cannot be derived directly from any system in the piloting phase but they
are more based on the assessment of use case experts. As the different use cases have
different objectives and aim for different business benefits, it is not possible to define BPIs
which are relevant for all the use cases. The method and process for BPI definition as well as
the first indicators are described in chapter 4.
In BPI measurement, the aim is to generate quantitative or semi-quantitative values for the
indicators while in the collection of use case experience the aim is to collect qualitative
information: the experience of the use cases, lessons learned and best practices, if they can be
identified. This information may be partly subjective, opinions and feelings. The approach for
experience collection is described in chapter 5.
Finally the DoA sets up the objective to assess the success of the use cases. The success is
demonstrated as the combination of the three areas: if the requirements are fulfilled, business
benefits good enough and there are positive experiences about the solutions.
Public
Copyright Manutelligence Consortium 2015-2018 Page 10 / 40
3 Requirement validation
3.1 Scope of Requirement validation in Task 6.5
The validation as a concept has been reviewed in Manutelligence D1.3 Requirements
validation. The objective of validation in general is “to ensure that the product actually meets
the user’s needs, the specifications are correct in the first place and the product fulfils its
intended use when placed in the intended environment” [IEEE 2012]. Validation can be
directed towards the final product / software or its preceding specifications. In Manutelligence
task T1.4 the validation was focused on the structured and aggregated requirements against
the needs of the involved use cases. In this task 6.5 the aim is to validate the implemented
solution/ platform using the use case pilots, against the aggregated requirements. These two
tasks constitute the validation “chain”:
T1.4 (D1.3): Validation of aggregated requirements against the original pilot stories and
scenarios.
T6.5 (D6.13-14): Validation of pilot solutions and the platform against the 21 aggregated
requirements.
Thus, with some restrictions it can be assumed that the validation chain leads to the validation
of the final solutions against the original pilot stories. The restrictions include that on one
hand not all scenarios described are implemented and on the other hand new scenarios have
been emerged during the development, both related to the changes in company needs and also
new identified potential of the platform.
In WP1 validation, basically two parallel methods were used: a research partner walk-through
and a workshop of all Manutelligence use case representatives. In the workshop, the end user
groups checked the sufficiency of the requirements using a common methodology, supported
by nominated research partners. Also some new needs of end users were identified.
In Task 6.5 Requirement validation, the solutions or the pilots, are checked against the
requirements. Here the use case partners are in the main role, but to follow a common
process, again support from technology and research partners is needed. The plan is to
perform the validation as a workshop. The workshop method is described in the next chapter.
3.2 Validation workshop planning
In a software development process both the provider and the user have their own opinions
about how the needs should be met with the implemented functionalities of the package.
Sometimes these opinions can be very close to each other, but often, especially in the
beginning of the process, they can differ significantly.
Public
Copyright Manutelligence Consortium 2015-2018 Page 11 / 40
This is often due to real differences but also different terminology and ways to pose a
question or a need makes communication challenging. Thus iterative discussion helps the
negotiating parties to better understand the real meaning behind the words and to understand
the quintessential differences. This iterative discussion has been initiated already in the earlier
requirements validation phases of the Manutelligence project and will continue in Task 6.5.
The validation takes place in a five-phase workshop as described below in Table 1.
Table 1. Steps for validation workshop (draft)
Phase Goal Graph
1
Recall of the
requirements and the
case
For all participants to jointly recall the
final 21 aggregated requirements. Short
“questions and discussion” session.
2
Separate evaluation
of the requirements
by Provider and User
For case groups to recall the case being
evaluated and to create a fresh perception
of requirement-platform functionality
matching. The scales used are Matching
(horizontal axis) and Importance
(vertical axis).
3
Iterative discussion
to compare the
evaluations
To understand other party’s opinions and
validation arguments. To adjust own
validation if seen necessary.
4
Shared
understanding
To combine the separate and adjusted
validations into a shared one. To
document reasoning behind differences
in validation. To agree what should be
done to still better match the opinions.
5
Cross case discussion
and summary about
the results
Final step is to present the results of each
case group and to summarize. It is
important to see that each case is
different and can have quite different
needs and opinions. Therefore, further
analysis of the results is done by
researchers after this group work.
Public
Copyright Manutelligence Consortium 2015-2018 Page 12 / 40
The validation is carried out for each use case by a group of people representing the case
owner (e.g. Meyer), platform providers (e.g. Dassault), and support partner. The assessment
is made from the use case viewpoint. Careful documentation is important especially for
tracking and understanding the changes made in validation as well as the reasoning behind.
Two assessment criteria are used. These are Matching (“Is the requirement fulfilled by the
functionalities of the platform?”) and Importance (“How important it is that this requirement
is fulfilled by the platform?”). In the following cross-case discussion, more emphasis is given
to those requirements having high importance.
Public
Copyright Manutelligence Consortium 2015-2018 Page 13 / 40
4 Assessment of Use case Business Benefits
4.1 Experience on business benefit assessment in
previous projects
Performance Indicators (PIs) are used to measure the progress or success of e.g. an
organization, a company, a part of a company or an activity or project. Often PIs are used and
measured several times, at certain interval to detect changes and trends in the development.
Key Performance Indicators (KPIs) are often used to measure the main or top indicators of a
company. KPI can be e.g. ROI (Return on Investment), overall sales, turnover, profit margins
etc. KPIs can be used to evaluate the Key Success Factors.
In a project like Manutelligence the development activities will have a hopefully positive
impact on the performance of the use case company. However the performance of the
company as a whole is influenced also by other internal decisions and activities as well as
factors coming from the outside world, partly out of the control of the company.
Consequently, it is not possible to assess the impact of the Manutelligence alone using high
level KPIs. The impact of Manutelligence development activities cannot simply be isolated
from other events and activities. Instead in Manutelligence there is a need for PIs that are
“closer” to the development activities and the development activities influence the selected PI,
in a more direct way. When PIs are used to assess the business performance they are called
business performance indicators (BPIs). Business performance indicators can be either
quantitative or qualitative. Quantitative indicators can be deployed using numbers while
qualitative indicators are often assessed using textual opinions. In Manutelligence, depending
on availability of data, qualitative, semi-quantitative or quantitative indicators will be used to
assess the business benefits of Manutelligence development activities.
There are several well-known approaches for assessing and measuring business benefits. It
has also been a task in several European projects. In FP7 project FITMAN [FITMAN] a
comprehensive methodology was developed, for verification and validation of the solutions,
in various aspects, coming from the adoption of new software in a number of use case trials.
One of the aspects was business performance using BPIs. In order to find the most suitable
approach, the project analysed in all the 37 most current business performance indicator
methods. It is out of the scope of this deliverable to give a comprehensive state-of-the-art
report on performance indicator methods; instead some of the results of FITMAN analysis are
used here [FITMAN D2.2 2012].
The following sections include short descriptions of three methods that are well known and
have been successfully used in previous projects. They are: Balanced Score Card, ECOGRAI
method, and Value Reference Model (VRM).
These methods were chosen because:
Public
Copyright Manutelligence Consortium 2015-2018 Page 14 / 40
- Balanced Score Card (BSC) is the most known and used by many organizations.
- ECOGRAI method has a specific approach and is based on the model of the
organization. ECOGRAI has some links with BSC.
- Experience of using: VRM used in two projects COIN and MSEE.
4.1.1 Balanced Score Card
The concept of the Balanced Score Card (BSC) was introduced by Kaplan and Norton in 1992
[Kaplan and Norton 1992]. It focuses on the strategy and the vision rather than on the control.
It provides the means to translate the vision into a list of objectives. For that, the strategy is
translated into a system of performance measures according to four perspectives (Figure 1):
- Financial (relating to the shareholders).
- Customers (to generate the turnover for the fulfilment of the financial objectives).
- Internal Processes (to satisfy the customers and to guarantee the financial output).
- Innovation and Learning (to develop the enterprise capacities to improve for its
viability).
If we succeed, how will we
look to our shareholders?
Financial Perspective
Objectives Measures Targets Initiatives
Customer Perspective
Objectives Measures Targets Initiatives
Internal Perspective
Objectives Measures Targets Initiatives
Learning/Growth Perspective
Objectives Measures Targets Initiatives
To achieve financial objectives, how
must I look to my customers?
To satisfy my customers, at
which processes must I excel?
To excel in processes, how must
my organization learn and improve?
Vision and Strategy
Ca
use
and E
ffe
ct
Vis
ion
Figure 1. The Balanced Score Card [Kaplan and Norton 1992]
The process to reach the performance measurement consists of:
- The development of performance indicators into line with the strategy.
- The communication of the strategy by the deployment of the performance
indicators.
- The strategic objectives measurement.
- The focusing on the key success factors.
- The consideration of the causal relationship between the four perspectives and the
performance indicators (the real keys of the BSC).
The BSC contains a mix of lag and lead indicators. Performance measures that represent the
consequences of actions previously taken are referred to as lag indicators or as key results
Public
Copyright Manutelligence Consortium 2015-2018 Page 15 / 40
indicators. They may focus on results at the end of a time period and characterize historical
performance. Leading Indicators are considered the "drivers" of lagging indicators. There is
an assumed relationship between the two which suggests that improved performance in a
leading indicator will drive better performance in the lagging indicator. For example:
Leading PI: Number of adjusted or changed parts in the products
Lagging PI: % of drop of customer’s complaints or decrease of the Numbers of products
returns.
As mentioned above, Balanced Scorecard has a wide strategic view, maybe too wide to assess
the business benefits of specific scenarios in the platform development. However, the
principles, like measuring the objectives, are suitable also for platform business benefit
assessment: what you assess, is dependent on what you are striving for.
4.1.2 ECOGRAI
ECOGRAI [Doumeingts G., Clave F., Ducq Y. 1995, Bitton M., Doumeingts G. 1990] is a
method to design and to implement Performances Indicator Systems in any kind of
application domains and it is applied with the implication of the decision makers of the
production management systems. ECOGRAI seeks to find a number of customized and
limited indicators in agreement with the objectives of decision makers.
Main characteristics:
- A logical process of analysis / design using a top-down approach, and allowing to
decompose the objectives of the strategic levels into objectives for tactical and
operational levels.
- A concrete participative process for the design and the implementation, creating a
dialogue between the various levels of the hierarchy, and favouring the
identification of indicators by the future users involved in the study: it is a bottom-
up implementation.
- The use of a number of tools and graphic supports: GRAI grids [G. Doumeingts, F.
Marcotte, H. Rojas. 1995], GRAI nets, splitting up diagrams, coherence panels,
specification sheets.
- A coherent distribution of Performances Indicators covering the various functions
and the various decision levels (strategic / tactical / operational).
The logical structured approach of the method consists of six phases (Figure 2) and described
shortly below.
Public
Copyright Manutelligence Consortium 2015-2018 Page 16 / 40
Figure 2: The six phases of the structured approach of ECOGRAI
The first phase: The modelling of the control system and the physical system (controlled
part) of the Enterprise (or a part of the enterprise). The determination of the decision centres
in which the PIs will be defined. It is also necessary to model the physical system by the
processes to define performance indicators at operational level.
The second phase: The identification of the Decision Centre objectives and coherence
analysis by a top-down approach:
- Identifying the Enterprise System objectives.
- Identifying the global objectives of each function belonging to the enterprise.
- Defining the objectives of each Decision Centre inside the functions.
These identifications are based on the notion of contribution: each objective must contribute
to the achievement of the objectives identified at an upper level.
The third phase: The identification of the DC Drivers (Decision Variable (DV) or Action
Variable (AV)) and analysis of the conflicts. This identification must be interpreted as one of
the steps leading to the building of the triplets {Objectives / Drivers / PIs}. This notion of
triplet expresses the controllability principle.
The fourth phase: Identification of the DC Performances Indicators and internal coherence
analysis: it consists in the identification of the PIs which is validated only by an internal
Public
Copyright Manutelligence Consortium 2015-2018 Page 17 / 40
coherence analysis inside each Decision Centre in terms of triplet {Objectives / Drivers / PIs
}. A triplet is coherent if:
- It is composed of one objective, one or several drivers and one or several
performances indicators.
- The performance indicators allow to measure the efficiency of an activity or a set of
activities of the considered function in the process to reach the objective, and they
are influenced by actions on the drivers.
The fifth phase: Design of Performances Indicators information system: it consists in the
specification of each PI. It means to define clearly each indicator with fundamental
parameters. The tool which guides these definitions is the specification sheet for each
indicator which contains:
- The identification of the indicator (name, decision centre, horizon, period).
- The objectives and the drivers related to the indicator.
- The perverse effects which have been identified,
- The identification of the data required for the implementation of the indicator,
- The definition of the associated processing,
- The way of representing the indicators, determined by the future users (using
graphics for most of the time).
The sixth phase: Integration of the PI information system inside the production information
system.
The whole ECOGRAI process with the above steps seems quite complex for assessing the
business benefits of platform developments. In the generic method, the approach is to support
decision making between different actions. However, in Manutelligence the actions are
predefined and the aim is not to select between platform adoption or something else. Thus the
six-step approach should be simplified. This was also done in FP7-FITMAN-project in which
“Simplified ECOGRAI”-method was used to assess business benefits of use case solutions [
FITMAN, Project ID: 604674]
4.1.3 Business benefit measurement based on process
improvements
In EU FP7 project COIN [COIN] the assessment of business benefits of the use cases was
performed in two phases:
In the first phase, the business benefits of the solutions were assessed in the user requirement
analysis phase using three concepts [COIN D6.1.1b 2009]:
- Changes in process parameters, which can be regarded as some kind of variables
named “Process Xs”.
Public
Copyright Manutelligence Consortium 2015-2018 Page 18 / 40
- Improvements in process results/capabilities, which represent the process
performance in the TO-BE use cases perceived by the end users and which are
based on the previous changes in the process parameters. These improvements can
be regarded as “Process Ys”.
- Expected benefits are representing the economic and environmental effects that
should be of positive nature for the end users and are accompanied by (mostly)
measurable performance indicators, which can be regarded as “Business Ys”.
These three concepts underlie the following logic: “Y is a function of X”. Thus the focus was
on process improvements.
In the use case testing phase, a common approach for defining the business benefit indicators
for the process improvements was taken. The approach Value Reference Model (VRM) was
selected; including a process map, metrics dimensions and key performance indicators (KPIs).
The idea was to compare “as-is” and “to-be” metrics for the processes which are supported by
the developed services. The preliminary metrics were checked to see if they were still valid
and how they could be consolidated with the VRM approach.
Value Reference Model is offered by Value Chain Group [VRM 2016]. The model offers a
framework to identify processes, measure and improve performance. The main processes are
Plan, Govern and Execute. These main processes and sub-processes are identified below in
Figure 3. For each sub-process a group of potential metrics, with different dimensions, has
been identified: Cost, Velocity, Asset, Reliability, Adaptability and Innovation.
Public
Copyright Manutelligence Consortium 2015-2018 Page 19 / 40
Figure 3. VRM process map [VRM 2016].
The VRM approach: going from selection of processes and dimensions to metrics was used
also in EU FP7 project MSEE [MSEE] as described in [Wiesner et al. 2014]. The end users
considered the approach understandable and useful. However, it was not as suitable for all
the cases. The best fit was with partners representing supply chains. Additionally, some
benefits may come more from processes affected by the supported processes (for example: the
main benefits of solutions supporting project management are not coming from lower cost of
project management but from projects carried out with improvements in costs, time, quality
and with lower risks). This, maybe too mechanical and still laborious approach was the
reason why VRM was not selected to be used in Manutelligence BPI definition. In
Manutelligence, it could be challenging to present all the benefits as process improvements.
Public
Copyright Manutelligence Consortium 2015-2018 Page 20 / 40
4.2 Simplified approach for indicator definition in
Manutelligence
In Manutelligence a method is needed to assess the business benefits of the use case scenario
developments when using the platform. It is clear that a harmonized indicator system with the
same indicators for all the use cases is not possible as the use cases have different scopes and
varying objectives. Thus it is essential to link the assessment to the objectives of each use
case. This is one principle behind the ECOGRAI approach presented above. In
Manutelligence consortium there is also experience about the application of ECOGRAI in a
simplified form in previous projects. Thus in Manutelligence it was decided to use a
simplified method based on ECOGRAI and its application in EU FP7 project FITMAN [
FITMAN, Project ID: 604674].
The ECOGRAI method involves a step covering the identification of the improvement actions
(Decision Variable (DV) or Action Variable (AV)). In Manutelligence the improvement
actions are already known, that is: the implementation of the Manutelligence Platform in the
use cases. Thus for the Manutelligence application a simplified version of ECOGRAI can be
used, with only three phases. These steps lead to the building of the triplets Objectives /
Drivers / BPIs, Figure 4.
Set objectives
Define actions (= implementation of Manutelligence Platform)
Measure progress
o Select performance indicators
o Definition of success
o Measure progress using performance indicators
Figure 4. Assessment method to be used in Manutelligence.
Setting Objectives
Each of the use case needs to define the objectives that the use case implementation wants to
reach. Good starting points are the Pilot Stories and the Use Case Descriptions (Deliverables
Objectives
Action Measure
Public
Copyright Manutelligence Consortium 2015-2018 Page 21 / 40
D6.1, D6.4, D6.7 & D6.10) that have been produced earlier in Manutelligence. These
documents include the first definition of objectives in Manutelligence, benefits and
preliminary performance indicators. The use case descriptions were further enhanced and
detailed in the Deliverables D6.2, D6.5, D6.8 & D6.11.
Definition of a Business Performance Indicator
As described above BPIs (Business Performance Indicators) measure the business benefits of
actions. It is not realistic to assess how a specific use case scenario may impact the company
high level objectives (like profitability); thus a proper level for the measurement needs to be
identified. A useful BPI has the following characteristics:
- Easy to be interpreted, to put in work, to use or to exploit
- Easily measurable, quantifiable
- Representative of the objective of which it measures
For the purpose to collect information from use case partners, the following template was
designed, Figure 5. The idea is to fill the template for each use case scenario which has been
implemented. For each scenario, several indicators may be defined. Each indicator is given a
unique name which is recommended to be somehow descriptive. Additionally a description of
the indicator is given. Indicator unit means the unit in which it is measured (for example
weeks/ hours/ euros/ %...). The next columns are about the information needed for the
measurement and where this information/ data can be retrieved from. BPI owner means the
person/ role in the company who is responsible for the BPI measurement.
The template was used in the Manutelligence consortium meeting in Lugano, November
2016, in order to collect the first definitions of the use case BPIs. The results are described in
the next chapter.
Figure 5. BPI definition template.
Public
Copyright Manutelligence Consortium 2015-2018 Page 22 / 40
4.3 Lugano workshop for BPI definition
The workshop for Use case Business Performance Indicator definition was organized in a
consortium meeting 30.11.2016 at Supsi in Lugano. The idea was to make the first definition
in this phase and check it later in the project. In the workshop, one group for each use case
was organized: the research and technology partners supported the end users in the task in the
following way:
Meyer + VTT
FabLab + SUPSI & DAPP
Lindbäck + Balance & BIBA
Ferrari + Das & Holonix & Polimi
The template presented above in figure 5 was applied. As all the use cases have more than one
scenario and the scenarios may have different objectives, the BPIs had to be defined
separately for each scenario. To focus on the most important objectives, the goal was to
define max 3 BPIs for the most important scenarios. VTT had prefilled the tables with
objectives and benefits based on the documentation of the Use case scenarios in previous
WP6 deliverables.
Thus the task given for the groups was the following:
- Check the prefilled use cases and scenarios. Are they OK? Is any important use case
missing?
- Check and update the objectives of each scenario.
- Check and update the business benefits of each scenario.
- Define 1-3 BPIs for each scenario.
The results of the workshop are summarized below and the filled templates can be found in
Appendix 1. As it is possible that during the final year the use cases may identify new
important scenarios, the BPIs will be checked again after M30.
4.4 First Definition of Business Benefit Indicators
4.4.1 Automotive use case
Two use case scenarios were selected: “Communication of the virtual and physical test results
to the designers” and scenario: “Requirements verification and validation and generation of
traceability report documents”. The main business benefit identified is “Improved design”
(including for example quality and driving comfort) and speed in the design process and data
acquisition. Thus most of the BPIs defined were related to time:
- Time to get all the needed info (in the search activity)
Public
Copyright Manutelligence Consortium 2015-2018 Page 23 / 40
- Time to insert the physical test data into the platform
- Time to find the structured data in the platform
- Time to execute further test using virtual instead of physical test
Other BPIs defined were the following:
- Number of successful search info activities
- Accuracy of the comparison between simulated and physical tests.
4.4.2 Ship use case
Three use cases and scenarios were selected. Two of them were related to enhancing feedback
to design, one from the customer and another from the production. The third scenario
“Change management” is linked to these two: the aim is to manage the change process based
on the received feedback.
By enhancing the customer feedback during the design the aim is on one hand to design more
attractive ships and on the other hand to prevent delays in the design and manufacturing
phase. The feedback from the production is aimed for identifying collisions or other problems
in the design early enough, from the manufacturing point of view. The change management is
linked to the feedback processes, thus no separate BPIs for it were defined in this phase.
Most of the BPIs defined were related to time delays:
- Cost complexity
- Architectural schedule
- Change order
Additionally BPIs related to design quality were defined:
- Production scores (4-10) to ship model quality
- Number of collisions (errors) in the model
4.4.3 Fablab use case
In Fablab use case the business benefits of three use cases and scenarios were analysed:
“Knowledge sharing within the network” with scenario “Shared repository of projects”, two
scenarios in “Developing the production cycle” and “Sustainability assessment”. The
objectives of the scenarios were quite different from each other, but also here one important
indicator type was time:
- Design time from idea to production start
- Average waiting time of the user for an available machine
- Machine downtime
One BPI is related to Energy consumption and two BPIs to number/ percentage of activities:
Public
Copyright Manutelligence Consortium 2015-2018 Page 24 / 40
- Number of shared projects
- Percentage of projects launching the sustainability assessment
4.4.4 Smart house use case
The smart house use case is focused on two use cases and three scenarios: “Design and
production support” – “Condition monitoring for product feedback and failure management
improvement” and “Service for smart home end users” with two scenarios: “Consumption
data monitoring to optimise energy consumption according to the individual lifestyle” and
“Offer of new services for living as an additional business area”. The BPI types are multiple,
related to time, costs, quality and number of events:
- Number of failure service events
- Costs of damages
- Reaction time for critical events
- Energy delta: difference between design and actual values
- Number of complaints from tenants
- Comfort level, including different physical measurements
- Income per costs of new services
4.4.5 Summary of indicators
The Lugano workshop for BPI definition ended up with 23 BPIs, each use case having its
own indicators, because of different business objectives. The use cases will check the
indicators again after M30, in the final phase of the project as it is assumed that at that time all
the scenarios implemented are known and there are no more any big changes.
The indicators defined were related to different types according to Table 2 where the numbers
of different types of indicators for each use case are given. The most important type is time.
This is understandable since the availability of the platform and ease of information
management in different phases contributes to shortening the time needed for different
phases. It is clear that efficiency and preventing delays also contributes to cost reduction, even
if only two of the defined indicators are directly related to costs.
Another important BPI type is quality. Quality is very important to reach customer
satisfaction and competitiveness in innovative areas and it is also related to costs. The
platform can help to increase the quality through better design and feedback mechanisms.
Public
Copyright Manutelligence Consortium 2015-2018 Page 25 / 40
Table 2. Number of Use case indicators for each indicator type (as defined in November 2016).
Use case Time Costs Quality Number (of
successful
activity)
Physical
measurement
(energy,
comfort)
Sum
Automotive 4 1 1 6
Ship 3 2 5
Fablab 3 2 5
Smart house 1 2 2 2 7
Sum 11 2 5 3 2 23
4.4.6 Measurement method & process
The template for the BPI definition used in Lugano workshop also included columns to define
the information needed, source of data and the owner of the indicator. Partly these were
defined by the use cases, but they will still need check. Some data is expected to be found at
the Manutelligence platform, for some company’s own existing tools were identified.
Sometimes the final user may be the source of information.
As a whole, the availability of the data is always challenging when defining indicators, and as
the use case solutions are at the level of demos and prototypes, the measurements cannot
usually be taken from any systems during the project. Thus assessment based on expected
improvements in the future are probably needed.
Public
Copyright Manutelligence Consortium 2015-2018 Page 26 / 40
5 Use case experience – lessons learnt
5.1 Objectives
In R&D projects, like Manutellligence, from the use case viewpoint, it is important not just to
develop the pilots, but to be able to learn from the pilot implementation and experimentation.
The experience can be used by different stakeholders:
- The technology providers can use it to develop the platform and the tools further
and to bring the new functionalities to market.
- The use cases can learn about the benefits, suitability and challenges of the platform
and what is needed to implement it in full scale in the company.
- Companies outside the project can learn about potential for new practices and
solutions.
- Research partners can use the experience to identify new topics that require
research.
To make full benefit of the experience, it is not enough to report it as a Manutelligence
deliverable D6.14. Also, it should be useful to be disseminated through public use case
success stories, conference presentations and research papers (WP8).
5.2 Experience collection
From the viewpoint of use case assessment different types of experience can be seen. The
experience may be related to the use case solutions provided by Manutelligence platform.
Different dimensions can be used, like suitability for the company, technological
performance, usability, ease of implementation, level of innovation, improvement proposals,
identification of needs in upscaling the solution etc. Another viewpoint is the Manutelligence
process for use case development, from requirements and scenarios to pilots and the
experience gained in the process.
Two methods for experience collection will be used:
- Questionnaires where opinions of the partners are collected for pre-defined
questions in a semi-structured way. The information may be collected either through
interviews or as written. Part of the questionnaires may be anonymous.
- Experience workshop, which may also be an event to discuss the findings from
questionnaires.
Public
Copyright Manutelligence Consortium 2015-2018 Page 27 / 40
6 Defining success of use cases
The task T6.5 description in Manutelligence DoA mentions: “the assessment methodology is
used to assess the success of the use cases”. It is clear that there is no single definition of
“success” but it must be defined for each context. How could the success be interpreted for
Manutelligence use cases:
- Achieving the desired outcomes and results: objectives, benefits, functionalities,
solving problems?
- Finishing the project on time and budget?
- Learning new approaches and solutions, getting experience of the potential and
feasibility of P-S platform?
- Creating and strengthening relationships with other organizations which may
support the use case in further developments?
For a H2020 project, schedule and budget are predefined and fixed and most often the
projects are not prolonged. The monitoring of project performance is part of project
management activities and thus it is left out here. The other criteria mentioned above are
already included in the three scope areas of Manutelligence use case assessment:
- Functionalities/ problem solving: validation of solutions against requirements
(chapter 3)
- Achieving objectives and benefits: Business Benefit Indicators (chapter 4)
- Learning and relationships: use case experience (chapter 5)
Thus assessing success in Manutelligence use cases can be seen as a combination of these
three aspects.
Public
Copyright Manutelligence Consortium 2015-2018 Page 28 / 40
7 Next steps: Plan for use case assessment
activities
Draft plan for the assessment is presented in Table 3. The three main scopes of assessment
are organized to be performed step by step. The current understanding is that the requirement
validation could be the first task and the assessment of the BPI values and experience
collection could take place in parallel. Finally the results are consolidated into D6.14
Manutelligence Assessment reporting, due M36.
The roles of the partners in the assessment continue the same as before in WP6. Thus also in
the assessment phase the use case partners are supported by the technical and research
partners in the following way:
Ferrari + Das & Holonix & Polimi
Meyer + VTT
FabLab + SUPSI & DAPP
Lindbäck + Balance & BIBA
Table 3. Assessment plan
M23
-24
M25
-26
M27
-28
M29
-30
M31
-32
M33
-34
M35
-36
T6.5 Assessment
Assessment planning; D6.13 released
First BPI definition & checking
BPI measurement
Requirements validation (workshop)
Experience collection
Consolidation to D6.14
Public
Copyright Manutelligence Consortium 2015-2018 Page 29 / 40
References
Bitton M., Doumeingts G., (1990), "Conception de systèmes de mesures de performances : la méthode
ECOGRAI", ECOSIP, Gestion industrielle et mesure économique, approches et applications nouvelles,
Economica , pp. 251-274.
COIN D6.1.1b, 2009, COIN Deliverable D6.1.1b, 2nd User Requirements Analysis for EI and EC. Available at:
http://analytics.ijs.si/~mitja/Courses/Coin_deliverables/COIN_D6.1.1b_Final_UR_analysis_for_EI_and_EC.pdf
COIN, Project ID: 216256 Funded under FP7-ICT, COllaboration and INteroperability for networked
enterprises. http://cordis.europa.eu/project/rcn/85550_en.html
Doumeingts G., Clave F. and Ducq Y., (1995), ECOGRAI - A method to design and to implement Performance
Measurement Systems for industrial organizations - Concepts and application to the Maintenance function, in
Benchmarking — Theory and Practice. A. Rolstadås (ed.), Springer Science+Business Media New York 1995
Doumeingts, G., Marcotte F., Rojas, H. GRAI Approach : A Methodology for Re-Engineering the
Manufacturing Enterprise, in Re-engineering the Enterprise, Proceedings of the IFIP TC5/WG5.7 Working
Conference on Re-engineering the Enterprise, Galway, Ireland, 1995, ISBN: 978-1-4757-6387-4
FITMAN, Project ID: 604674, Funded under: FP7-ICT, Future Internet Technologies for MANufacturing.
http://cordis.europa.eu/project/rcn/109803_en.html
FITMAN D2.2, FITMAN Deliverable D2.2, Verification & Validation, Business and Technical Indicators.
Available at: http://www.fitman-fi.eu/resources/deliverables.
IEEE. “1012-2012 - IEEE Standard for System and Software Verification and Validation.” 2012.
Kaplan, R. S., Norton, D. P. The Balanced Scorecard - Measures that drive performance," Harvard Business
Report, January-February 1992.
MSEE Project ID: 284860 Funded under: FP7-ICT , Manufacturing SErvice Ecosystem.
http://cordis.europa.eu/project/rcn/100080_en.html
VRM Model http://www.value-chain.org/vrm, accessed 23.11.2016
Wiesner S., Guglielmina, C., Gusmeroli, S., Doumeingts, G. (Eds.). 2014 Manufacturing Service Ecosystem.
Achievements of the European 7th Framework Programme FoF-ICT Project MSEE: ISBN 978-3-95886-003-2
Copyright Manutelligence Consortium 2015-2018 Manutelligence N°636951
Appendix 1: Draft Business Performance Indicators for Use cases
Automotive Case (Ferrari):
Definition of Business Performance Indicators
Use Case: Automotive
C: Communication of the virtual and physical test results
to the designers
Scenario: C1: Requirements verification and validation and generation of
traceability report documentations:
Objective of
Use Case:
Map all the prototype info, from test data to BOM to
methodology to execute and requirements in order to
trace and keep relationships between them.
Generate traceability documents in order to close the
design loop and provide feedback to the designers.
Business Benefit
of Use Case:
Improved design of existing or new car, based on data coming from
the real usage of the car.
Captured data simulating a car customer testing on a circuit, to
improve the design to obtain an enhanced "drive line" accuracy /
driving comfort.
Speed up the approval process (“delibera”) for choosing the best
design alternatives based on the results of virtual and physical test. .
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
SR Successful search info Number Search acti
vity Ferrari
TiT Time to get all the needed info about tests Time Search acti
vity Ferrari
Public
Copyright Manutelligence Consortium 2015-2018 Page 31 / 40
Definition of Business Performance Indicators
Use Case: Automotive
C: Communication of the virtual and physical test results
to the designers
Scenario: C1: Requirements verification and validation and generation of
traceability report documentations:
Objective of
Use Case: Physical test data acquisition IoT driven, elaboration and
sharing saving automatically in enterprise platform.
Give feedback to the design process through comparison
between virtual model and experimental tests, leveraging
on the possibility to virtually simulate the mechanism
behaviour in different conditions.
Implementation of a decision support tool for choosing
the best design alternatives based on the results of virtual
and physical test.
Business Benefit
of Use Case:
Improved design of existing or new car, based on data coming from
the real usage of the car.
Captured data simulating a car customer testing on a circuit, to
improve the design to obtain an enhanced "drive line" accuracy /
driving comfort.
.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
TtoINS Time to insert the physical test data into the
enterprise platform Time Test execution Ferrari
TtoFND Time to find the structured data into the
platform Time
Search data
into the
Platform
Ferrari
TtoSIM Time to execute further test using virtual
instead of physical object Time
Simulation
analyst Ferrari
ACC Accuracy of the comparison between
simulated and physical test
Standard
deviation Platform
Public
Copyright Manutelligence Consortium 2015-2018 Page 32 / 40
Ship use case (Meyer) Definition of Business Performance Indicators
Use Case: Ship
A: Enhancing customer feedback to design
Scenario: A2: Immersive 3D walking experience to sales.
Objective of
Use Case:
Increased customer feedback through interaction based
on Virtual Reality and Ship model.
(Service to customer in the design phase.)
To understand and “feel” the travelling experience of
the final customer (traveller) and functionality of the
design, especially when designing new innovative
solutions.
Business Benefit
of Use Case:
The travelling experience of the final customer (the traveller) can be tested
already in the ship design phase.
Better travelling experience,
More attractive ships,
Increased shipyard competitiveness.
Indicator
Name Description
Indicator
unit
Information
needed Source of data The owner Comment
Cost
complexity:
The ship design is frozen, the level of pilot
ship design reuse is known, construction
schedule keeping estimated.
Delays due to changes to the frozen pilot
(prototype?) ship design.
SAFRAN
weeks
delayed
General
arrangement
frozen
decision
Ship’s general
arrangement
Tapani Hietasaari,
head of interior
design department
Customer’s decision making
efficiency and schedule has major
impact
Architect
schedule:
Delays due to architectural changes.
Delays are costly and propagate in
construction phase
SAFRAN
weeks
delayed
schedule
changes
Architect’s schedule
& plan
Tapani Hietasaari,
head of interior
design
department
Customer’s decision making
efficiency and schedule has major
impact
Change order Delays due to change orders?
SAFRAN
weeks
delayed
schedule
changes
Information is e.g. in
work orders (several
medias like emails),
format pdf or on paper.
One of Manutelligence
aims is to fix this.
ship project
manager
Customer service enabling change
orders messes up construction
schedule. This is biggest problem
after architectural schedule
changes. Sales approves changes
too easily.
Public
Copyright Manutelligence Consortium 2015-2018 Page 33 / 40
Definition of Business Performance Indicators
Use Case: Ship
B: Intelligent use of ship model for manufacturing
Scenario: B1: Easy-to-use 3D walking experience to floor level
(Linking production feedback to ship model)
Objective of
Use Case:
Ensuring that the feedback from production reaches the
design fast enough and that the feedback is handled
properly.
Business Benefit
of Use Case:
In the manufacturing phase, better management of a change request
from shipyard about a component of current and next ships.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Model Quality Production gives feedback/scores to design
about the quality of ship model. 4-10 Excel data Excel
Ship area
production
planning
coordinators.
This feedback process has been
started as part of Manutelligence
project.
Collision
Number of collision in model (e.g. between
HVAC systems and cable racks at design
area borders)
Collision detection (like between HVAC
models). An existing quality tool finds the
number of errors and wrong material.
quantity Final report,
Excel data
Design
quality tool
Atte Peltola,
design quidance
Public
Copyright Manutelligence Consortium 2015-2018 Page 34 / 40
Definition of Business Performance Indicators
Use Case: Ship
A-B-D: Change management
Scenario: Change management and documents management
Objective of
Use Case:
The use case supports the process of design change when
a need for it has been identified.
Based on a need a change issue is created. In different
life cycle phases the change process may be different. In
the operation phase the change issue may be related to
future “sister” ships.
Business Benefit
of Use Case:
Improved communication
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
See two previous tables
Public
Copyright Manutelligence Consortium 2015-2018 Page 35 / 40
Fablab case (FCIM)
Definition of Business Performance Indicators
Use Case: FabLab
B: Knowledge sharing within the FabLab/ADF
network
Scenario: B.2: Shared repository of projects
Objective of
Use Case:
The objective of this scenario is to make the
collaboration, communication and sharing of
information between customers and designers easier.
To create an accessible portfolio of projects.
Business Benefit
of Use Case:
Make the design easier.
Achieve better quality of final products.
Promote ideas sharing.
Indicator
Name Description
Indicator
unit
Information
needed Source of data The owner Comment
Design time The time it takes from the idea to
production start Hours Actual time User FCIM
No. of shared
projects
Number of times a new product is
designed starting from an existing design
available in the shared repository
Number Use of file User/3Dexperience FCIM
Public
Copyright Manutelligence Consortium 2015-2018 Page 36 / 40
Definition of Business Performance Indicators
Use Case: FabLab
C: Developing the production cycle
Scenario: C4: Estimation of time and costs of production
C5: Statistics about use of machines
Objective of
Use Case:
The objective of this scenario is to make the
Manutelligence platform capable of supporting
production stages, being capable of gathering real data
of digital manufacturing machinery and to share the
data with the Fablab/ADF operators and/or with the
customers.
When the production cycle is defined, the
customer/user is provided with the estimation of the
production time and production cost.
Data related to the production cycle are stored in the
system so that it is possible to carry out analysis about
the use of machines.
Business Benefit
of Use Case:
Easier development of the production cycle.
Identification of resources to be used and their setting.
With the production data, it is possible to make statistics about the
most used machines, allowing the manager of the Fablab/ADF to
get better comprehension of his/her clients.
With this information, the manager can plan more activities related
to the usage of the most demanded machines.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Average
waiting time
The time the user has to wait for having
the machine available hours
Status of
machines/duration
of tasks
3D experience FCIM
Downtime Number of hours machines are not
available due to failures or stops hours Status of machines Technicians FCIM
Energy
consumption Energy consumed by machines Kwh
Energy
consumption 3D experience FCIM
Public
Copyright Manutelligence Consortium 2015-2018 Page 37 / 40
Definition of Business Performance Indicators
Use Case: FabLab
E: Sustainability assessment
Scenario: E1: Evaluation of environmental impact
E2: Life cycle cost analysis
E3: Comparison of alternative concepts
Objective of
Use Case:
The purpose of this use case is to integrate into the
design process sustainability considerations (LCA and
LCC) aimed at making designers more aware of
environmental and economic issues when designing
and manufacturing products in a Fablab/ADF.
Providing the designers with indicators that reflect the
environmental impact of the product designed.
Providing the designers with economic perspectives.
Provides data in (almost) real time so that it is possible
to compare different concepts of the same product on
the basis of its sustainability impacts.
Business Benefit
of Use Case:
Increasing the level of sustainability of the developed products.
Increase awareness of sustainability impacts of public.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Assessment
usage
Percentage of projects launching the
assessment %
Number of opening
of the assessment
tools
3D experience FCIM
Public
Copyright Manutelligence Consortium 2015-2018 Page 38 / 40
Smart House case (Lindbäcks)
Definition of Business Performance Indicators
Use Case: Smart House
B: Design and production support
Scenario: B‐1: Condition monitoring for product feedback and failure
management improvement
Objective of
Use Case:
Collect use-data throughout the life cycle to improve the
next generation of smart homes based on the real
measurements.
Improve failure management by collecting data from
more sensors at different locations in the apartment.
Monitor tenants attendance in the flat to optimize the
ventilation utilization, which means that the ventilation
can be reduced when the tenants are out of the apartment
for a longer period of time.
Business Benefit
of Use Case:
Improve quality of the houses.
Extend house life-time by faster reaction of technical -problems.
Reduced energy consumption by influencing the end users.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Events Events to be recognised (request for
repair, amounts of service activities, etc.) Number
Amount of event
reports
Service
reports LIND
Insurance costs Costs for repair of damages (e.g. water
leakages) €
Invoices for
damage repairs Invoices LIND
The higher the costs for
damage repairs the higher the
insurance fee
Reaction time Reaction time for critical events (e.g.
water leakage) hours
Time between the
critical event is
recognised and the
damage is repaired
Leakage
Measurements LIND
The faster the reaction the less
is the repair effort
Public
Copyright Manutelligence Consortium 2015-2018 Page 39 / 40
Definition of Business Performance Indicators
Use Case: Smart House
C: Service for smart home end users
Scenario: C-1: Consumption data monitoring to optimise energy
consumption according to the individual lifestyle
Objective of
Use Case:
Integrate energy consumption data into the
Manutelligence database for cross-data analysis
(compare measurement of similar apartment at different
sites).
Easy access to energy consumption monitoring for the
customer including forecast via web user interface.
Collect use-data throughout the life cycle for long term
evaluations.
Business Benefit
of Use Case:
Develop future flats and building designs by using real measured data
instead of assumptions.
Better calculation of e.g. the energy demand of an apartment than
theoretical assumption can do.
Optimise insulations in selected areas of the apartment might be
possible.
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Energy delta Delta between design specification and
actual measurement
Energy
data
(temp.
etc.)
Temperature &
energy
calculation
Energy
measurements
LIND &
tenant
Number of
complains Complains from the tenants Number
Complains
reports
Facility
management
reports
LIND
Comfort level Calculate the operative comfort level and
compare it with the measurements
Physical
units
Noise, smell,
draft, vibration,
temperature loss
Energy
measurements
Public
Copyright Manutelligence Consortium 2015-2018 Page 40 / 40
Definition of Business Performance Indicators
Use Case: Smart House
C: Service for smart home end users
Scenario: C-2: Offer of new services for living as an additional
business area
Objective of
Use Case:
Enable additional services for the end user of the house.
Instead of buying and paying for house features the user
can pay-peruse for selected features.
Business Benefit
of Use Case:
Apartment installations and functionalities can be standardized on a
high level.
Payback might be lower in a first phase but higher on a longer term –
continuous cash flow by selling functionalities on demand.
New services can be offered for additional business/services;
Users can order additional services and functionalities on a pay-per-
use basis (spend only money for needed functionalities)
Indicator
Name Description
Indicator
unit
Information
needed
Source of
data The owner Comment
Income per
costs
Cost for the installation of additional
functionalities (e.g. floor heating
equipment) must be compared with the
income of the additional services (heating
service)
€
Additional
investment costs;
Income from the
service
Service
requests LIND