• Non ci sono risultati.

Temporal Decision Trees for Diagnosis

N/A
N/A
Protected

Academic year: 2021

Condividi "Temporal Decision Trees for Diagnosis"

Copied!
30
0
0

Testo completo

(1)

Temporal Decision Trees for Diagnosis

Claudia Picardi

Dipartimento di Informatica Università di Torino

Model-Based Diagnosis Model-Based Diagnosis

µ

Model = description of a system to be diagnosed

 “objective” description.

 usually component oriented.

 may describe only correct behaviour or both correct and faulty behaviour.

µ

Diagnostic problem:

 discrepancy between values predicted by the model and values observed in the real world

µ

Model-Based Diagnosis:

 find source of discrepancy (component or fault)

 by reasoning on the model.

(2)

Temporal Decision Trees 3 for Diagnosis Claudia Picardi

Università di Torino

Decision Trees Decision Trees

µ

A decision tree can be seen as a classification procedure.

µ

Classification problem:

 X belongs to a class among C1,…,Cn

 You can carry out some tests T1,…,Tk on X.

 You must decide which class X belongs to.

µ

A decision tree encodes a procedure to classify X.

Example: Wine Classification Example: Wine Classification

µ Inner node: a test T to carry out on X.

µ Branches: a subtree for each possible outcome of T.

µ Leaves: final decision, i.e. a class C.

A1 Are you having exotic spicy food?

a if yes: aromatic wine (e.g. gewurtztraminer) S2 if no: are you having red meat?

b if yes: strong red wine (e.g. nebbiolo) c if no: are you having fish?

S3

(3)

Temporal Decision Trees 5 for Diagnosis Claudia Picardi

Università di Torino

Building Decision Trees Building Decision Trees

µ

Most common approach:

 Learn the decision tree from a set of examples

µ

Each example describes a situation where

 The outcome of each test is known

 The correct classification is known

 In other words, a table:

Test T1 Test T2 Test T3 Class

1 2 3

Test T1 Test T2 Test T3 Class

1 2 3

Test T1 Test T2 Test T3 Class

1 2 3

Generic algorithm Generic algorithm

Recursive function BuildTree Input: Examples; Output: Tree

µ if all examples belong to the same class C

µ return C as tree leaf

µ else

µ choosea test T as root of the tree

µ for each possible outcome v of T

µ select only those examples where T has outcome v

µ recursively call BuildTree on them

µ the resulting tree becomes subtree of the current one.

µ return the tree.

(4)

Temporal Decision Trees 7 for Diagnosis Claudia Picardi

Università di Torino

Choosing a test: the goal Choosing a test: the goal

µ

Goal: build the best decision tree.

µ

Best decision tree = tree with minimum average depth.

µ

Why?

 Efficient decision process: Min. depth = Min.

number of tests to carry out = Quicker classification.

 Generality: Avoid looking at unnecessary information that may depend on the particular set of examples.

ID3

[Quinlan, Machine Learning 1986]

ID3

[Quinlan, Machine Learning 1986]

µ

Choice based on the quantity of information carried by a test T.

 Quantity of information = −entropy

µ

Measures the significance of a test T wrt the final classification:

 If all classes are equiprobable wrt the outcome of T then entropy is max.

 If exactly one class corresponds to each

possible outcome of T then entropy is 0 (min).

µ

Choose the test with min. entropy.

(5)

Temporal Decision Trees 9 for Diagnosis Claudia Picardi

Università di Torino

Recursive function BuildTree Input: Examples; Output: Tree

µ if all examples belong to the same class C

µ return C as tree leaf

µ else

µ choose the test T with min. entropy as root

µ for each possible outcome v of T

µ select only those examples where T has outcome v

µ recursively call BuildTree on them

µ the resulting tree becomes subtree of the current one.

µ return the tree.

Generic algorithm Generic algorithm

On-Board Diagnosis On-Board Diagnosis

µ

Two categories of diagnostic software:

Off-line:

Runs after the device has stopped operating.

Its purpose is repair.

On-line:

Runs while the device is operating.

Its purpose is recovery.

real time requirements

(6)

Temporal Decision Trees 11 for Diagnosis Claudia Picardi

Università di Torino

On-Board Diagnosis On-Board Diagnosis

µ

Two categories of diagnostic software:

Off-line:

Runs after the device has stopped operating.

Its purpose is repair.

On-line:

Runs while the device is operating.

Its purpose is recovery.

On-board: On-line + Runs embedded within the device.

real time requirements memory constraints

MBD and On-Board Diagnosis MBD and On-Board Diagnosis

µ

Current model-based engines are not suited to most on-board applications:

 they need too much memory

 they are not swift enough

µ

They can however be exploited to…

 generate specifications for on-board software

 generate directly on-board diagnostics.

µ

A possible approach:

 use an MBD engine to generate decision trees

(7)

Temporal Decision Trees 13 for Diagnosis Claudia Picardi

Università di Torino

Decision Trees for Diagnosis Decision Trees for Diagnosis

µ Inner Nodes: sensors to read

µ Children: possible sensor values

µ Leaves: recovery actions to perform

S1 read sensor S1

a if value is x then do action a S2 if value is y then read S2

b if value is u then do action b c if value is v then do action c S3 if value is z then readS3

S1 read sensor S1

a if value is x then do action a S2 if value is y then read S2

b if value is u then do action b c if value is v then do action c S3 if value is z then readS3

Generating Decision Trees Generating Decision Trees

1. Exploit an MBD engine to generate a table:

 the table associates combinations of sensor values to recovery actions.

 can be built by simulating the device exhaustively for all relevant fault combinations.

 or by diagnosing a significant set of real cases.

Model-Based Diagnostic

Engine

S1 S2 S3 … … x

y y z

u u v v

r r t r

a b c

S1 S2 S3 … …

x y y z

u u v v

r r t r

a b c

(8)

Temporal Decision Trees 15 for Diagnosis Claudia Picardi

Università di Torino

Generating Decision Trees Generating Decision Trees 2. Use the table to generate a

decision tree.

 most used algorithm: ID3.

 choose most discriminating sensor as a new node

 create a children for each possible value

 if a value allows to select an action create a leaf.

 otherwise create an inner node.

 proceed recursively.

S1 S2 S3 … … x

y y z

u u v v

r r t r

a b c

S1 S2 S3 … …

x y y z

u u v v

r r t r

a b c

Generating Decision Trees Generating Decision Trees 2. Use the table to generate

a decision tree.

 most used algorithm: ID3.

 choose most

discriminating sensor at each step

Model-Based Diagnostic

Engine

S1 S2 S3 … … x

y y z

u u v v

r r t r

a b c

S1 S2 S3 … …

x y y z

u u v v

r r t r

a b c

(9)

Temporal Decision Trees 17 for Diagnosis Claudia Picardi

Università di Torino

Temporal On-Board Diagnosis?

Temporal On-Board Diagnosis?

µ This approach does not exploit temporal information.

With a temporal MBD engine it is possible to generate a table with temporal information

Temporal Model-Based

Diagnostic Engine

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

Temporal On-Board Diagnosis?

Temporal On-Board Diagnosis?

µ This approach does not exploit temporal information.

1. The tree cannot express temporal information.

¨temporal decision trees

Temporal Model-Based

Diagnostic Engine

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

(10)

Temporal Decision Trees 19 for Diagnosis Claudia Picardi

Università di Torino

Temporal On-Board Diagnosis?

Temporal On-Board Diagnosis?

µ This approach does not exploit temporal information.

2. the ID3 algorithm does not deal with temporal information.

¨temporal ID3

Temporal Model-Based

Diagnostic Engine

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

S1 S2 S3

x 1

b c a

2 … 1 2 … 1 2 … 1 2 … 1 2 … y …

x x … x y …

u u … v v … v u …

t r … t r … t t …

… … … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

… … …

Temporal Example Set Temporal Example Set

µ

An example of correct diagnosis:

 Sensors: s1,…,sk

 Snapshots: 0,1,…,T

 for each pair (s,t) the value of s at time t.

 the correct recovery action to perform.

µ

Additional information:

 Deadline for performing the action.

 Partial order defined on actions: recovery capability.

 Cost function defined on actions, must respect the partial order.

(11)

Temporal Decision Trees 21 for Diagnosis Claudia Picardi

Università di Torino

Temporal Decision Trees Temporal Decision Trees

µinner nodes have also time labels

µwhen the tree is executed a counter is activated

µtime labels express the value the counter must have in order to read the sensor.

S1,1 wait until time 1and read sensor S1 a if value is x then do action a

S2,4 if value is y then wait until time 4and readS2 b if value is u then do action b

c if value is v then do action c

S3,3 if value is z then wait until time 3and readS3

S1,1 wait until time 1and read sensor S1 a if value is x then do action a

S2,4 if value is y then wait until time 4and readS2 b if value is u then do action b

c if value is v then do action c

S3,3 if value is z then wait until time 3and readS3

TDT Properties TDT Properties

µ

Time Labels must be increasing from root to leaves

 The diagnostic software cannot go back in time to read previous sensor values

 Sensor values cannot be stored (memory restrictions)

µ

The time label cannot exceed the

deadline for any of the current candidates.

 rather than exceeding the deadline perform a recovery action immediately

 select the one with highest recovery capability (or combine several actions)

(12)

Temporal Decision Trees 23 for Diagnosis Claudia Picardi

Università di Torino

Why Waiting Is Good Why Waiting Is Good

µ

Decision Trees suggest a sequence of steps where the steps are either reading a sensor or performing a recovery action.

µ

Temporal Decision Trees introduce waiting.

µ

Loss of time? No.

 avoid introducing unnecessary nodes in the tree

 keep the tree smaller.

 of course must guarantee to select a recovery action in time.

Why Not Using ID3?

Why Not Using ID3?

µ

For Decision Trees: ID3 selects the most discriminating sensor.

µ

For Temporal Decision Trees: it could

select the most discriminating sensor and time label.

µ

Problem:

Moving ahead in time causes loss of information.

It is not possible to go back and read

previous sensor values.

(13)

Temporal Decision Trees 25 for Diagnosis Claudia Picardi

Università di Torino

Why Not Using ID3?

Why Not Using ID3?

Moving ahead in time causes loss of information.

It is not possible to go back and read previous sensor values.

S1 S2 S3

b c a

x

1 2 … y … x x … x y …

… … …

1 2 … u u … v v … v u …

… … …

1 2 … t r … t r … t t …

… … …

1 2 …

… … …

… … …

… … …

… … …

1 2 …

… … …

… … …

… … …

… … …

S1 S2 S3

b c a

x

1 2 … y … x x … x y …

… … … x

1 2 … y … x x … x y …

… … …

1 2 … u u … v v … v u …

… … … 1 2 … u u … v v … v u …

… … …

1 2 … t r … t r … t t …

… … … 1 2 … t r … t r … t t …

… … …

1 2 …

… … …

… … …

… … …

… … … 1 2 …

… … …

… … …

… … …

… … …

1 2 …

… … …

… … …

… … …

… … … 1 2 …

… … …

… … …

… … …

… … …

Measuring Information Loss Measuring Information Loss

µ

Max discrimination degree: given by example set.

 Tree that uses all sensors at all snapshots = max discrimination degree.

µ

Loss of information:

 The example set is reduced

 The resulting discrimination degree could be lower.

 Loss of discrimination capability.

µ

How can we quantitatively measure it?

(14)

Temporal Decision Trees 27 for Diagnosis Claudia Picardi

Università di Torino

Expected Cost of a TDT Expected Cost of a TDT

µ

What happens when it is not possible to discriminate?

 Choose the action with the highest recovery capability.

 In some cases I perform a stronger action than needed.

 Stronger actions cost more.

 I pay a higher cost.

µ

Idea: measure discrimination capability with the expected cost of the tree.

Expected Cost Function Expected Cost Function

µ

Expected cost of a tree:

µ

Recursively computed from subtrees.

µ

Expresses the expected cost of a recovery action selected by the tree.

child of

(Act( ))

( ) ( ( ) | (

if is a leaf

otherw

)) ( ) ise

c n

n

n P n c

n c

 χ

=

Χ ⋅

 Χ

  ∑ E E

(15)

Temporal Decision Trees 29 for Diagnosis Claudia Picardi

Università di Torino

Properties of Expected Cost Properties of Expected Cost

1.

It correctly measures discrimination capabiltiy:

µ A more discriminating tree has a lower cost.

2.

When I choose a pair (s,t) with higher t, it may decrease but it does not increase.

µ It measures also information loss.

3.

It depends only from the tree leaves

µ It does not depend on the internal structure of the tree.

New Goal New Goal

µ

When generating the tree…

µ

…choose a pair (s,t) of a sensor s and snapshot t such that:

1.

The resulting tree has the minimum possible expected cost.

2.

The resulting tree is reasonably small.

(16)

Temporal Decision Trees 31 for Diagnosis Claudia Picardi

Università di Torino

Temporal ID3 (sketch) Temporal ID3 (sketch)

µ

When choosing a sensor s and a snapshot t:

 moving ahead in time I may increase the cost.

time

Temporal ID3 (sketch) Temporal ID3 (sketch)

µ

When choosing a sensor s and a snapshot t:

 moving ahead in time I may increase the cost.

1. find the point in time where the loss becomes critical (the cost increases)

time

(17)

Temporal Decision Trees 33 for Diagnosis Claudia Picardi

Università di Torino

Temporal ID3 (sketch) Temporal ID3 (sketch)

µ

When choosing a sensor s and a snapshot t:

 moving ahead in time I may increase the cost.

1. find the point in time where the loss becomes critical (the cost increases)

2. choose ID3-like a sensor and a time label, but only among safe time labels.

time

Temporal ID3 (sketch) Temporal ID3 (sketch)

µ

Moving ahead in time I lose more and more information.

1. find the point in time where the loss becomes critical (the cost increases)

2. choose ID3-like a sensor and a time label, but only among safe time labels.

time

(18)

Temporal Decision Trees 35 for Diagnosis Claudia Picardi

Università di Torino

Temporal ID3 Temporal ID3

µ

How to determine whether the the cost dicreases or not for a given time label t?

 build a lowest-cost decision tree using only information at times ≥ t.

µ

Problem: if I build this tree recursively, I am performing exhaustive search!!!

 It is like looking explicitly at all the possibilities and selecting the best one.

 Not clever, not efficient.

First Step towards a Solution First Step towards a Solution

µ

These trees are needed only for their cost.

 The cost does not depend on the internal structure of the tree.

 In a sense, the cost depends only on the example set used for building the tree.

 These trees do not need to be small.

 Why bother with recursive calls? Build the simplest tree.

µ

Build the exhaustive tree.

 The tree using all sensors and snapshots remaining in the data set.

(19)

Temporal Decision Trees 37 for Diagnosis Claudia Picardi

Università di Torino

Second Step towards a Solution Second Step towards a Solution

µ

I must build TREE(t0), TREE(t0+1),…,TREE(T)

 TREE(i) uses only information at times ≥ i

 TREE(i-1) uses all the information of TREE(i) and some more.

 The cost does not depend on the inner structure of the tree.

µ

Build TREE(i-1) by adding nodes to TREE(i).

 I can build all the exhaustive trees…

 …while I am building TREE(t0)!

µ

Same complexity as standard ID3!

 O(N2MT) in the worst case and O(NMTlogN) in the best case

An Extension An Extension

µ

Deadlines are hard.

 A deadline either is met or not.

 There are situations in which there is a tradeoff between the risk of waiting and the risk of doing the “wrong” action.

µ

Example: preventive diagnosis.

 Waiting: risk that the fault occurs.

 Not discriminating properly: risk to stop the device when it is not necessary.

(20)

Temporal Decision Trees 39 for Diagnosis Claudia Picardi

Università di Torino

An Extension An Extension

µ

Deadlines are hard.

 A deadline either is met or not.

 There are situations in which there is a tradeoff between the risk of waiting and the risk of doing the “wrong” action

µ

Introduce soft deadlines:

 There is an additional cost associated to waiting

 This additional cost depends also on the fault situation.

The new cost function The new cost function

µ

The cost of a leaf now depends:

 on the action it selects.

 on the fault situations it corresponds to.

 on the time it is reached at.

( )

child of

( | ( )) [ (Act( ), ) ( ,Time( ))]

( ) ( ( ) | ( )) ( ) otherwi se

e n

c n

P e n n e e n

W n P c n W c

γ δ

 

=  

⋅ +

E

E

E E

if is a leaf n

(21)

Temporal Decision Trees 41 for Diagnosis Claudia Picardi

Università di Torino

The Goal The Goal

µ

Build a temporal decision tree:

 where sensor readings are used in the order they are read (same as in the original TID3).

 that has the minimum cost (same as in the original TID3 but with a different cost function)

 trying to keep the tree small in order to comply with the space constrain.

µ

Can we use TID3 as it is?

 No.

Generating the Tree:

a Problem Generating the Tree:

a Problem

µ

Conditions on the cost function for TID3 to work:

1. The cost of a tree must depend only on how the fault situations split among the tree leaves.

2. A more discriminating tree must not have an higher cost than a less discriminating one.

µ

Function W violates both conditions!

1. The cost depends also on the time at which tree leaves are reached.

2. A more discriminating tree could have an higher cost due to the time at which it selects recovery actions.

(22)

Temporal Decision Trees 43 for Diagnosis Claudia Picardi

Università di Torino

Generating the Tree:

Solution (outline) Generating the Tree:

Solution (outline)

µ

Idea: Transform the problem with soft deadlines into a problem with hard

deadlines so that:

1. The solution (a tree) to the transformed problem is also a solution to the original problem.

2. The transformation does not increase the complexity of the algorithm.

µ

We introduce a preprocessing step.

Preprocessing 1:

the Exhaustive Tree Preprocessing 1:

the Exhaustive Tree

µ

First step: build the exhaustive tree.

 Temporal Decision Tree that uses all information in the example set.

 It is definitely the most discriminating tree.

 It is not the cheapest.

 It is not small.

µ

This step requires visiting once the whole example set:

 Complexity = O(NMT)

 N=#rows; M=#sensors; T=#snapshots

 NMT = size (# of cells) of the example set.

(23)

Temporal Decision Trees 45 for Diagnosis Claudia Picardi

Università di Torino

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

Do I get a cheaper subtree by replacing this node with a leaf?

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

Do I get a cheaper subtree by replacing this node with a leaf?

(24)

Temporal Decision Trees 47 for Diagnosis Claudia Picardi

Università di Torino

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

(25)

Temporal Decision Trees 49 for Diagnosis Claudia Picardi

Università di Torino

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

(26)

Temporal Decision Trees 51 for Diagnosis Claudia Picardi

Università di Torino

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

(27)

Temporal Decision Trees 53 for Diagnosis Claudia Picardi

Università di Torino

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

Do I get a cheaper subtree by replacing this node with a leaf?

Preprocessing 2:

Pruning Preprocessing 2:

Pruning

µ Visit the tree from the leaves to the root.

µ For each inner node ask:

µ NO: Don’t prune.

µ YES: Prune.

µ Complexity: O(NMT) Do I get a cheaper

subtree by replacing this node with a leaf?

(28)

Temporal Decision Trees 55 for Diagnosis Claudia Picardi

Università di Torino

Theorem: let Tp denote the pruned tree.

For every other temporal decision tree T for the same set of examples:

1) W(T) ≥ W(Tp)

2) if there is a fault situation e for which Tp selects a recovery action at time t, while T selects it at time t’ > t, then W(T)>W(Tp)

Properties of the Pruned Tree Properties of the Pruned Tree

1. The pruned tree has the minimum possible cost for the given set of examples.

2. Any tree exploiting information which has been pruned will not have minimum cost.

Why is it so?

[Proof Sketch: part 1]

Why is it so?

[Proof Sketch: part 1]

µ

The pruning step selects one of the cheapest among all prunings.

 It tries out all prune operations and carries out only those that lower the cost of the tree.

 The obtained pruning has thus the lowest possible cost.

 Moreover, less invasive prunings are all more expensive.

 This shows that no unnecessary prunings are carried out.

µ

The theorem holds for prunings.

(29)

Temporal Decision Trees 57 for Diagnosis Claudia Picardi

Università di Torino

Why is it so?

[Proof Sketch: part 2]

Why is it so?

[Proof Sketch: part 2]

µ

Given any temporal decision tree for an example set:

 we can complete it into a pruning with the same or lower cost.

 The completed tree selects recovery actions at the same time as the initial one.

µ

This allows us to extend the result

 from the set of prunings…

 …to the set of all temporal decision trees.

The Transformed Problem The Transformed Problem

µ

Transforming the problem:

 The pruned tree can be used to define hard deadlines.

 The hard deadline (hd) for a fault situation is

( ) Time(Leaf ( ))

hd e = e

(30)

Temporal Decision Trees 59 for Diagnosis Claudia Picardi

Università di Torino

Recap Recap

µ

We have defined a preprocessing step that:

 Transforms the problem with soft deadlines, not solvable by TID3, into a problem with hard deadlines, that TID3 can solve.

 Has a complexity of O(NMT).

 Since the complexity of TID3 is O(N2MT) in the worst case and O(NMTlogN) in the best case, the overall complexity is unchanged.

Riferimenti

Documenti correlati

In this paper the problem of Connected Components Label- ing (CCL) in binary images using Graphic Processing Units (GPUs) is tackled by a different perspective.. In the last

In this work we have presented the results of numerical simulations of a fully Eulerian model for a two-way coupled particle-laden turbulent Kolmogorov flow at varying the inertia

So far segmentation methods for classification and regression trees have been proposed as supervised approach to analyze data sets where a response variable and a set of

In particular, the actions of PEPA are used to express conditions on input, output (and state) variables.. In the following we discuss in detail one component,

Sulla base delle specifiche di progetto precedentemente enunciate è stato studiato il layout dell’intera macchina, presentato in figura 2, dove è possibile identificare la zona

A seguito di un finanziamento del- la fondazione CRT (progetto CARBO- STRESS, https://www.plantstresslab. unito.it/projects/carbostress) è stata progettata e costruita

In the analysis of a building curtain, the technical elements are divided into units, areas or homogeneous volumes by type of construction identifying: the type of material (the

The Case Study of Sedad Hakki Eldem (Doctorate School in Architecture, design and history of the Arts at the Univer- sity of Florence), results partially published in S.