• Non ci sono risultati.

Network Topologies / Architectures

N/A
N/A
Protected

Academic year: 2021

Condividi "Network Topologies / Architectures"

Copied!
14
0
0

Testo completo

(1)

Network Topologies / Architectures

Feedforward only vs. Feedback loop (Recurrent networks)

Fully connected vs. sparsely connected

Single layer vs. multilayer

Multilayer perceptrons, Hopfield network, Boltzman machines, Kohonen network

(a) A feedforward network and (b) a recurrent network

(2)

Classification Problems

Given :

1) some “features” ( ) 2) some “classes” ( )

Problem :

To classify an “object” according to its features

n 2

1

f f

f , ,....,

m

1

c

c ,....,

(3)

Example # 1

To classify an “object” as :

= “ watermelon ”

= “ apple ”

= “ orange ”

According to the following features :

= “ weight ”

= “ color ”

= “ size ”

Example :

weight = 80g

color = green size = 10 cm³

c

1

c

2

c

3

f

1

f

2

f

3

“ apple ”

(4)

Example # 2

Problem : Establish whether a patient got the flu

Classes : { “ flu ” , “ non-flu ” }

(Potential) Features :

: Body temperature

: Headache ? (yes / no)

: Throat is red ? (yes / no / medium) :

f

1

f

2

f

3

f

4

(5)

Example # 3

Classes = { 0 , 1 }

Features = x , y : both taking value in [ 0 , +∞ [

Idea : Geometric Representation

(6)

Neural Networks for Classification

A neural network can be used as a classification device .

Input ≡ features values Output ≡ class labels

Example : 3 features , 2 classes

(7)

Thresholds

We can get rid of the thresholds associated to neurons by adding an extra unit permanently clamped at -1 .

In so doing, thresholds become weights and can be adaptively adjusted during learning.

(8)

Simple Perceptrons

A network consisting of one layer of M&P neurons connected in a feedforward way (i.e. no lateral or feedback connections).

Capable of “learning” from examples (Rosenblatt)

They suffer from serious computational limitations (Minsky and Papert, 1969)

(9)

Decision Regions

It’s an area wherein all examples of one class fall . Examples :

(10)

Linear Separability

A classification problem is said to be linearly separable if the decision regions can be separated by a hyperplane .

Example : AND

1 1

1

0 0

1

0 1

0

0 0

0

X AND Y Y

X

(11)

Limitations of Perceptrons

It has been shown that perceptrons can only solve linearly separable problems (Minsky and Papert , 1969) .

Example : XOR (exclusive OR)

0 1

1

1 0

1

1 1

0

0 0

0

X XOR Y Y

X

(12)

A View of the Role of Units

(13)

Convergence of Learning Algorithms

If the problem is linearly separable, then the learning rule converges to an appropriate set of weights in a finite number of steps (Nilsson 1965)

In practice, one does not know whether the problem is linearly separable or not.

So decrease η with the number of iterations, letting η 0 .

The convergence so obtained is artificial and does not necessarily yield a valid weight vector that will classify all patterns correctly

Some variations of the learning algorithm, e.g. Pocket algorithm, (Gallant, 1986)

(14)

Multi–Layer Feedforward Networks

Limitation of simple perceptron: can implement only linearly separable functions

Add “ hidden ” layers between the input and output layer. A network with just one hidden layer can represent any Boolean functions including XOR

Power of multilayer networks was known long ago, but algorithms for training or learning, e.g. back-propagation method, became available only recently (invented several times, popularized in 1986)

Universal approximation power: Two-layer network can approximate any smooth function (Cybenko, 1989; Funahashi, 1989; Hornik, et al.., 1989)

Static (no feedback)

Riferimenti

Documenti correlati

Solution proposed by Roberto Tauraso, Dipartimento di Matematica, Universit`a di Roma “Tor Vergata”, via della Ricerca Scientifica, 00133 Roma,

The walk begins at vertex 0 and continues until every vertex has been visited and the walk returns to vertex 0.. Indeed the already visited verteces are contiguous, the starting

[r]

If the training set is linearly separable, the perceptron learning algorithm always converges to a consistent hypothesis after a finite number of epochs, for any η > 0.. If

The autonomous motivation (RAI), sex, cognitive reasoning and school achievement were used as covariates to control their effects on measured knowledge variables (correct answers

Government Printing Office , Social Security Administration (US)..

In the last section, restricting ourselves to functional structure sheaves, we prove that the product of differential triads can be considered instead of the fibre product, as

One can easily verify that every element of ~ is a right tail of 5 preordered withrespect to relation defined 1n (j).. A CHARACTERIZATION DF V-PRIME ANO 5TRONGLY V-PRIME ELHlrtHS OF