Systems Engineering Assignment #6 Solution

$30.00 $24.00

1. [5 points] Consider the following training set x 1 = 0 ,t 1 = 0 , x 2 = 0 ,t 2 = 1 , x 3 = 1 ,t 3 = 1 , x 4 = 1 ,t 4 = 1 0 1 0 1 a) Plot the training samples in the feature…

5/5 – (2 votes)

You’ll get a: zip file solution

 

Description

5/5 – (2 votes)

1. [5 points] Consider the following training set

x 1 =

0

,t 1 = 0 , x 2 =

0

,t 2 = 1 , x 3 =

1

,t 3 = 1 , x 4 =

1

,t 4 = 1

0

1

0

1

a) Plot the training samples in the feature space.

  1. Apply the perceptron learning rule to the training samples one-at-a-time to obtain weights w1, w2, and bias w0 that separate the training samples. Use w = [w0, w1, w2] = [0, 0, 0] as initial values (consider bias input x0 = 1, and learning rate = 1). Write the expression for the resulting decision boundary and draw it in the graph. [Hint: You can use Excel / OO Calc to implement the learning rule for perceptron, such as the spreadsheet of InClass_09 posted on eClass].

Epoch

Inputs

Desired

Initial weights

Actual

Error

Updated weights

output t

output y

x1

x2

w0

w1

w2

w0

w1

w2

1

0

0

0

0

0

0

0

1

1

1

0

1

1

1

1

2

0

0

0

0

1

1

1

0

1

1

1

1

3

0

0

0

0

1

1

1

0

1

1

1

1

2. [5 points] Consider the following training set

x 1 =

0

,t 1 = 0 , x 2 =

0

,t 2

= 1 , x 3 =

1

,t 3 = 1 , x 4 =

1

,t 4 = 0

0

1

0

1

which describes the exclusive OR (XOR) problem.

  1. Establish mathematical (not graphical) proof that this problem is not linearly separable. [Hint: Start with assumption that these patterns are linearly separable, write down equations/inequalities corresponding to this assumption and examine them for conflict; first such inequality is provided below as an example.]

Suppose that the problem is linearly separable. The decision boundary can be represented as:

20 = 0 or (expanded) 0 0 + 1 1 + 2 2 = 0 This assumption means that either

+ + < 0 ( , ) = (0,1) ∧ ( , ) = (1,0)

  1. 0 0 + 1 1 + 2 2 ≥ 0 ( 1, 2) = (0,0) ∧ ( 1, 2) = (1,1),

or 0 0 1 1 2 21 21 2

+ + > 0 ( , ) = (0,1) ∧ ( , ) = (1,0)

  1. 0 0+ 1 1+ 2 2 ≤0 ( 1, 2)=(0,0)∧( 1, 2)=(1,1).0011221212

must be satisfied. Following one of the cases and putting the values ( 1, 2) under variables, one obtains

  1. 0 0+ 2<0

(2)

(3)

(4)

  1. Apply the perceptron learning rule following the same procedure as in Problem 1. Describe your observation.

Epoch

Inputs

Desired

Initial weights

Actual

Error

Updated weights

output t

output y

x1

x2

w0

w1

w2

w0

w1

w2

1

0

0

0

0

0

0

0

1

1

1

0

1

1

1

0

2

0

0

0

0

1

1

1

0

1

1

1

0

3

0

0

0

0

1

1

1

0

1

1

1

0

Systems Engineering Assignment #6 Solution
$30.00 $24.00