Contribution Separators Margin Processing Of Remote Sensing Data Biology Essay

Published: November 2, 2015 Words: 1550

The separators Wide Margin (SVMs) and Support Vector Machines in English, are supervised classification methods developed during the 90s by V. Vapnik. The SVMs aim to separate in a space dimension n appropriate, a set of vector data belonging to different classes of linear separators (right, plan or hyper plane). Among the multitude of linear separators that could be used, SVMs seek the most optimal separator, with a maximum distance between classes. According to the severability of data, SVMs are distinguished by two models: the linear SVM (where the data is linearly separable) and non-linear (in case the data are not linearly separable). We have therefore introduced the techniques of learning SVMs to the segmentation classification contained a satellite.

2 SEPARATE THEM WIDE MARGIN

The SVMs have been designed to solve the problems of binary classification supervised, and then extended through the strategies and approaches, the multi-class classification

2.1 The binary separators Wide Margin

Where the data is linearly separable and as part of a binary classification, it is a set of m vector data which is associated labels ti (-1, 1) representative of their class. The linear separator of this data set is defined by

h (x) = wt.x + b = 0

Where w = (w1, ..., wn) is the normal vector,

And x = (x1 ,..., xn) a vector splitter b linearly and the threshold.

The function h (x) will represent the border separating the two classes, the classes' ti another example x is defined as:

ti = sign (+ wt.x b)

For Maximizing the distance between classes through maximizing the margin, we must close distance to vehicles known vectors of support.

As the margin is inversely proportional to the standard w [2],The search for optimal

linear

separator

is:

Min 1 w 2

2

(1)

i:t .h(x ) ≥1

i.

i

The system obtained (1) represents the expression of primal optimization problem in SVMs. In order to simplify the constraints we solve the problem by its dual and using the method of Lagrange, we

get

L

(w,

b,

α):

1

m

2

t

Min

w −∑αi.[(w .xi +b).ti −1]

(2)

2

i=1

≥0 i =1...m

αi

As the αi represent the contribution of a xi element in the design of linear separator h (x), where only the αi corresponding to support vectors are not zero. The Lagrangian term obtained will aim to minimize L (w, b, α) over a: w and b, and maximize L (w, b, α) by providing a: α. So the research on extremum of L (w, b, α):

δL(w,b,α)

m

*

=0 w = αi.ti.xi

(3)

δw

i=1

δL(w,b,α)

m

=0∑αi.ti =0

(4)

δb

i=1

By replacing (3) and (4) in L (w, b, α)

gives

the

following

function:

m

1 m

t

F(α) = αi − ∑t i.t j.αi.αj.x i .x j

(5)

i=1

2 i,j=1

The equation (5) allows us to recreate the primal optimization problem expressed in

(6)

by

its

dual:

m

1

m

t

Max∑αi −

.∑ti..ti.αi.αj.xi.x j

i=1

2 i,j=1

m

=0

α .t

∑i

i

(6)

i=1

α ≥0

i =1..m

i

Thus, the search for optimal returns linear separator has a quadratic programming problem where and αi are reckonable b and w can be deducted.

The function associated decision becomes:

m

t

h (x )= αi.t i.x i .x +b

(7)

i =1

Where the data are not linearly separable, it is usually possible to find a linear separator performing an optimal projection of a space E to another area F larger and using functions called projection Ф (x) as:

φ: E →F

(8)

x =(x1,..,xn ) →φ(x) =(φ1(x),..,φn (x),..)

Avec,CARD(E) <CAED(F)

This allows us to translate the optimization problem expressed (6) by:

m

1 m

t

Max α− . t .t .α.α.x .x

∑i

∑i.

i

i

j

i

j

i=1

2 i,j=1

m

α.t =0

∑i

i

(9)

i=1

0≤α≤C i =1..m

i

This transformation can be done implicitly by the kernel K (x, y) = Ф (x) t.Ф (y) including the following:

3 CORE: 4 generic form: 5 parameters: 6:

Standard

deviation7 :laplacian

polynomial

8:

agenda

polynomial

9:Gaussian

based

radial

10:

standard

deviation

Thus, the optimization problem formulated by (8) relation can be rewritten as follows: In practice it is almost impossible to classify all the data perfectly. That is why

V. Vapnik proposes to introduce new variables known springs, to ease the

constraints,

we

get:

i : ti. h(xi)

≥1-ξι

(10)

The optimization problem becomes:

1

m

Min

w

2

+C ∑ξi

(11)

2

i =1

m

α −

1 m

t

Max

t .t .α.α.φ(x ) .φ(x )

∑i

∑i

j

i

j

i

j

i=1

2 i,j=1

m

=0

(13)

α .t

∑i i

i=1

α ≥0 i=1..m

i

3.

SVMS

MULTI-CLASSES:

As C is a constant adjustment between the margin and errors Thus, the dual expression remains the same (for SVMs linear and nonlinear), the only difference is that all Lagrange multipliers αi must be bounded by the constant Super C.

The application of SVMs to a classification with class k, with k> 2 requires a generalization of SVMs bi-level, an increase of binary classifiers. Among the different approaches that can be used for this purpose there is the approach one against one against one and all.

The dual system optimization in the case of

3.1

The

approach

one

against

One

linear

SVMs

becomes:

m

1 m

This

is to

design every possible

binary

Max∑αi −

∑ti.t j.αi.αj.K(xi,x j)

classifiers, and, for k classes will k. (k-1) /

i=1

2 i,j=1

2 classifiers. To

assign

an

element

to a

m

class

e,

e

must

be

tested

with all

the

∑αi.ti =0

(12)

classifiers

designed,

whenever

e

is

i=1

assigned to a class, we increment a counter

i associated

with

it

(i

is initially

set to

0 ≤αi ≤C i=1..m

zero). e will be awarded to the class that

Where: the table .1 is shown

has

a

meter

at

maximum

value.

3.2 A aproach against all approaches

In this approach and a classification with k

CORE :

generic form :

Parameter

classes, each class i is opposed to k-1 and

:

Laplacian

K (x , y )=exp ( − x −y

δδ:Ecart

other

classes

will

develop

k

binary

type

classifiers. To assign an element to a class

Polynomi

K (x , y )=( x . y +1) p

p :polynô

e, e must be tested against all classifiers

al

me Order

designed,

then

assigned

to

the

class

that

Gaussian

2

K(x, y)=exp(−x −y / 2

δδ:Ecart

presents

a

decision

point

e

maximum.

base

type

radial

4.

APPLICATION

OF

THE

SEGMENTATION SVMS CONTENT

Table .1

The dual system optimization in the case of non-linear SVMs becomes:

OF A SATELLITE IMAGE

We have applied SVMs non-linear approach to one against all the segmentation of the content of satellite

images type Landsat5 TM (Thematic Maps) from the region west Bejaia city dated 15 March 1993 at 9am 45mn. This study area was chosen for its varied landscape may be relevant to the application of SVMs to problems of classification multi-classes. We started our application by loading three images corresponding to the three channels TM1 TM2 and TM3. To ease the use of these images, we first made a contrast enhancement and a composition combining colored blue filter channel TM1, the filter TM3 green channel and red channel filter TM4. On the image we conducted sampling to build our base of learning, the latter representing different classes of pixels identified using knowledge themes:

Mer

Ressac

Sable & sol nu

Céréaliculture

Jachère

Urbain

Brûlis

Maquis

Forêt

Maraîchage

Sebkha 1

Sebkha 2

Figure 1: Identification of classes

Once the basic learning built we conducted our segmentation with the following entries: we set the compromise C to 500 and used a Gaussian kernel as well, we achieved the following results (figure.2):

Figure 2: Images of the resulting test 1 and 3 Gaussian kernel (from left to right)

The first test with a standard deviation of a Gaussian 0.1 gave us the best recognition rate (TR) but also a large number of confusion mainly located at the classroom level Sebkha 2, urban, sand and bare ground. The increase in the

standard

deviation gaussian

level

test 2 and test 3, has resulted in the

decrease of recognition and the

disappearance of

classes Sebkha2,

sand and bare ground. The Tests

Settings

percent

is

shown

as:

Tests

Settings

TR

(%)

1

δ

=

0.1

74.51

2

=

0.5

64.98

3

δ

=

1

62.82

Table 2: Results with the Gaussian kernel

5. SIMULATIONS AND RESULTS

To ensure the recovery of the flow of data without learning sequence equalizer based on maximum likelihood is running a two-step procedure to identify the channel by the algorithm that we described earlier, and then we search the data using the results of this identification.

To assess the performance of the equalizer, we begin by presenting the accuracy of the identification algorithm for different SNR

5-1 Identification of Images of Gaussian kernel:

The identification of the Images of the resulting test 1 and 3 Gaussian kernel (figure.2) is used to detect information symbols.

We worked with a comment that contains ten information symbols (N = 10). The estimated coefficients are represented according to the index iteration, for reports RSB equal to 0, 5 and 10 dB.