0% found this document useful (0 votes)
301 views

Icciet Proceedings

International Association of Engineering and Technology for Skill Development

Uploaded by

iaetsdiaetsd
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
301 views

Icciet Proceedings

International Association of Engineering and Technology for Skill Development

Uploaded by

iaetsdiaetsd
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 369

PROCEEDINGS

ICCIET - 2014
INTERNATIONAL CONFERENCE ON CURRENT
INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Sponsored By
INTERNATIONAL ASSOCIATION OF ENGINEERING &
TECHNOLOGY FOR SKILL DEVELOPMENT

Technical Program
31 August, 2014
Hotel Pavani Residency, Nellore

Organized By
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL
DEVELOPMENT

www.iaetsd.in

Copyright 2014 by IAETSD


All rights reserved. No part of this publication may be reproduced, stored in a
retrieval system, or transmitted, in any form or by any means, electronic,
mechanical, photocopying, recording, or otherwise, without the prior written
consent of the publisher.

ISBN: 378 - 26 - 138420 - 5

http://www.iaetsd.in

Proceedings preparation, editing and printing are sponsored by


INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY
FOR SKILL DEVELOPMENT COMPANY

About IAETSD:
The International Association of Engineering and Technology for Skill
Development (IAETSD) is a Professional and non-profit conference organizing
company devoted to promoting social, economic, and technical advancements
around the world by conducting international academic conferences in various
Engineering fields around the world. IAETSD organizes multidisciplinary
conferences for academics and professionals in the fields of Engineering. In order
to strengthen the skill development of the students IAETSD has established.
IAETSD is a meeting place where Engineering students can share their views,
ideas, can improve their technical knowledge, can develop their skills and for
presenting and discussing recent trends in advanced technologies, new educational
environments and innovative technology learning ideas. The intention of IAETSD
is to expand the knowledge beyond the boundaries by joining the hands with
students, researchers, academics and industrialists etc, to explore the technical
knowledge all over the
world, to publish proceedings. IAETSD offers
opportunities to learning professionals for the exploration of problems from many
disciplines of various Engineering fields to discover innovative solutions to
implement innovative ideas. IAETSD aimed to promote upcoming trends in
Engineering.

About ICCIET:
The aim objective of ICDER is to present the latest research and results of
scientists related to all engineering departments topics. This conference provides
opportunities for the different areas delegates to exchange new ideas and
application experiences face to face, to establish business or research relations and
to find global partners for future collaboration. We hope that the conference results
constituted significant contribution to the knowledge in these up to date scientific
field. The organizing committee of conference is pleased to invite prospective
authors to submit their original manuscripts to ICCIET 2014.
All full paper submissions will be peer reviewed and evaluated based on
originality, technical and/or research content/depth, correctness, relevance to
conference, contributions, and readability. The conference will be held every year
to make it an ideal platform for people to share views and experiences in current
trending technologies in the related areas.

Conference Advisory Committee:

Dr. P Paramasivam, NUS, Singapore


Dr. Ganapathy Kumar, Nanometrics, USA
Mr. Vikram Subramanian, Oracle Public cloud
Dr. Michal Wozniak, Wroclaw University of Technology,
Dr. Saqib Saeed, Bahria University,
Mr. Elamurugan Vaiyapuri, tarkaSys, California
Mr. N M Bhaskar, Micron Asia, Singapore
Dr. Mohammed Yeasin, University of Memphis
Dr. Ahmed Zohaa, Brunel university
Kenneth Sundarraj, University of Malaysia
Dr. Heba Ahmed Hassan, Dhofar University,
Dr. Mohammed Atiquzzaman, University of Oklahoma,
Dr. Sattar Aboud, Middle East University,
Dr. S Lakshmi, Oman University

Conference Chairs and Review committee:

Dr. Shanti Swaroop, Professor IIT Madras


Dr. G Bhuvaneshwari, Professor, IIT, Delhi
Dr. Krishna Vasudevan, Professor, IIT Madras
Dr.G.V.Uma, Professor, Anna University
Dr. S Muttan, Professor, Anna University
Dr. R P Kumudini Devi, Professor, Anna University
Dr. M Ramalingam, Director (IRS)
Dr. N K Ambujam, Director (CWR), Anna University
Dr. Bhaskaran, Professor, NIT, Trichy
Dr. Pabitra Mohan Khilar, Associate Prof, NIT, Rourkela
Dr. V Ramalingam, Professor,
Dr.P.Mallikka, Professor, NITTTR, Taramani
Dr. E S M Suresh, Professor, NITTTR, Chennai
Dr. Gomathi Nayagam, Director CWET, Chennai
Prof. S Karthikeyan, VIT, Vellore
Dr. H C Nagaraj, Principal, NIMET, Bengaluru
Dr. K Sivakumar, Associate Director, CTS.
Dr. Tarun Chandroyadulu, Research Associate, NAS

ICCIET - 2014 CONTENTS


1

PLACEMENT OF SUPER CONDUCTING FAULT CURRENT LIMITERS TO


MITIGATE FAULT CURRENTS IN SMART GRIDS WITH DIFFERENT TYPES OF
DISTRIBUTED GENERATION SOURCES

IMPLEMENTATION OF HDLC PROTOCOL USING VERILOG

A SMART ECO-FRIENDLY CAR SYSTEM WITH TRAFFIC CONJECTURE USING


INDUCTIVE LOOPS AND TOUCH LESS MOBILE CONTROL ALONG WITH
ALCOHOL PREVENTION IN AUTOMOBILE

A REVIEW ON MODIFIED ANTI FORENSIC TECHNIQUE FOR REMOVING


DETECTABLE TRACES FORM DIGITAL IMAGES

18

FPGA IMPLEMENTATION OF FAULT TOLERANT EMBEDDED RAM USING


BISR TECHNIQUE

24

ADAPTIVE MODULATION IN MIMO OFDM SYSTEM FOR 4G WIRELESS


NETWORKS

28

VLSI BASED IMPLEMENTATION OF A DIGITAL OSCILLOSCOPE

33

APPEARANCE BASED AMERICAN SIGN LANGUAGE RECOGNITION USING


GESTURE SEGMENTATION AND MODIFIED SIFT ALGORITHM

37

BAACK: BETTER ADAPTIVE ACKNOWLEDGEMENT SYSTEM FOR SECURE


INTRUSION DETECTION SYSTEM IN WIRELESS MANETS

46

10

PERFORMANCE ANALYSIS OF MULTICARRIER DS-CDMA SYSTEM USING


BPSK MODULATION

51

11

ITERATIVE MMSE-PIC DETECTION ALGORITHM FOR MIMO OFDM SYSTEMS

57

12

COMPUTATIONAL PERFORMANCES OF OFDM USING DIFFERENT PRUNED


RADIX FFT ALGORITHMS

62

13

IMPROVEMENT OF DYNAMIC AND STEADY STATE RESPONSES IN


COMBINED MODEL OF LFC AND AVR LOOPS OF ONE-AREA POWER
SYSTEM USING PSS

69

14

ENHANCED CRYPTOGRAPHY ALGORITHM FOR PROVIDING DATA SECURITY

76

15
16

SINTER COOLERS
CHAOS CDSK COMMUNICATION SYSTEM
A REVIEW ON DEVELOPMENT OF SMART GRID TECHNOLOGY IN INDIA AND
ITS FUTURE PERSPECTIVES

81
83

17

88

18

DESIGN AND ANALYSIS OF WATER HAMMER EFFECT IN A NETWORK OF


PIPELINES

95

19

A SECURED BASED INFORMATION SHARING SCHEME VIA SMARTPHONES


IN DTN ROUTINGS

107

20

STORAGE PRIVACY PROTECTION AGAINST DATA LEAKAGE THREADS IN


CLOUD COMPUTING

113

21

REAL TIME EVENT DETECTION AND ALERT SYSTEM USING SENSORS

118

22

AN EFFICIENT SECURE SCHEME FOR MULTI USER IN CLOUD BY USING


CRYPTOGRAPHY TECHNIQUE

123

23

DESIGN AND IMPLEMENTATION OF SECURE CLOUD SYSTEMS USING META


CLOUD

127

24

INCREASING NETWORK LIFE SPAN OF MANET BY USING COOPERATIVE


MAC PROTOCOL

134

25

SCALABLE MOBILE PRESENCE CLOUD WITH COMMUNICATION SECURITY

142

26

ADAPTIVE AND WELL-ORGANIZED MOBILE VIDEO STREAMING PUBLIC


NETWORKS IN CLOUD

146

27

IDENTIFYING AND PREVENTING RESOURCE DEPLETION ATTACK IN MOBILE


SENSOR NETWORK

155

28

EFFECTIVE FAULT TOERANT RESOURCE ALLOCATION WITH COST


REDUCTION FOR CLOUD

166

29

SECURED AND EFFICIENT DATA SCHEDULING OF INTERMEDIATE DATA


SETS IN CLOUD

30

SCALABLE AND SECURE SHARING OF PERSONAL HEALTH RECORDS IN


CLOUD USING MULTI AUTHORITY ATTRIBUTE BASED ENCRYPTION

172

31

LATENT FINGERPRINT RECOGNITION AND MATCHING USING STATISTICAL


TEXTURE ANALYSIS

178

32
33
34

RELY ON ADMINISTRATION WITH MULTIPATH ROUTING FOR INTRUSION


THRESHOLD IN HETEROGENEOUS WSNS
ELECTRIC POWER GENERATION USING PIEZOELECTRIC CRYSTAL
INTEGRATION OF DISTRIBUTED SOLAR POWER GENERATION USING
BATTERY ENERGY STORAGE SYSTEM

186
193
197

35

ESTIMATION OF FREQUENCY FOR A SINGLE LINK-FLEXIBLE MANIPULATOR


USING ADAPTIVE CONTROLLER

206

36

MODELLING OF ONE LINK FLEXIBLE ARM MANIPULATOR USING TWO


STAGE GPI CONTROLLER

219

37

DESIGN OF A ROBUST FUZZY LOGIC CONTROLLER FOR A SINGLE-LINK


FLEXIBLE MANIPULATOR

227

38
39
40

FPGA IMPLEMENTATION OF RF TECHNOLOGY AND BIOMETRIC


AUTHENTICATION BASED ATM SECURITY
DESIGN AND SIMULATION OF HIGH SPEED CMOS FULL ADDER
FPGA IMPLEMENTATION OF VARIOUS SECURITY BASED TOLLGATE SYSTEM
USING ANPR TECHNIQUE

237
245
251

41

MINIMIZATION OF VOLTAGE SAGS AND SWELLS USING DVR

256

42

POWER-QUALITY IMPROVEMENT OF GRID INTERCONNECTED WIND


ENERGY SOURCE AT THE DISTRIBUTION LEVEL

264

43

AN ENHANCEMENT FOR CONTENT SHARING OVER SMARTPHONE-BASED


DELAY TOLERANT NETWORKS

271

44

EFFICIENT RETRIEVAL OF FACE IMAGE FROM LARGE SCALE DATABASE


USING SPARSE CODING AND RERANKING

275

45

ELIMINATING HIDDEN DATA FROM AN IMAGE USING MULTI CARRIERITERATIVE GENERALISED LEAST SQUARES

278

46

IMPLEMENTATION OF CONTEXT FEATURES USING CONTEXT-AWARE


INFORMATION FILTERS IN OSN

282

47

ENHANCEMENT OF FACE RETRIVAL DESIGEND FOR MANAGING HUMAN


ASPECTS

289

48

DESIGN AND IMPLEMENTATION OF SECURE CLOUD SYSTEMS USING META


CLOUD

293

49

ASYNCHRONOUS DATA TRANSACTIONS ON SoC USING FIFO BETWEEN


ADVANCED EXTENSIBLE INTERFACE 4.0 AND ADVANCED PERIPHERAL BUS
4.0

300

50

APPLIANCES OF HARMONIZING MODEL IN CLOUD COMPUTING


ENVIRONMENT

307

51

SIMILARITY SEARCH IN INFORMATION NETWORKS USING META-PATH


BASED BETWEEN OBJECTS

317

52

PINPOINTING PERFORMANCE DEVIATIONS OF SUBSYSTEMS IN


DISTRIBUTED SYSTEMS WITH CLOUD INFRASTRUCTURES

323

53

SECURE DATA STORAGE AGAINST ATTACKS IN CLOUD ENVIRONMENT


USING DEFENCE IN DEPTH

326

54

IMPROVED LOAD BALANCING MODEL BASED ON PARITIONING IN CLOUD


COMPUTING

331

55

EFFECTIVE USER NAVIGABILITY THROUGH WEBSITE STRUCTURE


REORGANIZING USING MATHEMATICAL PROGRAMMING MODEL

334

56

AN EFFECTIVE APPROACH TO ELIMINATE TCP INCAST COLLAPSE IN


DATACENTER ENVIRONMENT

339

57

SECURED AND EFFICIENT DATA SCHEDULING OF INTERMEDIATE DATA


SETS IN CLOUD

347

58

KEY RECONSTRUCTION AND CLUSTERING OPPONENT NODES IN MINIMUM


COST BLOCKING PROBLEMS

352

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Placement of Super Conducting Fault Current Limiters to mitigate


fault Currents in Smart Grids with different types of
Distributed Generation Sources
1

Sai Sowmya Varanasi


Ramachandra college of Engineering,4th B.Tech [email protected]
from all power sources and has no negative effect on the integrated
Sources was suggested.

Abstract

The important application of superconducting fault


current limiters (SFCL) for upcoming smart grid is its possible effect on
the reduction of abnormal fault current and the suitable location in the
micro grids. Due to the grid connection of the micro grids with the
current power grids, excessive fault current is a serious problem to be
solved. In this paper, a resistive type SFCL model was implemented
using Simulink and SimPowerSystem blocks in Matlab. The designed
SFCL model could be easily utilized for determining an impedance level
of SFCL according to the fault-current-limitation requirements of
various kinds of the smart grid systems. In addition, typical smart grid
model including generation, transmission and distribution network with
different types of DGs having different ratings such as wind, solar,
diesel was modeled to determine the location and the performance of
the SFCL. A 10 MVA wind farm, 675 KW solar and 3MVA diesel
generator was considered for the simulation. Three phase fault has been
simulated at distribution grid and the effect of the SFCL and its
location on the wind farm, solar, diesel generator fault currents was
evaluated. The SFCL location in Smart Grid with different types
Distributed Generation Sources have been proposed and their
performances were analyzed.

2. SIMULLINK MODELS SET-UP


Matlab/Simulink/SimPowerSystem was selected to design and
implement the SFCL model. A complete smart grid power network
including generation, transmission, and distribution with an
integrated wind farm, solar, diesel generator model [6] was also
implemented in it. Simulink /SimPowerSystem have number of
advantages over its contemporary simulation software (like EMTP,
PSPICE) due to its open architecture, a powerful graphical user
interface and versatile analysis.
2.1. Power System Model
The modeled power system consists of an electric transmission
and distribution power system. Newly developed micro grid model
was designed by integrating a 10 MVA wind farm, 675 KW solar,
3MVA diesel generator with the distribution network. Fig.1 shows
the power system model designed in Simulink/SimPowerSystem.
The power system is composed of a 100 MVA conventional power
plant, composed of 3-phase synchronous machine, connected with
200 km long 154 kV distributed-parameters transmission line
through a step-up transformer TR1. At the substation (TR2),
voltage is stepped down to 22.9 kV from 154 kV. High power
industrial load (6 MW) and low power domestic loads (1 MW each)
are being supplied by separate distribution branch networks. The 10
MVA wind farm, 675 KW solar, 3MVA diesel generator are
connected directly through Step up transformer to the grid in Fig.1
artificial fault three-phase-to-ground fault and locations of SFCL
are indicated in distribution grid. Three prospective locations for
SFCL installation are marked as Location 1, Location 2 and
Location 3. Generally, conventional fault current protection devices
are located in Location 1 and Location 2. The output current of
wind farm, solar, diesel measured at the output of DG for various
SFCL locations have been measured and analyzed.

Index TermsFault current, micro grid, smart grid, SFCL, wind farm,
solar, diesel.

1.

INTRODUCTION

Conventional protection devices installed for protection of


excessive fault current in electric power systems allows initial two
or three fault current cycles to pass through before getting activated.
But, superconducting fault current limiter (SFCL) is innovative
electric equipment [1] which has the capability to reduce fault
current level within the first cycle of fault current. The first-cycle
suppression of fault current by a SFCL results in an increased
transient stability of the power system carrying higher power with
greater stability.
Smart grid is the novel term used for future power grid which
integrates the modern communication technology and renewable
energy resources to supply electric power which is cleaner, reliable,
resilient and responsive than conventional power system. One of the
important aspects of the smart grid is decentralization of the power
grid network into smaller grids, which are known as micro grids,
having distributed generation sources (DG) connected with them.
These micro grids need to integrate various kinds of DGs and loads
with safety should be satisfied. The direct connection of DGs with
the power grid are the excessive increase in fault current and the
islanding issue which is caused, when despite a fault in the power
grid, DG keeps on providing power to fault-state network . Solving
the problem of increasing fault current in micro grids by using
SFCL technology is the main topic of this work. In this paper, the
effect of SFCL and its position was investigated considering a wind
farm, solar [5], diesel generator [6] integrated with a distribution
grid model as one of typical configurations of the smart grid. The
impacts of SFCL on the wind farm, solar, diesel generator and the
strategic location of SFCL in a micro grid which limits fault current

2.2. Resistive SFCL Model


The three phase resistive type SFCL was modeled and
considering four fundamental parameters of a resistive SFCL.
These parameters and their selected values are 1) Transition or
response time 2) minimum impedance and maximum impedance
3) Triggering current and 4) recovery time .Its working voltage is
22.9 kV. Fig. 2 shows the SFCL model developed in
Simulink/Simpower System. The SFCL model works as follows.
First, SFCL model calculates the RMS value of the passing current
and then compares it with the Lookup table. Second, if a passing
Current is larger than the triggering current level, SFCLs resistance
increases to maximum impedance level in a pre-defined response
time. Finally, when the current level falls below the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig.1 Power system model designed in Simulink/SimPowerSystem,Fault and SFCL locations are indicated in the diagram

triggering current level the system waits until the recovery time and
then goes into normal state. SFCL is located at different locations.

2.4. Solar model


The terminal equations of the Voltage, Current and Power [5]
used for the devolpement of Solar Simulink model are given by
equations 1,2 and 3.

Fig.2 Single phase SFCL model developed in Simulink/SimPowerSystem.

2.3. Wind Turbine Generator Model


Where Iph is the light generated current, Is is the cell saturation of
dark current, q is the unit electric charge, k is the Boltzmans
constant t is the working temperature of p-n junction, A is the ideal
factor. Iph is the current based on cells working temperature and
solar irradiation level described in equation.2 . Isc is the cells short
circuit current Kt is the temperature coefficient and S is the solar
irradiation level in KW/m2. Power delivered by PV array is given
by equation(3) and PV array is modeled as single diode circuit . The
subsystem can be implemented by using equations 1, 2 and 3.

A 10 MVA wind farm[1] is composed of 5 fixed-speed inductiontype wind turbines,each having a rating of 2MVA at the time of
fault the domestic load is being provided with 3MVA out of which
2.7 MVA is provided by the wind farm. The wind farm is directly
connected with the branch network B6 through the transformer
TR3.Simulink model of wind turbine generator is shown.

Fig.3 Three phase wind turbine generator subsystem model.

Fig.4 Solar model developed in Simulink/SimPowerSystem.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The output power of pv array is 675KW which is given to the


inverter and then the ouput of model is connected directly to the
branchnetwork B6 through transformer T3. The PV simulink model
is shown in Fig.4.

ISBN: 378 - 26 - 138420 - 5

800

600

Current (Amp)

400

2.5. Diesel Generator Model

200

-200

A lower order model [6] is used for studying dynamics of internal


combustion engines. A self tuning PID controller is developed for
a small size diesel engine and generator set.The output power of
diesel generator is 3.125MVA.The ouput of model is connected
directly to the branchnetwork B6 through transformer T3.The
Simulink model of Governor and diesel engine is shown in Fig.5.

-400

-600
0.25

0.3

0.35

0.4

0.45

0.5

0.45

0.5

Time(Sec)

A. Fault current without SFCL


1200
1000
800

Current(Amp)

600
400
200
0
-200
-400
-600
0.25

Fig.5 Governor and diesel engine Simulink model developed

0.3

0.35

0.4
Time(Sec)

B.Fault current with SFCL at Location 1

The Simulink model of Voltage control and Speed control of diesel


generator shown in Fig.6.
1200
1000
800

Current(Amp)

600
400
200
0
-200
-400
-600
0.25

0.3

0.35

0.4

0.45

0.5

Time(Sec)

C. Fault current with SFCL at Location 2

Fig.6 Voltage and Speed control Simulink model developed


600

3. RESULTS AND CONCLUSIONS

400

Three scenarios of SFCLs possible locations were analyzed for


distribution grid fault occurring in the power system. The
simulation is carried by placing SFCL in Location 1,2 and 3 by
considering one DG at a time at the position shown in fig.1.

Current(Amp)

200

-200

-400

3.1. Wind Turbine Generator as DG


-600
0.25

0.3

0.35

0.4

0.45

0.5

Time(Sec)

Fault currents of one phase where the severity is more and


different locations of SFCL are shown in Fig.7. When a threephase-to-ground fault was initiated in the distribution grid (Fig.1)
SFCL located at Location 1 or Location 2, fault current contribution
from the wind farm was increased . These critical observations
imply that the installation of SFCL in Location 1 and Location 2
instead of reducing, has increased the DG fault current. This sudden
increase of fault current from the wind farm is caused by the abrupt
change of power systems impedance. The SFCL at these locations
(Location 1 or Location 2) entered into current limiting mode and
reduced fault current coming from the conventional power plant
due to rapid increase in its resistance. Therefore, wind farm which
is the other power source and also closer to the Fault is now forced
to supply larger fault current to fault point. In the case when SFCL
is installed at the integration point of wind farm with the grid,
marked as Location 3 in Fig.1 wind farm fault current has been
successfully reduced. SFCL gives 68% reduction of fault current
from wind farm and also reduce the fault current coming from
conventional power plant because SFCL is located in the direct
path of any fault current flowing towards Fault 1.

D.Fault current with SFCL at Location 3

Fig.7 Fault Currents of Wind Turbine Generator.

Comparison of fault currents when SFCL is placed at different


Locations in Fig.1 are shown in Fig.8

Fig.8.Comparison of fault currents at different locations

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

3.2. Solar model as DG


Fault currents of one phase where the severity is more and
different locations of SFCL are shown in Fig.9.When a three-phaseto-ground fault was initiated in the distribution grid (in Fig. 1).The
fault without SFCL is greater when compared with SFCL placed at
different Locations the following results were observed. Once again
the best results are obtained when a single SFCL is located at
Location 3, when compared with the remaining two locations. Fault
current has been successfully reduced.SFCL gives 70% reduction of
fault current from Solar module and also reduce the fault current
coming from conventional power plant because SFCL is located in
the direct path of any fault current flowing towards Fault 1.

Fig.10.Comparison of fault currents at different locations

3.3. Diesel Generator as DG

3000

2500

Fault currents of one phase where the severity is more and


different locations of SFCL are shown in Fig.11. In this case when a
three-phase-to-ground fault was initiated in the distribution grid (in
Fig. 1).The fault without SFCL is greater when compared with
SFCL placed at different Locations the following results were
observed. Once again the best results are obtained when a single
SFCL is located at Location 3, when compared with the remaining
two locations. Diesel fault current has been successfully reduced.
SFCL gives 84% reduction of fault current from diesel engine and
also reduce the fault current coming from conventional power plant
because SFCL is located in the direct path of any fault current
flowing towards Fault 1.

2000

C
urrent(am
p)

1500

1000

500

-500

-1000
0.25

0.3

0.35

0.4

0.45

0.5

Time(s ec)

A. Fault current without SFCL


1500

1000

1000

Current(amp)

500

800

-500
C
urrent(Am
p)

600

-1000

-1500
0.25

0.3

0.35

0.4

0.45

0.5

400

200

Time(Sec)

B. Fault current with SFCL at Location 1

1500
-200
0.25

0.3

0.35

0.4

0. 45

0.5

0.45

0.5

Time(Sec)

1000

Current(amp)

A. Fault current without SFCL


500

1000

800

0
600

-500

0.3

0.35

0.4

0.45

Current(Amp)

-1000
0.25

400

0.5

Time(Sec)

C.Fault current with SFCL at Location 2

200

-200

-400

1000
-600

800
-800
0.25

0.3

0.35

600
400
Current(amp)

0.4
Time(Sec)

B. Fault current with SFCL at Location 1

200
1000

0
800

-200
600

-400
400

-800
0.25

0.3

0.35

0.4

0.45

C
urrent(Am
p)

-600

0.5

200

Time(Sec)
-200

D. Fault current with SFCL at Location 3


-400

-600

Fig.9 Fault Currents of Solar model.

-800
0.25

0.3

0.35

0.4

0. 45

0.5

Time(Sec)

Comparison of fault currents when SFCL is placed at different


Locations in Fig.1 are shown in Fig.10.

C.Fault current with SFCL at location 2

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

and diesel. Multiple SFCLs will not used in micro grid because of
their inefficiency both in performance and cost. The strategic
location of SFCL in a power grid which limits all fault currents and
has no negative effect on the DG source is the point of integration
of the DG sources with the power grid.

500

400

300

200

Current(Amp)

ISBN: 378 - 26 - 138420 - 5

100

REFERENCES

-100

-200

[1] Umer A.Khan, J.K.seong, S.H.Lee Feasibility analysis of the


Positioning of Super Conducting Fault Current Limiters for the
Smart Grid Application using Simulink & Simpower Systems
IEEE Transactions on applied superconductivity, vol.21, No.3
June-2011.
[2] S.Sugimoto, J.Kida, H.Aritha Principle and Characteristics of
Fault Current limiters with series compensation IEEE Trans.
on Power Delivery, vol.11,No.2,pp 842-847,Apr 1996.
[3] J.Driensen, P.Vermeyen and R.Belmans Protection issues in
micro grids with multiple distributed generation units in Power conversion conference., Nagoya, April 2007, pp 646-653.
[4] K.Maki ,S.Repo and P.Jarventausta Effect of wind power baSed distributed generation on protection of distribution netwrk
Work in IEEE Developments in Power systems protection,
Dec.2004, vol.1, pp 327-330.
[5] N.Pandiarajan, Ranganath muth Modeling of Photo Voltaic
Module with Simulink international conference on electrical
Energy system (ICEES-2011), 3-5 Jan-2011.
[6] Takyin Takychan Transient analysis of integrated solar/diesel
hybrid power System using MATLAB/Simulink .
[7]
S.N.B.V.Rao Asst Prof. in Ramachandra college of
engineering

-300

-400
0.25

0.3

0.35

0.4

0.45

0.5

Time(Sec)

D.Fault current with SFCL at Location 3

Fig.11 Fault Currents of Diesel Engine.

Comparison of fault currents when SFCL is placed at different


Locations in Fig.1

Fig.12.Comparison of fault currents at different locations

When the SFCL was strategically located at the point of integration


Location 3, the highest fault current reduction was achieved
whatever may be the DG source. The multiple SFCLs in a micro
grid are not only costly but also less efficient than strategically
located single SFCL. Moreover, at Location 3, fault current coming
from the conventional power plant was also successfully limited.
The results of fault currents at three locations with different DG
units are summarized in Table I for comparison.
Table I: Percentage change in fault currents of different DG units
due to SFCL locations
3-Ph Distribution grid fault
DG
type

No
SFCL

Location 1

Location 2

Location 3

Wind

787A

1020A
30% increased

1020A
30% increased

265A
67% decreased

Solar

2675A

1100A
59% decreased

1100A
59% decreased

800A
70% decreased

Diesel

916A

798A
13% decreased

798A
13% decreased

148A
84% decreased

4. CONCLUSION
This paper presented a feasibility analysis of positioning of the
SFCL in rapidly changing modern power grid. A complete power
system along with a micro grid (having a wind farm, solar, diesel
connected with the grid) was modeled and transient analysis for
three-phase-to-ground faults at different locations of the grid were
performed with SFCL installed at different locations of the grid.
This placement of SFCL results in abnormal fault current
contribution from the wind farm and other two sources such as solar

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Implementation of HDLC Protocol Using Verilog


K. Durga Bhavani
Dept. of ECE
RGUKT-Basar
[email protected]

B. Venkanna
Dept. of ECE
RGUKT-Basar
[email protected]

Abstract A protocol is required to transmit data successfully


over any network and also to manage the flow at which data is
transmitted. HDLC protocol is the high-level data link control
protocol established by International Organization for
Standardization (ISO), which is widely used in digital
communications. High-level Data Link Control (HDLC) is the most
commonly used Layer2 protocol and is suitable for bit oriented
packet transmission mode. This paper discusses the Verilog
modeling of single-channel HDLC Layer 2 protocol and its
implementation using Xilinx.

II.

HDLC PROTOCOL

The HDLC Protocol Controller is a high-performance


module for the bit-oriented packet transmission mode. It is
suitable for Frame Relay, X.25, ISDN B-Channel (64 Kbits/s)
and D-Channel (16 Kbits/s) The Data Interface is 8-bit wide,
synchronous and suitable for interfacing to transmit and
receive FIFOs. Information is packaged into an envelope,
called a FRAME [4]. An HDLC frame is structured as follows:

Keywords- High Level Data link Control (HDLC), Frame Check


Sequence (FCS), and Cyclic Redundancy Check (CRC)

I.

K. Gayathri
Dept. of ECE
Intell Engg. College-Anatapur
[email protected]

FLAG

ADDRESS

CONTROL

INFORMATION

FCS

FLAG

8 bits

8 bits

8 /16 bits

variable

8 bits

INTRODUCTION

HDLC protocol is the high-level data link control protocol


established
by
International
Organization
for
standardization (ISO), which is widely used in digital
communication and are the bases of many other data link
control protocols [2]. HDLC protocols are commonly
performed by ASIC (Application Specific Integrated Circuit)
devices, software programming and etc.
The objective of this paper is to design and implement a
single channel controller for the HDLC protocol which is the
most basic and prevalent Data Link layer synchronous, bitoriented protocol. The HDLC protocol (High Level Data link
Control) is also important in that it forms the basis for
many other Data Link Control protocols, which use the
same or similar formats, and the same mechanisms as
employed in HDLC.
HDLC has been so widely implemented because it
supports both half duplex and full duplex communication
lines, point to point(peer to peer) and multi-point
networks[1]. The protocols outlined in HDLC are designed
to permit synchronous, code-transparent data transmission.
Other benefits of HDLC are that the control information is
always in the same position, and specific bit patterns used
for control differ dramatically from those in representing
data, which reduces the chance of errors.

Table 1. HDLC Frame


A. Flag
Each Frame begins and ends with the Flag Sequence which
is a binary sequence 01111110. If a piece of data within the
frame to be transmitted contains a series of 5 or more 1s,
the transmitting station must insert a 0 to distinguish this
set of 1s in the data from the flags at the beginning and
end of the frame. This technique of inserting bits is called
bit-stuffing [3].
B. Address
Address field is of programmable size, a single octet or a
pair of octets. The field can contain the value programmed
into the transmit address register at the time the Frame is
started.
C.

Control

HDLC uses the control field to determine how to control


the communications process. This field contains the
commands, responses and sequences numbers used to
maintain the data flow accountability of the link, defines
the functions of the frame and initiates the logic to control
the movement of traffic between sending and receiving
stations.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

D. Information or Data
address detect unit, FCS calculator unit, zero detection unit,
This field is not always present in a HDLC frame. It is only flag detection unit, receive control and status register unit
present when the Information Transfer Format is being and frame timer and synchronization logic unit.
used in the control field. The information field contains the
A. Transmitter Module
actually data the sender is transmitting to the receiver.
The Transmit Data Interface provides a byte-wide
E. FCS
interface between the transmission host and the HDLC
Controller. The Transmit data is loaded into the controller
The Frame Check Sequence field is 16 bits. The FCS is
on the rising edge of Clock when the write strobe input is
transmitted least significant octet first which contains the
asserted. The Start and End bytes of a transmitted HDLC
coefficient of the highest term in the generated check
polynomials. The FCS field is calculated over all bits of the
Frame are indicated by asserting the appropriate signals
addresses, control, and data fields, not including any bits
with the same timing as the data byte.
inserted for the transparency. This also does not include
The HDLC Controller will, on receipt of the first byte of a
the flag sequence or the FCS field itself. The end of the data
new packet, issue the appropriate Flag Sequence and
field is found by locating the closing flag sequence and
transmit the Frame data calculating the FCS. When the last
removing the Frame Check Sequence field (receiver section)
byte of the Frame is seen the FCS is transmitted along with
[5].
a closing Flag. Extra zeros are inserted into the bit stream
to avoid transmission of control flag sequence within the
III. HDLC MODULE DESIGN
Frame data.
In this design, HDLC procedures contain two modules, i.e.
The Transmit Data is available on TxD pin with
encoding-and-sending module (Transmitter) and receivingappropriate to be sampled by Clk. If TxEN is de-asserted,
and-decoding module (receiver). The function diagram is
transmit is stalled, and TxD pin is disabled.
shown as below.
A transmit control register is provided which can enable
or disable the channel. In addition it is possible to force the
transmission of the HDLC Abort sequence. This will cause
the currently transmitted Frame to be discarded. The
transmit section can be configured to automatically restart
after an abort, with the next frame, or to remain stalled
until the host microprocessor clears the abort.
B. Receiver Module
The HDLC Controller Receiver accepts a bit stream on
port RxD. The data is latched on the rising edge of Clock
under the control of the Enable input RxEN. The Flag
Detection block searches the bit stream for the Flag
Sequence in order to determine the Frame boundary. Any
stuffed zeros are detected and remove and the FCS is
calculated and checked. Frame data is placed on the
Receive Data Interface and made available to the host. In
addition, Flag information is passed over indicating the
Start and the End byte of the HDLC Frame as well as
showing any error conditions which may have been
detected during receipt of the Frame.
In normal HDLC protocol mode, all Receiver Frames are
presented to the host on the output register. A status
register is provided which can be used to monitor status of
the Receiver Channel, and indicate if the packet currently
being received includes any errors.

Fig.1. HDLC Block Design


Form this diagram we know that, transmitter module includes
transmit register unit, address unit, FCS generation unit, zero
insertion unit, Flag generation unit, control and status
register unit and transmit frame timer and synchronization
logic unit. Receiver module includes receive register unit,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

International Organization for Standardization, pp 10-17, July


2002.
[4] A.Tannenbaum, Computer Networks, Prentice Hall of India,
1993.
[5] Mitel Semiconductor, MT8952B HDLC Protocol Controller,

IV. RESULTS

Mitel Semiconductor Inc., pp 2-14, May 1997.

Fig.2. Simulation Waveform


Device Utilization Report: Clock Frequency: 78.2 MHz
Resource
Used
Avail
Utilization
IOs
60
180
33.33%
Function Generators
205
1536
13.35%
CLB Slices
103
768
13.41%
Dffs or Latches
108
1536
7.03%
Table 2. Synthesis Report

V. CONCLUSION
We designed HDLC protocol sending and receiving RTL
level modules in Verilog and had them tested successfully,
which has the following advantages like easy to program
and modify, suitable for different standards of HDLC
procedures, match with other chips with different interfaces.
So this proposed method can be more useful for many
applications like a Communication protocol link for
RADAR data processing.

REFERENCES
[1] Implementation of HDLC protocol Using FPGA, [IJESAT]
International Journal of Engineering Science & Advanced
Technology, ISSN: 2250-3676, Volume-2, Issue-4, 1122
1131.
[2] M.Sridevi, DrP.Sudhakar Reddy / International Journal of
Engineering Research and Applications (IJERA) ISSN: 22489622 www.ijera.com Vol. 2, Issue 5, September- October
2012, pp.2217-2219.
[3] ISO/IEC
13239,
Information
technology
-Telecommunications and Information exchange between
systems High-level data link control (HDLC) procedures,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

AUTOMOBILE
INNOVATION

stuck in the traffic jams mainly due to no

(A Smart Eco-Friendly Car


system with Traffic Conjecture
using inductive loops and touch
less mobile control along with
Alcohol Prevention in
Automobile)

driver/commuter about the level of traffic at

proper estimation on the traffic on their way.


So, we thought that we can eliminate this
problem by intimating or instructing the

any junction all over the city. Such that the


commuter can escape from the traffic jam
and he/she can reach the destination in time.
So to intimate the density of traffic at any
junction, we are going to use Induction
loops to calculate the density of traffic at
the junctions and to display them to the

Presented by:

commuter. In addition to this we are also


using Google maps to forecast the traffic

G.RANA PHANI SIVA SUJITH

through GPS to the driver at any place


Electronics and communications engineering

regarding the traffic at the junctions. And

Sri Vasavi Institute of Engineering and technology

also we are using these induction loops to

Ph no: 9030633480

reduce the noise pollution at the traffic

Email Id: [email protected]

intersections. We also quoted about one of


the smart car experience to the commuter

Abstract:

ie., a touch less mobile controlling system


The traffic jams in cities are most common

that enables the driver to operate his/her

and remained as a uncontrollable problem.

mobile with out even touching it. This touch

This idea we developed is the result of our

less mobile system is more useful to

observation in one of the major traffic

eliminate accidents that occur due to usage

intersection at Vijayawada (Benz circle). In

of mobile phones while driving.when a

this paper we are going to present you the

driver or any passengers of car consumes

best solution to eradicate heavy traffic jams

alcohol then the car automatically stops in a

without using any constructions like bridges

safe place side of a road using traffic

or additional lanes on the roads. We

conjecture & the front wheels of the car will

observed that the commuters are getting

be locked out.Through GPS technology a

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

complaint will be sent to the nearest police

1. Placing the induction loops at the

station.If driver talks in a mobile for more

junctions beneath the road.

than 8 secs then the system of the car

2. Setting up a detector to detect the sata

activates the jammers to cut the call signals

given by each and every loop placed on the

& It stops the car.We are also Placing a

road.

piezo-electric devices on roads & zebra


energy

3. Setting up a control system that enables

through the mechanical stress applied by the

us to display the density and give the

pedestrians & vehicles while going on the

instructions to the driver.

crossing

will

generate

electric

road. We are going to use this electric

4. Further using the data given by the

energy as the power supply to the Induction

detectors to indicate the Google map with

Loops,Traffic Signals & Traffic Street


Lightening

System.We

are

using

traffic density.

this

technique with less amount of investment


and more Accurate Output..

Flow chart:

Approach:

Installation of
inductive loops

The traffic can be controlled by using


various advanced technologies such as
digital signal processing and various other.
But here we are using a simple and cost
effective tool to estimate the density of

Detecting the data given by


the loops

traffic. And after getting the data on density


we are going to control the traffic depending
up on the density value.
Steps involved for traffic conjecture:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

10

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

metal objects in metal detectors or vehicle


presence indicators. A common modern use

Estimating the traffic and


calculation of time required
to reach the required
destination

for induction loops is to provide hearing


assistance to hearing-aid users. Vehicle
detection loops, called inductive-loop traffic
detectors, can detect vehicles passing or
arriving at a certain point, for instance
approaching a traffic light or in motorway
traffic. An insulated, electrically conducting
loop is installed in the pavement. The
electronics unit transmits energy into the
wire loops at frequencies between 10 kHz to

Displaying the data and


instructions

200 kHz, depending on the model. The


inductive-loop system behaves as a tuned
electrical circuit in which the loop wire and
lead-in cable are the inductive elements.
When a vehicle passes over the loop or is
stopped within the

loop, the vehicle

The main element that is needed to

induces eddy currents in the wire loops,

implement this idea is the inductive loop.

which

And also proper controlling system to attain

decreased inductance actuates

much accuracy.

electronics unit output relay or solid-state

decrease

their

inductance.

The
the

optically isolated output, which sends a

Inductive loops:

pulse

An induction

system

which

the

traffic

signal

controller

loop is

signifying the passage or presence of a

or

vehicle. Parking structures for automobiles

may use inductive loops to track traffic

an electromagnetic communication
detection

to

uses

moving magnet to induce an electrical

(occupancy) in and out or may be used by

current in a nearby wire. Induction loops are

access gates or ticketing systems to detect

used for transmission and reception of

vehicles while others use Parking guidance

communication signals, or for detection of

and information systems. Railways may use

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

11

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

an induction loop to detect the passage of


trains

past

given

point,

as

an

An increasingly common application is for

electronic treadle.

providing hearing aid-compatible "assistive

The relatively crude nature of the loop's

listening" telecoil. In this application a loop

structure means that only metal masses

or series of loops is used to provide an audio

above a certain size are capable of triggering

frequency oscillating magnetic field in an

the relay. This is good in that the loop does

area where a hearing aid user may be

not thus produce very many "false positive"

present. Many hearing aids contain a telecoil

triggers (say, for example, by a pedestrian

which allows the user to receive and hear the

crossing the loop with a pocket full of loose

magnetic field and remove the normal audio

metal change) but it sometimes also means

signal provided from the hearing aid

that bicycles, scooters, and motorcycles

microphone site. These loops are often

stopped at such intersections may never be

referred to as a hearing loop or audio

detected by them (and therefore risk being

induction loop.

ignored by the switch/signal). Most loops

An anti-submarine

can be manually adjusted to consistently

device used to detect submarines and surface

detect

vessels using specially designed submerged

the

presence

of

scooters

and

motorcycles at the least. A different sort of


"induction

loop"

is

applied

indicator loop was a

cables connected to a galvanometer.

to metal

Application of this idea to control

detectors, where a large coil, which forms

traffic:

part of a resonant circuit, is effectively


"detuned" by the coil's proximity to a
conductive object. The detected object may
be metallic (metal and cable detection) or
conductive/capacitive

(stud/cavity

detection). Other configurations of this


equipment use two or more receiving coils,
and the detected object

modifies the

inductive coupling or alters the phase angle


of the voltage induced in the receiving coils
relative to the oscillator coil.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

12

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Our Intelligent Traffic Light System is


capable of changing priority level of the
roads according to their traffic level. To
measure the traffic level we have several
mechanisms.

Image Processing

Pressure sensors, give reading when

pressure changes by the vehicles

Inductive Loops

From the above methods we choose


inductive loop.It is build on the concept of
inductance change of a coil when a metal
object come closer.You know that when you
send electrical current through a wire, it
generates a magnetic field. For a coil this
electromagnetic

The above is the case study that we had

field

is high.You can

change the inductance of the coil and change

conducted. The red colored lane indicates

the electromagnetic flux by introducing

the heavy traffic area. And the green lane

additional conductive materials into the

indicates the free path without any traffic.

loop's magnetic field. This is what happens

The vehicle and the destination are indicated

when a car pulls up to the intersection. The

above. At the first junction before the main

huge mass of metal that makes up your car

junction, we are going to place a board that

alters the magnetic field around the loop,

indicates the driver about the status of the

changing its inductance.

traffic and also instruct him to take diversion


or not. This makes the commuter to reach
his destination as fast as possible. To
achieve this we need to use the inductive
loop detectors and to be placed beneath the
road to detect the vehicle density.

Working:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

13

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

from the same point. The two ends of the


loop wire are connected to the loop
extension cable, which in turn connects to
the vehicle detector. The detector powers the
loop causing a magnetic field in the loop
area. The loop resonates at a constant
frequency that the detector monitors. A base
frequency is established when there is no
vehicle over the loop. When a large metal
object, such as a vehicle, moves over the
loop, the resonate frequency increases. This
increase

in

frequency is sensed and,

depending on the design of the detector,


So we have made a coil develop the

forces a normally open relay to close. The

inductive loop.We have some kind of

relay will remain closed until the vehicle

metering device to meter the voltage level

leaves the loop and the frequency returns to

change in the coil.

the base level. The relay can trigger any


number of devices such as an audio

An inductive loop vehicle detector system

intercom system, a gate, a traffic light, etc.

consists of three components: a loop

In general, a compact car will cause a

(preformed or saw-cut), loop extension cable

greater increase in frequency than a full size

and a detector. When installing or repairing

car or truck. This occurs because the metal

an inductive loop system the smallest detail

surfaces on the under carriage of the vehicle

can mean the difference between reliable

are closer to the loop. Figures 3 and 4

detection and an intermittent detection of

illustrate how the under carriage of a sports

vehicles. Therefore, attention to detail when

car is well within the magnetic field of the

installing or troubleshooting an inductive

loop compared to the sports utility vehicle.

loop vehicle detection system is absolutely

Notice that the frequency change is greater

critical. The preformed or saw-cut loop is

with the smaller vehicle.

buried in the traffic lane. The loop is a


continuous run of wire that enters and exits

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

14

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

frequency. For example, a one square foot


piece of sheet metal positioned in the same
plane of the loop has the same affect as a
hunk of metal one foot square and one foot
thick. Another way to illustrate the point is
to take the same one square foot piece of
sheet metal, which is easily detected when
held in the same plane as the loop, and turn
it perpendicular to the loop and it becomes
impossible to detect. Keep this principle in
mind when dealing with inductive loop
detectors.

Some detectors provide PC diagnostics via a


communication

port

on

the

detector.

Diagnostic software gives you a visual


picture of what is happening at the loop, and

Also, it is interesting to note that the

will help you troubleshoot any problems you

frequency change is very consistent between

may experience during installation or in the

two vehicles of the same make and model,

future. Detectors with this feature are

so much so that a detector can almost be

usually in the same price range as other

designed to determine the type of vehicle

detectors and can help you save time solving

over the loop. There is a misconception that

a detection problem. The PC software and

inductive loop vehicle detection is based on

cable is usually additional, however keep in

metal mass. This is simply not true.

mind that if you have multiple installations

Detection is based on metal surface area,

you need only buy the software and cable

otherwise known as skin effect. The greater

setup once. Diagnostics software can also

the surface area of metal in the same plane


as the loop, the greater the increase in

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

15

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

help determine the depth and position of the

ISBN: 378 - 26 - 138420 - 5

Alcohol prevention in Automobiles:

loop in the pavement.


we

are

seeing

daily

in

newspaper,newschannels etc., that many


people are dieing in accidents due to
Alcohol consumption or negligence of the
driver while driving. My solution to this
problem

is

Alcohol

Prevention

in

Automobiles.In this system firstly,we will


fix Alcoholic Sensors or Air Bleezers in
stearing of vehicle,near dashboard,back part
of front seats(rear & back),so we can easily
detect alcohol consumption of driver and
passengers in the cars,trucks,lorrys,buses
etc,. In Bikes,we will place sensors near the
speedometer to detect the alcohol. After

Mapping the traffic density and enabling

detection,the

GPS:

system

will

give

three

warnings with voice response saying you


By using the data given by the induction

are alcohol consumed please stop the

loop detector we are going to enable the

vehicle. If driver does not listen then

Google maps to give the updates of high

car(any vehicle) stops in a safe place on

traffic prone areas through GPS.

other side of the road by observing Traffic


conjecture on the screen in his/her car,or if
we have Wi-Fi in our mobile,then by
clicking on Traffic Conjecture App on your
Mobile u can see the Traffic Ahead of you.
After car has stopped, the Front Wheels of
vehicle will be locked, Car Engine will be
OFF and the position of the vehicle(through
Google map) will be given as a Compliant
to the nearest Police Station through wide

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

16

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

area

differential

global

positioning

LAUNCH

camera in front of driver seat,so if driver

and

then

stops

the

ERA

OF

Attachments:

falls asleep or became less sensitive then it


first

NEW

AUTOMOBILE..

system(WADGPS).we will also place a

warns

OF

ISBN: 378 - 26 - 138420 - 5

car

Videos:

Automatically after warning.


1.Induction Loops Practical Functioning
Video on Roads Practically.

Eco-Friendly Autonomous Car:

2.Future Alcoholic Drunk Drive System


Prevention Video Practically.

We are planning to make our autonomous


more Eco-Friendly one.For his,we are going

3.Power Point Presentation of Automobile

to

Innovation.

place

Piezo-Electric

Devices

on

Roads,Zebra Crossing,Footpath etc.,thereby

4.Autonomous Car Project Abstract.

we are going to generate Electric energy


from the mechanical stress applied by the

5.Power Point Presentation of Autonomous

pedestrians,vehicles while going on the

Car.

Roads.we are going this developed electric

References:

energy as a power supply to the Induction


Loops,GPS

system,Traffic

System,Street

Lightening

Signalling

1. http://www.fhwa.dot.gov/publication
s/publiroads/98septoct/loop.cfm

System,Traffic

Conjecture Display Screens etc,.Thus we are

2. http://auto.howstuffworks.com/cardriving-safety/safety-regulatorydevices/red-light-camera1.htm

renewable energy.

Conclusion:

3. http://www3.telus.net/chemelec/Proj
ects/Loop-Detector/LoopDetector.htm

At last, My Aim is to Develop an EcoFriendly Autonomous Car with Traffic

4. www.wikipedia.org

Conjecture using Induction Loops and


Alcoholic Prevention System. Being an
Electronics I did my maximum what I can
do for Automobile Field. Hence it is a

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

17

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

A REVIEW ON MODIFIED ANTI FORENSIC


TECHNIQUE FOR REMOVING DETECTABLE
TRACES FORM DIGITAL IMAGES
1

M.GOWTHAM RAJU.

N.PUSHPALATHA,

M.Tech (DECS) Student, Assistant professor


Department of ECE, AITS
Department of ECE, AITS
Annamacharya Institute of Technology and Sciences, Tirupati, India-517520
1

[email protected]
[email protected]

cameras produce instant images which can be viewed


without delay of waiting for film processing. it does not
require external development they can be store easily. And
there should not be taken any time delay. Images can be
processed in different ways. They are processed as jpeg
images, in some other cases they are processed in bit mat
format. When they are used in bitmap format it does need
to use without any information of past processing. to know
about the past processing information it is desirable to
know the artifacts of image. These techniques are capable
of finding the earlier processing information. Therefore
forensic researches need to examine the authenticity of
images to find how much the trust can be put up on the
techniques and this can also be used to find out the
drawback of this techniques. Person with good knowledge
in image processing can do undetectable manipulation. it is
also desirable to find the draw backs of these techniques.
For this purpose research has to develop both forensic and
anti-forensic techniques to understand the weaknesses.
Consider the situation that already tried to remove the
artifacts of compression. The forensic experts can easily
find out the existing techniques such as quantized
estimation. It is useful when image processing unit receives
compression details and quantization table used for
processing and compression. Some of the existing
techniques like detection of blocking signature estimation
of quantization table this allow the mismatches and
forgeries in jpeg blocks by finding the evidence of
compression. To solve this problem of image forensic the
research has to develop tools that are capable of fooling the
existing methodologies. Even though the existing methods
have advantages some limitations too. The main drawback
of these techniquesis that they do not report for the risk that
new technique may be design and used to conceal the
traces ofmanipulations. As mention earlier it may possible
for an image forger to generate undetectable compression
and other image forgeries. This modified anti-forensic
technique approach is presented which is capable of hiding
the traces of earlier processing including both compression
and filtering. This concept is that adding specially designed
noise to the images blocks will help to hide the proof of
tampering.

Abstract: The increasing attractiveness and trust on


digital photography has given rise to new acceptability
issues in the field of image forensics. There are many
advantages to using digital images. Digital cameras
produce immediate images, allowing the photographer
to outlook the images and immediately decide whether
the photographs are sufficient without the
postponement of waiting for the film and prints to be
processed. It does not require external developing or
reproduction. Furthermore, digital images are easily
stored. No conventional "original image" is prepared
here like traditional camera. Therefore when forensic
researchers analyze the images they dont have access
to the original image to compare. Fraud through
conventional photograph is relatively difficult,
requiring technical expertise. Whereas significant
features of digital photography is the ease and the
decreased cost in altering the image. Manipulation of
digital images is simpler. With some fundamental
software, digitally-recorded image can easily be edited.
The most of the alterations include borrowing, cloning,
removal and switching parts of a digital image. A
number of techniques are available to verify the
authenticity of images. But the fact is that number of
image tampering is also increasing. The forensic
researchers need to find new techniques to detect the
tampering. For this purpose they have to find the new
anti-forensic techniques and solutions for them. In this
paper a new anti-forensic technique is considered,
which is capable of removing the evidences of
compression and filtering. It is done by adding a
specially designed noise called tailored noise to the
image after processing. This method can be used to
cover the history of processing in addition to that it can
be also used to remove the signature traces of filtering.
Keywords: Digital forensic, jpeg compression, image
coefficients, image history, filtering, Quantization, DCT
coefficients.

Introduction
Digital images become very popular for transferring visual
information. And there are many advantages using these
images instead of traditional camera film. The digital

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

18

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

which we refer to as anti-forensic dither, to it value


according
totheequation
Z=Y+D
The segment length is equal to the length of the
quantization interval the probability that the quantized
coefficient value is qk is given by.

1. RELATED TO PROJECT WORK:


1.1. ANTI FORENSIC OF DIGITAL IMAGE
COMPRESSION:
As society has become increasingly reliant upon digital
image to communicate visual information, a number of
forensic techniques have developed. Among the most
successful of these are techniques that make use of an
images compression history and its associate compression
finger prints. Anti-forensic techniques capable of fooling
forensicAlgorithms this paper represents set of antiforensic techniques designed to remove forensically
significant indicators of compression of an image. in this
technique first distributes the image transform coefficients
before compression then adding anti-forensic transform
coefficients of compressed image so that distribution
matches estimation one. When we use these frame work of
anti-forensic techniques specially targeted at erased finger
prints left by both JPEG and wavelet based coders.
1.1.1. ANTI-FORENSIC FRAMEWORK:
All image compression techniques are subbing band
coders, which are themselves a subset of transform coders.
Transform coders are mathematically applying to the
signals of compressing the transforms coefficients. Sub
band coders are transform coders that decompose the signal
in to different frequency bands. By applying two
dimensional invertible transform, such as DCT to as image
as a whole that has been segmented into a series of disjoint
sets. Each quantized transform coefficient value can be
directly related to its corresponding original transform
coefficient value by equation.
=
<
+ 1 (1)

( =

( , )

(2)

The anti-forensic dithers distribution is given by the


formula
P (D=d)=

, )
( . )

<

+ 1)(3)

1.1.2. JPEG ANTI-FORENSICS:


Brief over view of JPEG compression then present our
anti-forensic technique designed to remove compression
finger prints from JPEG compressed image DCT
coefficients. For gray scale image, JPEG compression
begins by segmenting an image into a series of non over
lapping 8x8 pixel blocks then computing the two
dimensional DCT of each block. Dividing each coefficient
value by its corresponding entry in predetermined
quantization matrix rounding the resulting value to the
nearest integer. First image transformed from the RGB to
the YCBCrcolorspace. After this can been performed,
compression continues as if each color layer were an
independent gray scale image.
1.1.3. DCT
Removal:

If the image was divided into segment during compression,


another compression finger print may arise. Because of the
loss

Coefficient

Quantization

Fingerprint

Anti-forensic frame work which we


outlined in section 2.we begins by modeling the
distribution of coefficients values with in a particular ac
sub band using the Laplace distribution.
( = ) =
x
(4)
Using this model and the quantization rule
described above the coefficient values of an ac sub band of
DCT coefficients with in a JPEG compressed image will be
distributed according to the discrete Laplace distribution.
1
P(Y=y)=

if y = 0
if y=kQi,j
sin ( 2)
0

(5)

Fig1: anti forensic of digital image compression


When the anti-forensically modify each quantized
transform coefficient by adding specially designed noise,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

19

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Fig2: Histogram of perturbed DCT coefficient values from


a DCT sub band in which all coefficients were quantized to
zero during JPEG compression.
Wavelet-Based Compression Overview:
Through several wavelet-based image compression
techniques exists such as SPIHT,EZW,and most popularly
JPEG 2000.they all operate in a similar fashion and leave
behind similar compression finger prints.JPEG 2000 begins
compression by first segmenting an image into fixed sized
non over lapping rectangular blocks known as tiles while
other operate on the image as a whole.
Two dimensional DWT of the image or
each image tile is computed these sub bands of the wave let
coefficient.
Because of these sub bands corresponding
to either high or low frequency DWT coefficients in each
spatial dimension, the four sub bands are referred to using
the notation LL, LH, HL, and HH.
Image compression techniques achieve loss compression
through different processes they each introduce DWT
coefficient quantization finger prints into an image
Quantization and dequantization process causes DWT
coefficient in image compression in the multiples of their
respective sub bands.

ISBN: 378 - 26 - 138420 - 5

Fig3: Top: Histogram of wavelet coefficient from an


uncompressed image.
Bottom: wavelet coefficient from same image after SPIHT
compression.

As a result only the n most significant bits of


each DWT coefficients are retained. This is equivalent to
applying the quantization rule. Where X is a DWT
coefficient from an uncompressed imager y is the
corresponding DWT coefficient in its SPIHT compressed
counterpart.

Fig4: Top: peg compressed image using quality factor.


Bottom: Anti forensically modified version of same image.
2. UNDETECTABLE IMAGE TAMPERING
THROUGH JPEG COMPRESSION
Number of digital image forensic techniques have been
developed which are capable of identifying an images
origin, tracing its processing history, and detecting image
forgeries. Though these techniques are capable of
identifying standard image manipulation, they do not
address the possibility t be that anti forensic operations may
be designed and used to hide evidence of image tampering
.we propose anti-forensic operation capable of removing
blocking artifacts from a previously JPEG compressed
image. We can show that by help of this operation along
another anti-forensic operation we are able to fool forensic

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

20

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

methods designed to detect evidence of JPEG compression


in decoded images, determine an images origin, detect
double JPEG compression, and identify cut and paste
image forgeries.
A digital image forgery has resulted in an
environment where the authenticity of digital images
cannot be trusted. Many of these digital forensic techniques
rely on detecting artifacts left in image by JPEG
compression. Because most of the digital cameras make use
of proprietary quantization tables, an image compression
history can be used to help identify the camera used to
capture it. These techniques are quite adept at detecting
standard image manipulation, they do not account for the
possibility that anti-forensic operation designed to hide
traces of image manipulation may applied to an image.
Recent work as shown such operations can be constructed
to successfully fool existing image forensic techniques.
Back Ground:
When an image is subjected to JPEG compression, it is first
segmented into 8X8 pixel blocks. The DCT of each block
is computed and resulting set of DCT coefficients are
quantized by dividing each coefficient by its corresponding
entry in a quantization table then rounding the result to the
nearest integer. The set of quantized coefficients read into a
single bit stream and lossless encoded. so decompressed
begins by bit stream of quantized DCT coefficients and
reforming into a set of 8X8 pixel blocks.
As a result two forensically significant
artifacts are left in an image by JPEG compression. That is
DCT coefficient quantization artifact sand blocking
artifacts. Blocking artifacts are the discontinuities which
occur across 8X8 pixel block boundaries because of
JPEGs loss nature antiforensic technique capable of
removing DCT coefficient artifacts from a previously
compressed image.

ISBN: 378 - 26 - 138420 - 5

A measure of blocking artifacts strength is


obtained by calculating the difference between the
histograms of Z and Z values denoted by H1 and H2
respectively, using the equation.
K=|HI (Z= n) HII (Z= n)|.
The values of K lying above a fixed detection threshold
indicate the presence of blocking artifacts.

Fig5: Histogram of DCT coefficients from an image before


compression (top left), after JPEG compression (top right),
and after addition anti-forensic dither to the coefficients of
the JPEG compressed image.
2.2. IMAGE TAMPERING THROUGH ANTIFORENSIC:
We show that anti-forensic dither and our proposedantiforensic deblocking operation can be used to deceive
several existing image forensic algorithms that rely on
detecting JPEG compression artifacts.

2.1. ANTI-FORENSIC DEBLOCKING OPERATION


JPEGblocking artifacts must be removed from an image
after anti-forensic dither has been applied to its DCT
coefficients. Number of de blocking algorithms proposed
since the introduction of JPEG compression, these are all
suited for anti-forensic purposes. To be successful it must
remove all visual and statistical traces of block anti-facts.
We found that light smoothing an image followed by
adding low-power white Gaussiannoise. Able to remove
statistical traces of JPEG blocking artifacts without causing
the image DCT coefficient distribution to deviate from the
Laplace distribution. in the anti-forensically deblocked
image according to the equation.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

21

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig 7: Histogramof (3, 3) DCT coefficients from an image


JPEG compressed once using a quality factor of 85(left),
image after being double JPEG compressed using a quality
factor of 75 followed by 85(center),and the image after
being JPEG compressed using a quality factor of
75,followed by the application of anti forensic dither, then
recompressed using a quality factor of 85(right).
3. PROPOSED METHOD:

Tothe best knowledge increased in the field of antiforensics. Most of the methods of this an forensics is
to find out the process that which the image
compression is takes places, such of that methods
involves in like JPEG detection and quantization table
estimation.in this method of anti-forensic the JPEG
compression of an image history also produces the
information of camera used to produce an image.

Fig 6:Result of the proposed anti-forensic deblocking


algorithm applied to a typical image after it has been
JPEG compression using a quality factor of 90 (far left),
70(center left), 30(center right),and 10 (far right) followed
by the addition of anti-forensic dither to its DCT
coefficients.

Although it can be used to discover the forged areas


along with in the picture.in case of image
compression this technique is also developed to use as
evidence of image manipulation.so in this anti
forensic technique traces left by compression and
other processing are discussed

2.3. Hiding Traces of Double JPEG compression:


An image forger may wish to remove evidence of
corresponding a previously JPEG compressed image. Such
image forger wishes to alter a previously compressed
image, and then save the altered image as JPEG.Several
methods have been proposed to detect recompression of
JPEG compressed image commonly known as double
JPEG compression.
2.4. Falsifying an Images Origin:
In some scenarios, an image forger may wish to falsify the
origin of digital image simply altering the Mata data tags
associated with an images originating device is insufficient
to accomplish this because several origin identifying
features are intrinsically contained with a digital
image.Anti-forensic dither of an images DCT coefficient,
then re-compressing the image using quantization tables
associated with another device. by doing an image in this
manner, we are able to insert the quantization signature
associated with a different camera into an image while
preventing the occurrence of double JPEG compression
artifacts that may alert forensic investigators of such a
forgery.

4. CONCLUSION:

By the above two existing methods, one of the


method of anti-forensic method of digital image
compression it has increasingly up on digital images
to communicate and this method is considered anti
forensics method is fooling forensic algorithms. This
technique is designed to remove forensically
significant indicators of compression of an image.
First developing frame work its design the antiforensic techniques to remove compression finger
prints from image transform coefficients. This anti
forensic dither to the transform coefficient of
compressed image distribution matches the estimated
one. When we use this frame work it specifically
targeted at erasing compression finger prints left by
both JPEG and wavelet based coders. These
techniques are capable of removing forensically
detectable traces of image compression without
significantly impacting an images visual quality.
The second method of undetectable image tampering
through JPEG compression anti forensics digital

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

22

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[8] W. S. Lin, S. K. Tajo, H. V. Zhao, and K. J. Ray Liu,


Digital image source coder forensics via intrinsic
fingerprints, IEEE Trans. InformationForensics and
Security, vol. 4, no. 3, pp. 460475, Sept. 2009.
[9] H. Farid, Digital image ballistics from JPEG
quantization, Tech. Rep.TR2006-583, Dept. of Computer
Science, Dartmouth College, 2006.
[10] A.C. Popes cu and H. Farid, Statistical tools for
digital forensics, in 6th International Workshop on
Information Hiding, Toronto, Canada,
2004.
[11] T. Pevny and J. Fridrich, Detection of doublecompression in JPEG images for applications in
steganography, IEEE Trans. InformationForensics and
Security, vol. 3, no. 2, pp. 247258, June 2008.
[12] M. Kirchner and R. Bohme, Hiding traces of
resampling in digital images, IEEE Trans. Information
Forensics and Security, vol. 3, no.4, pp. 582592, Dec.
2008.

forensics are developed which are capable of


identifying an images origin. Thesetechniques are
capable of identifying standard image manipulations.
This anti forensic technique capable of removing
blocking artifacts from previously JPEG compression
image.in this method we are able to fool forensic
methods to designed to detect evidence of JPEG
compression in decoded images, determine an
images origin.
When comparing above two existing methods, the
anti-forensic method of removing detectable traces
from digital images has advanced technique increases
attractive ness and more over trust in the digital
images it has capable of removing evidences of
compression and filtering of in digital images history
processing.by adding tailored noise in the image
processing we can find out the where the images is
tampered and compressed, weather its fake or original
this can be used in the medical department as well as
in the police department cases. This method is to be
used to cover history of processing and it can be also
used to remove the signature traces of filtering.
REFERENCES
[1]M.chen, J.fridrich, M.goljan and Lukas, Determining
image origin and integrity using sensor noise IEEE
trans.inf.forensic security, vol.3, no.1, pp.74-90,
march.2008.
[2]A.swaminathan,M.Wu,andK.R>Liu,Digital
image
forensics
via
intrinsic
finger
prints,IEEEtrans.inf.forensicssecurity, vol.3, no.1, pp.101117, mar.2008.
[3] M.Kirchner and R.Bohme,Hiding traces of resampling
in digital images,IEEEtrans.inf.forensics security, vol.3,
no.4, pp.582-592, Dec.2008.
[4] I. Ascribes, S. Bayram, N.Memon, M. Ram Kumar, and
B. Sankur, A classifier design for detecting image
manipulations, in Proc. IEEE Int.Conf. Image Process.
Oct. 2004, vol. 4, pp. 26452648.
[5] M. C. Stamm and K. J. R. Liu, Forensic detection of
image manipulation using statistical intrinsic fingerprints,
IEEE Trans. Inf. ForensicsSecurity, vol. 5, no. 3, pp. 492
506, Sep. 2010.
[6] Z. Fan and R. de Queiroz, Identification of bitmap
compression history: JPEG detection and quantizer
estimation, IEEE Trans. ImageProcess. vol. 12, no. 2, pp.
230235, Feb. 2003.
[7] M. C. Stamm and K. J. R. Liu, Wavelet-based image
compressionanti-forensics, in Proc. IEEE Int. Conf. Image
Process., Sept. 2010,
pp. 17371740.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

23

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

FPGA Implementation of Fault Tolerant Embedded


RAM using BISR Technique
Swathi Karumuri, Suresh Chowdary Kavuri
Dept., of VLSI and Embedded systems, Dept., of Electronics and Communication Engineering, JNTU Hyderabad, India
[email protected]
[email protected]
Abstract-- Embedded memories with fault tolerant capability can
effectively increase the yield. The main aim of this project is to
design a fault tolerant memory which is capable of detecting stuck
at faults and to repair them automatically. It uses BISR with
redundancy scheme in order to achieve this. The architecture
consists of a multiplexer (MUX), memory (circuit under
test-CUT), Built in self test (BIST) module, Built in self diagnosis
(BISD) module. Besides adding spare rows and columns user can
select normal words as redundancy to repair the faults. So, it
reduces the area. High repairing speed is possible with one to one
mapping between faults and redundancy locations. It provides
test mode and access mode to the RAM users. Many repairing
schemes were implemented where repairing is not possible if the
redundancy locations have faults. This proposed architecture
shows the number of redundant locations and depending on that
it has the advantage of repairing faults in redundancy locations
also. The proposed model is simulated and synthesized using
Xilinx and tested using Spartan3E development board.

time-saving and less cost compared to the ones controlled by


the external tester (ATE) [8]. However, BIST doesnt repair the
faults. BISR techniques tests embedded memories, saves the
fault addresses and replaces them with redundancy. [5]
Presents multiple redundancies scheme and [9] proposes BISR
strategy applying two serial redundancy analysis (RA) stages.
All the previous BISR techniques have the ability to repair
memories, but they can not avoid storing fault addresses more
than once. [10] proposes a solution to this but it cannot repair
redundant location faults. This paper proposed an efficient
strategy that can repair redundancy faults also.
The rest of the paper is organized as follows. Section II briefs
different fault models, March algorithms and BIST. Section III
introduces the proposed BISR technique. Section IV shows the
experimental results. Section V concludes this paper. Section
VI gives the future work.
II.

Keywords-- Built in self diagnosis (BISD), Built in self repair


(BISR), Built in self test (BIST), FPGA, VHDL.

I.

FAULT MODELS, MARCH ALGORITHMS AND BIST

A fault model is a systematic and precise representation of


physical faults in a form suitable for simulation and test
generation [11]. Different RAM faults are as follows and can
be referred in [12].
AF: Address Fault
ADOF: Address Decoder Open Faults
CF: Coupling Faults
DRF: Data Retention Faults
SAF: Stuck-at Faults
SOF: Stuck Open Faults
TF: Transition Faults
Memory test algorithms are two types where traditional tests
are either simple, fast but have poor fault coverage or have
good fault coverage but complex and slow. Due to this
imbalance conflicts, the these algorithms are losing popularity.
The details of these algorithms and comparison can be referred
in [13]-[14]. An Efficient memory test should provide the best
fault coverage in the shortest test time. March tests are the most
common in use. They have good fault coverage and short test
time. In order to verify whether a given memory cell is good, it
is necessary to conduct a sequence of write(w) and read(r)
operations to the cell. The actual number of read/write
operations and the order of the operations depend on the target
fault model. March element is a finite sequence of read/write
operations applied to a cell in memory before processing the
next cell. The next cell address can be ascending or descending
in order. The comparison of different march algorithms are
tabulated in Table I [15]. MarchC- algorithm has better fault

INTRODUCTION

With improvement in VLSI technology, more and more


components are fabricated onto a single chip. It is estimated
that the percentage of area occupied by embedded memories on
an SOC is more than 65% and will rise up to 70% by 2017[1].
The increasing use of large embedded memories in SOCs
require automatic reconfiguration to improve yield and
performance. Memory fabrication yield is limited to
manufacturing defects, alignment defects, assembly faults,
random generated defect patterns, random leakage defects and
other faults and defects [2]. To detect defects in a memory,
many fault models and test algorithms have been developed
[3]. Most of these algorithms have been in use. To increase the
yield and efficiency of embedded memories, many redundancy
mechanisms have been proposed in [4]-[7]. In [5]-[6] both
redundant rows and columns are incorporated into the memory
array. Repairing the memory by adding spare words, rows and
columns into the word oriented memory core as redundancy is
proposed in [7]. All these redundancy mechanisms increase the
area and complexity in designing of embedded memories. A
new redundancy mechanism is proposed in this paper in order
to solve the problem. Some normal words in embedded
memories can be considered as redundancy instead of adding
extra rows and columns. It is necessary to test the memory
before using redundancy mechanisms to repair the faults. In
1970s, circuit under test is tested by external equipments
called ATEs. BIST controlled DFT circuitry is more efficient,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

24

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

coverage, less operations and less test length. It has 6 elements


and 10 operations. So it is used in this paper to test the memory.
The steps in MarchC- algorithm are as follows
TABLE I
MARCH ALGORITHMS AND FAULT COVERAGE

1 up - W0
2 up (R0, W1)
3 up (R1, W0)
4 down (R0, W1)
5 down (R1, W0)
6 down - R0
In above steps, W represents write, R represents read, up
represents executing RAM addresses in ascending order while
down in descending order. Write0(1) represents writing
0(1)into the memory, read 0(1) represents expected data(0 or 1)
from the memory in read operation.
The BIST module contains a test pattern generator and an
output response analyzer (ORA). MarchC- algorithm is used to
generate the test patterns and a comparator is used as an ORA.
The output from the memory in read operation is compared
with expected data in the corresponding March element and
indicates whether there are any faults in memory or not. It can
also indicate if the memory test is done by activating test_done
to 1.
III.

A.

Fig. 1 Proposed Redundancy mechanism in RAM

redundancy limit it activates overflow signal. It shows the


number of faults in redundant locations. When repairing is
activated, faulty addresses are replaced with redundant
addresses. MUX module controls the access between test mode
and access/user mode. In test mode (test_h=1) BIST generates
the inputs to the memory and in access mode (test_h=0) the
inputs are equal to the system inputs or user inputs.

PROPOSED BISR TECHNIQUE

Proposed Redundancy Mechanism

The proposed architecture is flexible. Some normal words in


RAM can be selected as redundancy if it needs to repair itself.
To distinguish them from the normal ones we name these
words as Normal-Redundant words. We take a 64x4 RAM as
shown in Fig.1 as an example. There are 58 normal words and 6
Normal-Redundant words. When repairing is not used, the
Normal-Redundant words are accessed as normal ones. The
Normal-Redundant words can be accessed, when there are
faults in normal words. This kind of selectable redundancy
architecture can increase efficiency and save area.
B.

Proposed BISR Architecture


The architecture of Proposed BISR is shown in Fig.2. It
consists of 3 parts named as BIST module, MUX module, and
BISD module. BIST module produces test inputs to the
memory using MarchC- algorithm. It detects the failures in
memory by comparing RAM output data with expected data of
MarchC- algorithm. compare_q=1 for faulty addresses. It
asserts test _done to 1 when testing is completed. BISD module
stores the faulty address in Fault_A_Mem. It maintains a
counter to count the number of faulty addresses. When the
count exceeds the maximum

Fig. 2 Architecture of proposed BISR

C.

Built in Self Repair Procedure


Fig. 3 shows the flowchart for testing and repairing
mechanism used. The BISR starts by resetting the system
initially (rst_l=0).After that the system should be tested to find
the faults in the memory. In this phase BIST and BISD modules
works in parallel. BIST detects the fault address and it will be
sent to the BISD. It checks it with already stored addresses .If

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

25

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

the faulty address has not been stored it stores in Fault_A_Mem


and the counter increments. It will be ignore otherwise. After
the test phase there will be two conditions. If there are no faults
or there are too many faults that overflows the redundancy
capacity, BISR goes into COMPLETE phase. If there are faults
in memory but without overflows then BISR goes into
REPAIR and TEST phase. As in TEST phase, BIST and BISD
works in parallel. BISD replaces the fault addresses with
redundant ones. It can repair the faults in redundancy locations
also. BIST module tests the memory again. There will be two
results, repair pass or fail. Fig. 4 shows the storing and
repairing fault addresses.

ISBN: 378 - 26 - 138420 - 5

but the fault addresses should not be stored thrice. So, one to
one mapping is proposed in BISD module.
Second feature is, BISR is flexible. In access mode user can
decide whether to use repairing or not. If BISR, i.e. repairing is
activated redundant locations are used for repairing .Table II
shows the operating modes of RAM.
Third feature is, its repairing speed is high. With one-one
mapping it can replace the faulty locations address with
corresponding redundant address very quickly.
TABLE II
OPERATING MODES OF RAM

E. Proposed Repairing Mechanism to Repair Faults in


Redundant Locations
In order to repair the faults, there should not be any faults in
redundant locations. After test phase Fault_A_Mem contains
all the faulty addresses. If it contains all normal location faults,
then the mapping shown in Fig. 4 works good. If the
Fault_A_Mem contains redundant location addresses then
mapping faulty address with corresponding faulty redundant
location results in faulty output again. To solve this problem
the repairing mechanism is modified slightly and is shown in
Fig. 5 where there is one redundant and four normal faults.

Fig. 3 Flow chart of testing and repairing mechanism

Fig. 5 Mapping procedure in presence of redundant fault

Fault-address 3 is mapped with redundant address 5 as


redundant address 3 is in Fault_A_Mem. Fig. 6 shows the flow
chart. If there are redundant faults then the resultant redundant
address should be checked in Fault_A_Mem .If there are any
unmapped redundant locations, then mapping is done to that
redundant location, if the new redundant location is fault free.
If it is faulty again then next unmapped redundant location is
considered and checking continuous until the new redundant
location is fault free. Otherwise repairing fails.

Fig. 4 Flow of storing fault addresses & repairing mechanism

D.

BISR Features
The first feature is, BISR is efficient. Normal redundant
words can be used when repairing is not activated. It saves chip
area. The fault addresses are stored only once. March algorithm
has 6 steps. The address will be read five times in one test.
Some faulty addresses will be detected mpore than one time.
Take stuck-at-1 fault, it will be detected in 2nd, 4th and 6th steps

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

26

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig.7 Output showing repairing of redundant location faults(rounded), normal


faults
REFERENCES
[1]
Fig. 6 Flow chart for repairing redundant and normal faults
IV.

[2]

EXPERIMENTAL RESULTS

The proposed BISR is simulated and synthesized using


Xilinx and is checked on Spartan 3E FPGA kit. A stuck-at-0
and stuck-at-1 faults are set into the memory to verify the
function. It is observed that in presence of faults, the data
written into and data read from the faulty location is not same if
repairing is not activated. Repairing is activated and is
observed that the locations are repaired. It is also verified that
the faults in redundant locations are repaired. Fig. 7 shows the
output when redundant and normal faults are repaired.

[3]
[4]

[5]

[6]

[7]

V. CONCLUSION

A fault tolerant embedded RAM using BISR technique has


been presented in this paper. It uses selectable redundancy and
is designed flexible such that user can select the operating
modes. BISD module avoids storing the fault addresses more
than once. The mechanism for repairing redundant locations
faults has been proposed and is observed that it efficiently
repairs the redundant faults along with normal faults.

[8]
[9]

[10]

VI. FUTURE WORK


The proposed architecture is mainly focused on
stuck-at-faults in the memory using MarchC- algorithm. There
are many fault models such as address decoder faults, transition
faults etc; and there are different test algorithms. Therefore a
future work can improve the proposed architecture to repair the
above mentioned faults in the memory using different test
algorithms.

[11]

[12]
[13]
[14]
[15]

Farzad Zarrinfa Optimizing embedded memory for the latest ASIC and
SOC design (2012) -chip design tools, technologies and methodologies
by Mentor Graphics.
C. Stapper, A. Mclaren, and M. Dreckman, Yield model for Productivity
Optimization of VLSI Memory Chips with redundancy and partially good
Product, IBM Journal of Research and Development, Vol. 24, No. 3, pp.
398-409, May 1980.
Ad J.van de Goor, Testing Semiconductor memories: Theory and
Practice, 1999, ISBN 90-80 4276-1-6 .
P. Mazumder and Y. S. Jih, A new built-in self-repair approach toVLSI
memory yield enhancement by using neural type circuits, IEEE
transactions on Computer Aided Design, vol. 12, No. 1, Jan, 1993.
W. K. Huang, Y. H. shen, and F. lombrardi, New approaches for repairs
of memories with redundancy by row/column deletion for yield
enhancement, IEEE Transactions on Computer-Aided Design, vol.
9,No. 3, pp. 323-328, Mar. 1990.
H. C. Kim, D. S. Yi, J. Y. Park, and C. H. Cho, A BISR (built-in self
repair) circuit for embedded memory with multiple redundancies, VLSI
and CAD 6th International Conference, pp. 602-605, Oct. 1999.
Shyue-Kung Lu, Chun-Lin Yang, and Han-Wen Lin, Efficient BISR
Techniques for Word-Oriented Embedded Memories with Hierarchical
Redundancy, IEEE ICIS-COMSAR, pp. 355-360, 2006.
C. Stroud, A Designers Guide to Built-In Self-Test, Kluwer Academic
Publishers, 2002.
I.Kang, W. Jeong, and S. Kang, High-efficiency memory BISR with two
serial RA stages using spare memories, IET Electron. Lett. vol. 44, no.
8, pp. 515-517, Apr. 2008.
Huamin Cao, Ming Liu, Hong Chen, Xiang Zheng, Cong Wang and
Zhihua Wang, Efficient Built-in Self-Repair Strategy for Embedded
SRAM with Selectable Redundancy,
Consumer Electronics,
Communications and Networks (CECNet), 2012 2nd International
Conference .
M. Sachdev, V. Zieren, and P. Janssen, Defect detection with transient
current testing and its potential for deep submicron CMOS ICs, IEEE
International Test Conference, pp. 204-213, Oct. 1998.
Mentor Graphics,MBIST Architect Process Guide,Software Version
8.2009_3, Aug 2009, pp. 113-116.
Pinaki Mazumder, Kanad Chakraborty,Testing and Testable Design of
High-Density Random-Access Memories.
Vonkyoung kim and Tom Chen,Assessing SRAM Test Coverage for
SubMicron CMOS Technologies,VLSI test symposium,1997,15th IEEE.
L.Dharma Teja,K. Kiruthika and V. Priyanka Brahmaiah,Built in self
repair for Embedded RAMS with efficient fault coverage using PMBIST,
International Journal of Advances in Engineering & Technology, Nov.
2013.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

27

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

ADAPTIVE MODULATION IN MIMO OFDM SYSTEM FOR4G


WIRELESS NETWORKS

POLI SURESH, Student M.Tech-ECE Dept,


Siddartha Educational Academy Group of Institutions,
Tirupati,Andhra Pradesh,[email protected]

M.VINOD,
Assistant Professor-ECE Dept,
Siddartha Educational Academy Group of Institutions,
Tirupati,Andhra Pradesh,India-517505.

[email protected]
ABSTRACT
This paper presents the strategy of applying
Hybrid adaption techniques in MIMO OFDM system.
With the rapid growth of digital communication in recent
years, the need for high-speed data transmission is increased.
Multiple-input-multiple-output (MIMO) antenna architecture
has the ability to increase capacity and reliability of a wireless
communication system. Orthogonal frequency division
multiplexing (OFDM) is another popular technique in
wireless communication which is famous for the
efficient high speed transmission and robustness to
frequency selective channels. Therefore, the integration of the
two technologies probably has the potential to meet the ever
growing demands of future communication systems. First
focusing on OFDM in which the bit error rate (BER) of
multilevel quadrature amplitude modulation (M-QAM) in
flat Rayleigh fading channel for 128,256,512 subcarriers
was calculated and also channel estimation can be done by
using different algorithms which is carried out through
Matlab software. The channel estimation of MIMO OFDM
system is calculated by using Minimum mean square error
algorithm (MMSE) and compares the actual value and
estimation errors using Matlab simulation.
Then take the feedback from the channel estimation
and apply hybrid adaptation techniques to improve the
spectral efficiency and to reduce the transmit power. This
system is used in wireless LAN i.e. IEEE 802.11a/g,
HYPERLAN etc.,
Keywords: MIMO, OFDM, BER, M-QAM, MMSE,
MIMO OFDM

I.

INTRODUCTION

Physical limitations of the wireless medium create a technical


challenge for reliable wireless communication. Techniques that
improve spectral efficiency and overcome various channel
impairments such as signal fading and interference have made an
enormous
contribution
to
the
growth
of
wireless
communications. Moreover, the need for high-speed wireless Internet
has led to the demand for technologies deliveringhigher capacities
and link reliability than achieved by current systems. Multipleinput multiple-output (MIMO) based communication systems are
capable accomplishing these objectives. The multiple antennas
configuration exploits the multipath effect to accomplish the
additional spatial diversity.

However, the multipath effect also causes the negative effect of


frequency selectivity of the channel. Orthogonal frequency division
multiplexing (OFDM) is a promising multi-carrier modulation
scheme that shows high spectral Efficiency and robustness to
frequency selective channels. In OFDM, a frequency-selective
channel is divided into a number of parallel frequency-flat sub
channels, Thereby reducing the receiver signal processing of the
system. The combination of OFDM and MIMO is a promising
technique to achieve high bandwidth efficiencies and System
performance. In fact, MIMO-OFDM is being considered for the
upcoming IEEE 802.11n standard, a developing standard for high
data rate WLANs

II. PAPER REVIEW


C.Poongodi,P.Ramya,A.Shanmugam,(2010) BER Analysis of
MIMO OFDM System using M-QAM over Rayleigh Fading
Channel Proceedings of the International Conference on
Communication and Computational Intelligence, explains the
BER of MIMO OFDM over the Rayleigh fading channel for MQAM Modulation. and also the estimation of channel at high
frequencies with conventional least squares(LS) and Minimum
Mean Square(MMSE) estimation algorithms which is carried out
through MATLAB simulation. The performance of MIMO
OFDM is evaluated on the basics of Bit Error Rate (BER)
andMeansquareError (MSE) level.
Dr.JayaKumari.J,(2010)
MIMO OFDM for 4G Wireless Systems, International
Journal of Engineering Science and Technology VOL.2 (7).,
explains the OFDM may be combined with antenna arrays at the
transmitter and receiver to increase the diversity gain and/or to
enhances the system capacity on time variant and frequencyselective channels, resulting in MIMO Configuration. As a
promising technology for the future broadband communication,
and the simulation results show that this is a promising
technology for next generation wireless systems and used in
applications such as HYPERLAN,WLAN and DSL etc. Pallavi
Bhatnagar ,Jaikaran Singh,Mukesh Tiwari (2011) Performance
of MIMO-OFDM System for Rayleigh fading channel (ISSN
2221-8386) Volume No 3 May 2011 explains the efficient
simulation for MIMO OFDM system with channel
equalization. BPSK modulation is used to detect the behavior of
the Rayleigh fading channels in presence of additive white
Gaussian noise and performance is evaluated. This paper shows
that the addition of equalizer reduces the BER and the channel
output becomes more pronounced.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

28

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

III. METHODOLOGY
Orthogonal frequency division multiplexing (OFDM) transforms
a frequency selective channel into a large set of individual
frequency non-selective narrowband channels, which is suited
for a multiple-input multiple-output (MIMO) structure that
requires a frequency non-selective characteristic at each channel
when the transmission rate is high enough to make the
whole channel frequency selective. Therefore, a MIMO system
employing OFDM, denoted MIMO-OFDM, is able to achieve
high spectral efficiency. However, the adoption of multiple
antenna elements at the transmitter for spatial transmission
results in a superposition of multiple transmitted signals at
the receiver weighted by their corresponding multipath
channels and makes the reception more difficult. This imposes a
real challenge on how to design a practical system that can offer a
true spectral efficiency improvement. If the channel is frequency
selective, the received signals are distorted by ISI, which makes
the detection of transmitted signals difficult. OFDM has emerged
as one of most efficient ways to remove such ISI. The delay
spread and Doppler spread are the most important factors to
consider in thee characterizing the SISO system. In the MIMO
system which employs multiple antennas in the transmitter and/or
receiver, the correlation between the transmitter and receiver
antenna is an important aspect of the MIMO channel.it depends
on the angle of arrival of each multipath component In fact,
MIMO technique is an essential means of increasing capacity in
the high SNR regime, providing at most N spatial degrees of
freedom. A typical multi-user MIMO communication
environment in which the multiple mobile stations are served by
single base station in the cellular system. Fig 5.1 and fig 5.2
shows the block diagram of MIMO OFDM transmitter and
receiver. This system is modification of OFDM, which provides
high BER and used in many applications such as DAB, DVB,
DSL.

Figure 2: Receiver block diagram of MIMO OFDM


Regardless of the type of MIMO system, most of the
equalization/detection schemes require knowledge of the channel
information in order to recover the signal. Hence, developing
an efficient method of approximating the transmission channel
between the transmitter and receiver is an essential component of
the receiver design. In this chapter the channel estimation for

MIMO-OFDM is briefly explained fundamentally. First, a


general overview on classical estimation theory is provided then
OFDM channel estimation is briefly explained in the next step
MIMO-OFDM channel estimation is investigated. The problem of
MIMO-OFDM system channel estimation in frequency domain is
addressed, and the solution for this problem is well interpreted.
Finally The MMSE algorithm as an alternative to decrease
the computational complexity of LS channel estimation is
investigated.

3.2 LS CHANNEL ESTIMATION:


The least-square (LS) channel estimation method finds the
channel estimate

in such a way that the following cost function

is minimized:
( ) =Y-X
= (Y-X ) ^H(Y-X )
=

Y-

X -

Y+

By setting the derivative of the function with respect to


( )

Figure 1: Transmitter block diagram of MIMO OFDM

= -2(

)+2(

to zero,

) = 0

We have
=
channel estimation as

, which gives the solution to the LS

3.1 CHANNEL ESTIMATION


=(

X)

by
The ultimate goal at the receiver is to recover the signal that was Let us denote each component of the LS channel estimate
originally transmitted. A variety of equalization and signal
[ ], k = 0,1,2,3,..N-1.since X is assumed to be diagonal due
detection techniques has been developed for MIMO systems to the ICI-free condition,the LS channel estimate
can be
depending on whether it is a diversity or spatial multiplexing written for each subcarrier as
system.
[ ] = [ ]/ [ ],k = 0,1,2,N-1
The mean square error(MSE) of this LS channel estimation is
given as

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

29

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

=E{(
=

) (H-

)}

ISBN: 378 - 26 - 138420 - 5

3.4 CHANNEL ESTIMATION OF


OFDM SYSTEM

MIMO

The problem of channel estimation for OFDM has been well


researched however, the results are not directly applicable to
MIMO-OFDM systems. In MIMO systems, the number of
channels increases by M.Nr-folds, where M and Nr is the number
of transmit and receive antenna, respectively. This significantly
increases the number of unknowns to be solved. Using the
MIMO-OFDM system model described in Chapter 4, the channel
3.3MMSE CHANNEL ESTIMATION
estimator for MIMO-OFDM can be developed. For doing so,
MIMO-OFDM 2 by 2 antenna configuration is assumed. Similar
Consider the LS solution in the equation,
=
. Using the to SISO the least square channel estimation is given as
weight matrix W.define ==W , which corresponding to the
( , )
MMSE estimate. Referring to the figure, MSE of the channel
= ( ( ))
,
estimate is given J ( ) = E { } = E { }.
Note that the MMSE in the equation is inversely proportional to
the SNR / , which implies that it may be subject to noise
enhancement, especially when the channel is in a deep null. Due
to its simplicity, however, the LS method has been widely used
for channel estimation.

And the MMSE channel estimation for MIMO OFDM system for
nth transmit antenna and nth receiver antenna is
( , )
( )
given
=
Where
( , )

=
( ( )) ,
( , )
= ( )
( ( )) +
,Where n = 1, 2.NT, m
= 1, 2.NR and NT, NR are the numbers of transmit and receive
antennas, respectively, X(n) is an N X N diagonal matrix whose
Then, the MMSE channel estimation method finds a better diagonal elements correspond to the pilots of the nth transmit
(linear) estimate in terms of W in such a way that the MSEin antenna and Y(m) is N length received vector at receiver antenna
equation is minimized. The orthogonality principle states that the m.
estimation error vector e =
is orthogonal to ,such that

E{

3.5ADAPTIVE MODULATION

}= E{( )

Adaptive modulation is a powerful technique for


maximizing the data throughput of subcarriers allocated to
a user. Adaptive modulation involves measuring the SNR
Where
is the cross-correlation matrix of N*N matrices A and of each subcarrier in the transmission, then selecting a
B, and is the LS channel estimate given as
modulation scheme that will maximise the spectral
efficiency, while maintaining an acceptable BER. This
=
technique has been used in Asymmetric Digital Subscriber
Line (ADSL), to maximise the system throughput. ADSL
Solving equation for the W yields
uses OFDM transmission over copper telephone cables.
The channel frequency response of copper cables is
W =
relatively constant and so reallocation of the modulation
scheme does not need to be performed very often, as a
Where
is the autocorrelation matrix of
given as
result the benefit greatly out ways the overhead required
for measuring of the channel response. Using adaptive
modulation in a wireless environment is much more
=E{H }
difficult as the channel response and SNR can change very
rapidly, requiring frequent updates to track these changes.
=E{H }+ / I
Adaptive modulation has not been used extensively in
And
is the cross correlation matrix between the true channel wireless applications due to the difficulty in tracking the
channel
effectively.
In the effectiveness of a
vector and temporary channel estimate vector in the frequency radio
domain. Using the equation the MMSE channel estimation multiuser OFDM system using an adaptive subcarrier, bit and
power allocation was investigated.
follows as
=W .
=

-W

=0

3.6ADAPTIVEMODULATOR
ANDDEMODULATOR

=
=

( _(

) +

The elements of
and
in equation are E{ , , } =
[ ] [ ] Where k and l denote the subcarrier
frequency index and OFDM symbol
(time) index,
respectively . In an exponentially decreasing multipath PDP
(power delay profile), the frequency- domain correlation [ ] is
given as [ ] = 1/1+j2i

At the transmitter the adaptive modulator block consists of


different modulators which are used to provide different
modulation orders. The switching between these modulators will
depend on the instantaneous SNR. The goal of adaptive
modulation is to choose the appropriate modulation mode for
transmission depending on instantaneous SNR, in order to
achieve good trade-off between spectral efficiency and overall
BER. Adaptive modulation is a powerful technique for

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

30

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

maximizing the data throughput of subcarriers allocated to a user. Table.2 Simulation


Adaptive modulation involves measuring the SNR of each Estimation
subcarrier in the transmission, then selecting a modulation
scheme that will maximize the spectral efficiency ,while
maintaining an acceptable BER.

Parameters

SYTEM

IV.SIMULATION RESULTS
FFT Size, Guard
Band,
OFDM
Symbol and No.of
Symbols used

Table. 1 Simulation parameters in OFDM system


System

OFDM

FFT size

128

Guard band size

32

Symbol duration

160

Channel

Rayleigh Fading

No of symbols used

96

Modulation

QAM

ISBN: 378 - 26 - 138420 - 5

for

OFDM

Channel

OFDM
CHANNEL
ESTIMATION
Nfft=128;
Ng=Nfft/8;
Nofdm=Nfft+Ng;
Nsym=100;

Pilot
Spacing,
Numbers of Pilots
and
Data
per
OFDM Symbol

Nps=4;
Np=Nfft/Nps;
Nd=Nfft-Np;

Number of Bits Per


(modulated)
Symbol
Algorithm
Calculate

Nbps=4; M=2^Nbps;

LS and MMSE
No of symbol Errors and
MSE value

Figure 4.OFDM channel estimation


Figure 3.BER analysis of OFDM system
THE SIMULATION RESULTS ARE
Signal power= 5.860e-003,

Table.3. The
Estimation

EbN0 = 0[dB], BER= 157/1152 = 1.363e-001

EbN0= 15[dB], BER= 27/115200000 = 2.344e-007

of

OFDM

Channel

MSE OF
LSLINEAR

MSE
OF
LSSPLITTER

MSE OF
MMSE

NO.
OF
SYMBOL
ERRORS

25

5.8523e003

7.3677e-003

1.4212e003

84

30

1.8578e003

2.3317e-003

5.2873e004

32

35

5.9423e004

7.3929e-004

1.9278e004

15

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

31

Results

SNR

EbN0 = 5[dB], BER= 154/3456 = 4.456e-002


EbN0= 10[dB], BER= 104/47232 = 2.202e-003

Simulation

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

40

1.9446e004

2.3583e-004

6.5509e005

11

V.CONCLUSION

45

6.7919e005

7.6674e-005

2.1960e005

50

2.7839e005

2.6372e-005

7.8805e006

Hence the each and every block of OFDM is studied and plotted
the BER analysis under AWGN channel as well as Rayleigh
fading channel and compare the simulation results for different
EbN0[dB] values and BER value.

Table.4: Simulation Parameters for MIMO OFDM Channel


Estimation

SYSTEM
No of receive antennas
No
of
transmit
antennas
Channel
Algorithm

MIMO-OFDM CHANNEL
ESTIMATION
3
2
Rayleigh fading
MMSE

Figure 5: Channel Estimation of MIMO OFDM System.

The channel estimation of MIMO OFDM system by using MMSE


algorithm, which is quite complicated, and the simulation results
shown that as the signal to noise ratio increases the error value
slightly reduces. The actual estimation, estimation 1 and
estimation 2 are plotted and the difference between estimation 1
and estimation 2 are also plotted. The simulation results shown
that the channel condition is worst at 20 dB.
Then the hybrid adaptation techniques to the channel estimation
of MIMO OFDM to improve the spectral efficiency and to reduce
the transmission power of the system.

REFERENCES
[1]. C.Poongodi, P.Ramya ,A. Shanmugam,(2010) BER Analysis of
MIMO OFDM System using M-QAM over Rayleigh Fading
Channel Proceedings of the International Conference on
Communication and Computational Intelligence.
[2]. Dr.JayaKumari.J,(2010) MIMO OFDM for 4G Wireless
Systems, International Journal of ngineering Science and
Technology VOL.2 (7).
[3].PallaviBhatnagar,JaikaranSingh,MukeshTiwari(2011)
Performance of MIMO-OFDM System for Rayleigh fading
channel International Journal of Science and Advanced
Technology(ISSN 2221-8386) Volume No 3 May 2011.
[4].PuLi,Haibin Zhang, Job Oostveen, Erik Fledderus(2010)
MIMO OFDM Performance in Relation to Wideband Channel
Properties The Netherlands 21 st IEEE International Symposium
on Personal,Indoor and Mobile Radio Communications .
[5].Jia Tang and Xi Zhang(2010), Hybrid-daption-Enhanced
Dynamic Channel Allocation for OFDM Wireless Data Networks ,
Networking and information systems laboratory, department of
electrical engineering.
[6]. Viet- Ha Xian in Wang, MD.JahidurRahman, and Jay
Nadeau(2010) , Channel Prediction Based Adaptive Power
Control for Dynamic Wireless Communications Department of
electrical and computer engineering the
university of
westernOntario, London, on, Canada N6A 5B8.
[7]Ohno and iwaosasase(2010) Adaptive Transmission Power
Control for MIMO Diversity Employing Polarization Diversity in
OFDM Radio Access department of information and computer
science,kieo university JAPAN.
[8]Andrea Goldsmith Wireless Communications
[9]TheodereS.Rappaport Wireless Communications Principles and
Practices Second edition
[10] http://www.cambridge.org/9780521837163

Figure 6: BER for Adaptive modulation in MIMO OFDM

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

32

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

VLSI Based Implementation of a digital


Oscilloscope
MS. DHANSHRI DAMODAR PATIL
Department of Electronics,
Amravati University,
SSGM College of Engineering,
Shegaon 444203,
District: Buldhana, State: Maharashtra
Country:India
[email protected]
Mobile No.9096584765

PROF. V. M. UMALE
Department of Electronics,
Amravati University,
SSGM College of Engineering,
Shegaon 444203,
District: Buldhana, State: Maharashtra
Country:India
[email protected]

Semi-standard Oscilloscope look and feel


VGA display; drives a standard computer monitor
PS/2 Mouse User Interface
9-bit input data width
Developed specifically for the Spartan III development kit
from Xilinx.

Abstract In todays fast-paced world, engineers need the best


tools available to solve their measurement challenges quickly and
accurately. There are many types of oscilloscope available in the
market. The main types of oscilloscopes are analog oscilloscope,
digital oscilloscope and PC based oscilloscope. From which the
digital oscilloscope are widely used now a days due there accuracy
portability, high speed, high resolution, data storing capability etc.
here we provide an alternative solution which is basically a digital
oscilloscope with almost all the control options which any
standard digital oscilloscope has. It has basically three large
blocks, first is the ADC part which has the analog to digital IC
which is controlled by the STK500 AVR kit. Second is the
oscillator control part which is implanted on the FPGA Spartan 3
kit. It has the entire storage element and the user input control and
driver part of IC and VGA. The third large block is display device
which is a CRT monitor along this the LEDs and seven segment
display part on FPGA is also used to display information. So this is
a cheap alternative to expensive oscilloscopes; using a VGA
display and a simple mouse interface, a user can use this scope to
look at and measure signals up to about 80 MHz if an extra high
frequency clock will provide to the design.
Keywords- ADC(Analog To Digital IC), FPGA Spartan 3 kit,
VGA Display.

Fig 1: Diagram of Digital oscilloscope


2) Logical diagram:

I INTRODUCTION
This digital oscilloscope provides a cheap alternative to
expensive oscilloscopes; using a VGA display and a simple
mouse interface, a user can use this scope to look at and
measure signals up to about 80Mhz. this kind of scope would
be ideal for hobbyists and students looking to learn and debug
circuits. Development is based on the Spartan III Starter Kit
from Xilinx. The ADC is simply controlled by an MCU
(another starter kit: the ATK400 from Atmel) but will soon
be controlled by the FPGA (to achieve the faster speeds).
In the future, schematics and PCB layout binaries will be
available.
1. Features
Timescale selection
Selectable trigger (Rising, Falling / Normal, Single, Auto)
Horizontal and Vertical position offsets
Grid Display On/Off/Outline

Fig 2: Block diagram of Digital oscilloscope

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

33

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig 4: Block diagram of ADC Data Buffer Part


Character Display Driver:
The signal is display on VGA. But some information
is
given on the seven segment display about the signal and the
working mode thus the character display part is used for this
purpose.

3) Functional Description:
The digital oscilloscope processes the digital data and shows
it on the VGA. This is thus have three large block, these are:
1. ADC converter 2. Display device 3. controller part (FPGA
part)

Mouse Driver:
The user define values are needed to control the oscilloscope
such as time scale, vertical offset, triggering style etc. thus a
input device is needed to give the input in This project this
block use mouse as PS2 format to give user input.

ADC converter:
This block has an analog to digital IC to convert the input
analog signal into its equivalent digital Signal. The ADC is
simply controlled by an MCU (another starter kit: the
ATK500 from Atmel) but will soon be controlled by the
FPGA (to achieve the faster speeds). The digital data from
ADC then goes to the FPGA Spartan 3 kit

Fig 5: Block diagram of Mouse Driver Part


VGA Driver:
The output is shown on the VGA. Thus this block need the
VGA driver to control the CRT monitor and show the wave
for on the monitor. So this block have to generate some
signals as vertical synchronizing and horizontal
synchronizing for scanning and RGB for color except this
there are many other character and user line which divide the
screen for measuring purpose thus this block have to also
generate these fixed lines.

Fig 3: Block diagram of ADC Driver part


Display (CRT) Monitor:
This is the display devise at which the signal waveform is
displayed .in this project I use the CRT monitor as display
device.
FPGA Spartan Kit Block:
In this block the digital data is processed and synchronizing
with ADC IC this block is very important and has many other
processing blocks which are: ADC Data Buffer: This block
has buffer the data so that the continuous waveform can be
seen on VGA. It also has the time scale option so that the data
can be read from the RAM at different frequency speed
according to the user or signal.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

34

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

III. FUTURE SCOPE


The purposed design is having great future possibility.
Following are the some feature that can be added in this
design
FFT display
Measurement Display (amplitude, frequency)
Cursors
Vectors
Multi-channel display (up to 8)
Channel Math
UART or USB computer communication (data export)

IV. CONCLUSION
Fig 6: Block diagram of VGA Driver Part

The usefulness of an oscilloscope is not limited to the world


of electronics. With the proper transducer, an oscilloscope
can measure all kinds of phenomena. Oscilloscopes are used
by everyone from physicists to television repair technicians.
An automotive engineer uses an oscilloscope to measure
engine vibrations. A medical researcher uses an oscilloscope
to measure brain waves. The digital oscilloscopes are
generally having very high costs. The proposed design
provides a cheap alternative to expensive oscilloscopes;
using a VGA display and a simple mouse interface, a user can
use this scope to look at and measure signals up to about
80Mhz.this kind of scope would be ideal for hobbyists and
students looking to learn and debug circuits.

Seven Segment Driver:


The analog to digital converted data can also see on FPGA
board this is done by seven segment display present on
Spartan 3 kit. The digital data is shown on two seven segment
block in hexadecimal format.
VGA Data Buffer:
The signal which is shown on VGA display is stored in RAM
in digital from. Thus for continuous viewing of the waveform
it is necessary to store the data and retrieve it time to time.
This is done by this block in this block there RAM is built in
FPGA and data is stored in it.
Mouse user Input Driver:
In this oscilloscope the input is given by mouse so this block
is control the user input and gives the signals to display and
control the waveform showing on the VGA display.

V. REFERENCE
[1] Pereira, J.M.D., The history and technology of
oscilloscopes, Instrumentation & Measurement Magazine,
IEEE Volume 9, Issue 6, 2006
[2]Oscilloscope
Types:
http://www.radioelectronics.com/info/t_and_m/oscilloscope/
oscilloscope_types.php
[3] Hawkins Electrical Guide, Theo. Audel and Co., 2nd
ed. 1917, vol. 6, Chapter 63: Wave Form Measurement, pp.
1841-2625
[4]
XYZ
of
Oscilloscopes
Tutorial
http://www.tek.com/Measurement/
programs/301913X312631/?lc=EN&PRODUCT=&returnUr
l=ct=TI &cs=pri&ci=2280&lc=EN

DCM (Digital Clock Manager):


The function of this block is to control the clock and removes
the problem of clock skew.

II. SYNTHESIS REPORT


Device utilization summary: Selected Device: 3s200ft256-4

[5] Bhunia C, Giri S, Kar S, Haldar S, Purkait P, A low-cost


PC-based
virtual
oscilloscope,Education,
IEEE
Transactions on Volume 47, Issue 2, May 2004.
[6] Moschitta A, Stefani F, Petri D, Measurements of
Transient Phenomena With Digital Oscilloscopes,
Instrumentation and Measurement, IEEE Transactions on,
Volume 56, Issue6, Dec. 2007 Page(s):2486 - 2491
[7] English W.O., Digital Storage Oscilloscope vs. Digital
Multimeter, Industrial and Commercial Power Systems
Technical Conference, 2006 IEEE.
[8] Hengkietisak S, Tipyakanont S, Tangsiriworakul C,
Manop C, Senavongse W, Laboratory digital signal analysis
with virtual
[9] Rapid Prototyping of Digital Systems, book by James o.
Hamblen and Michael d. Furman, page(s) 134-151

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

35

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[10] Kuenzi C.D., Ziomek C.D., Fundamentals of


Oscilloscope Measurements in Automated Test Equipment
(ATE), Systems Readiness Technology Conference, IEEE,
18-21 Sept. 2006 Page(s):244 252
[11] Lembeye Y., Keradec J.P., Cauffet G., Improvement
in the linearity of fast digital oscilloscopes used in averaging
mode Instrumentation and Measurement, IEEE Transactions
on Volume 43, Issue 6, Dec. 1994 Page(s):922 928
[12] Moschitta A., Stefani F., Petri D., Measurements of
transient phenomena with digital
oscilloscopes
Instrumentation and Measurement Technology Conference,
2003. IMTC '03. Proceedings of the 20th IEEE Volume 2,
20-22 May 2003 Page(s):1345 - 1349 vol.2

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

36

4
www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

APPEARANCE BASED AMERICAN SIGN LANGUAGE RECOGNITION


USING GESTURE SEGMENTATION AND MODIFIED SIFT ALGORITHM
Author1 prof.P.Subba Rao
Professor of ECE
SRKR College of Engineering
Dept. of Electronic and communications
Bhimavaram, India

Author2 Mallisetti Ravikumar


M.E. (Communication Systems)
SRKR College of Engineering
Dept. of Electronic and communications
Bhimavaram, India

Abstract the work presented in this paper is to develop a system


for automatic Recognition of static gestures of alphabets in American Sign Language. In doing so three feature extraction methods
and neural network is used to recognize signs. The system recognizes
images of bare hands, which allows the user to interact with the system in a natural way. An image is processed and converted to a feature vector that will be compared with the feature vectors of a training set of signs. Further work is to investigate the application of the
Scale-Invariant Feature Transform (SIFT) to the problem of hand
gesture recognition by using MATLAB.The algorithm uses modified SIFT approach to match key-points between the query image
and the original database of Bare Hand images taken. The extracted
features are highly distinctive as they are shift, scale and rotation
invariant. They are also partially invariant to illumination and affine transformations. The system is implemented and tested using
data sets of number of samples of hand images for each signs. Three
feature extraction methods are tested and best one is suggested with
results obtained from ANN. The system is able to recognize selected
ASL signs with the accuracy of 92.33% using edge detection and
98.99 using sift algorithm.

expressions are extremely important in signing. (www.nidcd.nih.gov


(US government)). ASL also has its own grammar that is different
from other sign languages Such as English and Swedish. ASL consists of approximately 6000 gestures of common words or proper
Nouns. Finger spelling used to communicate unclear Words or
proper nouns. Finger spelling uses one hand and 26 gestures to
communicate the 26 letters of the alphabet.

Index Terms ASL using MATLAB, Orientation Histogram, SIFT,


ASL Recognition, ASL using ANN, ASIFT Algorithm

There are two types of gesture interaction, communicative gestures


work as symbolic language (Which is the focus in this project) and
manipulative gestures provide multi-dimensional control. Also, gestures can be divided into static gestures (hand postures) and dynamic
gestures (Hong et al., 2000). The hand motion conveys as much
meaning as their posture does. A static sign is determined by a certain configuration of the hand, while a dynamic gesture is a moving
gesture determined by a sequence of hand movements and configurations. Dynamic gestures are sometimes accompanied with body and
facial expressions.The aim of sign language alphabets recognition is
to provide an easy, efficient and accurate mechanism to transform
sign language into text or speech. With the help of computerized
digital image processing and neural network the system can interpret
ASL alphabets.

The 26 alphabets of ASL are shown in Fig.1.

I.INTRODUCTION

A.

American Sign language:

The sign language is the fundamental communication method between the people who suffer from hearing defects. In order for an
ordinary person to communicate with deaf people, a translator is
usually needed the sign language into natural language and vice
versa. International Journal of Language and Communication Disorders, 2005) Sign language can be considered as a collection of gestures, movements, posters, and facial expressions corresponding to
letters and words in natural languages.
American Sign Language (ASL) National Institute on Deafness
&
Other communication Disorders, 2005) is a complete language that
employs signs made with the hands and other facial expressions and
postures of the body. According to the research by Ted Camp found
on the Web site www.silentworldministries.org, ASL is the fourth
most used language in the United States only behind English, Spanish and Italian (Comp). ASL is a visual language meaning it is not
expressed through sound but rather through combining hand shapes
through movement of hands, arms and facial expressions. Facial

Figure. 1. The American Sign Language finger spelling alphabet

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

37

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

.
B. Related Work
Attempts to automatically recognize sign language began to appear
in the 90s. Research on hand gestures can be classified into two cate
Gories. First category relies on electromechanical devices that are
used to measure the different gesture parameters such as hands position, angle, and the location of the fingertips. Systems that use such
devices are called glove-based systems. A major problem with such
systems is that they force the signer to wear cumbersome and inconvenient devices. Hence the way by which the user interacts with the
system will be complicated and less natural. The second category
uses machine vision and image processing techniques to create visual based hand gesture recognition systems. Visual based gesture
recognition systems are further divided into two categories: The first
one relies on using specially designed gloves with visual markers
called visual-based gesture with glove-markers (VBGwGM) that
help in determining hand postures. But using gloves and markers do
not provide the naturalness required in humancomputer interaction
systems. Besides, if colored gloves are used, the processing complexity is increased. The second one that is an alternative to the second kind of visual based gesture recognition systems can be called
pure visual-based gesture (PVBG) means visual-based gesture
without glove-markers. And this type tries to achieve the ultimate
convenience naturalness by using images of bare hands to recognize

1. Feature extraction, statistics, and models


1. The placement and number of cameras used.
2. The visibility of the object (hand) to the camera for simpler extraction of hand data/features.
5. The efficiency and effectiveness of the selectedalgorithms to provide maximum accuracy and robustness
3. The extraction of features from the stream of Streams of raw Image data.
4. The ability of recognition algorithms to extracted features.
5. The efficiency and effectiveness of the selected algorithms to provide maximum accuracy and robustness.
II.SYSTEM DESIGN AND IMPLEMENTATION
The system is designed to visually recognize all static signs of the
American Sign Language (ASL), all signs of ASL alphabets using
bare hands. The user/signers are not required to wear any gloves or
to use any devices to interact with the system. But, since different
signers vary their hand shape size, body size, operation habit and so
on, which bring more difficulties in recognition. Therefore, it realizes the necessity for signer independent sign language recognition
to improve the system robustness and practicability in the future. The
system gives the comparison of the three feature extraction methods
used for ASL recognition and suggest a method based on recognition
Rate. It relies on presenting thegesture as a feature vector that is
translation, rotation and scale invariant. The combination of the feature extraction method with excellent image processing and neural
networks capabilities has led to the successful development of ASL
recognition system using MATLAB. The system has two phases: the
feature extraction phase and the classification as shown in Fig.2.
Images were prepared using portable document format (PDF) form
so the system will deal with the images that have a uniform background.

Types of algorithms can be used for image recognition

1. Learning algorithms.
a. Neural network (e.g. research work of Banarse, 1993).
b. Hidden Markov Models (e.g. research work of Charniak, 1993).
c. Instance-based learning(research work of Kadous,1995)
2. Miscellaneous techniques.
a. The linguistic approach (e.g. research work of Hand, Sexton, and
mullan, 1994)
b. Appearance-based motion analysis (e.g. research Work of Davis
and Shah, 1993).
c. Spatio-temporal vector analysis (e.g. research Work of Wuek,
1994
a. Template matching (e.g. research work Darrell and Pentland,
1993)
b. Feature extraction and analysis, (e.g. research work of Rbine,
1991)
c. Active shape models Smart snakes (e.g. research work of Heap
and Samaria, 1995)
d. Principal component analysis (e.g. research work of Birk, Moeslund and Madsen, 1997)
e. Linear fingertip models (Research work of Davis and shah, 1993)
f. Causal analysis (e.g. research work of Brand and Irfan, 1995).
Among many factors, five important factors must be considered for
the successful development of a visionbased solution to collecting
data for hand posture and gesture recognition

Fig 2. Designed System block diagram

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

38

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Figure. 4. System Overview.

A. Feature Extraction Phase

Images of signs were resized to 80 by 64, by default imresize uses


nearest neighbor interpolation to determine the values of pixels in the
output image but other interpolation methods can be specified. Here
Bicubic method is used because if the specified output size is
smaller than the size of the input image, imresize applies a low
pass filter before interpolation to reduce aliasing. Therefore we get
default filter size 11- by11. To alleviate the problem of different
lighting conditions of signs taken and the HSV (Hue, Saturation,
Brightness) non-linearity by eliminating the HSV information while
retaining the luminance. The RGB color space (Red, Green and Blue
which considered the primary colors of the visible light spectrum) is
converted through grayscale image to a binary image. Binary images
are images whose pixels have only two possible intensity values.
They are normally displayed as black and white. Numerically, the
two values are often 0 for black and either 1 or 255 for white. Binary
images are often produced by thresholding a grayscale or color image from the background. This conversion resulted in sharp and clear
details for the image. It is seen that the RGB color space conversion
to HSV color space then to a binary image produced images that lack
many features of the sign. So edge detection is used to identify the
Parameters of a curve that best fir a set of given edge points. Edges
are significant local changes of intensity in an image. Edges typically
occur on the boundary between two different regions in an image.
Various physical events cause intensity changes. Goal of edge detection is to produce a line drawing of a scene from an image of that
scene. Also important features can be extracted from the edges. And
these features can be used for recognition. Here canny edge detection
technique is used because it provides the optimal edge detection Solution. Canny edge detector results in a better edge detection co
pared to Sobel edge detector. The output of the edge detector defines
where features are in the image. Canny method is better, but in
some cases it provides extra details more than needed. To solve this
Problem a threshold of 0.25 is decided after testing different threshold values and observing results on the overall recognition system.

Prepared image
-----------------------------------------Image resizing

Rgb to gray conversion

Feature
extraction

Edgedetetion

Featureextraction
-----------------------------------------Feature vector

Classification

Neuralnework

Classified sign

1. Feature Extraction Methods Used.


a. Histogram Technique
b. Hough
c. OTSUs segmentation algorithm
d. Segmentation and Extraction with edge detection

III.MODIFIED SIFT ALGORITHM

B.Classification Phase

A complete description of SIFT can be found in [1].An overview of


the algorithm is presented here. The algorithm has the major stages
as mentioned below:
Scale-space extrema detection: The first stage searches over scale
space using a Difference of Gaussian function to identify potential
interest points.
Key point localization: The location and scale of each candidate
point is determined and key points are selected based on measures of
stability.
Orientation assignment: One or more orientations are assigned to
each key point based on local image gradients.
Key point descriptor: A descriptor is generated for each keypoint
from local image gradients information at the scale found in the second stage.
Each one of the above-mentioned stages is elaborated further in the
following sections.

The next important step is the application of proper feature extraction method and the next is the classification stage, a 3-layer, feedforward back propagation neural network is constructed.
The classification neural network is shown (see figure 3).
It has 256 instances as its input. Classification phase includes network architecture, creating network and training the network. Network of feed forward back propagation with supervised learning is
used.

Fig. 3: Classification network.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

39

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The goal is to design a highly distinctive descriptor for each interest point to facilitate meaningful matches, while simultaneously
ensuring that a given interest point will have the same descriptor
regardless of the hand position, the lighting in the environment,
etc. Thus both detection and description steps rely on invariance
of various properties for effective image matching. It attempts to
process static images of the subject considered, and then matches
them to a statistical database of preprocessed images to ultimately
recognize the specific set of signed letter.

ISBN: 378 - 26 - 138420 - 5

that this step can't be elim nated. In this algorithm, the orientation
is in the range [- PI, PI] radians.

D. KEYPOINT DESCRIPTORS
First the image gradient magnitudes and orientations are calculated around the key point, using the scale of the key point to select the level of Gaussian blur for the image. The coordinates of
the descriptor and the gradient orientations are rotated relative to
the key point orientation. Note here that after the grid around the
key point is rotated, we need to interpolate the Gaussian blurred
image around the key point at non-integer pixel values. We found
that the 2D interpolation in MATLAB takes much time. So, for
simplicity, we always approximate the grid around the key point
after rotation to the next integer value. By experiment, we realized that, this operation increased the speed much and still had
minor effect on the accuracy of the whole algorithm. The gradient
magnitude is weighted by a gaussian weighting function with ,
equal to one half of the descriptor window width to give less
credit to gradients far from center of descriptor. Then, these magnitude samples are accumulated into an orientation histogram
summarizing the content over 4x4 subregion. Fig. 4 describes the
whole operation. Trilinear interpolation is used to distribute the
value of each gradient sample into adjacent bins. The descriptor is
formed from a vector containing the values of all the orientation
histogram entries. The algorithm uses 4x4 array of histograms
with 8orientation bins in each resulting in a feature vector of 128
elements. The feature vector is then normalized to unit length to
reduce the effect of illumination change. The values in unit length
vector are thresholded to 0.2 and then renormalized to unit length.
This is done to take care of the effect of nonlinear illumination
changes.

A. FINDING KEYPOINTS
The SIFT feature algorithm is based upon finding locations
(called key points) within the scale space of an image which can
be reliably extracted. The first stage of computation searches over
all scales and image locations. It is implemented efficiently by
using a difference-of-Gaussian function to identify potential interest points that are invariant to scale and orientation. Key points
are identified as local maxima or minima of the DoG images
across scales. Each pixel in a DoG image is compared to its 8
neighbours at the same scale, and the 9 corresponding neighbours
at neighbouring scales. If the pixel is a local maximum or minimum, it is selected as a candidate key point.
We have a small image database, so we don't need a large number
of key points for each image. Also, the difference in scale between large and small bare hands is not so big.

Figure5: Detected key points for Image representing Y Character

B.

KEYPOINT LOCALIZATION

In this step the key points are filtered so that only stable and more
localized key points are retained. First a 3D quadratic function is
fitted to the local sample points to determine the location of the
maximum. If it is found that the extremum lies closer to a different sample point, the sample point is changed and the interpolation performed instead about that point. The function value at the
extremum is used for rejecting unstable extrema with low contrast.The DoG operator has a strong response along edges present
in an image, which give rise to unstable key points. A poorly defined peak in the DoG function will have a large principal curvature across the edge but a small principal curvature in the perpendicular direction.

C.

Figure6: Gaussian & DoG pyramids (Source: Reference 1)

ORIENTATION ASSIGNMENT

In order for the feature descriptors to be rotation invariant, an


orientation is assigned to each key point and all subsequent operations are done relative to the orientation of the key point. This
allows for matching even if the query image is rotated by any
angle. In order to simplify the algorithm, we tried to skip this part
and assume no orientation for all key points. When tested, it gave
wrong results with nearly all the images where the bare hand image is rotated with an angle of 15 to 20 or more. We realized

Figure 7: 2x2 descriptor array computed from 8x8 samples (Source: Reference 1)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

40

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

tions are P-by-4 matrix, in which each row has the 4 values for a
key-point location (row, column, scale, orientation). The orientation is in the range [-PI, PI] radians

E. SIMPLIFICATIONS TO SIFT ALGORITHM

The distance of one feature point in first image and all feature
points in second image must be calculated when SIFT algorithm
is used to match image, every feature point is 128- dimensional
data, the complexity of the calculation can well be imagined.A
changed Comparability measurement method is introduced to
improve SIFT algorithm efficiency. First, Euclidean distance is
replaced by dot product of unit vector as it is less computational.
Then, Part characteristics of 128-dimensional feature point take
part in the calculation gradually. SIFT algorithm time reduced.
Euclidean Distance is distance between the end points of the two
vectors. Euclidean distance is a bad idea because Euclidean distance is large for vectors of different lengths. This measure suffers from a drawback: two images with very similar content can
have a significant vector difference simply because one is much
longer than the other. Thus the relative distributions may be identical in the two images, but the absolute term frequencies of one
may be far larger. So the key idea is to rank images according to
angle with query images. To compensate for the effect of length,
the standard way of quantifying the similarity between two images d1 and d2 is to compute the cosine similarity of their vector
representations V (d1) and V (d2)

Fig8. SIFT Key-points Extraction, Image showing matched


keypoints between input image and database image.

Algorithmblockdiagram

Sim (d1, d2) = V (d1). V (d2) / |V (d1) ||V (d2)|


Where the numerator represents the dot product (also known as
the inner product) of the vectors V (d1) and V (d2), while the
denominator is the product of their Euclidean lengths.
F. KEYPOINT MATCHING USING UNIT VECTORS

1. Match (image1, image2). This function reads two images, finds


their SIFT [1] [6] features. A match is accepted only if its distance is less than dist Ratio times the distance to the second closest match. It returns the number of matches displayed. Where the
numerator represents the dot product (also known as the inner
product) of the vectors V (d1) and V (d2), while the denominator
is the product of their Euclidean lengths.
2. Find SIFT (Scale Invariant Fourier Transform) Key points for
each image. For finding the SIFT Key points specify what are its
locations and descriptors.
3. It is easier to compute dot products between unit vectors rather
than Euclidean distances. Note that the ratio of angles acos of dot
products of unit vectors is a close approximation to the ratio of
Euclidean distances for small angles.
4. Assume some distance ratio for example suppose distance ratio=0.5 it means that it only keep matches in which the ratio of
vector angles from the nearest to second nearest neighbour is less
than distance Ratio.
5. Now for each descriptor in the first image, select its match
To second image.
6. Compute matrix transpose, Computes vector of dot
Products, Take inverse cosine and sort reproducts, Take inverse
cosine and sort results. Check if nearest neighbour has angle less
than dist Ratio times second.
7. Then create a new image showing the two images side by side.
Using this algorithm we read image and calculate key- points,
descriptors and locations by applying threshold. Descriptors given
as P-by-128 matrix where p is number of key-points and each row
gives an invariant descriptor for one of the p key-points. The descriptor is a vector of 128 values normalized to unit length. Loca-

Now apply these steps in our previous image from which SIFT
features are extracted.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

41

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

G
H
I
J
K
L
M
N
O
P
Q
R
S
T
U
V
W
X
Y
Z
TOTAL

7
7
8
8
7
7
8
7
7
8
8
7
7
8
8
8
8
6
8
6
193

ISBN: 378 - 26 - 138420 - 5

1
1
0
0
1
1
0
1
1
0
0
1
1
0
0
0
0
2
0
2
15

66.66
66.66
100
100
66.66
66.66
100
66.66
66.66
100
100
66.66
66.66
100
100
100
100
33.33
100
33.33
92.78

Table 1 Results of training 8 samples for each sign with (0.25) Canny
Threshold

IV. EXPERIMENTAL RESULTS AND ANALYSIS

The network is trained on 8 samples of each sign. Samples of


same size and other features like distance rotation and lighting
effect and with uniform background are taken into consideration
while discarding the others.

The performance of the recognition system is evaluated by testing


its ability to classify signs for both training and testing set of data.
The effect of the number of inputs to the neural network is considered.
A.Data Set

The data set used for training and testing the recognition system
consists of grayscale images for all the ASL signs used in the
experiments are shown see fig. 4. Also 8 samples for each sign
will be taken from 8 different volunteers. For each sign 5 out of 8
samples will be used for training purpose while remaining five
signs were used for testing. The samples will be taken from different distances by WEB camera, and with different orientations.
In this way a data set will be obtained with cases that have different sizes and orientations and hence can examine the capabilities
of the feature extraction scheme.
B.Recognition Rate

The system performance can be evaluated based on its ability to


correctly classify samples to their corresponding classes. The
recognition rate can be defined as the ratio of the number of correctly classified samples to the total number of samples and can
be given as
Recognition rate =

Figure. 9 Training chart for a network trained on 8 samples for each sign,
(0.25) Canny threshold
Figure 10 Percentage error recognition chart of neural network

no.ofcorrectlyclassifiedsigns
totaL n oofsigns

C.Experimental Results

SIGN
A
B
C
D
E
F

Recognized
samples
7
7
7
8
8
8

Misclassified
samples
1
1
1
0
0
0

Ecognition
rate
66.66
66.66
66.66
100
100
100

E.

GUI Simulating Results (sign to text)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

42

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Figure 11: comparison of key-points on given input with database For a


single input image for first cycle.

For testing the unknown signs we have created a GUI as shown in


Fig. 5 which provides the user an easy way to select any sigh
He/She wants to test and then after clicking on the Apply
pushbutton it will display the meaning of the selected sign.
F.

In Figure 8, we compare Database Images 1, 3, 7 with input image key points. So Database Image 3 is closest match with input
image.

GUI SIMULATING RESULTS (TEXT TO SIGN)

A text to sign interpreter means if the user types any sign or sentence corresponding signs are shown so that normal person to
deaf people communication can be done. Examples of showing
the sign to text converters are shown (See figure 6). Fig. 6 shows
when the user type the name BOB in the text box its corresponding signs are appear on the screen one by one above the text
or spelling.

Figure 12: comparison of key-points on given input with database For a


single input image after resetting threshold value and Distance ratio value

The problem now is how we can identify a 'No Match'. For this,
we saw that the 'No Match' query images are in many cases confused with the database images that have a large number of feature vectors in the feature vectors database. We decided to compare the highest vote (corresponding to the right image) and the
second highest vote (corresponding to the most conflicting image). If the difference between them is larger than a threshold,
then there is a match and this match corresponds to the highest
vote. If the difference is smaller than a threshold, then we declare
a 'No Match'. The values of THRESHOLD were chosen by experiment on training set images either with match or no match.

The approach described above has been implemented usingMATLAB. The implementation has two aspects: training and
inference. During the training phase locally invariant features
(key points, orientation, scales and descriptors) from all training
images are retrieved using the SIFT algorithm.During inference,
the objective is to recognize a test image. A set of local invariant
features are retrieved for the test image during the inference phase
and compared against the training feature-set using the metric
explained in section 4.The title of the closest match is returned as
the final output.
In order to prove the performance of our proposed system, we
Predefined the number of gestures from B, C, H, I, L, O, Y and
create a hand gesture database. Matching is performed between
images by unit vectors. The matching is accomplished for proposed method and the result shows that it produces 98% accuracy.
In Figure 7, we can easily identify Database Images 1, 3, 7 have
more number of key-points matched with input image key-points
.So Distance Ratio
Parameter and threshold are adjusted.

Figure 13: comparison of key-points on given input with database For a


single input image (No Match Case)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

43

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

(1.5 GHz AMD processor, 128 MB of RAM running under windows 2008.) WEB-CAM-1.3 is used for image capturing.
The system is proved robust against changes in gesture.Using
Histogram technique we get the misclassified results. Hence Histogram technique is applicable to only small set of ASL alphabets
or gestures which are completely different from each other. It
does not work well for the large or all 26 number of set of ASL
signs. For more set of sign gestures segmentation method is suggested. The main problem with this technique is how good differentiation one can achieve. This is mainly dependent upon the
images but it comes down to the algorithm as well. It may be enhanced using other image
Processing technique like edge detection as done in the presenting
paper. We used the well known edge detector like Canny, Sobel
and Prewitt operators to detect the edges with different threshold.
We get good results with Canny with 0.25 threshold value. Using
edge detection along with segmentation method recognition rate
of 92.33% is achieved. Also the system is made background independent. As we have implemented sign to text interpreter reverse
also implemented that is text to sign interpreter.

Figure14: Example of a no match Image not in training set for figure3

Gesture
Name
B
C
H
I
L
O
Y

Testing Number
150
150
150
150
150
150
150

Success
Number
149
148
148
149
148
148
149

Correct Rate

The Algorithm is based mainly on using SIFT features to match


the image to respective sign by hand gesture. Some modifications
were made to increase the simplicity of the SIFT algorithm. Applying the algorithm on the training set, we found that it was always able to identify the right sign by hand gesture or to declare
No Match in case of no match condition. The algorithm was
highly robust to scale
Difference, rotation by any angle and reflection from the test image. SIFT is a state-of-the-art algorithm for extracting locally
invariant features and it gave me an opportunity to understand
several aspects of application in image recognition. I believe this
effort resulted in a robust image recognition implementation,
which should perform quite well with the final test images. In
future I would like to work on improving the performance of the
SIFT for Global Features.
The local invariant features of SIFT can be augmented by computing global features of an image.

99.3
98.7
98.7
99.3
98.7
98.7
99.3

Table2. The results of classifier for the training set and Testing set

FUTURE SCOPE / CHALLENGES

The work presented in this project recognizes ASL static signs


only. The work can be extended to be able to recognize dynamic
signs of ASL. The system deals with images with uniform background, but it can be made background independent. It is overcome and it is made Background independent. The network can
be trained to the other types of images. It is important to consider
increasing data size, so that it can have more accurate and highly
performance system.

FIGURE15 RECOGNIZED STATIC SIGN USING PCA ALGORITHM

ACKNOWLEDGMENT

The authors wish to thank to guide Prof.P.SUBBA RAO for his


valuable guidance for this work. And the volunteers who has
given gesture images in the required format no of times whenever
required till to the results.

FIGURE 16 RECOGNIZED DYNAMIC SIGN USING PCA ALGRITHM

7. CONCLUSIONS:
REFERENCES

IV. HARDWARE AND SOFTWARE

a.

The system is implemented in MATALAB version R13.5. The


recognition training and tests were run on a modern standard PC

J.S. Bridle, Probabilistic Interpretation of Feedforward Classification


Network Outputs, with Relationships to Statistical Pattern Recognition, NeurocomputingAlgorithms, Architectures and Applications, F.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

44

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

b.
c.

d.
e.
f.

g.

h.

i.

j.
k.
l.
m.

n.

o.

ISBN: 378 - 26 - 138420 - 5

Fogelman-Soulie and J. Herault, eds., NATO ASI Series F68, Berlin:


Springer-Verlag, pp. 227-236, 1989. (Book style with paper title and
editor)
W.-K. Chen, Linear Networks and Systems. Belmont, Calif.:
Wadsworth, pp. 123-135, 1993. (Book style)
Poor, A Hypertext History of Multiuser Dimensions, MUD
History,
http://www.ccs.neu.edu/home/pb/mudhistory.html. 1986. (URL link *include year)
K. Elissa, An Overview of Decision Theory," unpublished.
(Unplublished manuscript)
R. Nicole, "The Last Word on Decision Theory," J. Computer
Vision, submitted for publication. (Pending publication)
J. Kaufman, Rocky Mountain Research Laboratories, Boulder, Colo., personal communication, 1992. (Personal communication)
D.S. Coming and O.G. Staadt, "Velocity-Aligned Discrete
Oriented Polytopes for Dynamic Collision Detection," IEEE
Trans. Visualization and Computer Graphics, vol. 14, no. 1, pp.
1-12, Jan/Feb 2008, doi:10.1109/TVCG.2007.70405. (IEEE
Transactions )
S.P. Bingulac, On the Compatibility of Adaptive Controllers, Proc. Fourth Ann. Allerton Conf. Circuits and Systems
Theory, pp. 8-16, 1994. (Conference proceedings)
David G. Lowe. Distinctive Image Features from ScaleInvariant Key points. International Journal of Computer Vision, 60, 2 (2004), pp.91- 110.
Sven Siggelkow. Feature Histograms for Content-Based Image Re trieal.
PhD Thesis, Albert-Ludwigs-University Frieiburg, December
2002.
Mikolajczyk, K., Schmid, C.: An Affine Invariant Interest
Point Detector. In: ECCV, (2002) 128-142
Schaffalitzky, F., Zisserman, A.: Multi-view Matching for
Unordered Image Sets, or How Do I Organize My Holiday
Snaps? In: ECCV, (2002) 414-431
Van Gool, T. Moons, and D. Ungureanu. Affine photometric
invariants for planar intensity patterns. In ECCV, pp. 642651, 1996.
[6] D. Lowe, Object Recognition from Local Scale-Invariant Features, Proceedings of the Seventh IEEE International Conference.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

45

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

BAACK: Better Adaptive


AcknowledgementSystemfor Secure Intrusion
Detection System in Wireless MANETs
Mr. G.Rajesh M.tech

Parvase Syed

Assistant Professor, Department of CSE


Audisankara College of Engineering & Technology
Gudur, Nellore, Andhra Pradesh, India

PG Student, Computer Science & Engineering


Audisankara College of Engineering & Technology
Gudur, Nellore, Andhra Pradesh, India

network [1][2]. MANETs consist of a self-healing, peerto-peer, self-organization network in contrast to a mesh
topology has a central controller (to optimize,
determine, and distribute the routing table). MANETs
circa 2000-2015 habitually communicate at radio
frequencies (30 MHz - 5 GHz).

Abstract: -In recent years, the use of mobile


adhoc networks (MANETs) has been well-known in
various applications, including some mission acute
applications, and as such security has become one of the
most important concerns in MANETs. MANETs have
some unique characteristics, Due to that prevention
methods alone are not enough to make them secure;
therefore, detection should be added as an additional
defense before an attacker can breach the device. In
general, the intrusion detection techniques for usual
wireless networks are not well-matched for MANETs. In
this paper, A novel intrusion detection system named
Better Adaptive Acknowledgement (BAACK) especially
designed for MANETs. By the adoption of MRA scheme,
BAACK is proficient of detecting nodes in spite of the
existence of false misbehavior report and it compared with
other popular mechanisms in different scenarios. This
scenarios gives an outline for enhancing security level of
IDS architecture in MANETs based on secure attributes
and then various algorithms, namely RSA and DSA.

One of the major benefits of wireless networks


arebetween
different
parties
allows
data
communicationand still their mobility will be
maintained. However, this communicationto the range
of transmitters is partial. This means if the distance
between the two nodes is outside the range they cannot
communicate with each other. MANET answers this
problem by allowing middle parties to relay data
transmissions. This is accomplished by dividing
MANET into two kinds of networks, namely, multihop
and single-hop. In a single-hop network, Nodes can
directlycommunicate with other nodes if all nodes are
within the same radio range. On the other side, in a
multihop network, if the destination node was out of
their radio range then they take the help of intermediate
nodes to transmit or communicate with each other in a
network.

Keywords:Digital signature, MANET, DSR, AODV

I.INTRODUCTION
A mobile ad hoc network (MANET) is a
continuously
self-configuring,
infrastructureless network of mobile devices connected without
wires.In MANET devices can move in any direction
independently, and will therefore nodes frequently
change its links to other nodes. Unrelated to its use
every node must forward traffic in network, and
therefore be a router. The key challenge in constructing
a MANET is equipping each node to regularly maintain
the information required to route trafficproperly. Such
networks maybe connected to the larger Internet or
mayoperate by themselves. In between nodes contain
one or multiple and different transceivers. This results in
a highlydynamic, autonomous topology.
MANETs are a kind of Wireless ad hoc
network that usually has a routable networking
surroundings on top of a Link Layer ad hoc

Fig 1: Architecture of MANET

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

46

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

By observing above described characteristics a MANET


was mainly developed for military purpose by observing
above described characteristics, in the battle nodes are
spread across the battle field and there is no particular
infrastructure to service them form a network.In last few
years, rapidly MANETs are developed and the usage of
MANETs gradually more in many applications, ranging
from military tocommercial uses and civilian, without
the help of human interaction or any infrastructure
MANETs can setup easily by their own behavior. Some
examples are: emergency services,in collecting data,and
virtual classrooms and conferences where PDA, laptops
or other mobile devices distribute their data through
wireless medium and communicate with each other. As
MANETs become extensively used, the primary
concern about the MANETs is security issues. For
example, the proposed protocols for MANETs most of it
think that every node is not malicious and cooperative
[1]. Therefore, in entire network any one compromised
node can cause failure of the whole network.

ISBN: 378 - 26 - 138420 - 5

techniques, IDS are alsobe categorized into three


modules as follows [2].
ADS (Anomaly detection systems): The normal
behaviors (or normal profile)of operators are kept in the
system. The captured data will be compared by the
system with these profiles, and after that treats any
action that differs fromthe baseline as a possible
intrusion by initializing a proper response or informing
system administrators.
Misuse detection systems: The system keeps
configurations (or signatures)of well-known attacks and
capture data will be compared with all known attacks.A
treat which matches with any accorded pattern is
considered as intrusion;new kinds of attacks cannot be
identified by it.
Specification-based detection: The correct operation of
protocol or program was described by the system
defined constraints set.According to the defined
constraints it displays the execution of the application.
III.EXISTING SYSTEM

Digital Signature
A digital
signature or digital
signature
scheme was a type of asymmetric cryptography. For
messages sent through an insecure medium, a fine
implementation of digital signature algorithm is the one
that gives confidence to the receiver the claimed sender
only sending the messages to it, and trust the message.

The Watchdog/Path rater is anexplanation to


the problem of selfish (or misbehaving) nodes in
MANET. The system initiates two extensions to the
DSR algorithm to moderate the effects of routing
misbehavior: the Watchdog, to detect the misbehaving
nodes and the Path rater, to respond to the intrusion by
isolating the misbehaving.

In many respects Digital signatures are equivalent to


traditional handwritten signatures;Digital Signature
implemented properly, then it is more difficult to copy
than the handwritten type. By using cryptography
Digital Signature was implemented. Digital signatures
can also provide acknowledgement, meaning thata
message cannot successfully claim by the signer if they
did not sign, while claiming its private key also remains
secret.

With the presence of following Watchdog fails to


identify malicious misbehaviors
false misbehavior report,
collusion,
ambiguous collisions,
receiver collisions,
limited transmission power,
partial dropping node from the network operation

II. RELATED WORKS


III. PROPOSED SYSTEM

Many historical events have exposedthose


intrusion prevention techniquesalone, such as
authentication and encryption, which are generally a
first levelof defense, are not sufficient. As the system
turn out to be more complex, thereare also
morelimitation, which may leads to more security
issues. Intrusiondetection can be used as a second level
of defense to protect the networkfrom such issues. If the
intrusion is identified, a response can be initiatedto stop
or reduce damage to the system.

Secure IDS architecture (BAACK) initiate to


improve the security stage of MANETs based on
various algorithmsandsecurity attributes, namely DSA
and RSA. BAACK is designed to deal with three out of
six weaknesses of Watchdog IDS, namely, 1) Receiver
collision, 2) Limited transmission power, 3) False
misbehavior.
Receiver collisions:Example of receiver collisions,
shown in Fig. 2, after node X sends Packet 1 to m
node Y, it tries to overhear if node Y forwarded this
packet to node Z; meanwhile, node F is forwarding
Packet 2 to node Z. In such case, node X overhears
that node Y has successfully forwarded Packet 1 to
node Z but failed to detect that node Z did not

Intrusion detection hadclassified based on audit


data as either network-based or host-based. A networkbased IDS captures and then analyzes packets from
network traffic while application logs in its study or a
based on hosts IDS uses OS. Based on detection

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

47

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

receive this packet due to a collision between


Packet 1 and Packet 2 at node Z.

Fig 2: Receiver collisions in MANETs

Limited transmission power: Example of Limited


power, shown in Fig. 3, in order to manage the
battery resources in MANETs, node Y limits its
transmission power so it is very strong to be
overheard by node X after transmitting the packet
(P1) to node Z , but too weak to reach node Z
because of transmission power can be reduced.

ISBN: 378 - 26 - 138420 - 5

Due to the open medium and remote distribution of


typical MANETs, attackers can easily capture and
compromise one or two nodes to achieve this false
misbehavior report attack. As discussed in earlier
sections, TWOACK and AACK resolve two of
these three problems, namely, receiver collision and
limited transmission power. However, two of them
are vulnerable to the false misbehavior attack. In
order to solves not only limited transmission power
and receiver collision but also the false misbehavior
problem to launch Secure IDS architecture
(BAACK) [1].

Secure IDS description


BAACK is consisted of three major parts, namely,
ACK, secure ACK (S-ACK), and misbehavior report
authentication (MRA). Sequentially to differentiate
different packet types in different schemes to include a
2-b packet header in EAACK. In accordance with the
Internet draft of DSR [7], there is 6 b reserved in the
DSR header. In BAACK, use 2 b of the 6 b to flag
different types of packets.

Fig5:EAACK protocol in MANETs


In these secure IDS, It is assumed that the link between
each device in the network is bidirectional.Moreover,
for each communication process, both the destination
node and the source node are not malicious. All packets
of acknowledgment are required to be digitally signed
by its sender and verified by its receiver.

Fig 3: Limited Transmission power in MANET

False misbehavior: Example of false misbehavior


in MANETs, shown in Fig. 4, Even though node X
and Y forwarded Packet 1 to node Z successfully,
node X still inform node Y as misbehaving, as
shown in Fig. .

Fig 4: False Misbehavior in MANET

ACK (Acknowledgment): ACK is fundamentally an


end-to-end ACK IDS. It performs as a part of the
hybrid IDS in BAACK, aiming to cut network
overhead when no misbehavior is detected
overnetwork. Consider the scenario source first
sends out an ACK data packet to the target node D.
If all the intermediary nodes along the route
between nodes S and D are mutual cooperative and
node D successfully receives packet, required to
returnpacket with an ACK acknowledgment by
node D along the unchanged route but in aninverse
order.Within a predefined time span, if node S
receives packet, then the transmission of
packettonode D from node S is successful. If not,
node S will change to S-ACK mode by sending out
an S-ACK data packet to identify the malicious
nodes in the route.

S-ACK (Secure-Acknowledgment): It is an improved


version of the TWOACK IDS [6]. The principle of S-

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

48

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ACK is to let every three successive nodes work in a


group to identify misbehaving nodes. For every three
consecutive nodes in the route, the every third node is
required to send an S-ACK acknowledgment packet
to the first node. Introducing S-ACK mode main
intention is to identify misbehaving nodes in the
occurrence of receiver collision or limited
transmission power.

MRA (Misbehavior Report Authentication):Unlike


the TWOACK IDS, where the target node
immediately believes the misbehavior report,
BAACK requires the target node to shift to MRA
mode and approve this misbehavior report. This is a
crucial step to identify false misbehavior.
Watchdog fails to detect the misbehavior nodes
with the presence of false misbehavior because of
its limitations to resolve this MRA field is designed.
Themalicious
attackersmay
generate
false
misbehavior report to falsely report innocent nodes
as malicious. The core of MRA field is to validate
whether the target node has received the reported
missing packet through another route. To initiate
the MRA mode, the source node primarilylooks its
local knowledge base and search for an alternative
route to the target node. If there is no other that
exists, the source node finds another route by using
DSR routing request algorithm. Due to the nature of
MANETs, it enables to find out various routes
between the two nodes. If MRA packet is received
by target node, then hunts its local knowledge base
and compares the reported packet which was
already received by the target node. If they both are
matched, then it is safe to determine that this is a
false misbehavior report and the false report
generated node will be mark as malicious. Else, the
misbehavior report wasaccepted andtrusted. By the
enhancing of MRA scheme,In spite of the existence
of false misbehavior report BAACK is capable of
detecting malicious nodes.

ISBN: 378 - 26 - 138420 - 5

Secure IDS in DSA and RSA:The signature size of


DSA is much smaller than the signature size of
RSA. So the DSA scheme always produces slightly
less network overhead than RSA does. However, it
is interesting to observe that the Routing Overhead
differences between RSA and DSA schemes vary
with different numbers of malicious nodes[16]. The
more malicious nodes there are, the more ROs the
RSA scheme produces. Assume that this is due to
the fact that more malicious nodes require more
packets, thus increasing the ratio of digital signature
in the whole network overhead. With respect to this
result, find DSA as a more desirable digital
signature scheme in MANETs [1]. The reason is
that data transmission in MANETs consumes the
most battery power. Although the DSA scheme
requires more computational power to verify than
RSA, considering the tradeoff between battery
power and performance, DSA is still preferable.
IV. CONCLUSION

In this paper, a comparative study of Secure


Intrusion- Detection Systems (SIDS) for discovering
malicious nodes and attacks on MANETs is presented.
Due to some special characteristics of MANETs,
prevention mechanisms alone are not adequate to
manage the secure networks. In this case detection
should be focused as another part before an attacker can
damage the structure of the system. We study about
secure IDS named BAACK protocol specially designed
for MANETs and in future it is required to compare
against other popular mechanisms. Security is major
part in MANETs;hybrid cryptography architecture will
tackle the issue in an efficient manner. This way we can
better preserve battery and memory space of mobile
nodes.
References
[1] EAACK A Secure Intrusion Detection System for
MANETs Elhadi M. Shakshuki, Senior Member,
IEEE, Nan Kang and Tarek R. Sheltami, Member,
IEEE
[2] Investigating Intrusion and Detection Systems in
MANET and Comparing IDSs for Detecting
Misbehaving Nodes Marjan Kuchaki Rafsan,Ali
Movaghar and Faroukh Koroupi,World Academic
of Science Engineering and Technology 44 2008.
[3] L. Zhou, Z.J. Haas, Cornell Univ., Securing ad hoc
networks, IEEE Network, Nov/Dec 1999, [4]
Mishra Amitabh, Nadkarni Ketan M., and Ilyas
Mohammad,2009.Chapter 30:Security in wireless
ad- hoc networks, the handbook of Ad hoc wireless
network. CRC PRESS Publisher

Digital
Signature:
BAACK
is
an
acknowledgement-based IDS schema. All three
modules of BAACK, namely, ACK, S-ACK, and
MRA, are acknowledgment-based detection
schemes. They all rely on ACK packets to detect
misbehaviors in the network. Thus, it is very
important to confirm that all acknowledgment
packets in BAACK wereuntainted andauthentic.
Otherwise, if the attackers are veryclever to forge
acknowledgment packets, all of the three systems
will be vulnerable. To overcome this problem, need
to incorporate digital signature in secure IDS. In
order to guarantee the integrity of the IDS, in
BAACK before sending any message to destination
it must be digitally signed and verified till they are
accepted [1].

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

49

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[4] S. Marti, T. J. Giuli, K. Lai, and M. Baker,


Mitigating routing misbehavior in mobile ad hoc
networks, in Proc. 6th Annu. Int.Conf. Mobile
Comput. Netw., Boston, MA, 2000, pp. 255265.
[5] A study of different types of attacks on multicast
in mobile ad hoc networks Hoang Lan Nguyen,
Uyen
Trang
Nguyen,
Elsevier
AdHoc
Networks(2008) 32-46.
[6] D. Johnson and D. Maltz, Dynamic Source
Routing in ad hoc wireless networks, in Mobile
computing. Norwell, MA: Kluwer, 1996, ch. 5, pp.
153181.
[7] T. Sheltami, A. Al-Roubaiey, E. Shakshuki, and A.
Mahmoud, Video transmission enhancement in
presence ofmisbehaving nodes inMANETs, Int. J.
Multimedia Syst., vol. 15, no. 5, pp. 273282, Oct.
2009.
[8] K. Stanoevska-Slabeva and M. Heitmann, Impact
of mobile ad-hoc networks on the mobile value
system, in Proc. 2nd Conf. m-Bus., Vienna,
Austria, Jun. 2010
[9] A. Tabesh and L. G. Frechette, A low-power
stand- alone adaptive circuit for harvesting energy
from a piezoelectric micro power generator, IEEE
Trans. Ind. Electron., vol. 57, no. 3, pp. 840849,
Mar. 2010.
[10] Misbehavior Nodes Detection and Isolation for
MANETs OLSR ProtocolAhmed M. Abdulla,
Imane A. Saroitb, Amira Kotbb, Ali H. Afsaric a*
2010 Published by Elsevier Ltd.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

50

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

PERFORMANCE ANALYSIS OF MULTICARRIER DS-CDMA


SYSTEM USING BPSK MODULATION
Prof. P. Subbarao 1, Veeravalli Balaji2,
1
2

MSc (Engg), FIETE, MISTE, Department of ECE, S.R.K.R Engineering College, A.P, India
B.Tech (M.Tech), Department of ECE, S.R.K.R Engineering College, Bhimavarm, A.P, India

Abstract
In this paper we apply a multicarrier signalling technique to a direct-sequence CDMA system, where a data
sequence multiplied by a spreading sequence modulates multiple carriers, rather than a single carrier. The
receiver provides a correlator for each carrier, and the outputs of the correlators are combined with a maximalratio combiner. This type of signalling has the desirable properties of exhibiting a narrowband interference
suppression effect, along with robustness to fading, without requiring the use of either an explicit RAKE
structure or an interference suppression filter.We use band limited spreading waveforms to prevent selfinterference, and we evaluate system performance over a frequency selective Rayleigh channel in the presence
of partial band interference. There is no interference from the CDMA signals to the existing microwave systems.
Thus, there is no need for either a narrowband suppression filter at the receiver or a notch filter at the
transmitter. This paper specially analyses the BER performance under Rayleigh fading channel conditions of
multicarrier DS-CDMA in presence of AWGN (Additive White Gaussian Noise) using BPSK modulation for
different number of subcarrier, different number of users using MATLAB program

Keywords:CDMA, multicarrierDS-CDMA, AWGN, BER, Rayleigh fading channel


-----------------------------------------------------------***----------------------------------------------------------system has a narrowband interference suppression
1.INTRODUCTION
effect and finally, a lower chip rate is required,
since, in a multicarrier DS system with M carriers,
the entire bandwidth of the system is divided into
M (not necessarily contiguous) equi-width
frequency bands, and thus each carrier frequency is
modulated by a spreading sequence with a chip
duration which is M times as long as that of a
single-carrier system. In other words, a multicarrier
system requires a lower speed, parallel-type of
signal processing, in contrast to a fast, serial-type
of signal processing in a single carrier RAKE
receiver [6]. This, in turn, might be helpful for use
with a low power consumption device.

Direct sequence spread spectrum (DS-SS)


techniques to multiple access communications [1].
This is partly due to its multiple access capability,
robustness against fading, and anti-interference
capability.
In direct sequence spread spectrum, the stream of
information to be transmitted is divided into small
pieces, each of which is allocated across to a
frequency channel across the spectrum. A data
signal at the point of transmission is combined with
a higher data-rate bit sequence (also known as a
chipping code) that divides the data according to a
spreading ratio. The redundant chipping code helps
the signal resist interference and also enables the
original data to be recovered if data bits are
damaged during transmission.

In fact, multicarrier DS systems have already been


proposed, and these proposed techniques can be
categorized into two types, a combination of
orthogonal frequency division multiplexing
(OFDM) and CDMA or a parallel transmission
scheme[8] of narrowband DS waveforms in the
frequency domain. In the former system, a
spreading sequence is serial-to-parallel converted,
and each chip modulates a different carrier
frequency. This implies that the number of carriers
should be equal to the processing gain, and each
carrier conveys a narrowband waveform, rather
than a DS waveform. In other words, the resulting
signal has a PN coded structure in the frequency

1.1 Multicarrier DS-CDMA


In this paper, we propose a multicarrier DS SS
system [2] [3] in which a data sequence multiplied
by a spreading sequence modulates M carriers,
rather than a single carrier. The receiver provides a
correlator for each carrier, and the outputs of the
correlators are combined with a maximal-ratio
combiner. This type of system has the following
advantages: First, a multicarrier DS SS system is
robust to multipath fading, second, a multicarrier

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

51

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

domain. In the latter system, the available


frequency spectrum is divided into M equi-width
frequency bands, where M is the number of
carriers, typically much less than the processing
gain, and each frequency band is used to transmit a
narrowband DS waveform. In fact, both systems
show a similar fading mitigation effect over a
frequency selective channel. However, the latter
system requires only M adaptive gain amplifiers in
the maximal ratio combiner, which may simplify
the receiver. The system described in this paper
belongs to the second group.

multipath. A multicarrier system can be considered


as one realization of such a wideband DS system.

In a multicarrier system [9], carrier frequencies are


usually chosen to be orthogonal to each other, i.e.,
carrier frequencies satisfy the following condition:

Fig 1. (a) PSD of Single Carrier (b) PSD of


Multicarrier DS waveform

Tc

cos( t ) cos( t )dt 0, for i j.(1)


i

Fig 1. (a) Shows a band limited single-carrier


wideband DS waveform in the frequency domain,
where the bandwidth, BW1, is given by

Where

ISBN: 378 - 26 - 138420 - 5

Tc is the chip duration, i and

j are, respectively, the i thand j th carrier


frequencies, and

BW1 (1 )

i and j are arbitrary carrier

phases, respectively. This is done so that a signal in


the j th frequency band does not cause interference

In (2),

1
..(2)
Tc

0 1, and T, is the chip duration of

the single carrier system. In a multicarrier system,


we divide BW1 into M equi-width frequency bands
as shown in Fig. l(b), where all bands are disjoint.

in the correlation receiver for the i th frequency


band. However, in an asynchronous CDMA system
[5], signals from other users are no longer
orthogonal to the desired signal, even if (1) is
satisfied. In addition, orthogonality might be lost
because of multipath propagation or Doppler
frequency shift even for a synchronous system.
This implies that co-channel interference [4] in one
frequency band causes interference not only at the
output of the correlator for that frequency band, but
also in the signals out of all the other correlators. In
this paper, we use band limited multicarrier DS
waveforms to minimize such unnecessary selfinterference, and so orthogonality among carriers is
not required. Also, this signalling scheme prevents
narrowband waveforms from causing interference
to all frequency bands

Then the bandwidth of each frequency band,


BWM, is given by

BWM

BW1
M

1
=(1+ )
MTC

.(3)

2.1 TRANSMITTER
The transmitter has input a random binary sequence
representing data, and pseudo-random spreading
signature sequences are given to multiplier. We
assume that there are N chips per symbol, and that
each user has a different signature sequence. The
sequence modulates an impulse train, where the
energy per chip is E,. After passing through a chip
wave-shaping filter, the signal out of the filter
modulates the multiple carrier signals and is
transmitted. Note that the transmitter and receiver
block diagrams for the proposed multicarrier

2. SYSTEM MODEL
In recent years, several wideband CDMA systems
have been proposed either to realize an overlay
system [5], where DS CDMA waveforms are
overlaid onto existing narrowband signals to
enhance the overall capacity, or to combat

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

52

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

system can be effectively implemented by using a


DFT technique

ISBN: 378 - 26 - 138420 - 5

Hence the spectral density of such a waveform is a


sinc function squared, with first zeros at 1/TC
PN sequences are periodic sequences that have a
noise like behavior. They are generated using shift
registers, modulo-2 adders (XOR gates) and
feedback loops. The following diagram illustrates

Fig 2. Block Diagram of transmitter

2.2 Pseudo-Noise Sequences:

this:

PN sequences are periodic sequences that have a


noise like behaviour. They are generated using shift
registers, modulo-2 adders (XOR gates) and
feedback loops.

Fig 4. Generation of PN sequence


The maximum length of a PN sequence is
determined by the length of the register and the
configuration of the feedback network. An N bits
register can take up to 2N different combinations of
zeros and ones. Since the feedback network
performs linear operations, if all the inputs (i.e. the
content of the flip-flops) are zero, the output of the
feedback network will also be zero. Therefore, the
all zero combination will always give zero output
for all subsequent clock cycles, so we do not
include it in the sequence. Thus, the maximum
length of any PN sequence is 2N-1 and sequences of
that length are called Maximum-Length Sequences
or m-sequences.

So far we haven't discussed what properties we


would want the spreading signal to have. This
depends on the type of system we want to
implement. Let's first consider a system where we
want to use spread spectrum to avoid jamming or
narrow band interference.
If we want the signal to overcome narrow band
interference, the spreading function needs to
behave like noise. Random binary sequences are
such functions. They have the following important
properties:

Impulse Modulator: It is used to modulate the


impulse train coming from the multiplier, where the
energy per chip is E.

Balanced: they have an equal number of


1's and 0's
Single Peak auto-correlation function

Wave shaping filter: This filter is used to modify


the shape of the waveform i.e., it is used to adjust
the energy levels of the waveform before the
sequence modulates the multiple carriers.

In fact, the auto-correlation function of a random


binary sequence is a triangular waveform as in the
following figure, where TC is the period of one
chip:

3. CHANNEL MODEL
3.1 RAYLEIGH CHANNEL MODEL
Rayleigh fading is a statistical model [9] for the
effect of a propagation environment on a radio
signal, such as that used by wireless devices.
Rayleigh fading models assume that the magnitude
of a signal that has passed through such a
transmission medium will vary randomly, or fade,
according to a Rayleigh distribution the radial

Fig 3. Auto Correlation function

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

53

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

component of the sum of two uncorrelated


Gaussian random variables. Rayleigh fading is
viewed as a reasonable model for tropospheric and
ionospheric signal propagation as well as the effect
of heavily built-up urban environments on radio
signals. Rayleigh fading is most applicable when
there is no dominant propagation along a line of
sight between the transmitter and receiver Rayleigh
fading is a reasonable model when there are many
objects in the environment that scatter the radio
signal before it arrives at the receiver, if there is
sufficiently much scatter, the channel impulse
response will be well modelled as a Gaussian
process irrespective of the distribution of the
individual components. If there is no dominant
component to the scatter, then such a process will
have zero mean and phase evenly distributed
between 0 and 2 radians. The envelope of the
channel response will therefore be Rayleigh
distributed.

ISBN: 378 - 26 - 138420 - 5

Serial to parallel converter: At the transmitter


end N bits are sent simultaneously over N subcarriers. Each sub-carrier transmit different symbol
with spreading code in time-domain. In order
toseparate different symbol of each subscriber we
use this converter.
Maximal

ratio

combiner:

Maximal

Ratio

Combining is defined as all paths co-phased and


summed with optimal weighting to maximize
combiner output SNR MRC is the optimum linear
combining technique for coherent reception with
independent fading. Its main advantage is the
reduction of the probability of deep fades

3.2 AWGN CHANNEL MODEL


Additive White Gaussian Noise channel model [9]
as the name indicate Gaussian noise get directly
added with the signal and information signal get
converted into the noise in this model scattering
and fading of the information is not considered.
Additive white Gaussian noise (AWGN) is a
channel model in which the only impairment to
communication is a linear addition of wideband or
white noise with a constant spectral density
(expressed as watts per hertz of bandwidth) and a
Gaussian distribution of amplitude. The model does
not account for fading, frequency selectivity,
interference etc. However, it produces simple and
tractable mathematical models which are useful for
gaining insight into the underlying behaviour of a
system before these other phenomena are
considered.

Fig 6. Maximum ratio combiner


2

f 1 H ( f ) x(t )

H ( f ) df 1

.. (4)

Also, we assume that H ( f ) is band limited to W,


where W is

W ' BWM / 2 ( fi 1 f i ) / 2 ..(6)


and fi is the

i th carrier frequency. This implies

that the DS waveforms do not overlap.

4. SIMULATION RESULTS
3.3 RECEIVER:

Bit error rate (BER) of a communication system is


defined as the ratio of number of error bits and total
number of bits transmitted during a specific period.
It is the likelihood that a single error bit will occur
within received bits, independent of rate of
transmission. There are many ways of reducing
BER. In our case, we have considered the most
commonly used channel: the Additive White
Gaussian Noise (AWGN) channel where the noise
gets spread over the whole spectrum of frequencies.

Fig 5. Block Diagram of Receiver

BER has been measured by comparing the


transmitted signal with the received signal and

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

54

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

computing the error count over the total number of


bits. For any given BPSK modulation, the BER is
normally expressed in terms of signal to noise ratio
(SNR).

ISBN: 378 - 26 - 138420 - 5

Fig 9. BER for 8 user system and 6 user system


BER for BPSK modulation with Spread Spectrum thechniques
0.5
User 1 of 12 users
One user system

0.45
0.4

The figure shows simulated graph between BER


and SNR for 2 user system under Rayleigh fading
channel conditions of multicarrier DS-CDMA in
presence of AWGN (Additive White Gaussian
Noise) using BPSK modulation

Bit Error Rate

0.35
0.3
0.25
0.2
0.15

BER for BPSK modulation with Spread Spectrum thechniques


0.1

0.5
User 1 of 2 users
One user system

0.45

0.05

0.4

-50

-45

-40

-35

-30

-25
-20
SNR, dB

-15

-10

-5

Fig 10. BER for 12 user system and 10 user System

0.3
0.25

BER for BPSK modulation with Spread Spectrum thechniques

0.2

0.5

0.15

0.45

0.1

0.4

0.05

0.35

-50

-45

-40

-35

-30

-25
-20
SNR, dB

-15

-10

-5

Bit Error Rate

Bit Error Rate

0.35

User 1 of 14 users
One user system

0.3
0.25
0.2
0.15

Fig 7. BER for 2 users system and single user


system

0.1
0.05

BER for BPSK modulation with Spread Spectrum thechniques


0.5
User 1 of 4 users
One user system

0.45

-50

0.4

-45

-40

-35

-30

-25
-20
SNR, dB

-15

-10

-5

Fig 11. BER for 14 user system and 3 user system

Bit Error Rate

0.35
0.3

Computer simulations are done to simulate SNR vs.


BER performance of Multicarrier DS-CDMA for
different channel noise conditions

0.25
0.2
0.15
0.1

CONCLUSION

0.05

-50

-45

-40

-35

-30

-25
-20
SNR, dB

-15

-10

-5

In this paper the performance of Multicarrier DSCDMA in AWGN channel and Rayleigh channel
Using BPSK modulation technique is considered. It
can be evident that as Eb/N0 increases the BER
decreases. BER vs. SNR graphs for different
number of users under Rayleigh Fading channel in
presence AWGN is plotted and analysed
successfully by using MATLAB programming.

Fig 8. BER for 4 user system and single user 2


system
BER for BPSK modulation with Spread Spectrum thechniques
0.5
User 1 of 8 users
One user system

0.45
0.4

Bit Error Rate

0.35

FUTURE SCOPE:

0.3
0.25

Multicarrier DS-CDMA technology is more useful


for 3G and 4G mobile generations. There is a very
wide scope for future scholars to explore this area
of research in the field of Multicarrier DS-CDMA.

0.2
0.15
0.1
0.05

-50

-45

-40

-35

-30

-25
-20
SNR, dB

-15

-10

-5

Further works can be carried out

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

55

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

1. To evaluate the effect of timing jitter in a MC


DS CDMA system in the presence of fading.
2. To evaluate the performance of MC DS CDMA
system with Rake receiver to overcome the effect
of fading.
3. To find the performance limitations due to nonorthogonality among the subcarriers (due to
imperfect carrier synchronization).
4. To evaluate the performance improvement with
forward error correction coding like convolution
coding and Turbo coding etc.

REFERENCES
[1] G. L. Turin, Introduction to spread-spectrum
antkultipath techniques and their application to
urban digital radio, Proc. IEEE, vol. 68.
[2] R. E. Ziemer and R. L. Peterson, Digital
Communications and Spread Spectrum Systems.
New York Macmillan.
[3] R. L. Pickholtz, D. L. Schilling, and L. B.
Milstein,
Theory
of
spreadspectrum
communications-A
tutorial,
IEEE
Trans.
Commun., vol. COM-30, no. 5,.
[4]S. Kondo and L. B. Milstein, Multicarrier
CDMA system with cochanne1 interference
cancellation, in Proc. VTC 94, Stockholm,
Sweden, pp. 164C-1644.
[5] Multicarrier DS CDMA systems in the
presence of partial band interference, in Proc.
MILCOM , Fort Monmouth, NJ.
[6] J.Proakis, Digital Communications, McGrawHill
[7]R. E. Ziemer and R. L. Peterson, Digital
Communications and Spread Spectrum Systems,
Macmillan
[8] M. Schwartz, W. R. Bennet, and S. Stein,
Communication Systems and Techniques, McGrawHill
[9] G. Brindha, Performance Analysis Of McCdma System Using Bpsk Modulation,
International Journal Of Research In Engineering &
Technology (Ijret) Vol. 1, Issue 1, 45-52

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

56

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Iterative MMSE-PIC Detection Algorithm for


MIMO OFDM Systems
Gorantla Rohini Devi

K.V.S.N.Raju

Department of ECE
SRKR Engineering College
Bhimavaram, AP, India
[email protected]

Head of ECE Department


SRKR Engineering College
Bhimavaram, AP, India
[email protected]

Buddaraju Revathi
Asst. Professor, Department of ECE,

SRKR Engineering College


Bhimavaram, AP, India
[email protected]

the space dimension to improve wireless system capacity,


range, and reliability. MIMO system can be employed to
transmit several data streams in parallel at the same time and
on the same frequency but different transmit antennas.
MIMO systems arise in many modern communication
channels such as multiple user communication and multiple
antenna channels. It is well known that the use of multiple
transmit and receive antennas promises sub performance gains
when compared to single antenna system. The combination
MIMO-OFDM system is very natural and beneficial since
OFDM enables support of more antennas and large bandwidth
since it simplifies equalization in MIMO systems. In MIMOOFDM system offers high spectral efficiency and good
diversity gain against multipath fading channels [2][3].
In MIMO system depends on the different detection
techniques used at the MIMO receiver. The better detector that
minimizes the bit error rate (BER) is the maximum likelihood
(ML) detector. But the ML detector is practically difficult as it
has computational complexity is exponential. On the other
hand, linear detectors, such as zero-forcing (ZF) and minimum
mean square error (MMSE) receivers, have low decoding
complexity, but detection performance decrease in portion to
the number of transmit antennas.
Therefore, there has been a study on a low complexity
nonlinear receiver, namely, parallel interference cancellation
(PIC) receiver, which parallely decodes data streams through
nulling and cancelling. PIC algorithm [4] relies on a parallel
detection of the received block. At each step all symbols are
detected by subtracted from the received block. PIC detection
is used to reduce the complexity and prevents error
propagation. The PIC detection uses the reconstructed signal
to improve the detection performance by using iteration
process. Iterative MMSE-PIC detection algorithm [5][6] best
detection technique compared all nonlinear receivers. For
improving the performance of overall system, the output of
detector is regarded as input of the PIC detection to do again.
By exchanging information between the MIMO detection and
decoder, the performance of receiver may greatly be enhanced.
Where number of iteration increases to improve the bit error
rate (BER) performance.
PIC introduces parallely, which enables to reduce the
interference and therefore increases the reliability of the
decision process. The channel as a flat fading Rayleigh
multipath channel and the modulation as BPSK has been
taken. MIMO-OFDM technology has been investigated as the
infrastructure for next generation wireless networks.

Abstract- Wireless communication systems are required to


provide high data rates, which is essential for many services such
as video, high quality audio and mobile integrated services. When
data transmission is affected by fading and interference effects
the information will be altered. Multiple Input Multiple Output
(MIMO) technique is used to reduce the multipath fading.
Orthogonal Frequency Division Multiplexing (OFDM) is one of
the promising technologies to mitigate the ISI. The combination
of MIMO-OFDM systems offers high spectrum efficiency and
diversity gain against multipath fading channels. Different types
of detectors such as ZF, MMSE and PIC, Iterative PIC. These
detectors improved the quality of received signal in high
interference environment. Implementations of these detectors
verified the improvement of the BER v/s SNR performance.
Iterative PIC technique give best performance in noise
environment compared to ZF, MMSE and PIC.
Keywords: Orthogonal Frequency Division Multiplexing (OFDM),
Multiple Input Multiple Output (MIMO), Zero Forcing (ZF),
Minimum Mean Square Error (MMSE), Parallel Interference
Cancellation (PIC), Bit Error Rate (BER), Signal to Noise Ratio
(SNR), Inter Symbol Interference (ISI), Binary Phase Shift
Keying (BPSK).
I. INTRODUCTION

In wireless communication the signal from a transmitter will


be transmitted to a receiver along with a number of
different paths, collectively referred as multipath. These
paths may causes interference from one another and result in
the original data being altered. This is known as Multipath
fading. Furthermore wireless channel suffer from co-channel
interference (CCI) from other cells that share the same
frequency channel, leading to distortion of the desired signal
and also low system performance. Therefore, wireless system
must be designed to mitigate fading and interference to
guarantee a reliable communication.
High data rate wireless systems with very small symbol
periods usually face unacceptable Inter Symbol Interference
(ISI) originated from multi-path propagation and their inherent
delay spread. Orthogonal Frequency Division Multiplexing
(OFDM) has emerged as one of the most practical techniques
for data communication over frequency-selective fading
channels into flat selective channels. OFDM is one of the
promising technologies to mitigate the ISI. On the other hand,
to increase the spectral efficiency of wireless link, MultipleInput Multiple-Output (MIMO) systems [1]. It is an antenna
technology that is used both in transmitter and receiver
equipment for wireless radio communication. MIMO exploit

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

57

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

II. SYSTEM MODEL

A. Spatial Multiplexing

Consider a MIMO OFDM system with transmitting and


receiving antennas. When the MIMO technique of spatial
multiplexing is applied encoding can be done either jointly
over the multiple transmitter branches.

U
s
e
r

The transmission of multiple data stream over more than one


antenna is called spatial multiplexing. It yields linear (In the
minimum number of transmit and receive antenna) capacity
increases, compared to systems with a single antenna at one or
both sides of the wireless link, at no additional power or
bandwidth expenditure. The corresponding gain is available if
the propagation channel exhibits rich scattering and can be
realized by the simultaneous transmission of independent data
stream in the same frequency band. The receiver exploits
difference in the spatial signature induced by the MIMO
channel onto the multiplexed data stream to separate the
different signals, there by realizing a capacity gain.

X1
Modulation

IFFT

ISBN: 378 - 26 - 138420 - 5

MIMO
Channel

X2
XL

B. Diversity Schemes
Y1 Y2

YL

In which two or more number of signals sent over different


paths by using multiple antennas at the transmitting and
receiving side. The space is chosen, in such a way the
interference between the signals can be avoided. To improve
the link reliability we are using diversity schemes. Spatial
diversity improves the signal quality and achieves higher
signal to noise ratio at the receiver side. Diversity gain is
obtained by transmitting the data signal over multiple
independently fading dimensions in time, frequency, and space
and by performing proper combing in the receiver. Spatial
diversity is particularly attractive when compared to time or
frequency diversity, as it does not incur expenditure in
transmission time or bandwidth. Diversity provides the
receiver with several (ideally independent) replicas of the
transmitted signal and is therefore a powerful means to combat
fading and interference and there by improve link reliability.

P-element

U
s
e
r

Demodulation

ZF/MMSE/
PIC/
Iterative
PIC

F
F
T

Receiver
antenna
array

Fig1.Schematic of PIC detection for MIMO OFDM system

According to the block diagram in Figure1 consists of two


users, one user source while the other user as destination. The
two users interchange their information as source to different
instant of time. In MIMO channel model, L simultaneous
antennas having same data for transmission, while receiver has
P antennas.
The binary data are converted into digitally modulated signal
by using BPSK modulation technique and after that converted
from serial to parallel through convertor. The digitally
modulated symbols are applied to IFFT block. After the
transformation, the time domain OFDM signal at the output of
the IFFT. After that, Cyclic Prefix (CP) is added to mitigate
the ISI effect. This information is sent to parallel to serial
convertor and again, the information symbols are
simultaneously transmitted over the MIMO channel and later
AWGN noise added at receiver side.
At the receiver side, firstly serial to parallel conversion occurs
and cyclic prefix removed. The received signals samples are
sent to a fast Fourier transform (FFT) block to demultiplex the
multi-carrier signals and ZF / MMSE / PIC / Iterative-PIC
detectors is used for separating the user signals at each element
of the receiver antenna array. Finally demodulated outputs and
the resulting data combined to obtain the binary output data.

Two kinds of spatial diversities are considered, Transmitter


diversity and Receiver diversity. There are two famous space
time coding schemes. Space time block code (STBC) and
Space time trellis code (STTC).
III. PROPOSED DETECTION ALGORITHM FOR
MIMO-OFDM SYSTEMS
The co-channel interference is one of the major limitations in
cellular telephone network. In the case of cellular network
such as 3G or beyond 3G (4G), the co-channel interference is
caused by the frequency reuse. Our main idea is to reject the
co- channel interference in MIMO-OFDM cellular systems.
To eliminate the inter symbol interference (ISI) different types
of highly interference channel equalization techniques are
used. MIMO-OFDM detection method consists of linear and
nonlinear detection methods. Linear equalizers are ZF [7] and
MMSE [8] and nonlinear equalizers are PIC and Iterative PIC.
1.

Zero Forcing (ZF) equalizer:

Zero forcing Equalizer is a linear equalization algorithm used


in communication systems, it inverse the frequency response
of the channel. The output of the equalizer has an overall
response function equal to one of the symbol that is being
detected and an overall zero response for the other symbols. If
possible, this results in the removal of the interference from all
other symbols in the absence of the noise.
Zero Forcing is a linear method that does not consider the
effects of noise. In fact, the noise may be enhanced in the
process of eliminating the interference.

MIMO Techniques:
Current MIMO system includes MISO and SIMO system that
uses MIMO technique to improve the performance of wireless
system can be divided into two kinds. One is spatial
multiplexing which provides a linear capacity gain in relation
to the number of transmitting antenna and the other is spatial
diversity schemes which can reduce the BER and improve the
reliability of wireless link.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

58

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Consider a 2x2 MIMO system. The received signal on the first


antenna is given by:

y1 h1,1 x1 h1,2 x2 n1 h1,1

x
h1,2 1 n1
x2

estimator quality. The most important characteristic of MMSE


equalizer is that it does not usually eliminate ISI totally but
instead of minimizes the total power of the noise and ISI
components in the output. If the mean square error between
the transmitted symbols and the outputs of the detected
symbols, or equivalently, the received SNR is taken as the
performance criteria, the MMSE detector [9] is the optimal
detection that seeks to balance between cancelation of the
interference and reduction of noise enhancement.

(1)

The received signal on the second antenna is given by:

y2 h2,1 x1 h2,2 x2 n2 h2,1

x
h2,2 1 n2
x2

(2)

The received signal on the first receive antenna is,

Where,
y1 and y2 are the received symbol on the first and second
antenna, h1,1 is the channel from 1st transmit antenna to 1st
receive antenna, h 1,2 is the channel from 1st transmit antenna to
2nd receive antenna, h2,1 is the channel from 2nd transmit
antenna to 1st receive antenna, h2,2 is the channel from 2nd
transmit antenna to 2nd receive antenna, x1, and x2 are the
transmitted symbols and n1 and n2 are the noise on 1st and 2nd
receive antennas respectively.
The sampled baseband representation of signal is given by:

x
y1 h1,1 x1 h1,2 x2 n1 h1,1 h1,2 1 n1
x2

H 1, 2

H 2 ,2

H N R ,2

H 1, N T

H 2 ,NT

H N R , N T

(8)

The received signal on the second antenna is,

x1
y2 h2,1x1 h2,2 x2 n2 h2,1 h2,2 n2
x2

(9)

Where,
y1, y2 are the received symbol on the 1st and 2nd antenna
respectively, h1,1 is the channel from 1st transmit antenna to 1st
receive antenna, h1,2 is the channel from 1st transmit antenna to
2nd receive antenna, h2,1 is the channel from 2nd transmit
antenna to 1st receive antenna, h2,2 is the channel from 2nd
transmit antenna to 2nd receive antenna, x1, x2 are the
transmitted symbols and n1, n2 is the noise on 1st , 2nd receiver
antennas.
The above equation can be represented in matrix notation as
follows:

y= Hx+n
(3)
Where,
y = Received symbol matrix,
H = Channel matrix,
x = Transmitted symbol matrix,
n = Noise matrix.
For a system with NT transmit antennas and NR receiver
antennas, the MIMO channel at a given time instant may be
represented as NT x NR matrix:

H 1,1

H 2 ,1
H

H N R ,1

ISBN: 378 - 26 - 138420 - 5

y1 h1,1 h1,2 x1 n1
y h h x n
2 2,1 2,2 2 2

(4)

Equivalently,

(10)

y = Hx+n

To solve for x, we find a matrix W which satisfies WH = I.


The Zero Forcing (ZF) detector for meeting this constraint is
given by,
W = (HHH)-1 HH
(5)

To solve for x, we know that we need to find a matrix W


which satisfies WH=I. The Minimum Mean Square Error
(MMSE) linear detector for meeting this constraint is given
by,

Where,
W= Equalization matrix
H= Channel matrix
This matrix is known as the pseudo inverse for a general m x n
matrix where

W=[HHH+NoI]-1HH

h*
H 1*,1
h1 , 2

h 2* ,1 h1 ,1

h 2* , 2 h 2 ,1

h1 , 2

h2 , 2

Using MMSE equalization, the receiver can obtain an estimate


of the two transmitted symbols x1, x2, i.e.

x1
H
-1 H
x = (H H+N0I) H
2

(6)

y1
y
2

(12)

3. Parallel Interference Cancellation (PIC):

It is clear from the above equation that noise power may


increase because of the factor (HHH)-1. Using the ZF
equalization approach, the receiver can obtain an estimate of
the two transmitted symbols and x1 and x2 i.e.

x1
H
-1 H y 1
x = (H H) H y
2
2

(11)

Here the users symbols are estimated in a parallel manner.


This detects all layers simultaneously by subtracting
interference from other layers regenerated by the estimation
from ZF or MMSE criteria.
PIC detection is used to reduce the complexity and prevents
error propagation. The parallel MMSE detector consists of two
or more stages. The first stage gives a rough estimation of
substreams and the second stage refines the estimation. The
output can also be further iterated to improve the performance.
The first stage will be implemented by using either ZF or
MMSE detection algorithm. The MMSE detector minimizes

(7)

2. Minimum Mean Square Error (MMSE) Equalizer:


A MMSE estimator is a method in which it minimizes the
mean square error (MSE), which is a universal measure of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

59

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

the mean square error between the actually transmitted


symbols and the output of the linear detector is
W=[HHH+No I]-1 HH

IV. SIMULATION RESULTS


In all simulation results shown by using four equalizers (ZF,
MMSE, PIC and Iterative PIC) in MIMO OFDM system.
Rayleigh fading channel is taken and BPSK modulation
scheme was used. Channel
estimation
as
well
as
synchronization is assumed to be ideal. We analyze the BER
performance of data transmission in Matlab software.

(13)

By using MMSE detector the output of the first stage is


d = Dec(W.y)

ISBN: 378 - 26 - 138420 - 5

(14)

Where, W is the parameter of Equalization matrix which is


assumed to be known and Dec(.) is the decision operation. In
each a vector symbol is nulled.
This can be written as
S=I.d

(15)

Where, I is identity matrix and d is rough Estimated symbols


of MMSE.
The PIC detection algorithm can be expressed as
R= y-H.S

(16)

Hence S is the estimated symbols of MMSE Equalizer. The


estimated symbol using the detection scheme of the
appropriate column of the channel matrix
Z= Dec (W.R)

(17)
Fig. 2. BER for BPSK modulation with ZF and MMSE equalizers in 2x2
MIMO-OFDM system.

Where,
R is the output of PIC Equalizer
W is the parameter of MMSE Equalization matrix
Z is the estimated symbols of PIC Equalizer

From the plot it is clear that 2x2 MIMO-OFDM system with


MMSE equalizer for case of pure equalization compared to ZF
equalizer. Modulation scheme employed here is BPSK.

4. Iterative PIC detection:


In which, the estimated signal by decoder is used to
reconstruct the transmitted code signal. The PIC detection uses
the reconstructed signal to improve the detection performance
by using iterative process.
PIC cancellation estimates and subtract out all the interference
for each user in parallel in order to reduce the time delay. At
iteration process the output of PIC detector is given it as input.
Combing MMSE detection with the PIC cancellation directly
impacts on the global performance of the systems and also on
the associated complexity. The complexity directly linked with
the number of iterations for the detection.
The Iterative PIC detection scheme based on MIMO system
algorithm is given by:

For i = 1: nT
nT - 1
c = y- H (: , J). Z
j=1
E = Dec (W. c)

Fig. 3. Performance comparison of PIC and Iterative PIC equalizers in 2x2


MIMO-OFDM system.

(18)
From the plot it is clear that 2x2 MIMO-OFDM system with
Iterative PIC equalizer for case of pure equalization compared
to PIC equalizer. The code BER of proposed scheme is
produced after iteration. when iteration increases the BER is
significantly improved. From simulation results the proposed
scheme Iterative PIC is quite effective compared to PIC.
Modulation scheme employed here is BPSK.

Where,
E is the estimation of transmitted symbols of iterative PIC
detector,
W is the MMSE equalization matrix,
c is the output of iterative PIC detector,
nT is the number of transmitting antennas.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

60

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

V. CONCLUSION
The combination of MIMO-OFDM systems are used to
improve the spectrum efficiency of wireless link reliability in
wireless communication systems. Iterative PIC scheme for
MIMO OFDM systems transmission including the feasibility
of using the priori information of the transmit sequence of
MMSE compensation. Performance of Iterative PIC detection
technique is better compared to ZF, MMSE, PIC using BPSK
modulation scheme in high interference environment. The
simulation result shows that the performance of proposed
scheme is greatly improved compared to other detection
receivers for MIMO-OFDM systems.
VI. FUTURE SCOPE
Any type of modulation techniques such as QPSK or QAM
will integrate the channel encoding part.

REFERENCES

Fig. 4. Performance comparison of ZF, PIC and Iterative PIC equalizers in


2x2 MIMO-OFDM system.
[1]

From the plot it is clear that 2x2 MIMO-OFDM system with


Iterative PIC equalizers for case of pure equalization compared
to ZF, MMSE, and PIC equalizer. The code BER of proposed
scheme is produced after iteration. when iteration increases the
BER is significantly improved. The Zero Forcing equalizer
removes all ISI and is ideal only when the channel is noiseless.
From simulation results the proposed scheme Iterative PIC is
quite effective compared to ZF and PIC. Modulation scheme
employed here is BPSK.

[2]

[3]

[4]

[5]

[6]

[7]

[8]

[9]

Fig .5. Performance comparison of ZF, MMSE, PIC and Iterative PIC
equalizers in 2x2 MIMO-OFDM system .

From the plot it is clear that 2x2 MIMO-OFDM system with


Iterative PIC equalizers for case of pure equalization compared
ZF, MMSE, and PIC equalizer. The code BER of proposed
scheme Iterative PIC is produced after iteration. when iteration
increases the BER is significantly improved. From simulation
results the proposed scheme is quite effective in all simulation
configurations. However, Iterative PIC detection scheme is
better in the diversity gain and when the intefrence comes
from the other layers is completely cancelled. Modulation
scheme employed here is BPSK.

I. E. Telatar, Capacity of multiple-antenna Gaussian


channels, Eur. Trans. Telecommun., vol. 10, no. 6, pp.
585595, Nov/Dec. 1999.
G. J. Foschini and M. J. Gans, On limits of wireless
communications in a fading environment when using
multiple antennas, Wirel. Pers. Commun., vol. 6, no. 3,
pp. 311335, Mar. 1998.
A. Paulraj, R. Nabar, and D. Gore, Introduction to Space
Time Wireless Communications, 1st ed. Cambridge, U.K.:
Cambridge Univ. Press, 2003
Junishi Liu, Zhendong Luo,Yuanan Liu, MMSEPIC
MUD for CDMA BASED MIMO OFDM System, IEEE
Transaction Communication., vol.1, oct.2005 .
Hayashi,H.Sakai , Parallel Interference Canceller with
Adaptive MMSE Equalization for MIMO-OFDM
Transmission, France telecom R&D Tokyo.
Z.Wang, Iterative Detection and Decoding with PIC
Algorithm
for
MIMO
OFDM
System
,Int.J.communication, Network and System Science,
published august 2009.
V.JaganNaveen, K.MuraliKrishna, K.RajaRajeswari
"Performance analysis of equalization techniques for
MIMO systems in wireless communication" International
Journal of Smart Home, Vol.4, No.4, October, 2010
Dhruv Malik, Deepak Batra "Comparison of various
detection algorithms in a MIMO wireless communication
receiver" International Journal of Electronics and
Compute Science Engineering, Vol.1, No 3, page no16781685.
J.P.Coon and M. A. Beach, An investigation od MIMO
single-carrier frequency-domain MMSE equalizer in
Proc London comm.Symposium, 2002,pp. 237-240.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

61

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Computational Performances of OFDM using


Different Pruned Radix FFT Algorithms
Alekhya Chundru

P.Krishna Kanth Varma

M.tech Student,
Department Of Eelectronics and Communications,
SRKR Engineering College,
ANDHRA PRADESH, INDIA.
Email: [email protected]

Asst Professor,
Department Of Eelectronics and Communications,
SRKR Engineering College,
ANDHRA PRADESH, INDIA.
Email: [email protected]

Abstract- The Fast Fourier Transform (FFT) and its


inverse (IFFT) are very important algorithms in signal
processing, software-defined radio, and the most
promising modulation technique i.e. Orthogonal
Frequency Division Multiplexing (OFDM). From the
standard structure of OFDM we can find that
IFFT/FFT modules play the vital role for any OFDM
based transceiver. So when zero valued inputs/outputs
outnumber nonzero inputs/outputs, then general
IFFT/FFT algorithm for OFDM is no longer efficient in
term of execution time. It is possible to reduce the
execution time by pruning the FFT. In this paper we
have implemented a novel and efficient input zero
traced radix FFT pruning (algorithm based on radix-2
DIF FFT, radix-4 DIF FFT, radix-8 DIF FFT). An
intuitive comparison of the computational complexity of
orthogonal frequency division multiplexing (OFDM)
system has been made in terms of complex calculations
required using different radix Fast Fourier transform
techniques with and without pruning. The different
transform techniques are introduced such as various
types of Fast Fourier transform (FFT) as radix-2 FFT,
radix-4 FFT, radix-8 FFT, mixed radix 4/2, mixed radix
8/2 and split radix 2/4. With intuitive mathematical
analysis, it has been shown that with the reduced
complexity can be offered with pruning, OFDM
performance can be greatly improved in terms of
calculations needed.

FFT's spectrum resolution and computational time


consumption limits its application. To match with the
order or requirement of a system, the common
method is to extend the input data sequence x(n) by
padding number of zeros at the end of it and which is
responsible for a increased value of computational
time. But calculation on undesired frequency is
unnecessary. As the OFDM based cognitive radio [2]
has the capability to nullify individual sub carriers to
avoid interference with the licensed user. So, that
there could be a large number of zero valued
inputs/outputs compare to non-zero terms. So the
conventional radix FFT algorithms are no longer
efficient in terms of complexity, execution time and
hardware architecture. Several researchers have
proposed different ways to make FFT faster by
pruning the conventional radix FFT algorithms.
In this paper we have proposed an input zero
traced radix DIF FFT pruning algorithm for different
radix FFT algorithms, suitable for OFDM based
transceiver. The computational complexity of
implementing radix-2, radix-4, radix-8, mixed radix
and split radix Fast Fourier Transform with and
without pruning has been calculated in an OFDM
system and compared their performance. Result
shows IZTFFTP of radix algorithms are more
efficient than without pruning.

Index terms- OFDM (Orthogonal frequency division


multiplexing), Fast Fourier Transform (FFT), Pruning
Techniques, MATLAB.

I.

II.

INTRODUCTION

OFDM SYSTEM MODEL

OFDM is a kind of FDM (Frequency Division


Multiplexing) technique in which we divide a data
stream into a number of bit streams which are
transmitted through sub-channels [3].
The characteristics of these sub-channels are that
they are orthogonal to each other. As the data that are
transmitted through a sub-channel at a particular time
are only a portion of the data transmitted through a
channel so bit rate in a sub-channel can be kept much
low. After splitting the data in N parallel data streams
each stream is then mapped to a tone at a unique
frequency and combined together using the Inverse

Orthogonal Frequency Divisional Multiplexing


(OFDM) is a modulation scheme that allows digital
data to be efficiently and reliably transmitted over a
radio channel, even in multi-path environments [1].
In OFDM system, Discrete Fourier Transforms
(DFT)/Fast Fourier Trans- forms (FFT) are used
instead of modulators. FFT is an efficient tool in the
fields of signal processing and linear system analysis.
DFT isn't generalized and utilized widely until FFT
was proposed. But the inherent contradiction between

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

62

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Fast Fourier Transform (IFFT) to yield the time


domain waveform to be transmitted [4]. After IFFT is
done, the time domain signals are then converted to
serial data and cyclic extension is added to the signal.
Then the signal is transmitted. At the receiving side
we do the reverse process to get original data from
the received one [4,5].
In case of deep fade, several symbols in single
carrier is damaged seriously, but in parallel
transmission each of N symbol is slightly affected. So
even though the channel is frequency selective, the
sub-channel is flat or slightly frequency selective.
This is why OFDM provide good protection against
fading [6].
In an OFDM system there are N numbers of subchannels. If N is high then it will be very complex to
design a system with N modulators and
demodulators. Fortunately, it can be implemented
alternatively using DFT/FFT to reduce the high
complexity. A detailed system model for OFDM
system is shown in Figure 1 [5,6].

ISBN: 378 - 26 - 138420 - 5

fast
Fourier
transform
(FFT)
algorithms.
Decomposing is an important role in the FFT
algorithms. There are two decomposed types of the
FFT algorithm. One is decimation-in-time (DIT), and
the other is decimation-in-frequency (DIF). There is
no difference in computational complexity between
these two types of FFT algorithm. Different Radix
DIF algorithms we used are
A. Radix-2 DIF FFT Algorithm
Decomposing the output frequency sequence
X[k] into the even numbered points and odd
numbered points is the key component of the Radix-2
DIF FFT algorithm [6]. We can divide X[k] into 2r
and 2r+1, then we can obtain the following equations

[2 ] =

[ ]

[2 + 1] =

[ ]

= 0,1,2, . . ,

(1)

(2)

Because the decomposition of the Equation (1)


and Equation (2) are the same, we only use Equation
(1) to explain as shown in Equation (3).

[2 ] =

[ ]

[ ]

(3)

Finally, by the periodic property of twiddle


factors, we can get the even frequency samples as

[2 ] =

( [ ] + [ + /2])

= 0,1,2, . . ,

(4)

Similarly, the odd frequency samples is

Figure1: OFDM System Model


III. FOURIER TRANSFORM ALGORITHM

[2 + 1] =

Discrete Fourier Transform (DFT) computational


complexity is so high that it will cause a long
computational time and large power dissipation in
implementation. Cooley and Tukey provided a lot of
ways to reduce the computational complexity. From
that, many fast DFT algorithms have been developing
to reduce the large number of the computational
complexity, and these fast DFT algorithms are named

[ ]

= 0,1,2, . . ,
1
(5)
From Equation (4) and (5), we can find out the same
components, x[n] and x[n+N/2], so we can combine
the two equations as one basic butterfly unit shown in
Figure 2. The solid line means that x[n] adds x[n + N

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

63

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

/ 2] , and the meaning of the dotted line is that x[n]


subtracts x[n + N / 2] .

Figure 3: The simplified butterfly signal flow graph


of radix-4 DIF FFT
Figure 2: The butterfly signal flow graph of radix-2
DIF FFT

This algorithm results in (3/8)N log


complex
multiplications and (3/2)N log
complex additions.
So the number of multiplications is reduced by 25%,
but the number of addition is increased by 50%.

We can use the same way to further decompose


N-point DFT into even smaller DFT block. So from
the radix-2 dif FFT, there is a reduction of number of
multiplications, which is about a factor of 2, showing
the significance of radix-2 algorithm for efficient
computation. So this algorithm can compute N-point
FFT in N/2 cycles.

C.Radix-8 DIF FFT


Comparing with the conventional radix-2 FFT
algorithm and radix-4 FFT algorithm, the advantage
of developing radix-8 FFT algorithm is to further
decrease the complexities, especially the number of
complex multiplications in implementation. We can
split Equation (2.1) and replace index k with eight
parts, including 8r, 8r+1,8r+2, 8r+3, 8r+4, 8r+5,
8r+6, and 8r+7. Hence, we can rewrite Equation (6)
and obtain the Equation (9).

B.Radix-4 DIF FFT


In case N-data points expressed as power of 4M,
we can employ radix-4 algorithm [9] instead of radix2 algorithm for more efficient estimation. The FFT
length is 4M, where M is the number of stages. The
radix-4 DIF fast Fourier transform (FFT) expresses
the DFT equation as four summations then divides it
into four equations, each of which computes every
fourth output sample. The following equations
illustrate radix-4 decimation in frequency.
( )=

( )

( )

(8 + ) =

( )

( )

[ ]
+2
+4

+

+
+6

+ 8
+
+3 8
+
+
+5 8
+7 8
+

(6)

( )

(7)

8
8
8

(9)
The butterfly graph can be simplified as shown in
Figure 4

Equation (7) can thus be expressed as

( )=

( ) + ( ) ( + 4) + (1)
( + 2) + ( ) ( + 3 4)

(8)
So, Equation (8) can then be expressed as four N/ 4 point
DFTs. The simplified butterfly signal flow graph of

Figure 4: The simplified butterfly signal flow graph


of radix-8 DIF FFT

radix-4 DIF FFT is shown in Figure 3.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

64

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

D. Mixed radix DIF FFT


There are two kinds of mixed-radix DIF FFT
algorithms. The first kind refers to a situation arising
naturally when a radix-q algorithm, where q = 2m > 2,
is applied to an input series consisting of N = 2k qs
equally spaced points, where1 k < m. In this case,
out of necessity, k steps of radix-2 algorithm are
applied either at the beginning or at the end of the
transform, while the rest of the transform is carried
out by s steps of the radix-q algorithm.

ISBN: 378 - 26 - 138420 - 5

complexities. If we use radix-2 DIF FFT algorithm


for the even frequency terms and the radix-22 DIF
FFT algorithm for the odd parts, we can obtain the
split-radix 2/4 algorithm [10,11] as shown in the
equation
in the Equation (10).
[2 ] =

[ ]+

(10)


= 0,1,2, . . ,

For example if N = 22m+1 = 2 4m, the mixedradix algorithm [7][8] combines one step of the
radix-2 algorithm and m steps of the radix-4
algorithm. The second kind of mixed-radix
algorithms in the literature refers to those specialized
for a composite N = N0 N1 N2 ... Nk. Different
algorithms may be used depending on whether the
factors satisfy certain restrictions. Only the 2 4m of
the first kind of mixed-radix algorithm will be
considered here.
The mixed-radix 4/2 butterfly unit is shown in
Figure5.

(4 + 1) =
( ) + ( ) ( + 4)
+(1) ( + 2) + ( ) ( + 3 4)

(11)
(4 + 3) =
( ) + ( ) ( + 4)
( + 2) + ( ) ( + 3 4)
(12)

Thus the N-point DFT is decomposed into one N/2 point DFT without additional twiddle factors and two
N/4 -point DFTs with twiddle factors. The N-point
DFT is obtained by successive use of these
decompositions up to the last stage. Thus we obtain a
DIF split-radix-2/4 algorithm. The signal flow graph
of basic butterfly cell of split-radix-2/4 DIF FFT
algorithm is shown in Figure 6

Figure 5: The butterfly signal flow graph of mixedradix-4/2 DIF FFT


It uses both the radix-22 and the radix-2 algorithms
can perform fast FFT computations and can process
FFTs that are not power of four. The mixed-radix 4/2,
which calculates four butterfly outputs based on
X(0)~X(3). The proposed butterfly unit has three
complex multipliers and eight complex adders.
E. Split-Radix FFT Algorithms
Split-radix FFT algorithm assumes two or more
parallel radix decompositions in every decomposition
stage to fully exploit advantage of different fixedradix FFT algorithm. As a result, a split-radix FFT
algorithm generally has fewer counts of adder and
multiplication than the fixed-radix FFT algorithms,
while retains applicability to all power-of-2 FFT
length.
More computational complexity of the odd
frequency terms than the even frequency terms, so we
can further decompose the odd terms to reduce

Figure 6: The butterfly signal flow graph of mixedradix-2/4 DIF FFT


(0) = ( ) +

we have

(2) =

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

65

+
+


3
4

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

(1) =

(3) =

( ) + ( ) ( + 4)
+(1) ( + 2) + ( ) ( + 3 4)

ISBN: 378 - 26 - 138420 - 5

B. Input Zero Traced Radix-4 DIF FFT Pruning


In radix-4 since we couple four inputs to obtain
four outputs, we therefore have 16 combinations of
those four inputs at radix-4 butterfly. Now therefore
for radix-4 pruning there exist five conditions only
based upon zeros at the input.

( ) + ( ) ( + 4)
( + 2) + ( ) ( + 3 4)

No zero at the input: No pruning takes place,


butterfly calculations are same as radix-4
Any one input zero: Output will be only the copied
version of remaining inputs available, butterfly
calculations are reduced compared to radix-4.
Any two inputs are zeros: Output will be only the
copied version of that remaining two inputs
available, butterfly calculations are reduced
compared to radix-4 pruning with one zero at input.
Any three inputs are zeros: Output will be only the
copied version of that remaining single input
available, butterfly calculations are reduced
compared to radix-4 pruning with two zero at
input.
All zeros input: Output is zero and is obtained from
mathematical calculations is zero.

(13)
As a result, even and odd frequency samples of
each basic processing block are not produced in the
same stage of the complete signal flow graph. This
property causes irregularity of signal flow graph,
because the signal flow graph is an L-shape
topology.
IV PRUNING TECHNIQUES
To increase the efficiency of the FFT technique
several pruning and different other techniques have
been proposed by many researchers. In this paper, we
have implemented a new pruning technique i.e.
IZTFFTP by simple modification and some changes
and also includes some tricky mathematical
techniques to reduce the total execution time.

C. Input Zero Traced Radix-8 DIF FFT Pruning


In radix-8 since we couple eight inputs to obtain
eight outputs, we therefore have 256 combinations of
those eight inputs at radix-8 butterfly. Now therefore
for radix-8 pruning there exist seven conditions only
based upon zeros at the input. Similarly to radix-4
pruning, output is the version of non zero input. The
more the number of zeros at input leads to less
mathematical calculations compared to radix-8.

Zero tracing- as in wide band communication system


a large portion of frequency channel may be
unoccupied by the licensed user, so no. of zero
valued inputs are much greater than the non-zero
valued inputs in a FFT/IFFT operation at the
transceiver. Then this algorithm will give best
response in terms of reduced execution time by
reducing the no. of complex computation required for
twiddle factor calculation. IZTFFTP have a strong
searching condition, which have an array for storing
the input & output values after every iteration of
butterfly calculation. In a input searching result
whenever it found zero at any input, simply omit
that calculation by considering useful condition
based on radix algorithm used.

D. Input Zero Traced Mixed radix DIF FFT Pruning


If we consider mixed radix 4/2, it uses the
combination of radix-2 pruning and radix-4 pruning.
Similarly mixed radix 8/2 uses the combination of
radix-2 pruning and radix8 pruning.
E. Input Zero Traced Split radix DIF FFT Pruning
If we consider spilt radix 2/4, it uses the
combination of radix-2 pruning and radix-4 pruning.

A Input Zero Traced Radix-2 DIF FFT Pruning


In radix-2 since we couple two inputs to obtain
two outputs, we therefore have 4 combinations of
those two inputs at radix-2 butterfly. Now there exist
three conditions only based upon zeros at the input.

V RESULTS
In order to compare the computational
complexities among the different radix DIF FFT
algorithms on OFDM, the calculations based on the
OFDM block sizes have been performed which are
given in Table 1 and with pruning comparison in
Table 2.
The speed improvement factors from without to
with pruning of different radix algorithms are seen in
Table 3.

No zero at input: No pruning happens in this case,


butterfly calculations are same as conventional
radix-2.
Any one input zero: Output will be only the copied
version of input available, butterfly calculations are
reduced compared to conventional radix-2.
All zero input: Output is zero and is obtained from
mathematical butterfly calculations is zero.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

66

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

OFDM
Block
Size

Radix -2
cm
1
4
12
32
80
192

2
4
8
16
32
64

cadd
2
8
24
64
160
384

Radix-4
cm
3
24
144

Radix-8

cadd
8
64
384

cm
7
112

cadd
24
384

Mixed
Radix-4/2
cm
10
28
64
160

cadd
24
64
160
384

Mixed
Radix-8/2
cm
22
60
152

cadd
64
160
384

ISBN: 378 - 26 - 138420 - 5

Split
Radix-2/4
cm
0
4
12
36
92

cadd
8
24
64
160
384

Table 2: Comparison of complex additions(cadd) and complex multiplications(cm) of different radix algorithms
without pruning

OFDM
Block
Size

Radix -2
cm
0
0
12
31
76
179

2
4
8
16
32
64

cadd
2
8
24
64
160
384

Radix-4
cm
3
24
141

Radix-8

cadd
8
64
384

cm
-

cadd
-

7
112

24
384

Mixed
Radix-4/2
cm
8
26
64
157

cadd
24
64
160
384

Mixed
Radix-8/2
cm
22
60
152

cadd
64
160
384

Split
Radix-2/4
cm
0
4
12
36
90

cadd
8
24
64
160
384

Table 2: Comparison of complex additions(cadd) and complex multiplications(cm) of different radix algorithms
with pruning

FFT
Size

Radix2

Radix4

Radix8

Mixed
Radix4/2

Mixed
radix8/2

Split
radix2/4

1.25

16
32

1.03
1.05

1
-

1.07
1

1
1

1
1

64

1.07

1.02

1.01

Table 3: Speed Improvement Factor without to with


pruning in terms of Multiplications
Output shows the significant reduction of
computational complexity by reducing the total no. of
complex operation i.e. both the multiplications and
additions compare to the ordinary radix FFT
operations. The complex multiplications and
additions are compared for different radix and pruned
algorithms.
The comparison of complex multiplications for
different radix DIF FFT algorithms is shown in
Figure 7 and for different input zero traced radix DIF
FFT pruned algorithms are shown in Figure 8.

Figure 7: Comparison of complex multiplications for


different radix DIF FFT

Figure 8: Comparison of complex multiplications for


different Radix DIF FFT pruned algorithms

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

67

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

VI CONCLUSION
The computational performance of an OFDM
system depends on FFT as in an OFDM system. FFT
works as a modulator. If the complexity decreases,
then the speed of OFDM system increases. Results
shows input zero traced radix DIF FFT pruned
algorithms are much efficient than the Radix DIF
FFT algorithms as it takes very less time to compute
where number of zero valued inputs/outputs are
greater than the total number of non zero terms, with
maintaining a good trade-off between time and space
complexity, and it is also independent to any input
data sets.
VII References
[1] B. E. E. P. Lawrey, Adaptive Techniques for
Multi-User OFDM, Ph.D. Thesis, James Cook
University, Townsville,2001, pp. 33-34.
[2] J. Mitola, III, "Cognitive Radio: An Integrated
Agent Architecture for Software Defined Radio,"
Thesis (PhD), Dept. of Teleinformatics, Royal
Institute of Technology (KTH), Stockholm
Sweden, May 2000.
[3] S. Chen, Fast Fourier Transform, Lecture
Note, Radio Communications Networks and
Systems, 2005.
[4] OFDM for Mobile Data Communications, The
International Engineering Consortium WEB
ProForum Tutorial, 2006. http://www.iec.org.
[5] Andrea Goldsmith, Wireless Communications
Cambridge university press, 2005, ISBN:
978052170416.
[6] J.G. Proakis and D.G. Manolakis, Digital Signal
Proc-essing: Principles, Algorithms and Edition,
2002, pp. 448-475.
[7] E. Chu and A. George, Inside the FFT Black Box
:Serial & Parallel Fast FourierTransform
Algorithms. CRC Press LLC, 2000.
[8] B. G. Jo and M. H. Sunwoo, New ContinuousFlow Mixed-Radix (CFMR) FFT Processor
Using Novel In-Place Strategy, Electron
Letters, vol. 52, No. 5, May 2005.
[9] Charles Wu, Implementing the Radix-4
Decimationin Frequency (DIF) Fast Fourier
Transform
(FFT)
Algorithm
Using
aTMS320C80 DSP, Digital Signal Processing
Solutions,January 1998.
[10] P. Duhamel and H. Hollmann, Split-radix FFT
Algorithm, Electron Letters, vol. 20, pp 14-16,
Jan. 1984.
[11] [4] H. V. Sorensen, M. T. Heideman and C. S.
Burrus, On Computing the Split-radixFFT,
IEEE Trans. Acoust., Speech, Signal Processing,
vol. ASSP-34, pp. 152-156,Feb. 1986.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

68

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Improvement of Dynamic and Steady State


Responses in Combined Model of LFC and AVR
Loops of One-Area Power System using PSS
1

1,2

Anil Kumar Sappa, 2Prof.Shyam Mohan S.Palli


Dept. of EEE, Sir C R Reddy College of Engg., Andhra Pradesh, India

AbstractThis paper describes the improvement in stability by


reducing the damping oscillations in one area power system using
a power system stabilizer. A power system model, which includes
both LFC and AVR loops, is considered. Low-frequency
oscillation studies are made.PSS is designed to improve the
stability of the system. The simulation results obtained indicate
that, adding a PSS to a combined model of LFC and AVR
improves the dynamic stability by reducing the low-frequency
oscillations.

LFC since many years [7]-[8], but these researches gave little
attention to AVR effects on the results. In fact, in LFC power
system control literature there is a lack of stability analysis for
AVR effects or the mutual effects between these loops.
Usually, these studies are based on the assumption that there is
no interaction between the power/frequency and the reactivepower/voltage control loops. But in practical systems some
interactions between these control channels do exist during
dynamic perturbations [9]. Also by neglecting the effect of
voltage deviation on load demand, an important interaction in
LFC systems is ignored. A combined model with LFC and
AVR loops and their mutual effects is considered.
In this paper the power system is designed by adding a
PSS to a combined model of LFC and AVR loops in order to
improve for more dynamic stability. The interaction of
coupling effects between LFC and AVR loops are shown [3]
and also the performance of proposed model is shown with
simulations. The results of the proposed method with PSS are
compared with a combined model without adding PSS and
also by separate models of LFC and AVR loops without any
interaction between those loops. The simulations are shown by
adding a PSS to a combined model of LFC and AVR loops. It
is observed that this proposed model can improve the dynamic
stability of a complete power system by reducing the damping
oscillations.

Index Terms Automatic Voltage Regulator, Power System,


Load Frequency control, Power system stabilizer, voltage,
deviations, stability.

I. INTRODUCTION
The change in operating conditions of a power system leads
to low frequency oscillations of small magnitude that may exist
for long periods of time. In some cases these oscillations will
limit the amount of power transmitted through interconnecting
lines. So a power system stabilizer is designed to provide an
additional input signal to the excitation system in order to
damp these power system oscillations [1].
The interconnected power system model for low frequency
oscillation studies should be composed of mechanical and
electrical loops. These oscillations can be damped by varying
the exciter and speed-governor control parameters [2].
Furthermore, it has been shown that the load-voltage
characteristic of the power system has a significant effect on
its dynamic responses, and suggestions have been made for
the proper representation of these characteristics in simulation
studies [3]-[5].For economic and reliable operation of a power
system, the two main control loops are required. The Load
Frequency Controller loop (LFC) and Automatic Voltage
Regulator loop (AVR) as shown in Figure 1. The turbine is fed
by speed governor whose steam rate can be controlled by
varying the internal parameters. Automatic Generation
Control method deals with frequency through the LFC loop
and with voltage through the AVR loop. The main purpose of
these two loops is to maintain frequency and voltage at
permissible levels [6]. Lot of studies have been made about

Fig 1. Automatic generation control with LFC and AVR loops.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

69

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig 2. Block diagram model of Load Frequency Control (LFC).

comprises of excitation control mechanism and its main aim is


to control the field current of the synchronous machine. Here
the controlling of field current is to regulate the voltage
generation by the machine. The maximum permissible limit for
voltage is about 5 6% .

II. LOAD FREQUENCY CONTROL


Load Frequency Control (LFC) loop is the basic control
mechanism loop in the operation of power system. As the
active power demand changes continuously with respect to
load changes ,steam input to turbo-generators (water input to
hydro-generators) must be continuously regulated in order to
maintain the constant active power demand, failing which the
machine speed changes with consequent change in frequency
which may be highly undesirable. The maximum permissible
change in power frequency is about 0.5Hz. Hence continuous
regulation is to be provided by the LFC system to the generator
turbine.
At the time of Load change, the deviation from the nominal
frequency is referred as frequency error (f), This symbol
indicates that there is a mismatch and it can be used to send the
appropriate command to change the generation by altering the
LFC system [6].From Figure 2, it is seen that there are two
control loops, one is primary control loop and the other one is
secondary control loop. With primary control loop, if there is
any change in system load, it will result in to a steady-state
frequency deviation, which depends on the governor speed
regulation. To nullify the frequency deviation, we should
provide a reset action. The reset action is accomplished by
secondary control loop which introduces an integral controller
to act on the load reference setting in order to change the speed
at the desirable operating point. Here the integral controller
increases the system type by 1 which forces and makes the
final frequency deviation to zero. Thus by using integral
controller action in LFC system a zero steady state frequency
error and a fast dynamic response will be achieved. The gain of
the integral controller should be adjusted such that its transient
response is satisfactory. The negative sign for gain of integral
controller shown in Figure 2 is to provide a negative or reduced
command signal for a positive frequency error.

Fig 3. Block diagram model of AVR system.

The Amplifier and Exciter block shown in Figure 3


regulates and amplifies input control signals to an appropriate
level which is convenient to provide DC power to the generator
field winding. This block must be expandable if the excitation
system has rotating exciter and voltage regulator [2].
Depending upon how the DC supply is given to the
generator field winding, the excitation systems are classified as
DC, AC, and Static excitation systems [1].
IV. MATHEMATICAL MODELLING OF POWER SYSTEM
In order to improve the dynamic stability of overall system,
modeling of major components of power system is required.
The study of low frequency oscillation studies is completely
based on a single machine connected to an infinite bus system
[2].
The single machine connected to an infinite bus system
through transmission lines with a local load is shown in Figure
4. Here Z is the series impedance of transmission line and Y is
the shunt admittance representing the local load.

III. AUTOMATIC VOLTAGE REGULATOR


The basic mechanism of Automatic Voltage Regulator
(AVR) is to regulate the system voltage magnitude. The AVR

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

70

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Iq

Vd
Xq

ISBN: 378 - 26 - 138420 - 5

(5)

Eq Vq X d I d

(6)

To determine the values of V0 and 0.From the definition of


torque angle (0) which is the angle between the infinite-bus
voltage (V0) and the internal voltage (Eq ) [2], we have

Fig 4. Single machine connected to an infinite bus model of


a power system.

A. Combined model of LFC and AVR of one-area power


system
A Combined model of LFC and AVR of one-area power
system with PI controller is shown in figure5 .The PI
controller used in figure 6 produce an output signal with a
combination of proportional and integral controllers and the
transfer function of that signal consists of a proportional
gain (KP) which is proportional to the error signal and the
term consisting of integral time constant (Ti) which is
proportional to the integral of error signal [10].
Transfer function of PI controller = K 1 1
p

Ti s

(1)

The block diagram shown in Figure 5 shows the coupling


effects between LFC and AVR loops [9].Here the gain
constants K1, K2, K3, K4, K5 and K6 are calculated using the
equations (2)-(27) [2]. Where K1 is the change in electrical
power for a change in rotor angle with constant flux linkages in
the d-axis, K2 is change in electrical power for a change in the
direct axis flux linkages with constant rotor angle, K3 is an
impedance factor, K4 is the demagnetizing effect of a change in
rotor angle, K5 is the change in terminal voltage with change in
rotor angle for constant Eq and K6 is the change in terminal
voltage with change in Eq for constant rotor angle.

Vt 2
2

Vd PV
P

e t
e
e X

1/2

V0 d C1Vd C2Vq Rx I d XI q

(7)

V0 q C2Vd C1Vq XI d Rx I q

(8)

V
0 tan 1 0d
V0q

(9)

V0 (V0 d 2 V0 q 2 )1/2

(10)

R1 Rx C2 X d

(11)

R2 Rx C2 X q

(12)

X 1 X C1 X q

(13)

X 2 X C1 X d

(14)

Z e 2 R1 R2 X 1 X 2

(15)

C1 1 Rx G XB

(16)

C2 XG Rx B

(17)

Yd

C1 X 1 C2 R2
Ze2

(18)

Yq

C1 R1 C2 X 2
Ze2

(19)

Fd

V0
( R2 cos 0 X 1sin 0 )
Ze2

(20)

Fq

V0
( X 2cos 0 R1sin 0 )
Ze2

(21)

(2)

1/ 2

Vq Vt 2 Vd 2
Id

Pe I qVq
Vd

(3)

(4)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

71

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig 5. PSS with combined LFC and AVR.

B. Combined LFC-AVR model with PSS

Finally, the constants K1 to K6 can be represented as

K1 Fd X q X d I q Fq Eq X q X d I d

The proposed model of combined LFC and AVR with PSS is


shown in figure 5 for dynamic improvement of overall system
response.

K 2 I q Yd X q X d I q Yq Eq X q X d I d

(24)

K3

1
1 X d X d Yd

(24)

K4

Xd Xd
Fd

(25)

( X d Vq )
X qVd
K 5 Fd
Fq

Vt
Vt

K6

( X d Vq )
X qVd
Yd
Yq

Vt
Vt

Vt

Vq

The basic function of a power system stabilizer (PSS) is to


provide an additional input signal to the regulator to damp the
Power System oscillations. This can be achieved by
modulating the generator excitation in order to produce a
component of electrical torque in phase with the rotor speed
deviations. Some of the commonly used input signals are rotor
speed deviation, accelerating power, and frequency deviation.
A block diagram model of PSS is shown in Figure 6.

(23)

(26)

(27)
Fig 6. Block diagram model of PSS.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

72

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The frequency deviation (f) is used as input signal for


PSS. Here the PSS will generate a pure damping torque at all
frequencies that is the phase characteristics of PSS must
balance the phase characteristics of GEP(s) at all frequencies.
As this is not practicable, the time constants of PSS are to be
adjusted in order to produce the phase compensation
characteristic which shows the best performance [11]. The
model of PSS shown in Figure 6 consists of three blocks, phase
compensation block, washout filter block and gain block. The
phase compensation block provides a suitable phase-lead
characteristic to compensate the phase lag between the input of
the exciter and the machine electric torque. The transfer
function GEP(s) of the block shown with dashed lines in in
figure 5 which represents the characteristics of the generator
excitation and power system is given by [11]

GEP s

K a K3 K 2
(1 sTa )(1 sK 3Td 0 ) K a K 3 K 6

ISBN: 378 - 26 - 138420 - 5

From the simulations results shown in Figures 7-10 it can


be observed that, by using PSS the settling time is reduced with
a better dynamic stability. Hence desirable damping for low
frequency oscillations in overall single-area power system is
achieved.

(28)

The washout block acts as a high-pass filter. For local


mode of oscillations, the washout time constant should be in
between 1 to 2s for desirable operating point [12]. The PSS
gain is chosen such that it is fraction of gain corresponding to
instability [11].The PSS complete transfer function is given by

sTW
PSS (s)
1 sTW

KS

1 sT1

1 sT2

Fig 7.Frequency deviations in a single-area power system


in pu.

(29)

Where GEP(s) is the plant transfer function through which


the stabilizer must provide compensation. TW is washout time
constant. KS is the gain of the stabilizer. T1 and T2 are the time
constants of phase compensator. An optimal stabilizer is
obtained by the proper selection of the time constants and gain
of the PSS.
V. SIMULATION RESULTS
In this paper a power system stabilizer is designed to show
the improvement in dynamic response for combined model of
LFC and AVR. In this study the performance of the proposed
model of PSS with combined LFC and AVR is compared with
combined model of LFC and AVR loops and also with classic
model of load frequency control system by separating AVR
loop (i.e. excitation system). The simulations shown in this
paper are carried out using MATLAB platform.
The simulations results shown in this paper are performed
by assuming real power as 0.9 pu, reactive power as 0.61 pu
and the machine terminal voltage as 0.91 pu.

Fig 8.Turbine output power deviations in a singlearea power system in pu.

Turbine and Governor System Parameters are given in


table-I, One Machine-Infinite Bus System Parameters are given
in table-II and AVR and Local parameters are given in tableIII.
The calculated gain parameter constants in AVR system for
load change in real power at 10% for LFC is given in Table IV.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

73

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

TABLE III.

ISBN: 378 - 26 - 138420 - 5

AVR AND LOCAL LOAD PARAMETERS IN PER UNIT VALUE.

Ka

Ta

Kr

Tr

20

0.05

0.05

0.89

0.862

TABLE IV.

CALCULATED GAIN PARAMETER CONSTANTS IN AVR SYSTEM


IN P ER UNIT VALUE.

K1

K2

K3

K4

K5

K6

0.40

1.85

0.38

0.22

0.07

1.02

VI. CONCLUSION
In this paper the Load Frequency Control loop and
Automatic Voltage Regulator loop are combined to show the
mutual effects between these two loops in a one-area power
system. A Power System Stabilizer is designed to improve the
dynamic and steady state responses for one-area power system.
Finally it is observed that a better dynamic stability is accessed
by using a PSS to the combined LFC-AVR model.

Fig 9.Internal electrical power deviations in a singlearea power system in pu.

NOMENCLATURE

Fig 10.Terminal voltage deviations in a single-area


power system in pu.

TABLE I.

T URBINE AND GOVERNOR SYSTEM PARAMETERS FOR LFC IN


PER UNIT VALUE.

KP

TP

TT

TG

102

20

0.32

0.06

1.7

TABLE II.

ONE MACHINE-INFINITE B US SYSTEM PARAMETERS IN P ER


UNIT VALUE.

Xd

Xq

X d'

Td 0 '

Rx

1.973

0.82

0.1

7.76

0.004

0.74

R
Ka

The overall gain of excitation system

Ta

The overall time constant of excitation system

Pm

Turbine output power deviation

PL

Load disturbance

KP

The equivalent overall gain of a power system

Speed regulation due to governor action

TP

The equivalent time constant of a power system

Kr

The equivalent gain of a sensor

Tr

The equivalent time constant of a turbine

TT

The equivalent time constant of a turbine

TG

The equivalent time constant of a governor

E q '

Deviation of internal voltage

Td 0 '

Deviation of torque angle


Transient time constant of generator field

Xd

Synchronous reactance of d-axis

Xd '

Transient reactance of d-axis

Vref

Reference input voltage

V0

Infinite bus voltage

Vt

Terminal voltage

VS

Stabilizer output

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

74

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[7] E.Rakshani and J.Sadeh,A Reduced-Order control with


prescribed degree of stability for Two-Area LFC System in a
deregulated environment, Proceeding 2009 IEEE PES Power
Systems Conference and Exhibition (PSCE09).
[8] J.Sadeh and E.Rakshani,Multi-area load frequency control in a
deregulated power system using optimal output feedback
method, International conf. on European Electricity Market,
pp.1-6, May 2008.
[9] Haadi Saadat, Power system analysis, Tata McGraw-Hill
Edition 2002, pp.528.
[10] K.Ogata, Modern Control Systems, 5th edition, Prentice Hall
Publications, 2002, pp.669-674.
[11] E.V.Larsen and D.A.Swan,Applying power system stabilizers
parts I, II and III, IEEE Trans. On Power Apparatus and
Systems, Vol.PAS-100, June-1981, pp.3017-3046.
[12] P.Kundur, M.Klein, G.J.Rogers and M.S.Zywno,Application of
power system stabilizers for enhancement of overall system
stability, IEEE Trans. on Power systems, Vol.4, No.2, May
1989, pp.614-626.

REFERENCES
[1] P.Kundur, Power System Stability and Control, McGraw-Hill
Inc., 1994, pp.766-770.
[2] Yau-NaN yu, Electrical Power System Dynamics, London
Academic Press, 1983, pp.66-72.
[3] E.Rakhshani, K.Rouzehi, S.Sadeh, A New Combined Model
for Simulation of Mutual Effects between LFC and AVR
Loops, Proceeding on Asia-Pacific Power and Energy
Engineering, Wuhan, China, 2009.
[4] S.C.Tripathy, N.D.Rao, and L.Roy,Optimization of exciter and
speed governor control parameters in stabilizing intersystem
oscillations with voltage dependent load characteristics.
Electric power and energy systems, vol.3, pp.127-133, July
1981.
[5] K.Yamashita and H.Miyagi,Multivariable self-tuning regulator
for load frequency control system with interaction of voltage on
load demand, IEEE Preceeedings-D, Vol.138, No.2, March
1991.
[6] D.P.Kothari, I.J.Nagrath Modern Power System Analysis,
Third Edition, pp.290-300

Authors Biography

Anil Kumar Sappa received his


B.Tech degree in Electrical and
Electronics
Engineering
from
JNTUK, Andhra Pradesh in 2011. He
is currently pursuing M.E degree in
Power Systems and Automation from
Sir C R Reddy college of Engineering affiliated to A.U
Visakhapatnam, Andhra Pradesh. His areas of interests
include Power Systems.

Prof Shyam Mohan S Palli received


his B.Tech degree in Electrical and
Electronics Engineering from JNTU
Kakinada in 1978; M.Tech degree in
Power systems from same institute in
1980. He joined in teaching
profession in 1981. He has put up an
experience of 33 years in teaching as lecturer, Asst.
Professor, Professor. Presently he is working as
Professor and HOD, EEE of Sir C.R. Reddy college of
Engineering, Andhra Pradesh. He is author of Circuits
and Networks, Electrical Machines Published by TMG
New Delhi. He published many papers in referred
journals and conferences. His areas of interests include
Machines and Power Systems.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

75

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Enhanced Cryptography Algorithm for Providing


Data Security
Dr. K.Kiran Kumar, K.Kusuma Kavya, K.R.L.Sindhura
Department of Information Technology, Bapatla Engineering College, Bapatla.
Abstract: Information security is the
process of protecting information. Due to
enormous
development
in
internet
technology the security of data has now
become a very important challenge in data
communication network. One cannot send
any confidential data in raw form from one
machine to another machine as any hacker
can intercept the confidential message. In
this paper we have developed a new
cryptography algorithm which is based on
block cipher concept. In this algorithm we
have used logical operations like XOR and
shift operations which is used to increase
security.

must consider many factors like: security,


time and space complexity.
A Simplified Model of Conventional
Encryption

This paper is further divided into three


sections. In section 2, we are presenting
detailed description of Information security
using cryptography and various algorithms.
In section 3, we are presenting our proposed
algorithm. In section 4, we have explained
the proposed algorithm by taking an
example of 128 bits (16 characters) and in
section 5 we have conclusion.

Keywords:Information security, Encryption,


Decryption, Cryptography.
1. Introduction
The main feature of the encryption/
decryption program implementation is the
generation of the encryption key. Now a
days, cryptography has many commercial
applications. The main purpose of the
cryptography is used not only to provide
confidentiality, but also to provide solutions
for other problems like:
data integrity,
authentication,
non-repudiation.
Cryptography is the method that allows
information to be sent in a secure form in
such a way that the only receiver able to
retrieve the information. However, it is very
difficult to find out the specific algorithm,
because we have already known that they

2. Information security using


cryptography
Here a newly developed technique named,
A new Symmetric key Cryptography
Algorithm using extended MSA method:
DJSA symmetric key algorithm is
discussed. In this a symmetric key method
where we have used a random key generator
for generating the initial key and the key is
used for encrypting the give source file and
we are using substitution method where we
take 4 characters from any input file and
then search the corresponding characters in
the random key matrix file after getting the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

76

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

encrypted message they store the encrypted


data in another file. To decrypt any file one
has to know exactly what is the key and to
find the random matrix theoretically. They
apply method on possible files such as
executable file, Microsoft word file, excel
file, access database, FoxProfile, text file,
image file, pdf file, video file, audio file,
oracle database and they have found in all
cases it giving 100% correct solution while
encrypting a file and decrypting a file. In
the following section we are going in detail.
Here another newly developed technique
named, Effect of Security Increment to
Symmetric Data Encryption through AES
Methodology is discussed. In this method
they describe about symmetric cipher
algorithm which is much more similar. The
difference is that, Rijndael algorithm start
with 128 bits block size, and then increase
the block size by appending columns[10],
whereas his algorithm start with 200 bits.

ISBN: 378 - 26 - 138420 - 5

Encryption Approach Used


Here we are using symmetric encryption
approach, and already we know that
symmetric encryption approach is divided
into two types:
1. Block
cipher
symmetric
cryptography technique
2. Stream
cipher
symmetric
cryptography.
But here we are choosing block cipher type
because its efficiency and security. In the
proposed technique we have a common key
between sender and receiver, which is
known as private key. Basically private key
concept is the symmetric key concepts
where plain text is converting into encrypted
text known as cipher text and is decrypted
by same private key into plane text. The
keys, in practice, represent a shared secret
between two or more parties that can be
used to maintain private information.
Proposed Key Generation Steps

3. Proposed Algorithm
1. Create any private key of Size 16
characters. It will be varying from 16 to 64
characters.
2. We can choose any character from 0 to
255 ASCII code.
3. Use of 16*8 key that means 128 bits in
length.
4. Divide 16 bytes into 4 blocks of KB1,
KB2, KB3 and KB4.
5. Apply XOR operation between KB1 and
KB3.Results will store in new KB13.
6. Apply XOR operation between KB2 and
KB13.Results will store in new KB213.
7. Apply XOR operation between KB213
and KB4. Results will store in new KB4213.
( KB:KEY BLOCK)

In this section we are presenting a new


block based symmetric cryptography
algorithm, and using a random number for
generating the initial key, where this key
will use for encrypting the given source file
using proposed encryption algorithm.
Basically In this technique a block based
substitution method will use. In the present
technique I will provide for encrypting
message multiple times. The proposed key
blocks contains all possible words
comprising of number of characters whose
ASCII code is from 0 to 255 in a random
order. The pattern of the key blocks will
depend on text key entered by the user. To
decrypt any file one has to know exactly
what the key blocks is and to find the
random blocks theoretically one has to
apply 2^256 trial run and which is
intractable.

Steps for proposed Algorithm


1. Initially select plain text of 16 bytes.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

77

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

2. Apply XOR operation between key


(KB4213) and plain text block. Result will
store in CB1.
3. Apply right circular shift with 3 values.
Result will store in CB2.
4. Apply XOR operation between CB2 and
KB2.
5. Result will store in new CB3.
6. Apply XOR operation between CB3 and
KB4. Result will store in CB4.

KB 13: 00011110
00011100

ISBN: 378 - 26 - 138420 - 5

00010100

11110011

Step:3
Apply the EX-OR operation between KB2
& KB13
KB2: 10000100
01110110
01101001
KB13:00011110
00010100
00011101
KB213: 10011010 01100010
01110100

(CB:CIPHER BLOCK)
4. Implementation example of the
Proposed Algorithm

01100101
11110011
10010110

Step:4
Encryption Method
Apply the EX-OR operation between KB4
& KB213:

Key Generation:

KB213:10011010 01100010 10010110


01110100
KB4: 01101001
01101001
10000010
10000011
KB4213: 11110011 00001011 00010100
11110111

Step:1
1.Take the 16 characters plain text
2.Plain text: BAPATLAENGINEERS
3. Divide the key in to 4 blocks
BAPA
KB 1

TLAE
KB2

NGIN
KB 3

EERS
The obtained key is:
11110011
00001011
11110111

KB 4

00010100

Step:2
Encryption:
Apply the EX-OR operation between KB1
& KB3

Step:1

BAPA
KB1

Apply the EX-OR operation between KB


4213 & KB1:

NGIN
KB3

BAPA- 66 65 80 65
NGIN- 78 71 73 78
KB1: 01100110
01100101
KB3: 01111000
01111000

KB4213: 11110011 00001011


11110111
KB1: 01100110 01100101
01100101
CB1: 10010101 01101110
10010010

01100101 10000000
01110001

01110011

00010100
10000000
10010100

Step:2

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

78

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

P1: 01100110
01100101

Apply right circular shift with 3bits of CB1.


Result will store in new CB2.
CB2: 01010010
10010010

10101101

10000000

10101101

11010010

CB 2: 01010010 10101101
10010010
CB 3: 11010110 11011011
11111011
P2: 10000100
01110110
01101001

01110110

01100101

Step:3

11011011

10110111

Apply the EX-OR operation between CB3


& CB4:

Apply the EX-OR operation between CB2


& KB2:

10110111

CB3: 11010110
11111011
CB4: 10111111
01111000
P4: 01101001
10000011

10000010

Step:4

00110101

Plain text 3 is :
Apply the EX-OR
KB4321 & P4

Step:4
Apply the EX-OR operation between CB3 &
KB4:
CB3: 11010110
11111011
KB4: 01101001
10000011
CB4: 10111111
01111000

01100101

Step:2
Apply the EX-OR operation between CB2&
CB3:

11010010

Step:3

CB2: 01010010
10010010
KB2: 10000100
01101001
CB3: 11010110
11111011

ISBN: 378 - 26 - 138420 - 5

11011011
01101001
10110010

Now CB1, CB2, CB3, CB4 are the cipher


blocks respected to KEY1, KEY2, KEY3,
KEY4

Step:1
Apply the EX-OR
KB4213 & CB1

operation

KB4213: 11110011 00001011


11110111
CB1: 10010101 01101110
10010010

between

10110111
01100101

11011011

10110111

10110010

00110101

01101001

10000010

operation

KB4213: 11110011 00001011


11110111
P4: 01101001
01101001
10000011
X1: 10011010
1100010
01110100

Decryption Method:

11010010

between

00010100
10000010
10010110

Step:5
Apply the EX-OR operation between X1 &
P2

00010100
10010100

X1: 110011010
01110100

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

79

01100010

10010110

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

P2: 10000100
01101001
X2: 00011110
00011101
Step:6

01110110

01100101

6. References

00010100

11110011

[1] Dripto Chatterjee, Joyshree Nath,


Suvadeep Dasgupta, Asoke Nath A new
Symmetric key Cryptography Algorithm
using extended MSA method: DJSA
symmetric key algorithm published in 2011
International Conference on Communication
Systems and Network Technologies, 978-07695-4437-3/11 $26.00 2011 IEEE.
[2] Yan Wang and Ming Hu Timing
evaluation of the known cryptographic
algorithms 2009 International Conference
on Computational Intelligence and Security
978-0-7695-3931-7/09 $26.00 2009 IEEE
DOI 10.1109/CIS.2009.81.
[3] Symmetric key cryptography using
random key generator, A.Nath, S.Ghosh,
M.A.Mallik, Proceedings of International
conference on SAM-2010 held at Las
Vegas(USA) 12-15 July,2010, Vol-2,P-239244.
[4] Data Hiding and Retrieval, A.Nath,
S.Das, A.Chakrabarti, Proceedings of IEEE
International conference on Computer
Intelligence and Computer Network held at
Bhopal from 26-28 Nov, 2010.
[5] Neal Koblitz A Course in Number
Theory and Cryptography Second Edition
Published by Springer-Verlag.
[6] T Morkel, JHP Eloff ENCRYPTION
TECHNIQUES:
A
TIMELINE
APPROACH published in Information and
Computer Security Architecture (ICSA)
Research Group proceeding.
[7] Text book William Stallings, Data and
Computer Communications, 6eWilliam 6e
2005.
[8] [Rijn99]Joan Daemen and Vincent
Rijmen, AES submission document on
Rijndael, Version 2, September 1999.

Apply the EX-OR operation between X2


& P1
X2: 00011110
00011101
P1: 01100110
01100101
P3: 01111000
01111001

ISBN: 378 - 26 - 138420 - 5

00010100

11110011

01100101

10000000

01110001

01110011

Here P1,P2,P3,P4 are the plain texts with


respect to key blocks KB1,KB2,KB3,KB4.
If we want to get P3, we are going to use
KB4321,P4,P2,P1.
Decrypted Text is:
BAPATLAENGINEERS
5. Conclusion
In this paper we have specified a new
algorithm based on block cipher principle. In
this algorithm we have used logical
operations like XOR and shift operations
which is used to increase security. The
algorithm is explained with the help of an
example clearly. Our method is essentially
block cipher method and it will take less
time if the file size is large. The important
thing of our proposed method is that it is
almost impossible to break the encryption
algorithm without knowing the exact key
value. We propose that this encryption
method can be applied for data encryption
and decryption in any type of public
application for sending confidential data.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

80

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Sinter coolers

Ramireddy. Pavankalyan Reddy

Telukutla. Harika Sivani

Dept. of electrical and electronics engineering


Lakkireddy Balireddy College of engineering
Mylavaram, Krishna district
Andhrapradesh, India
[email protected]

Dept. of electrical and electronics engineering


Lakkireddy Balireddy College of engineering
Mylavaram, Krishna district
Andhrapradesh, India
[email protected]

Abstract At present, distributed generation (DG) has been a


research focus all over the world. As a kind of DG system,
cogeneration system utilizing waste heat from sintering-cooling
process plays an important role in modern iron and steel
enterprises

the Blast Furnace. The vertical speed of sintering depends on


the suction that is created under the grate. At VSP, two
exhausters are provided for each machine to create a suction of
1500 mm water column under the grate.
There are several types of sintering machines based on their
construction and working, they are
a) Belt type
b) Stepping type
c) Air Draft type
d) Box type and so on.
Smelting is the term related to metallurgy and we use blast
furnaces for smelting. We can call blast furnaces differently in
different relations like bloomeries for iron, blowing houses for
tin, smelt mills for lead, sinter plants for base metals like steel,
copper, iron ultimately.
Iron ore cannot be directly charged in a blast furnace. In the
early 20th century sinter technology was developed for
converting ore fines into lumpy material chargeable in blast
furnaces, though it took time to gain acceptance in the iron ore
making domain but now places an important role in generating
steel, metallurgical waste generated in steel plants to enhance
blast furnace operation.

I. INTRODUCTION

The frequently used and most worrying thing now a days is


global warming which actually we know as it is the increase of
earths average temperature due to green house gases which
trap heat that would otherwise escape from earth but recent
studies specifying that waste heat produced from industries
(large scale industries like steel making plants, oil refinery
industries etc.,) is more speedily deteriorating the environment
now a days than above said green house gases so we are
converting that waste heat produced from steel making
industries into electricity in order to reduce the heat even in a
small quantity. Most of our steel plants are now using sinter
plants or sinter coolers to convert iron into steel and these are
producing the exhaust steam in a larger quantity.
.

III. WASTE HEAT RECOVERY IN SINTER PLANT


In sinter plant sensible heat can be recovered both
from the exhaust gases of the sinter machine and off-air of the
sinter cooler. Heat recovery can be in different forms.
Hot air steams from both sinter machine and sinter
cooler can be used for the generation of steam with
the installation of recovery boilers. This steam can be
used to generate power or can be used as process
steam. For increased heat recovery efficiency, a high
temperature exhaust section should be separated from
a low temperature exhaust section and heat should be
recovered only from high temperature exhaust
section.
Sinter machine exhaust section can be recirculated to
the sinter machine, either after going through a heat
recovery boiler or without it.

II. SINTER PLANT


Sintering is an agglomeration process of fine mineral
particles into a porous mass by incipient fusion caused by heat
produced by combustion within the mass itself. Iron ore fines,
coke breeze, limestone and dolomite along with recycled
metallurgical wastes are converted into agglomerated mass at
the Sinter Plant, which forms 70-80% of iron bearing charge in

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

81

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

V. RECOVERY OF WASTE HEAT EMITTED BY THE COOLING

Heat recovered from the sinter cooler can be recirculated to the sinter machine or can be used for pre
heating the combustion air in the ignition hood, for
pre heating of the raw mix to sinter machine. It can be
used to produce hot water for district heating.

PROCESS INTO STEAM

As like heat recover ventilators this system will work i.e.


whenever the heat recovery hoods take heat from sinter cooler
this will be directly given to the boiler which on high
temperature and water convert this into steam. The water tube
boiler which has water in its tubes heated by this hot recovered
air and this water will be converted into steam. This steam
drives the turbine and give mechanical energy which is input to
the generator and this generator will give electricity.

A. Features
Waste gas heat of a sintering plant is recovered as
steam or electric energy. The heat recovery efficiency
is 60% for waste gas from cooler and 34% for waste
gas from sintering machine proper.
Waste gas heat recovery from sintering machine
proper also leads to the reduction of coke
consumption.
Applicable whether the cooler is of a circular type or
linear type.
CO2 emissions can be reduced, leading to a
possibility of employing this system in a CDM
project.

IV. SINTER PLANT COOLER WASTE HEAT RECOVERY SYSTEM

Fig. 2 Recovery of waste heat emitted


VI. ADVANTAGES AND DISADVANTAGES
A. Advantages
1) Reduction in pollution: A number of toxic combustible
wastes such as carbon monoxide gas, sour gas, carbon black
off gases, oil sludge, Acrylo nitrile and other plastic chemicals
etc. releasing to atmosphere if/when burnt in the incinerators
serves dual purpose i.e. recovers heat and reduces the
environmental pollution levels.
2) Reduction in equipment sizes: Waste heat recovery
reduces the fuel consumption, which leads to reduction in the
flue gas produced. This results in reduction in equipment sizes
of all flue gas handling equipment such as fans, stacks, ducts,
burners, etc.
3) Reduction in auxiliary energy consumption: Reduction in
equipment sizes gives additional benefits in the form of
reduction in auxiliary energy consumption like electricity for
fans, pumps etc.
Recovery of waste heat has a direct effect on the efficiency
of the process. This is reflected by reduction in the utility
consumption & costs, and process cost.

Fig. 1 Block diagram of sinter cooler plant


This is a system for recovering the sinter coolers hightemperature exhaust gas as steam which can be used for power
generation. Furthermore reuse of the exhaust heat as the
thermal source of sintered ore production will improve the
productivity of sinter machines.
The main principle involved in this system is converting
heat into steam then we use the normal generation process
where the turbine rotates giving mechanical energy as input to
the generator in order to get electricity as output. The system
recovers sensible heat from hot air emitted by the cooling
process of two sinter coolers located downstream of two sinter
machines. The heat is captured by heat recovery hoods and
then routed to a heat recovery boiler to generate super-heated
steam, which is converted to electricity by a turbine connected
to a generator.

B. Disadvantages
1) Capital cost: The capital cost to implement a waste heat
recovery system may outweigh the benefit gained in heat
recovered. It is necessary to put a cost to the heat being offset.
2) Quality of heat: Often waste heat is of low quality
(temperature). It can be difficult to efficiently utilize the
quantity of low quality heat contained in a waste heat medium.
Heat exchangers tend to be larger to recover significant
quantities which increases capital cost.
3) Maintenance of Equipment: Additional equipment
requires additional maintenance cost.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

82

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Waste heat constitute almost 20% in global warming where


as in this maximum amount of heat is from large scale
industries and power plants. As 70% of our steel plants
containing sinter plants and the circulation system is for waste
heat, which has been emitted only to the atmosphere. The
system is expected to promote energy efficiency by utilizing
waste heat, thereby reducing CO2 emissions. It will enhance
environmental effects because cooling air is used in a closed
cycle without emitting a high concentration of dust into the
atmosphere and power shortages can also be overcome.

CONCLUSION
To meet the increasing world demand for energy, the rate of
depletion of non-renewable energy sources must be reduced
while developing alternative renewable sources. This can be
achieved by increasing the overall thermal efficiency of
conventional power plants. One way to do this is by waste heat
recovery. Most of the techniques currently available recover
waste heat in the form of thermal energy which is then
converted to electricity in a conventional steam power plant.
Another approach which has received little attention so far is
direct conversion of thermal waste energy into electricity.
soo In this article, a configuration of waste heat recovery
system is described
we studied the composition and characteristics of waste heat
resources and found out a typical process of energy recovery,
conversion and utilization.

ACKNOWLEDGMENT
The preferred spelling of the word acknowledgment in
America is without an e after the g. Avoid the stilted
expression, One of us (R. B. G.) thanks . . . Instead, try
R. B. G. thanks. Put applicable sponsor acknowledgments
here; DO NOT place them on the first page of your paper or as
a footnote.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

83

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Chaos CDSK Communication System


Arathi. C
M.Tech Student,
Department of ECE,
SRKR Engineering College,
Bhimavaram, India
[email protected]

performance according to chaos maps, and find a


chaos map that has the best BER performance [4]. In
addition, chaos users evaluate the BER performance
according to chaos modulation system [5][6], and
propose a new chaos map that has the best BER
performance.
In this paper, in AWGN and Rayleigh fading
channel, BER performances of chaotic CDSK system
is evaluated. At existing study, we proposed a novel
chaos map in order to improve the BER performance
[7], and we named a novel chaos map "Boss map".

Abstract: In recent years chaotic communication


systems have emerged as an alternative solution to
conventional spread spectrum systems. The chaotic
carrier used in this kind of modulation-demodulation
schemes, have unique properties that make them suited
for secure, and multi-user communications. The
security of chaos communication system is superior to
other digital communication system, because it has
characteristics such as non-periodic, wide-band, non predictability, easy implementation and sensitive initial
condition. In this paper, a new approach for
communication using chaotic signals is presented.
Keywords Chaos Communication System, CDSK

I.

II.

INTRODUCTION

CHAOTIC SYSTEM

A chaotic dynamical system is an


unpredictable, deterministic and uncorrelated system
that exhibits noise-like behavior through its sensitive
dependence on its initial conditions, which generates
sequences similar to PN sequence. The chaotic
dynamics have been successfully employed to
various engineering applications such as automatic
control, signals processing and watermarking. Since
the signals generated from chaotic dynamic systems
are noise-like, super sensitive to initial conditions and
have spread and flat spectrum in the frequency
domain, it is advantageous to carry messages with
this kind of signal that is wide band and has high
communication security. Numerous engineering
applications of secure communication with chaos
have been developed [8].

Previous digital communication technology


continually used a linear system. However, as this
technology reached basic limit, people started to
improve performance of nonlinear communication
systems applying chaos communication systems to
nonlinear systems [1]. Chaos communication systems
have the characteristics such as non - periodic, wideband, non-predictability and easy implementation.
Also, chaos communication system is decided by
initial conditions of equation, and it has sensitive
characteristic according to initial condition, because
chaos signal is changed to different signal when
initial condition is changed [2]. Chaos signal is
expressed as randomly and non-linearly generated
signal. If initial conditions of chaos signal is not
exact, users of chaos system are impossible to predict
the value of chaos signal because of its sensitive
dependence on initial conditions [1][3]. As these
characteristics, the security of chaos communication
system is superior to other digital communication
system.
Due to security and other advantages, chaos
communication systems are being studied
continuously. Look at existing research, in order to
solve disadvantage that bit error rate (BER)
performance of this system is bad, chaos
communication system is evaluated the BER

III.

CHAOTIC SIGNALS

A chaotic sequence is non-converging and


non-periodic sequence that exhibits noise-like
behavior through its sensitive dependence on its
initial condition [1]. A large number of uncorrelated,
random-like, yet deterministic and reproducible
signals can be generated by changing initial value.
These sequences so generated by chaotic systems are
called chaotic sequences [8].
Chaotic sequences have been proven easy to
generate and store. Merely a chaotic map and an

1
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

84

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

initial condition are needed for their generation,


which means that there is no need for storage of long
sequences. Moreover, a large number of different
sequences can be generated by simply changing the
initial condition. More importantly, chaotic
sequences can be the basis for very secure
communication. The secrecy of the transmission is
important in many applications. The chaotic
sequences help achieve security from unwanted
reception in several ways. First of all, the chaotic
sequences make the transmitted signal look like
noise; therefore, it does not attract the attention of an
unfriendly receiver. That is, an ear-dropper would
have a much larger set of possibilities to search
through in order to obtain the code sequences [4][8].
Chaotic sequences are created using
discrete, chaotic maps. The sequences so generated
even though are completely deterministic and initial
sensitive, have characteristics similar to those of
random noise. Surprisingly, the maps can generate
large numbers of these noise-like sequences having
low cross-correlations. The noise-like feature of the
chaotic spreading code is very desirable in a
communication system. This feature greatly enhances
the LPI (low probability of intercept) performance of
the system [4].
These chaotic maps are utilized to generate
infinite sequences with different initial parameters to
carry different user paths, as meaning that the
different user paths will spread spectrum based on
different initial condition [8].
IV.

CDSK transmitter is composed of sum in


which delayed chaos signal multiplied with
information bit is added to generated chaos signal
from chaos signal generator. Here, information bit
that is spread as much as spreading factor is
multiplied by delay chaos signal.
s = x + d x

(1)

Above equation (1) indicates transmitted


signal from transmitter.
r
y
d
r r
r
L
Figure 2: Receiver of CDSK system
CDSK receiver is correlator based receiver,
and it is performed in order to recover the symbol.
Received signal and delay received signal are
multiplied, and this signal is as much added as
spreading factor. Afterward the signal pass through
the threshold, and information signal recover through
decoding.
Information bits are possible to recover
when delay time and spreading factor have to use
exact value that is used in transmitted signal.
B. Chaos maps
In this paper, types of chaos map used are
Tent map and Boss map. At existing study, Boss map
means a novel chaos that we proposed for BER
performance improvement [8].

SYSTEM OVERVIEW

A. Correlation delay shift keying system


CDSK system has an adder in transmitter.
Existing modulation system than CDSK system
consists switch in transmitter, and problem of power
waste and eavesdropping occurs by twice
transmission. Technique that has been proposed for
overcoming these problems is CDSK system. And,
transmitted signal does not repeat by replacing an
adder with a switch in the transmitter [9].
Chaotic
signal

ISBN: 378 - 26 - 138420 - 5

Sk

d {+1, 1}
Figure 1: Transmitter of CDSK system

Figure 3: Trajectory of tent map

2
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

85

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Figure (3) shows trajectory of Tent map.


The x-axis and the y-axis of figure (3) mean xn and
xn+1, and Tent map has trajectory of triangular shape.
x

= b|x c| F(x )

ISBN: 378 - 26 - 138420 - 5

chaotic CDSK system with tent map and boss map is


observed. Here, we observe that the BER
performance of Boss map is better than Tent map at
each stage i.e. at different values of SNR we observe
that the Boss map shows better performance than
Tent map. We also observe that at initial values the
BER is same for both maps but as SNR increases the
BER for Boss map is less than Tent map.

(2)

Equation (2) of tent map is expressed as


above. Equation (2) of Tent map uses existing output
value as current input value, and it is indicated as
figure when initial value is 0.1 and parameter alpha is
1.9999.

Figure 5: BER analysis in AWGN channel


In Rayleigh fading channel, figure (6) shows
the BER performance of chaotic CDSK system. Here,
the performance is evaluated for both Tent map and
Boss map. We observe that at initial values of SNR
the BER performance is the same for both maps. But
as SNR value increases the BER performance of
Boss map is better than Tent map.

Figure 4: Trajectory of boss map


Figure (4) shows trajectory of Boss map, a
novel map that is proposed in order to improve the
BER performance. The x-axis and the y-axis of Boss
map mean xn and yn unlike the Tent map, it draws
trajectory like pyramid shape.
= 0.45|0.503
=
0.3

|
(3)

Equation (3) of Boss map is expressed as


above. Equation (3) form of Boss map is similar to
Tent map because Boss map was proposed by
transforming from Tent map. And, trajectory of Boss
map is indicated as figure (4) when initial value is 0.1
and parameter alpha is 2.5.

V.

PERFORMANCE EVALUATION
Figure 6: BER performance in Rayleigh fading
channel

In this paper, the BER performance of


chaotic CDSK system in AWGN (adaptive white
Gaussian noise) channel and Rayleigh fading channel
is evaluated for Tent map and Boss map.
In AWGN channel, figure (5) shows BER
performance of chaotic CDSK system is evaluated.
Looking at the figure (5), the BER performance of

VI.

CONCLUSION

In this paper, a new type of communication


system using chaos is proposed. Chaos sequences are
non periodic sequences which are sensitive to their

3
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

86

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

initial conditions. Chaos sequences are generated


using chaos map. CDSK system using chaos has
many advantages over other systems. But the BER
performance of chaos communication system is bad.
In order to improve this, we proposed a new chaos
map that has better BER performance than existing
map. In AWGN and Rayleigh fading channel the
chaotic CDSK system is evaluated and we observed
that the BER performance of chaos system with Boss
map has better than with Tent map which improves
BER of CDSK communication system.
VII.

Technical Conference on Circuit/System, Computers


and Communication (ITC-CSCC 2013), Yeosu,
Korea, pp. 775-778, July2013.
7. M.A. Ben Farah, A. Kachouri and M.
Samet, "Design of secure digital communication
systems using DCSK chaotic modulation," Design
and Test of Integrated Systems in Nano-scale
Technology,
2006.
DTIS
2006.International
Conference on, pp. 200-204, Sept. 2006.
8. Ned J. Corron, and Daniel W. Hahs A new
approach to communication using chaotic signals,
IEEE transactions on circuits and systemsI:
fundamental theory and applications, VOL. 44, NO.
5, MAY 1997.
9. Wai M. Tam, Francis C. M. Lau, and Chi K.
Tse, Generalized Correlation-Delay-Shift-Keying
Scheme for Non - coherent Chaos-Based
Communication Systems IEEE transactions on
circuits and systemsI: regular papers, VOL. 53,
NO. 3, MARCH 2006.

FUTURE SCOPE

Chaos communication system increases the


number of transmitted symbols by spreading and
transmitting information bits according to
characteristic of chaos maps. So the research that
improves data transmission speed is necessary for
chaos communication system. If many antennas are
applied to chaos communication system, the capacity
of data is proportional to the number of antenna. So it
is good way applying multiple-input and multipleoutput (MIMO) to the chaos communication system.
VIII.

ISBN: 378 - 26 - 138420 - 5

REFERENCE

1. M. Sushchik, L.S. Tsimring and A.R.


Volkovskii, "Performance analysis of correlationbased communication schemes utilizing chaos,"
Circuits and Systems I: Fundamental Theory and
Applications, IEEE Transactions on, vol. 47, no. 12,
pp. 1684-1691, Dec. 2000.
2. Q. Ding and J. N. Wang, "Design of
frequency-modulated correlation delay shift keying
chaotic communication system," Communications,
IET, vol. 5, no. 7, pp. 901-905, May 2011.
3. Chen Yi Ping, Shi Ying and Zhang Dianlun,
"Performance of differential chaos-shift-keying
digital communication systems over several common
channels," Future Computer and Communication
(ICFCC), 2010 2nd International Conference on, vol.
2, pp. 755- 759, May 2010.
4. Suwa Kim, Junyeong Bok and HeungGyoon Ryu, "Performance evaluation of DCSK
system with chaotic maps," Information Networking
(ICOIN), 2013 International Conference on, pp. 556559, Jan. 2013.
5. S. Arai and Y. Nishio, Noncoherent
correlation-based communication systems choosing
different chaotic maps, Proc. IEEE Int. Symp. On
Circuits and Systems, New Orleans, USA, pp. 14331436, June 2007.
6. Jun-Hyun Lee and Heung-Gyoon Ryu,
"New Chaos Map for CDSK Based Chaotic
Communication System," The 28th International

4
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

87

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

A Review on Development of Smart Grid


Technology in India and its Future Perspectives

fiscal year, which in the impending years is


set to reach double digit growth (10%+) [1].
But India suffers from serious power
shortage which is likely to worsen over the
next few decades. India has a power sector
characterized by deficient generation and
high distribution losses. In addition to that,
abhorrent geological and environmental
factors have encouraged carbon footprints
since its grass roots level of CO2emissions,
greenhouse effect and the adverse effect of
globalization in the country [2]. This may
cause instability in the power system and
problems like brownout and blackout might
arise. In order to prevent the occurrence of
instability, it is essential to upgrade the
prevailing power systems.
One of such incipient
technology, Smart Grid (SG) plays a very
vital role in achieving the key technical
benefits like power loss reduction; refining
quality of supply, peak reduction, economic
load dispatch etc. Smart Grid technology has
been a high priority topic of research and
development in many developing as well as
developed countries. This technology also
has a dynamic role in remodeling the energy
scenario of the global market. Factors like
policies, regulation, efficiency of market,
costs and benefits and services normalizes
the marketing strategy of the Smart Grid
technology.
Other
concerns
like
securecommunication,
standardprotocol,
advance database management and efficient
architecture with ethical data exchange add
to its essentials [3]. Such technology has a
potential to prolific other technologies like
Flexible AC Transmission System (FACTS)

Abstract
India is truculent to meet the
electric power demands of a fast expanding
economy. Restructuring of the power industry
has only increased several challenges for the
power system engineers. The proposed vision of
introducing viable Smart Grid (SG) at various
levels in the Indian power systems has
recommended that an advanced automation
mechanism needs to be adapted. Smart Grids are
introduced to make the grid operation smarter
and intelligent. Smart grid operations, upon
appropriate deployment can open up new
avenues and opportunities with significant
financial implications. This paper presents
various Smart grid initiatives and implications in
the context of power market evolution in India.
Various examples of existing structures of
automation in India are employed to underscore
some of the views presented in this paper. It also
Reviews the progress made in Smart grid
technology research and development since its
inception. Attempts are made to highlight the
current and future issues involved for the
development of Smart Grid technology for
future demands in Indian perspective.
Index Terms-- Smart Grid; Indian Electricity Act
2003; Availability Based Tariff (ABT); Demand
Side Management (DSM); Renewable Energy;
Rural Electrification (RE); Micro Grid.

I. INTRODUCTION
THE economic growth of
developing countries like India depends
heavily on reliability and eminence of its
electric power supply. Indian economy is
anticipated to grow at 8 to 9% in 2010- 2011

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

88

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

and Wide Area Monitoring (WAM) to


redefine the capability of power system
engineering and unite the necessity of the
rural, suburban and urban regions across the
globe under single roof [4]. In addition, the
technology employs the reduction of carbon
footprints and foot-dragging the greenhouse
gas emission. This paper designates about
the Smart Grid initiatives along with various
examples of existing structures of
automation in India. It also reviews the
encroachment made in Smart Grid
technology in R&D, initiated by various
public and private sector organizations
supported by prominent institutions across
the globe. Limelight on the current and
future issues involved for the development
of Smart Grid technology for future
demands has also been debated. The
organization of the paper is as follows: In
section II, an overview of the Indian Power
market along with its current strategy of
power system is presented. Section III
describes the vision of India on Smart Grid
(SG) technology along with section IV
debriefing about the prevailing units and its
future enactments. Section V reveals some
of the required focus areas and advent of
enhanced smart grid technologies. Section
VI is dedicated to general conclusion
followed by references.

ISBN: 378 - 26 - 138420 - 5

provision for national policy, Rural


Electrification (RE), open access in
transmission, phased open access in
distribution, mandatory state electricity
regularity commission (SERCs), license free
Generation and distribution, power trading,
mandatory metering, and stringent penalties
for theft of electricity. In
Addition to these guidelines, a concept
called as Availability Based Tariff (ABT)
has also been implemented to bring effective
day ahead scheduling and frequency
sensitive charges for the deviation from the
schedule for efficient real-time balancing
and grid discipline. Exclusive terms like
fixed cost and variable cost, and
unscheduled interchange (UI) mechanism in
ABT acts as a balancing market in which
real-time price of the electricity is
determined by the availability and its
capacity to deliver GWs on day-to-day
basis, on scheduled energy production and
system frequency . Indian power system has
an installed capacity of around 164 GW and
meets a peak demand of 103 GW.
According to the Current five year plan
(2007-2012) by the year 2012, the installed
capacity is estimated to be over 220 GW and
the peak demand is expected to be around
157 GW and is projected to reach about 800
GW by next two decades. However certain
complexities are envisaged in integrating
IPPs into grid such as, demarcation,
scheduling, settlement and gaming. But
these issues are being addressed by proper
technical and regulatory initiatives. In
addition to that, the transmission Sector has
progressed in a very subsequent rate,
currently at installed capacity of 325,000
MVA at 765, 400, 220kV Voltage levels
with 242,400 circuit kilometers (ckt-km) of
HVAC and HVDC transmission network,
including 765kV transmission system of
3810 ckt-km . On distribution sector, the
Ministry of Power has also maneuvered to
leverage the digital technology to transform

I. OVERVIEW OF INDIA POWER M


ARKET AND ITS STRATEGY
The re-evaluation of the
Indian Electricity Supply Act, 1948 and
Indian Electricity Act, 1910, has led the
Electricity Act 2003 which has facilitated
government and many non-government
organizations to participate and to alleviate
the electricity demand. The act redefines the
power market economy, protection of
consumers interest and provision of power
to urban, sub-urban and rural regions across
the country. The act recommends the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

89

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

and reshape the power sector in India to


make an open and flexible architecture so as
to meet the core challenges and burning
issues, and get the highest return on
investment for the technology .The
Electricity Act 2003, created a liberal and
competitive
environment,
facilitating
investments by removal of energy barriers,
redefining the role of system operation of
the national grids. New transmission pricing,
loss allocation schemes, introduction of
ULDC scheme and Short Term Open Access
(STOA) schemes have been introduced
based on distance and direction so that
power could be traded from any utility to
any utility across the nation on a nondiscriminatory basis currently; Indian
transmission grid is operated by a pyramid
of 1 NLDC, 5 RLDCs and 31 SLDCs,
monitoring round the clock with SCADA
system enabled with fish as well as bird
eye view, along with advance wideband
speech
and
data
communication
infrastructure. In addition, other key features
like
smart
energy
metering,
CIM,Component Interface Specification
(CIS), synchro phasor technology, Wide
Area Monitoring (WAM) system using
phasor
measurements,
enhanced
visualization and self-healing functions are
being exclusively employed
III. VISION
TECHNOLOGY

OF

INDIA

ON

ISBN: 378 - 26 - 138420 - 5

the scenario of the nations electric power


grid, by the convergence of information and
operational technology applied to electrical
grid, allowing sustainable option to the
customers and upgraded security, reliability
and efficiency to utilities The elite vision of
Smart Grid (SG) Technology allows energy
to be generated, transmitted, distributed and
utilized more effectively and efficiently.
Demand Side Management (DSM) is an
essential practice for optimized and effective
use of electricity, particularly in the
developing countries like India where the
demand is in excess of the available
generation. Such kind of non-technical
losses can be overcome by electricity grid
which focuses on advanced control and
communication protocols integrated with the
utility providing a complete package for the
requirement of Smart Grid. With the
introduction of the Indian Electricity Act
2003, the APDRP was transformed to
restructured APDRP (R-APDRP) which has
improvised the operation and control , , and
has attempted a seamless integration of
generation (including distributed energy
resources
(DER),
transmission
and
distributed system through usage of
intervening information technology (IT) that
uses high speed computers and advance
communication network, and employing
open standard with vendor-neutrality is
deemed a cornerstone for embracing the upand-coming conceptualization of Smart Grid
for India scenario. A vivid study of the
power scenario has been illustrated each
classified rendering to the timeline in brief.
Introducing with the power strategy
management in the past, the whole system
was monitored and controlled using
telephonic medium which was purely a bluecollar job. The system was solely dependent
on a single generation unit or the
interconnected substations. On further
progress in science and technology, the

SMART GRID

Due to the consequence of cutting


edge technology, buzzwords like energy
conservation and emission reduction, green
sustainable development, safety factor,
reduction ofT&D losses, optimal utilization
of assets, have turn out to be the core of
discussion. AsIndia is struggling to meet its
electricity demands, both in terms of Energy
and Peak Load, Smart Grids can help better
manage the shortage of power and optimize
the power grid status in the country. A
Smart Grid is a perception of remodeling

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

90

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

IV. SMART GRID INITIATIVES IN INDIA

system is monitored round the clock using


advance data communication protocols. As
well the substation has the islanding facility
with immediate power backups to maintain
the grid stable. India as a developing
country, the scenario of the power system
changes in exponential basis. Moreover the
system is expected to be more reliable and
flexible with its advancement in data
communication and data analysis facility.
Fig. 1 illustrates about the advancement and
it
immediate
results
during
its
implementation in future. The conclusive
approach for the Indian Smart Grid would
be visualized accordingly, with latest
technological advancement and extensive
features as shown in Fig. 2 .

As it has been acknowledged


earlier that, Smart Grid technology has a
widespread overview of transforming the
Indian power grid from technology based
standard to performance based standard. The
Ministry of Power (Mop) participated in the
SMART 2020 event with The Climate
Group and The Global Sustainability
Initiative (GeSI) in October 2008 which
aimed to highlight the reports relevant to
key stakeholders in India Unfortunately, the
possible way forward has not yet been
drilled out and is still a question mark for
the Government. But to facilitate demand
side management distribution networks has
been fully augmented and upgraded for IT
enabling, which has enhanced the grid
network with amended customer service.
Table-1 provides a brief analysis of some of
the initiative which has been taken under the
supervision of many government and private
bodies and allies. In the view of multitude
that could be accrued, it is suggested that
there should be ample Government
regulatory support and policy initiatives to
move towards Smart Grids. India is in its
nascent stage of implementing various other
controls and monitoring technology, one of
such is ADA Further researches are being
carried out in some of the elite institutes in
the country in collaboration with some of
the various multinational companies and
power sectors across the nation.

Hierarchy of super grid

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

91

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

With the perception of renewable energy,


the energy converges to; reduction in carbon
footprints, cleaner environment, plug-in EV,
decentralized power which increases the
quality of living standard and enhances the
power system quality along with the stability
of the grid network. But in contrary to that
the power quality also bids some of the
potential challenges such as; voltage
regulation, power system transient and
harmonics, reactive power compensation,
grid synchronization, energy storage, load
management and poor switching action etc.,
[27]. These problems are mainly visualized
for major renewable energy sources like
wind and solar energy. Other energy sources
like biomass, hydro and geothermal sources
have no such significant problem on
integration of grid.

V. ENHANCED SMART GRID TECHNOLOGY


Due to advent of advance
information and communication technology
(ICT) and proliferation of green energy, its
liable that Smart Grid technology transforms
to more superior and advanced form. Some
the newly innovated prospects like
renewable
energy
integration,
rural
electrification and micro grid are to be
featured in it [25].
. A. Renewable Energy Integration
Present-day
environmental
awareness, resulting from coal fired power
station, has fortified interest in the
development of the modern smart grid
technology and its integration with green
and sustainable energy. Table-2 provides
and brief analysis of the renewable energy
development in India which has been
planned according to Five year Plans by the
Indian Government and the Ministry of New
and Renewable Energy (MNRE)

Integration of renewable with


the Smart Grids makes the system more
reliable and flexible in economic load
dispatch, not only in a specified location but
in a wide area, even between the nations.
Nordic counties have practiced such grid
integration among its neighboring nations
and still future implementations are being
focused on [28].However, forecasting
approaches, design algorithm and other
models are being developed by many
research analysis teams and are to be
established in many regions across the
nationwide. Fig. 4 below represents a brief
analysis of solicitation of renewable in smart
grid technology in its whole network of
power system engineering.

SMART GRID INITIATIVES IN INDIA BY VARIOUS


ORGANIZATIONS

The volatility of fossil fuels has opened the


ground for new and renewable energy
sources. With the inherent unpredictability,
the implementations of the renewables need

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

92

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

to have motivating government policies and


well established standards. Proper financial
support is the governing factor for a
generation deficient and developing country
like India. Wind and the photo voltaic cell
should be supported by upcoming
technologies like Micro Grid and ICT [27].
Such emerging technologies will play a
major role in sustainable standard of living
with economical insolence. Large scale.

ISBN: 378 - 26 - 138420 - 5

Rajiv Gandhi Gramen Vidyutkaran Yojana


(RGGVY).Other schemes like, Pradhan
Mantri Garmodaya Yojana (PMGY), Three
phase feeders-single phasing and Smart
metering,KutirJyotiProgram(KJP),Accelerat
ed Rural Electrification Program (AREP),
Rural Electricity Supply Technology
Mission (REST), Accelerated Electrification
of one hundred villages and 10 million
households, Remote Village Renewable
Energy Program me (RVREP) and Gridconnected Village Renewable Programme
(GVREP) [5], [29-30]. Some of them have
got a remarkable success but some of them
got trapped in for their own interest due to
various non-technical issues [31], [32].
Some of the key features of such projects
are; to achieve 100% electrification of all
villages and habitation in India, provide
electricity access to all households, free-ofcost electricity to BPL households, DG
system, smart based metering, promote fund,
finance and facilitate alternative approaches
in rural electrification, single light solar
lightning system for remote villages and its
hamlets.
The present rural electrification
scenario in the nation is still uncertain, and
is yet to be put on more exploration and
verified by the Ministry of Power (Mop) and
Ministry of New and Renewable Energy
(MNRE). Over 500,000 thousand of Indias
600,000 thousand villages are deemed to be
electrified [33]. As in such case, the Indian
Government and Indian businesses sector
would need to invest on more such projects
and schemes, for low-footprint technologies,
renewable sources of energy, smart metering
and resource efficient infrastructure.

B.Rural Electrification
Technologies are advancing
day-by-day, Smart distribution technologies
allowing for increased levels of distributed
generation have a high potential to address
rural electrification needs and minimize the
erection costs, transmission losses and
maintenance costs associated with large
transmission grids. Rural Electrification
Corporation Limited (REC) is a leading
public infrastructure finance company in
Indias power sector which finances and
promotes rural electrification projects
across the nation, operating through a
network of 13 Project
offices and 5 Zonal offices. Along with the
government of India has launched various
programs and schemes for the successful
promotion and implementation of rural
electrification. One such major scheme is

Suggestions for Future Works


As the report only had pulled
the grid connection requirement for wind
power generation, which has been planned

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

93

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

to stretch upon to the study of photovoltaic


(PV) and its grid connection planning in
Indian scenario. Also, few more work
related to micro grids and hybrid energy
with energy storage system is premeditated
to complete by near future. Upon the
finalizing of the entire study, the further
research perspective would deliberately act
as an advocate to discover the rank and
strategy of nations development in power
and energy with respect to current and future
energy demand.

ISBN: 378 - 26 - 138420 - 5

regulatory affairs of India is also presented


here. In this connection, the paper should act
as advocate to bring forth the significance
and fortification of Smart Grid philosophy
and implanting it on the basis of proposed
ideology in Indian subcontinent.
REFERENCES
[1]Sinha, A.; Neogi, S.; Lahiri, R.N.; Chowdhury, S.;
Chowdhury, S.P.; Chakraborty, N.; , "Smart grid
initiative for power distribution utility in India," IEEE
Power and Energy Society General Meeting, 2011 , vol.,
no., pp.1-8, 24-29 July 2011
[2]The Green Grid: Energy Savings and Carbon
Emission Reductions Enabled by a Smart Grid,
EPRI Palo Alto, CA: 2008
[3]V.S.K. Murthy Balijepalli, S.A. Kharparde, R.P.
Gupta, Yemula Pradeep, Smart Grid Initiatives and
Power Market in India Proc. Of IEEE Power and
Energy Society General Meeting, pp.1-7, Jul. 2010.
[4]Bossart, S.J.; Bean, J.E.; , "Metrics and benefits
analysis and challenges for Smart Grid field projects,"
Energetic, 2011 IEEE, vol., no., pp.1-5, 25-26 May
2011.
[5]Electricity Act 2003, Govt. of India, New Delhi,
2003
[6]Central
Electricity
Authority,
2010.
[Online]Available:http://www.cea.nic.in/reports/electric
ity_act2003.pdf [7] Ministry of Power, Government of
India
Website.
[Online]
Available:
http://powermin.nic.in, Nov. 2009
[8]Pradeep, Y.; Thomas, J.; Sabari, C.L.;
Balijepalli,V.S.K.M.; Narasimhan, S.R.; Khaparde,
S.A.; , "Towards usage of CIM in Indian Power Sector,"
IEEE Power and Energy Society General Meeting, 2011
, vol., no., pp.1-7, 24-29 July 2011.
[9]Central Electricity Authority, 2010. [Online]
Available:http://www.cea.nic.in/reports/yearly/energy_g
eneration10_11.pdf.
[10]Raoot, M.G.; Pentayya, P.; Khaparde, S.A.; Usha,
S.; , "Complexities in integrating IPPs in Indian power
system," IEEE Power and Energy Society General
Meeting, 2010, vol., no., pp.1-9, 25-29 July 2010.
[11]Central Electricity Authority, 2010. [Online]
Available:
http://www.cea.nic.in/power
sec
reports/executivesummary/2010
08/index.htm. [12]Power Grid Corporation of Indian
Limited, Unified Load Despatch & Communications
pp.1-7, 25-29 July 2010.

CONCLUSIONS
The paper presents a
discussion on Indian Power Strategy along
with its pitfalls in various technical and nontechnical themes, with an organized
approach to evolve the conceptualization of
Smart Grid. An overview of Indian Power
Market along with brief analysis about the
power system units is described. Power
market in India is generally characterized by
the poor demand side management and
response for lack of proper infrastructure
and awareness. Smart Grid Technology can
intuitively overcome these issues. In
addition to that, it can acknowledge
reduction in line losses to overcome
prevailing power shortages, improve the
reliability of supply, power quality
improvement
and
its
management,
safeguarding revenues, preventing theft etc..
Model architecture as well as Indias Smart
Grid initiatives taken by the government and
many private bodies, are presented in the
paper. Further, various prospects of
sustainable energy and off-grid solutions,
Rural Electrification (RE) and evolution of
Micro Grid along with various policies and

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

94

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Design and Analysis of Water Hammer Effect in a Network


of Pipelines.
V. Sai Pavan Rajesh *
*

Department of Control Systems, St. Marys Group of Institutions, Jawaharlal Nehru Technological UniversityHyderabad, Main Road, Kukatpally Housing Board Colony, Kukatpally, Hyderbad, Telangana, India.
*

[email protected],

Abstract-There will be a chance for the destruction of the system due to transient if it is not provided with adequate
protection devices. Generally, transient takes place when parameters involving in conventional flow are distorted
with respect to the time. Rapid closing of valve in a pipe network will be resulting into hydraulic transient known as
water hammer occurs due to sudden change in pressure and velocity of flow with respect to time. Due to impulsive
action, pressure surges are induced in the system travel along the pipe network with the rapid fluid acceleration
leading to the dramatic effects like pipe line failure, damage to the system etc. Considering the importance of
hydraulic transient analysis, we design a system capable of verifying pipe network containing fluid flow.This paper
demonstrates design of different pipe structures in pipe line network and analysis of various parameters like excess
pressure distribution, velocity variations and water hammer amplitude with respect to time using COMSOL
Multiphysics v 4.3. The magnitude of water transient in pipe line network at different pressure points has been
discussed in detail.

Keywords- COMSOL, Pressure distribution, Velocity variation, Water Hammer.

Corresponding Author E-mail*: [email protected]

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

95

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

I.

ISBN: 378 - 26 - 138420 - 5

INTRODUCTION
The key to the conservation of water is good water measurement practices. As fluid will be running in
water distribution system, system flow control is dependent based on the requirement for opening or closing of valves, and
starting and stopping of pumps. When these operations are performed very quickly, they convert the kinetic energy carried
by the fluid into strain energy in pipe walls, causing hydraulic transient

[1]

phenomena to come into existence in the water

distribution system i.e., a pulse wave of abnormal pressure is generated which travels through the pipe network. Pressure
surges that are formed or fluid transients in pipelines are referred to as Water hammer. This oscillatory form of unsteady
flow generated by sudden changes results in system damage or failure if the transients are not minimized. So now the steady
state flow conditions are altered by this effect [2] resulting in the disturbance to the initial flow conditions of the system.
Where the system will tend to obtain a static flow rate by introduction of new steady state condition. The intensity of water
hammer effects will depend upon the rate of change in the velocity or momentum. Conventional water hammer analyses
provide information under operational conditions on two unknown parameters i.e., pressure and velocity within a pipe
system. Generally effects such as unsteady friction, acoustic radiation to the surroundings or fluid structure interaction are
not taken into account in the standard theory of water hammer, but were considered in general approach [3]. But mechanisms
acting all along the entire pipe section such as axial stresses in the pipe and at specific points in the pipe system such as
unrestrained valves will fall under fluid structure interaction extension theory for conventional water hammer method.

Figure 1. Pipe connected to control valve at the end with water inlet from reservoir.
In the past three decades, since a large number of water hammer events occurred in the light-water- reactor power plants [4], a
number of comprehensive studies on the phenomena associated with water hammer events have been performed. Generally
water hammer can occur in any thermal-hydraulic systems and it is extremely dangerous for the thermal-hydraulic system
since, if the pressure induced exceeds the pressure range of a pipe given by the manufacturer, it can lead to the failure of the
pipeline integrity. Water hammers occurring at power plants are due to rapid valve operation [5], void induced operation, and
condensation induced water hammer [6]. In existing Nuclear Power Plants water hammers can occur in case of an inflow of
sub-cooled water into pipes or other parts of the equipment, which are filled with steam or steam-water mixture [7].

The water hammer theory has been proposed to account for a number of effects in biofluids under mechanical stress, as in
the case of the origin of Korotkoff sounds during blood pressure measurement
cavity within the spinal cord

[8, 9]

, or the development of a fluid-filled

[10]

. In the voice production system, the human vocal folds act as a valve

[11

which induces

pressure waves at a specific point in the airways (the glottis), through successive compressing and decompressing actions
(the glottis opens and closes repeatedly). Ishizaka was probably the first to advocate in 1976 the application of the water
hammer theory, when discussing the input acoustic impedance looking into the trachea [12]. More recently, the water hammer
theory was invoked in the context of tracheal wall motion detection

[13]

. Generally Water utilities, Industrial Pipeline

Systems, Hydropower plants, chemical industries, Food, pharmaceutical industries face this water transient problem.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

96

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

The present work reports the design of different pipe channels and analysis of the pressure distribution and velocity
variation produced all along the pipe flow network when subjected to one pressure measuring point. Various parameters like
inlet input pressure, wall thickness and measurement point are changed for analysis.

II.

USE OF COMSOL MULTIPHYSICS


The software package selected to model and simulate the pipe flow module was COMSOL Multiphysics Version

4.3. It is a powerful interactive environment for modelling and Multiphysics were selected because there was previous
experience and expertise regarding its use as well as confidence in its capabilities. A finite element method based
commercial software package, COMSOL Multiphysics, is used to produce a model and study the flow of liquid in different
channels. This software provides the flexibility for selecting the required module using the model library, which consists of
COMSOL Multiphysics, MEMS module, micro fluidics module etc. Using tools like parameterized geometry, interactive
meshing, and custom solver sequences, you can quickly adapt to the ebbs and flows of your requirements, particle tracing
module along with the live links for the MATLAB. At present this software can solve almost problems in multi physics
systems and it creates the real world of multi physics systems without varying there material properties. The operation of this
software is easier to understand and easier to implement in various aspects for designers, in the form of finite element
analysis system.

Figure 2. Multiphysics modelling and simulation software-COMSOL


In this model as the valve is assumed to close instantaneously the generated water hammer pulse has a step function like
shape. To correctly solve this problem requires a well posed numerics. The length of the pipe is meshed with N elements
giving a mesh size dx = L/N. For the transient solver to be well behaved requires that changes in a time step dt are made on
lengths less than the mesh size. This gives the CFL number condition

CFL= 0.2= c.dt/dx

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

97

(1)

www.iaetsd.in

INTERNATIONAL
CONFERENCE
ON CURRENT
INNOVATIONS
ENGINEERING
378 - 26 - 138420 - 5
Meaning
that changes
during the time
dt maximallyINmove
20 % of theAND
meshTECHNOLOGY
length dx. Thus increasing the meshISBN:
resolution

also requires decreasing the time stepping. This advanced version of software helps in designing the required geometry using
free hand and the model can be analysed form multiple angles as it provides the rotation flexibility while working with it.

III.

THEORITICAL BACKGROUND
Water hammer theory dates to 19th century, where several authors have contributed their work in analyzing this effect.
Among them, Joukowsky

[14]

conducted a systematic study of the water distribution system in Moscow and derived a

formula that bears his name, that relates to pressure changes, p, to velocity changes, v, according to the equation

(2)

P = cU

Where is the fluid mass density and c is the speed of sound. This relation is commonly known as the Joukowsky
equation, but it is sometimes referred to as either the Joukowsky-Frizell or the Allievi equation.
For a compressible uid in an elastic tube, c depends on the bulk elastic modulus of the uid K on the elastic modulus of the
pipe E, on the inner radius of the pipe D, and on its wall thickness. The water hammer equations are some version of the
compressible uid ow equations. The choice of the version is problem-dependent: basic water hammer neglects friction and
damping mechanisms, classic water hammer takes into account uid wall friction, extended water hammer allows for pipe
motion and dynamic Fluid Structure Interaction [15, 16].

In water hammer at static condition pressure wave is a disturbance that propagates energy and momentum from one point to
another through a medium without significant displacement of the particles of that medium. A transient pressure wave,
subjects system piping and other facilities to oscillating at high pressures and low pressures. This cyclic loads and pressures
can have a number of adverse effects on the hydraulic system. Hydraulic transients can cause hydraulic equipments in a
pipe network to fail if the transient pressures are excessively high. If the pressures are excessively higher than the pressure
ratings of the pipeline, failure through pipe or joint rupture, or bend or elbow movement may occur. Conversely, excessive
low pressures (negative pressures) can result in buckling, implosion and leakage at pipe joints during sub atmospheric
phases. Low pressure transients are normally experienced on the down streamside of a closing valve. But when the valve is
closed energy losses are introduced in the system and are normally prescribed by means of an empirical law in
terms of a loss coecient. This coecient, ordinarily determined under steady ow conditions, is known as the valve
discharge coecient, especially when the pipeline is terminated by the valve. It enables to quantify the ow response in
terms of the valve action through a relationship between the ow rate and pressure for each opening position of the
valve. The discharge coecient provides the critical piece of missing information for the water hammer analysis. Because
the existing relationship between pressure and ow rate is often a quadratic law type, the empirical coecient is defined
in terms of the squared ow rate. When water distribution system comprising a short length of pipes (i.e., <2,000 ft. {600m})
will usually be less vulnerable to problems with hydraulic transient. This is because wave reflections e.g., at tanks,
reservoirs, junctions tend to limit further changes in pressure and counteract the initial transient effects. An important
consideration is dead ends, which may be caused by closure of check valves that lock pressure waves into the system in
cumulative fashion. Wave reflections will be both positive and negative pressures; as a result the effect of dead ends must be
carefully evaluated in transient analysis.

These pressure surges provide the most effective and viable means of identifying weak spots, predicting potentially
negative effects of hydraulic transient under a number of worst case scenarios, and evaluating how they may possibly be
avoided and controlled. The basic pressure surge modeling is based on the numerical conservation of mass and linear

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

98

www.iaetsd.in

momentum equations. For this Arbitary Lagrangian Elurian(LE)[17] numerical solution helps in providing the exact analytical

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

solution. On the other hand when poorly calibrated hydraulic network models results in poor prediction of pressure surges
thus leading to more hydraulic transients. In more complex systems especially, the cumulative effect of several types of
devices which influence water hammer may have an adverse effect. However, even in simple cases, for example in pumping
water into a reservoir, manipulations very unfavorable with regard to water hammer may take place. For example, after the
failure of the pump, the operator may start it again. Much depends on the instant of this starting. If it is done at a time when
the entire water hammer effect has died down, it is an operation for which the system must have been designed.

IV.

DESIGN PROCEDURE
The design and analysis of the hydraulic transient in a pipe flow includes geometry, defining the parameters for the
required geometry, providing mesh & inputs. The 3D model is constructed in the drawing mode of COMSOL Multiphysics.
In this, a pipe of length L = 20 m is constructed assuming that one end is connected to a reservoir, where a valve is placed at
the other end. The pipe with inner radius of 398.5mm, the thickness of the wall about 8 mm and Youngs modulus of
210GPa was designed. In order to verify pressure distribution, a pressure sensor measurement point at a distance of z 0 =
11.15 m from the reservoir was arranged and flow has been sent into pipe with an initial flow rate of Q 0= 0.5 m3/sec.

Figure 3. Single pipe line with closing vale at the output.

Figure 4. Three pipe line intersection in a network.

After designing the geometry for a flow channel in a pipe, materials are to be selected from the material browser. Water,
liquid and structural steel are selected from the built in section of material browser. Edges are selected for water flow and
steel pipe model sections. Now pipe properties are defined one by one by first selecting the round shape from the shape list
of pipe shape. Initially the reservoir acts as a constant source for pressure producing p 0 which is equal to 1 atm. As the fluid
is allowed to flow from the reservoir tank into the pipe model, the fluid enters the left boundary of the pipe first and leaves
the right boundary of the pipe with the valve in open condition. As the valve is open water is flowing at a steady flow rate, so
at time t = o seconds the valve on the right hand side is closed instantaneously creating disturbance in the normal flow rate
leading to a change in discharge at the valve. As a result of the compressibility of the water and the elastic behaviour of the
pipe a sharp pressure pulse is generated travelling upstream of the valve. The water hammer wave speed c is given by the
expression

1/ c2 =1/c2s + A

(3)

Where cs is the isentropic speed of sound in the bulk fluid and is 1481 m/s while the second terms represents the component
due to the pipe flexibility. The water density is and A the pipe cross sectional compressibility. Resulting in an effective
wave speed of 1037 m/s. The instantaneous closure of the valve will results in a water hammer pulse of amplitude P given by
Joukowskys fundamental equation [18]

(4)

P = cuo

Where u0 is the average fluid velocity before valve was closed. Exact solution can only be obtained based on the verification
of the pipe system and valve point [19]. Study was extended for different pipe line intersection models based on the reference

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

99

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

model

[20]

ISBN: 378 - 26 - 138420 - 5

. To design a three axis pipe line intersection, from geometry more primitives, three polygons are chosen where

first polygon coordinates corresponds to x in (1 z0 L), while y & z remain in (0 0 0). For second polygon y coordinates are (1
z0 L) and the remaining are left for (0 0 0), in a similar way z coordinates are (1 z0 L), and the resulting geometry is
depicted in figure 4.

V.

RESULTS AND DISCUSSION


Using this Multiphysics modelling and simulation COMSOL software three different kinds of studies were carried out i.e.,
Study 1 is the time dependent is based on the fluid flow and its corresponding interaction with the pipe which results in
pressure distribution along the pipe line along with the velocity variation. Where the Study 2 is not only based on the excess
pressure measurement point but also valve point. On the other hand Study 3 corresponds to the pressure profile in the pipe
line with respect to different times As it is a time dependent study, the ranges of the times are edited in the study setting
section from range (0, 0.005, and 0.75) to range(0, 1e-3, 0.24). Now from study 1 right click and select show default solver.
Then expand the solver window and click the time stepping section in which the maximum step check box is marked and in
the associated edit field type dt. Now right click study 1 to compute the results. So pressure along pipe line at t=0.24s is
obtained along with the velocity variation. Results are analysed by transforming the pipe line shapes to different geometrys.
From the analysis of results, it allow us to conclude that when more pipes lines are inter connected there is more chance for
water transient to tack place at an easy rate and cause damage to the system.

VI.

MESHING
Meshing can provide the information of required outputs anywhere on the proposed structure along with the input given.
Numerical ripples are visible on the point graphs of excess pressure history at pressure sensor and water hammer amplitude.
As the closure in the valve is instantaneous the pressure profile has a step like nature. This is difficult to resolve numerically.
The ripples can be reduced by increasing the mesh resolution parameter N. So the number of mesh points (N) selected in this
model is about 400. In this model meshing is done for the Edges of the pipe. Where the maximum element size of the
parameter is defined for L/N m (L= 20 m & N= 400) and the minimum element size is 1[mm].

VII.

SIMULATION
In this study, the simulations are performed using the fundamental equation in water hammer theory that relates to
pressure changes, p, to velocity changes, v, according to the equation (2). Simulation comprises of application of different
input initial pressure at the inlet portion for different pipe network sections. Pressure measurement points are changed along
the pipe length L and computed for time interval from T=0 to T=0.24 seconds. Both the velocity and pressure are measured
at the above time interval. Other parameters like water hammer amplitude, maximum and minimum pressure for two
different geometrys; along with velocity variations are listed in table 1.
Table 1. Pressure distribution and velocity variation values for single pipe line & three pipe lines geometry.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

100

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Parameters

Single pipe

Three pipes

Min. Pressure at T=0.07s

-9.579*105 Pa

-1.577*105 Pa

Max. Pressure at T=0.07s

4.047*105 Pa

1.1617*106 Pa

Velocity

Variation

Range

-4
at 4.257*10 to 1.1632 m/s

ISBN: 378 - 26 - 138420 - 5

0.2672 to 1.2742 m/s

T=0.23s
Excess

Pressure

distribution 1.35*106 Pa

-0.85*106 Pa

along the pipe for t = 0.24 s.

VIII.

CONCLUSION
Flow channel is designed in a pipe network and its reaction with the valve when closed was analysed using
COMSOL Multiphysics Version 4.3. Simulation for the proposed model is done by changing the initial flow rates along with
the pipe networks to explore the variations in the properties of fluid like pressure distribution and velocity variation with
respect to time. When inlet mean velocity is increased the magnitude of water hammer amplitude remains the same but the
chances for the water transient is more which results in easy breakdown of pipe section. When multiple pipe line was
connected the maximum pressure distribution and velocity variation were very less even though the water hammer amplitude
remains the same when compared for different cases. Positive pressure difference exists when multiple pipes were connected
where negative pressure difference exists for singe pipe line geometry, which states that network of pipe lines results in less
water transient effect. This study can be extended by observing the changes in the flow by inclining the pipe line, by using T,
L shaped piping geometries. Further extension was made for micro piping system by changing the dimensions of the
geometry from meters to micro meters. This study helps in building micro piping network systems that are used in bio
medical applications and Automobile industries.

IX.

ACKNOWLEDGEMENTS
The authors would like to thank NPMASS for the establishment of National MEMS Design Centre (NMDC) at
Lakireddy Bali Reddy Engineering College. The authors would also like to thank the Director and Management of the
college for providing the necessary facilities to carry out this work.

REFERENCES
1.

Avallone, E.A., T. Baumeister ID, "Marks' Standard Handbook for Engineers" McGraw-Hill, 1987, 9th Edition.
pp. 3-71.

2.

Moody, F. J., "Introduction to Unsteady Thermo fluid Mechanics" John Wiley and Sons, 1990,Chapter 9, page
405.

3.

A.S. Tijsseling, Fluid-structure interaction in liquid-lled pipe systems: A review, Journal of Fluids and
Structures, 1996,(10), PP. 109-146.

4.

Algirdas Kaliatka.; Eugenijus Uspuras.; Mindaugas Vaisnoras.; Analysis of Water Hammer Phenomena in
RBMK-1500 Reactor Main Circulation Circuit, International Conference on Nuclear Energy for New Europe
2006, Portoro, Slovenia, 2006 September 18-21.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

101

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

5.

ISBN: 378 - 26 - 138420 - 5

M. Giot, H.M. Prasser, A. Dudlik, G. Ezsol, M. Habip, H. Lemonnier, I. Tiselj, F. Castrillo, W. Van Hove, R.
Perezagua, & S. Potapov, Twophase flow water hammer transients and induced loads on materials and structures
of nuclear power plants (WAHALoads) FISA-2001 EU Research in Reactor Safety, Luxembourg, 2001,
November, pp. 12- 15.

6.

P. Griffith, Screening Reactor Steam/Water Piping Systems for Water Hammer, Report Prepared for U. S. Nuclear
Regulatory Commission NUREG/CR-6519, 1997.

7.

M. Giot, J.M. Seynhaeve, Two-Phase Flow Water Hammer Transients : towards the WAHA code, Proc. Int. Conf.
Nuclear Energy for New Europe '03, Portoro, Slovenia, 2003,Sept. 811, Paper 202, 8p.

8.

D. Chungcharoen, Genesis of Korotkoff sounds, Am. J. Physiol.; 1964, 207, pp. 190194.

9.

J. Allen, T. Gehrke, J. O. Sullivan, S. T. King, A. Murray , Characterization of the Korotkoff sounds using joint
time-frequency analysis, Physiol. Meas.; 2004, (25), Pp. 107117.

10. H. S. Chang, H. Nakagawa, Hypothesis on the pathophysiology of syringomyelia based on simulation of


cerebrospinal fluid dynamics, Journal of Neurology Neurosurgery and Psychiatry, 2003, (74), pp. 344347.
11. N. H. Fletcher, Autonomous vibration of simple pressure-controlled valves in gas flows, J. Acoust. Soc. Am, 1993,
93 (4), pp. 21722180.
12. K. Ishizaka, M. Matsudaira, T. Kaneko, Input acoustic-impedance measurement of the sub glottal system, J.
Acoust. Soc. Am, 1976, 60 (1), pp. 190197.
13. G. C. Burnett, Method and apparatus for voiced speech excitation function determination and non-acoustic assisted
feature extraction, U.S Patent 20020099541, (2002) A1.
14. N. Joukowsky, Uber den hydraulischen stoss in wasserleitungsrohren, Memoires de lAcademie Imperiale des
Sciences de St. Petersbourg Series 8, 9 (1900).
15. F. DSouza, R. Oldeburger, Dynamic response of uid lines, ASMEJournal of Basic Engineering, 1964, (86), pp.
589598.
16. D. J. Wood, A study of the response of coupled liquid ow-structural systems subjected to periodic disturbances,
ASME

Journal of Basic Engineering, 1968,(90) , pp.532540.

17. J. Donea, Antonio Huerta,J. Ph. Ponthot and A. Rodrguez-Ferran, Arbitrary Lagrangian Eulerian
Methods,UniversitedeLi`ege, Li`ege, Belgium.
18. M.S. Ghidaoui, M. Zhao, D.A. McInnis, and D.H. Ax worthy, A Review of Water
a. Hammer Theory and Practice, Applied Mechanics Reviews, ASME, 2005.
19. A.S. Tijsseling, Exact Solution of Linear Hyperbolic Four-Equation Systems in Axial Liquid-Pipe Vibration,
Journal Fluids and Structures, vol. 18, pp. 179196, 2003.
20. Model library path: pipe_flow_module /verification _models /
water_hammer_verification{.http://www.comsol.co.in/showroom/gallery/12683/}

FIGURES AND TABLES


CAPTION FOR FIGURES
FIGURE 1. Pipe connected to control valve at the end with water inlet from reservoir.
FIGURE 2. Multiphysics modelling and simulation software-COMSOL
FIGURE 3. Single pipe line with closing vale at the output.
FIGURE 4. Three pipe line intersection in a network.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

102

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

FIGURE 5. Pressure distribution at T=0.07s for single pipe line geometry.


FIGURE 6. Pressure distribution at T=0.07s for three pipe line intersection geometry.
FIGURE 7. Velocity variation at T=0.23s for single pipe.
FIGURE 8. Velocity variation at T=0.23s for three pipe line intersection geometry.
FIGURE 9. Excess pressure history measured at the pressure sensor for single pipe line geometry.
FIGURE 10. Excess pressure history measured at the pressure sensor for three pipe line geometry.
FIGURE 11. Excess pressure at the valve (green line) & Predicted water hammer amplitude (Blue line) for single pipe line.
FIGURE 12. Excess pressure at the valve (green line) & Predicted water hammer amplitude (Blue line) for three pipe lines.
FIGURE 13. Excess pressure distribution along the pipe for t= 30 s for single pipeline geometry.
FIGURE 14. Excess pressure distribution along the pipe for t= 30 s for three pipeline geometry.

CAPTION FOR TABLE


Table 1. Pressure distribution and velocity variation values for single pipe line & three pipe lines geometry.

Figure 5: Pressure distribution at T=0.07s for single pipe


line geometry.

Figure 6: Pressure distribution at T=0.07s for three


pipe line intersection geometry.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

103

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Figure 7: Velocity variation at T=0.23s for single pipe


line geometry

Figure 8: Velocity variation at T=0.23s for three


pipe line intersection geometry.

Figure 9. Excess pressure history measured at the


pressure sensor for single pipe line geometry.

Figure 10. Excess pressure history measured at the


pressure sensor for three pipe line geometry.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

104

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Figure 11. Excess pressure at the valve (green line) &


Predicted water hammer amplitude
(Blue line) for single pipe.

ISBN: 378 - 26 - 138420 - 5

Figure 12. Excess pressure at the valve (green line) &


Predicted water hammer amplitude
(Blue line) for three pipe lines.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

105

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Figure 13. Excess pressure distribution along the pipe for t= 30 s for single pipeline geometry.

Figure 14. Excess pressure distribution along the pipe for t= 30 s for three pipe line intersection geometry.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

106

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

A Secured Based Information Sharing Scheme via


Smartphones in DTN Routings
T. Mohan Krishna1 V.Sucharita2
2

PG Scholar1, M.Tech. Associate Professor


Department of Computer Science and Engineering,
[email protected]
Audisankara College of Engineering & Technology, Gudur.
A.P - India
Abstract :

With the growing range of smart phone users, peer-to-peer content sharing is anticipated to occur a lot of usually. Thus, new
content sharing mechanisms ought to be developed as ancient information delivery schemes aren't economical for content
sharing attributable to the isolated property between sensible phones. To achieve data delivery in such complex environments,
researchers have anticipated the use of epidemic routing or store-carry-forward protocols, in this a node stores a message may
be a note and carries it until a forwarding chance arises through an encounter with other node. Earlier studies in this field
focused on whether two nodes would come across each other and the place and time of encounter.During this paper, we tend to
propose discover-predict-deliver as associate economical content sharing theme for delay-tolerant Smartphone networks. In our
projected theme,. Specifically, our approach employs a quality learning formula to spot places inside and outdoors. A hidden
markov model and viterbi algorithm is employed to predict associate individuals future quality info. analysis supported real
traces indicates that with the projected approach, eighty seven p.c of contents is properly discovered and delivered among a pair
of hours once the content is out there solely in thirty p.c of nodes within the network .In order to decrease energy consumption
we are using asymmetric multi core processors and efficient sensor scheduling is needed for that purpose we are using POMPDs
in sensor scheduling.
Keywords:- Tolerant Network, ad hoc networks, Store and forward networks, peer-to-peer network
I.INTRODUCTION
Now a days Number of advanced smart phone
clients has rapidly expanded in the recent years. users can
make different sorts of substance effectively utilizing easy to
use interfaces accessible within the advanced smart phones
phones. However substance offering among shrewd
telephones is dreary as it obliges a few activities, for example,
transferring to concentrated servers, seeking and downloading
substance. One simple route is to depend on adhoc technique
for distributed substance imparting. Shockingly with the
current adhoc steering conventions, substance are most
certainly not conveyed if a system parcel exists between the
associates at the point when substance are imparted. Thus
Delay Tolerant System (DTN) steering conventions
accomplish better execution than customary adhoc steering
conventions. These conventions don't oblige an incorporated
server. Thus the substance are put away on advanced mobile
phones itself. Smartphone's consists of many network
interfaces like Bluetooth and Wi-Fi so ad hoc networks can be
easily constructed with them. The Connectivity among
Smartphone's is likely to be alternating because of movement
patterns of carriers and the signal transmission phenomena.
A wide variety of Store-carry-forward protocols have been
anticipated by researchers. Routing in delay-tolerant
networking concerns itself with the ability to route, data from

a source to a destination, which is a essential ability of all


communication networks must have. In these exigent
environments, mostly used or familiar ad hoc routing
protocols fail to establish routes. This is due to these
protocols first try to establish a complete route and then, once
the route has been established forwards the actual data. Still,
when immediate end-to-end paths are complicated or
unfeasible to institute, routing protocols should take to a
"store and then forward" method or approach, where data or a
message is moved and stored incrementally all over the
network in hops that it will finally arrive at its destination. A
general technique used to maximize the likelihood of a
message being effectively transferred is to duplicate many
copies of the message in hops that one will be successful in
reaching its destination. DTN workplace of forty buses and
simulations supported real traces. we tend to concentrate on
store-carry-forward networking situations, during which the
nodes communicate victimization DTN bundle design. Some
smart phones within the network store content that they're
willing to share with others. All smart phone users square
measure willing to get together and provide a restricted
quantity of their resources, like information measure, storage,
and process power, to help others. Our goal is to permit users
to issue queries for content hold on on alternative smart
phones anyplace within the network and to assess the
probabilities of getting the knowledge required. we tend to

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

107

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

assume that smart phones will perform searches on their


native storage, and that we realize the relevant results for a
given question to facilitate looking.
Delay Tolerant Network (DTN) routing
protocols attain enhanced performance than usual ad hoc
routing protocols. Over the proposed DTN routing protocols,
Epidemic routing is a essential DTN routing solution. In
Epidemic routing by vahdat et al[2], messages are forwarded
to each encountered node that does not have a replica of the
same message. This solution exhibits the finest performance
in terms of delivery pace and latency, but it involves abundant
resources, such as storage, bandwidth, and energy.
This paper spotlight mainly on efficiency of content
discovery and its delivery to the targeted destination. Here we
suggest recommendation based discover-predict-deliver
(DPD) as efficient and effective content sharing scheme for
smart phone based DTNs. DPD suppose that smart phones
can hook up when they are in close proximity that is where the
Smartphone users reside for a longer period. Earlier studies
have shown that Smartphone users stay indoors for a longer
period where GPS cannot be accessed. The objective of our
work is to discover solutions to the problems in content
sharing and to minimize the energy consumption using sensor
scheduling scheme and finally improve the performance and
want to reduce energy consumption by using asymmetric
multicore processors and efficient sensor scheduling my using
POMPDs in scheduling scheme.

ISBN: 378 - 26 - 138420 - 5

rule downside and conjointly gift many extensions to permit


OPF to use solely partial routing data and work with different
probabilistic forwarding schemes like ticket-based forwarding.
Implement OPF and several and a number of different and
several other other protocols and perform trace-driven
simulations. Simulation results show that the delivery rate of
OPF is barely five-hitter below epidemic, and two hundredth
bigger than the progressive delegation forwarding whereas
generating five-hitter a lot of copies and five-hitter longer
delay.
B.DTN Routing as a Resource Allocation Problem:
Many DTN routing protocols use a range of mechanisms, as
well as discovering the meeting possibilities among nodes,
packet replication, and network cryptography. The first focus
of those mechanisms is to extend the chance of finding a path
with restricted info, thus these approaches have solely
associate incidental result on such routing metrics as most or
average delivery latency. during this paper we have a
tendency to gift fast associate intentional DTN routing
protocol which will optimize a particular routing metric like
worst-case delivery latency or the fraction of packets that area
unit delivered inside a point. The key insight is to treat DTN
routing as a resource allocation drawback that interprets the
routing metric into per-packet utilities that confirm however
packets ought to be replicated within the system.
C.Resource Constraints:
RAPID (resource allocation protocol for intentional DTN)
routing a protocol designed to expressly optimize AN
administrator-specified routing metric. speedy routes a
packet by opportunistically replicating it till a replica reaches
the destination. Speedy interprets the routing metric to perpacket utilities that verify at each transfer chance if the utility
of replicating a packet justifies the resources used. Speedy
loosely tracks network resources through an impact plane to
assimilate a neighbor hood read of worldwide network state.
To the present finish speedy uses AN in-band management
channel to exchange network state data among nodes
employing a fraction of the obtainable information measure.
D. Epidemic routing:

Fig1: Finding meaning ful places and their simulation area


II.LITERATURE SURVEY
A. Optimal Probabilistic Forwarding Protocol in DTN:
To provide Associate in Nursing optimum forwarding
protocol that maximizes the expected delivery rate whereas
satisfying a definite constant on the amount of forwardings
per message The optimum probabilistic forwarding (OPF)
protocol, we have a tendency to use Associate in Nursing
optimum probabilistic forwarding metric derived by modeling
every forwarding as Associate in Nursing optimum stopping

Epidemic routing may be a easy answer for DTNs, during


which messages square measure forwarded to each
encountered node. Thus, Epidemic routing achieves the best
attainable delivery rate and lowest attainable latency, however
it needs vast information measure and storage resources.
Investigated a settled wave model for the progress of
Epidemic routing. Many approaches are projected to cut back
the overhead and to boost the performance of Epidemic
routing examined variety of various ways to suppress
redundant transmissions. projected conversation strategies,
during which a node chooses a random range between zero
and one, and therefore the message is forwarded to a different

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

108

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

node if the chosen range is higher than a user-predefined


chance. These works belong to resource- aware routing
protocols. different protocols square measure classified into
two groups. Opportunity-based schemes and prediction-based
schemes.

III.RELATED WORK
A delay tolerant network (DTN) is a mobile
network where a existing source-destination path may not
exist amid a pair of nodes and messages are forwarded in a
store-carry-forward routing hypothesis [6]. Vahdat et al [2]
anticipated Epidemic routing as fundamental DTN routing
protocol in which a node forwards a message to each and
every encountered node that does not have a replica of the
message. The solution shows the finest performance terms of
delivery pace and latency but wastes bulk of bandwidth.
A substitute solution was resource based [3], [4],where
systems utilize data mules as message carriers that straightly
delivers the message to the destination. Next, opportunitybased routing protocols make use of history of encounters to
convey a message to the destination [5], [6], [7]. Prediction
based schemes [8], [9], make use of complicated utility
functions to decide whether the forward a message to the
node.

ISBN: 378 - 26 - 138420 - 5

will be communicating using the DTN bundle protocol. A few


devices in the network store content which they are ready to
share with others. All nodes are willing to assist and provide a
restricted amount of their local system resources (bandwidth,
storage, and dispensation power) to aid other nodes. Our
objective is to permit users to issue queries for content that is
stored on the other nodes everywhere in the network and
consider the possibility of such a node to acquire the required
information. To ease searching, we suppose that nodes are
capable to carry out searches on their local storage and
uncover the appropriate results for a given query. The content
sharing process is characterized into two phases:
content discovery phase and the content delivery phase.
In the content discovery phase, the user inputs or enters in a
content sharing application requests for the content. The
application initially searches the content it its own or
individual database and if not found, the application then
creates a query that is forwarded based on the users request.
When the content is found, the content delivery phase is
initiated or commenced, and the content is forwarded to the
query originator.

Yuan et al precisely predicted encounter opportunities by


means of time of encounters. Pitkanen et al anticipated stateof-the-art content sharing scheme in DTNs. They mainly
focused on restraining search query propagation and
anticipated several query processing methods.
Chang et al anticipated a process for searching for a node or
an object in a bulky network and restraining search query
propagation. Here a class or a set of controlled flooding
search strategies where query/search packets are broadcasted
and propagated in the network in anticipation of a preset TTL
(time-to-live) value carried in the packet terminates. The
objective of our work is to discover the content sharing
problem in Smartphone based DTNs involves minimizing
energy consumption using sensor scheduling schemes.
In General, these works are outlined focused around contact
histories. when two or more smart phone clients are in the
same spot, their gadgets don't generally make correspondence
or recognize contact opportunities. Subsequently, the contact
history gives less exact data on future contact opportunities
than portability data. This result might be seen in the dissects
talked about
Content sharing in DTN'S involves the following problems:
A. Content sharing
In this segment we examine the problem of content
sharing in delay tolerant networks and depict substitute
solutions. As specified in the introduction, we spotlight on
mobile opportunistic networking scenarios where the nodes

Figure: Processing of incoming query.


B. Content discovery
In content discovery, mainly systems spotlight on how to
formulate queries, that depends on assumptions about the
format or layout of the content to be discovered. A common
protocol should sustain various types of queries and content,
but we abstract or summarize from the actual similar or
matching process in order to spotlight on discovering content
in the network. The easiest strategy to discover and deliver
the contents is Epidemic routing [2]. But, due to resource
limitations, Epidemic routing is often extravagant, so we have
to consider methods that limits the system resources used up
on both content discovery and delivery. Preferably, a query
should only be forwarded to neighbours that hold on the
matching contents or those are on the path to other nodes
having matching content . Different nodes should return no
overlapping responses to the requester. As total knowledge or
active coordination is not an alternative in our scenario, each

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

109

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

node can only make autonomous forwarding decisions. These


autonomous forwarding decisions should attain a fine trade off
amid discovery efficiency and necessary resources. Analogous
limitations pertain to content delivery. A few methods
anticipated by Pitkanen et al. may be used for restraining the
distribution of queries. Additionally, we study two substitutes
for restraining the distribution of queries: a split query lifetime
limit and a query distance limit. We employ the controlled
replication-based [9] routing scheme that performs a singlecopy scheme. This single-copy scheme turn both query
lifetime and distance limits into random walk, and the scheme
is not effectual when content-carrier nodes (i.e., destinations)
are not well-known. By distinguishing, the controlled
replication-based scheme dispenses a set of message replicas
and evade the excessive spread of messages.
C. Content delivery
When the query matching content is discovered, the
content carrying node should transmit only a subset of results.
This constraint is needed to limit the amount of resources
utilized both locally and globally for sending and storing the
responses, and to eliminate potential copies . The query
originator sets a limit for both the number of replications or
duplicates and the amount of content that should be produced.
When nodes require to forward a query message, the limits
incorporated in the query message are used to make the
forwarding decision. If the amount of the content go beyond
the response limit, the node wants to select which ones to
forward.
D. Mobility Prediction
Numerous studies have largely specified another
problem of content sharing: mobility learning and prediction.
Beacon Print discover meaningful places by constantly
determining constant scans for a time period. Place Sense
senses the arrival and exit from a place by utilizing invasive
RF-beacons. The system uses a radio beacons retort rates to
attain vigorous beacon conclusion. EnTracked is a position
tracking system for GPS-enabled devices. The system is
configurable to recognize different tradeoffs amid energy
consumption and heftiness. Mobility prediction has been
extensively studied in and out of the delay-tolerant networking
area. Markov-based schemes, make the problem as a Hidden
Markov or semi-Markov model and probabilistic prediction of
human mobility. In contrast, neural network based schemes
try to match the observed user behaviour with earlier
observed behaviour and estimate the prospect based on the
experimental patterns.Markov based schemes are suitable for
resource- restricted devices, like Smartphones, owing to their
low computation overhead and reserved storage requirements.
In our work, we have to develop a mobility learning and
prediction method. This method has been built to offer coarsegrained mobility information with a less computation
overhead. When the difficulty of mobility learning and

ISBN: 378 - 26 - 138420 - 5

prediction scheme can be mistreated, the schemes specified in


can be worn to offer fine-grained mobility information.

Problem in previous works


In the previous works, the energy consumption is more so
the battery lifetime will be reduced. By using asymmetric
multicore processors and efficient sensor scheduling
mechanisms energy consumption can be reduced and can
increase the lifespan of the batteries.
IV. PROPOSED WORK
In order to content sharing we can use the DPD Technique
and with that Here we are using asymmetric multicore
processors for performance expansion in sensing. Below fig
shows the daily energy utilization outline, which is calculated
using the composed stationary and movement time from
seven different users in four weeks . The scrutiny does not
comprise the energy utilization of content swap as it mainly
rely on the volume and communication pace of the nodes. The
typical energy consumptions of GPS, Wi-Fi, and the
accelerometer varies. The accelerometer do have
the
maximum energy consumption as it is used endlessly over 24
hours. Wi-Fi energy utilization is observed owed to the
scanning of neighbour APs for place recognition. GPS has a
huge discrepancy in energy consumption as this may not be
available in all places.
In this paper we examine the issue of following an item
moving random through a system of remote sensors. Our
destination is to devise techniques for planning the sensors to
advance the trade off between following execution and energy
consumption. We give the scheduling issue a role as a
Partially observable Markov Decision Process (POMDP),
where the control activities compare to the set of sensors to
initiate at each one time step. Utilizing a bottom-up approach
methodology, we consider different sensing, movement what's
more cost models with expanding levels of trouble
In order to minimize energy consumption or utilization
we use sensor scheduling schemes or mechanisms[10]. Sensor
systems have an wide-ranging diversity of prospective,
functional and important applications. In any case, there are
questions that have to be inclined for prolific procedure of
sensor system frameworks in right applications. Energy
sparing is one fundamental issue for sensor systems as most of
the sensors are furnished with no rechargeable batteries that
have constrained lifetime. To enhance the lifetime of a sensor
set up, one vital methodology is to attentively schedule
sensors' work sleep cycles (or obligation cycles). In addition,
in cluster based systems, grouping heads are usually selected
in a way that minimizes or reduces the aggregate energy
utilization and they may axle among the sensors to fine-tune
energy utilization. As a rule, these energy productive
scheduling components or mechanisms (furthermore called
topology arrangement components) required to accomplish

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

110

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

certain application requirements while sparing energy. In


sensor arranges that have various outline requirements than
those in conventional remote systems. Distinctive instruments
may make characteristic suspicions about their sensors
together with identification model, sense zone, transmission
scope, dissatisfaction or disappointment model, time
management, furthermore the capability to get area and
parting data.

ISBN: 378 - 26 - 138420 - 5

discovery rate. With an increasing query lifetime, both DPD


and Epidemic show a high discovery ratio because with a
longer duration, each query is forwarded to more nodes.

Fig 4: Discovery Efficiency


The influence of the query lifetime on hop-based discovery
methods is not significant. These observations are derived
from the limitation on the number of forwards.
C. Prediction Accuracy;
Figure: Mean energy consumption in a day
V.PERFORMANCE EVALUATION
A. Learning Accuracy
Learning accuracy shows how efficiently and correctly the
places were identified. The accuracy of place learning
influences the estimation of encounter opportunity between
two nodes. For example, If two different places are identified
as identical ones, we may incorrectly estimate that two nodes
will encounter each other when they visit two different
places.Also, the correct computation of and depends on the
geographical location information of nodes.

Mobility prediction is a key factor in the estimation of utility


function. Here, we evaluate our prediction method according
to trajectory deviation, prediction duration, and learning
period, as illustrated. Trajectory deviation indicates the
irregularity of a users mobility. For this evaluation, we
modify the existing mobility information with noise data.
Thus, 10, 20, and 30 percent of the meaningful places are
randomly chosen locations for trajectory deviations of 0.1,
0.2, and 0.3, respectively. As the trajectory deviation
increases, the prediction accuracy decreases. Prediction
accuracy is computed as the ratio of correctly predicted
locations to the total predicted locations.

Fig3: Learning Accuracy

Fig 5: Prediction Accuracy

B. Discovery Efficiency

D. Sharing Cost and Latency:

The discovery ratio is the ratio of discovered contents to the


generated queries within a given duration. DPDs discovery
performance is subjective to the two forwarding. In Epidemic,
queries are forwarded to every node. In hops-10 and hops-5, a
query message is forwarded until its hop count reaches 10 and
5, respectively. When a query matching content is available
only on a few nodes, the discovery methods show a low

Finally, we evaluate the protocols in terms of latency and cost,


as shown in Fig. 3f and Fig. 4. E-E uses Epidemic routing in
both the discovery and delivery phases. E-S&W uses
Epidemic routing for content discovery and Spray and Wait
for content delivery. The sharing latency is the sum of the
discovery latency and the delivery latency. E-E shows the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

111

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

lowest latency, and both DPD and E-S&W show the highest
latency. DPD exhibits such results due to the high latency of
the discovery phase. However, the delivery latency of DPD is
much smaller than that of E-S&W and is close to that of E-E.
E-E shows the highest overhead. The latency and overhead are
tradeoffs. In summary, DPD achieves good efficiency in the
delivery phase, whereas the efficiency of the discovery phase
can be improved. Content header caching on all nodes may be
a good solution, and this issue will be addressed in future
works.

ISBN: 378 - 26 - 138420 - 5

REFERENCE
[1]

T3I Group LLC, http://www.telecomweb.com, 2010.

[2]

A. Vahdat and D. Becker, Epidemic Routing for Partially Connected


Ad Hoc Networks, technical report, Dept. of Computer Science, Duke
Univ., Sept. 2000.

[3]

A. Balasubramanian, B.N. Levine, and A. Venkataramani, DTN


Routing as a Resource Allocation Problem, Proc. ACM SIGCOMM,
pp. 373-384, 2007.

[4]

R.C. Shah, S. Roy, S. Jain, and W. Brunette, Data Mules: Modelling a


Three-Tier Architecture for Sparse Sensor Networks, Elsevier Ad Hoc
Networks J., vol. 1, pp. 215-233, Sept. 2003.

[5]

A. Lindgren, A. Doria, and O. Schelen, Probabilistic Routing in


Intermittently Connected Networks, SIGMOBILE Mobile Computer
Comm. Rev., vol. 7, no. 3, pp. 19-20, 2003.

[6]

C. Liu and J. Wu, An Optimal Probabilistic Forwarding Protocol in


Delay Tolerant Networks, Proc. ACM MobiHoc, pp. 14, 2009.

[7]

J. Wu, M. Lu, and F. Li, Utility-Based Opportunistic Routing in MultiHop Wireless Networks, Proc. 28th Intl Conf. Distributed Computing
Systems (ICDCS 08), pp. 470-477, 2008.

[8] T. Spyropoulos, K. Psounis, and C.S. Raghavendra, Spray and Wait: An


Efficient Routing Scheme for Intermittently Connected Mobile
Networks, Proc. ACM SIGCOMM Workshop Delay-Tolerant
Networking (WDTN 05), pp. 252-259, 2005.

Fig 4: Sharing Cost and Latency


CONCLUSION

[9]

In this paper, we tend to planned associate degree economical


content sharing mechanism in Smartphone-based DTNs. We
tend to tried to utilize the benefits of todays Smartphones
(i.e., handiness of varied localization and communication
technologies) and suitably designed the protocol. In planning
a content sharing algorithmic rule, we tend to centered on 2
points: 1) individuals move around meaningful places, and 2)
the quality of individuals is certain. supported this proposition,
we tend to developed a quality learning and prediction
algorithmic rule to reckon the utility operate. We tend to
learned that contents so have geographical and temporal
validity, and that we planned a theme by considering these
characteristics of content. for instance, distributing queries for
content in a vicinity twenty miles from the placement of the
content searcher has solely a zero. Three percentage
probability to get the content whereas generating twenty
percentage additional transmission price. the time limitation
on question distribution reduces transmission price. most vital,
the planned protocol properly discovers and delivers eighty
seven percentage of contents among two hours once the
contents square measure on the market solely in thirty
percentage of nodes within the network.energy consumption
are reduced by using sensor scheduling and further works
have to done on user privacy.

T. Spyropoulos, K. Psounis, and C. Raghavendra, Efficient Routing in


Intermittently Connected Mobile Networks: The Single-Copy Case,
IEEE/ACM Trans. Networking, vol. 16, no. 1, pp. 63-76, Feb. 2008.

[10] T. Spyropoulos, K. Psounis, and C.S. Raghavendra, Efficient Routing in


Intermittently Connected Mobile Networks: The Multiple-Copy Case,
IEEE/ACM Trans. Networking, vol. 16, pp. 77-90, Feb. 2008.
[11] I. Cardei, C. Liu, J. Wu, and Q. Yuan, DTN Routing with Probabilistic
Trajectory Prediction, Proc. Third Intl Conf. Wireless Algorithms,
Systems, and Applications (WASA 08,), pp. 40-51, 2008.
[12] Q. Yuan, I. Cardei, and J. Wu, Predict and Relay: An Efficient
Routing in Disruption-Tolerant Networks, Proc. 10th ACM MobiHoc,
pp. 95-104, 2009.
[13]

E.M. Daly and M. Haahr, Social Network Analysis for Routing in


Disconnected Delay-tolerant MANETs, Proc. Eighth ACM MobiHoc,
pp. 32-40, 2007.

[14]

N.B. Chang and M. Liu, Controlled Flooding Search in a Large


Network, IEEE/ACM Trans. Networking, vol. 15, no. 2, pp. 436- 449,
Apr. 2007.

[15]

C. Avin and C. Brito, Efficient and Robust Query Processing in


Dynamic Environments Using Random Walk Techniques, Proc. Third
Intl Symp. Information Processing in Sensor Networks (IPSN 04), pp.
277-286, 2004.

[16]

M. Pitkanen, T. Karkkainen, J. Greifenberg, and J. Ott, Searching for


Content in Mobile DTNs, Proc. IEEE Intl Conf. Pervasive Computing
and Comm. (PERCOM 09), pp. 1-10, 2009.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

112

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

STORAGE PRIVACY PROTECTION AGAINST DATA


LEAKAGE THREADS IN CLOUD COMPUTING
K ABHINANDAN REDDY
M.Tech 2nd Year , CSE
[email protected]
__________________________________________________________________________
Abstract - Using Cloud Storage, users can

Index terms Cloud storage, privacy- protection,

remotely store their data and enjoy the on-

public auditability, cloud computing.

demand high quality applications and services


I. INTRODUCTION

from a shared pool of configurable computing


resources, without the burden of local data

Cloud computing promises lower costs, rapid

storage and maintenance.However, the fact that

scaling,

users no longer have physical possession of the

availability anywhere, anytime, a key challenge

outsourced

is how to ensure and build confidence that the

data makes the data integrity

protection in Cloud Computing a formidable

easier

maintenance,

and

service

cloud can handle user data securely.

task, especially for users with constrained


Cloud computing is transforming the very

computing resources. Moreover, users should be


able to just use the cloud storage as if it is local,
without worrying about the need to verify its

nature of how businesses use information


technology. From users prospective, including
both individuals and IT enterprises, storing data

integrity

remotely to the cloud in a flexible on-demand


This paper proposes privacy preserving public

manner brings appealing benefits: relief of the

auditability for cloud storage is of very

burden for storage management, universal data

importance so that users can arrange to a third

access with independent geographical locations,

party auditor (TPA) to check the integrity of

and

outside data and be worry-free. To securely

hardware, software, and personnel maintenances,

initiate an effective TPA, the auditing process

etc.

avoidance

of

capital

expenditure

on

should bring no new security challenges towards


user data privacy and no additional online
burden to user. This paper proposes a secure
cloud

storage

system

supporting

privacy-

preserving public auditing. We further extend the


TPA to perform audits for multiple users
simultaneously and efficiently.

While cloud computing makes these advantages


more appealing than ever, it also brings new and
challenging security threats towards users
outsourced data. Data outsourcing is actually
relinquishing users ultimate control over the
fate of their data. As a result, the correctness of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

113

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

the data in the cloud is being put at risk due to

would also beneficial for the cloud service

the following reasons.

providers to improve their cloud based service


platform.

i)

The infrastructures under the cloud are much


more powerful and reliable than personal

Recently, the notation of public auditability has

computing devices; they are still facing the

been proposed in the context of ensuring

broad range of both internal and external

remotely stored data integrity under different

threats for data integrity.

system and security models. Public auditability

ii) There do exist various motivations for Cloud

allows external party, in addition to the user

Service Provider to behave unfaithfully

himself, to verify the correctness of remotely

towards the cloud users regarding the status

stored data. From the perspective of protecting

of their outsourced data.

data privacy, the users, who own the data and


rely on TPA just for the storage security of their

The problem, if not properly addressed, may


impede the successful deployment of the cloud
architecture. As users have no longer physically

data,

do not

want this auditing process

introducing new vulnerabilities of unauthorized


information leakage towards their data security

posses the storage of their data, traditional


cryptographic primitives for the purpose of the

Exploiting data encryption before outsourcing is

data security protection cannot be directly

one way to mitigate this privacy concern, but it

adopted. Considering the large size of the

is only complementary to the privacy preserving

outsourced data and the users constrained

public auditing scheme to be proposed in this

resource capability, the tasks of auditing the data

paper. Without a properly designed auditing

correctness in cloud environment

can be

protocol, encryption itself cannot prevent data

formidable and expensive for the cloud users.

from flowing away towards external parties

Moreover, the overhead of using cloud storage

during the auditing process. Therefore, how to

should be minimized as much as possible, such

enable a privacy-preserving third-party auditing

that user does not need to perform too many

protocol, independent to data encryption, is the

operations to use the data.

problem we are going to tackle in this paper.

To fully ensure the data integrity and save the

To address these problems, our work utilizes the

cloud users computation resources as well as

technique of public key based homomorphic

online burden, it is of critical importance to

linear authenticator (or HLA for short), which

enable public auditing service for cloud data

enables TPA to perform the auditing without

storage, so that users may resort to an

demanding the local copy of data and thus

independent third party auditor (TPA) to audit

drastically reduces the communication and

the outsourced data when needed. In addition to

computation overhead as compared to the

help users to evaluate the risk of their subscribed

straightforward data auditing approaches. By

cloud data services, the audit result from TPA

integrating the HLA with random masking, our

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

114

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

protocol guarantees that the TPA could not learn

it integrity security. Then we present our main

any knowledge about the data content stored in

scheme and show how to extent our main

the cloud server during the efficient auditing

scheme to support batch auditing for the TPA

process.

upon delegations from multiple users. Finally we


discuss how to generalize our privacy-preserving

a) Design Goals

public auditing scheme and its support of data

To enable privacy-preserving public auditing for


cloud data storage our protocol design should
achieve the following security and performance
guarantees.

dynamics.
a) Our Framework & Definitions
We follow similar definition of previously
proposed schemes in the context of remote data

1) Public auditability: to allow TPA to verify


the correctness of the cloud data on demand

integrity checking and adapt the framework for


our privacy preserving public auditing system.

without retrieving a copy of the whole data


or introducing additional online burden to

A public auditing scheme consists of four

the cloud users.

algorithms

2) Storage correctness: to ensure that there

(KeyGen,

SigGen,

GenProof,

VerifyProof).

exists no cheating cloud server that can pass

KeyGen: is a key generation algorithm run by

the TPAs audit without indeed storing

the user to setup the scheme.

users data intact.

SigGen: is used by the user to generate

3) Privacy-preserving: to ensure that the TPA


cannot derive users data content from
information collected during the auditing

verification metadata, which may consist of


MAC, signatures, or other related information
that will be used for auditing.

process.
4) Batch Auditing: to enable TPA with secure

GenProof: is run by the CS (Cloud server) to

and efficient auditing capability to cope with

generate a proof of data storage correctness,

multiple auditing delegations from possibly

while VerifyProof is run by the TPA to audit the

large

proof from the cloud server.

number

of

different

users

simultaneously.
5) Lightweight: to allow TPA to perform

Running a public auditing system consists of two


phases, Setup and Audit:

auditing with minimum communication and


computation overhead.
II. PRPOSED SCHEMES

Setup: the user initializes the public and secret


parameters of the system by executing KeyGen,
and pre-processes the data file by using SigGen

This heading presents our public auditing

to generate verification metadata.

scheme which provides a complete outsourcing


solution of data -not only the data itself, but also

Audit: the TPA issues an audit message or


challenge to the cloud server to make sure that

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

115

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

the cloud server has retained the data file

scheme proposed by Boneh,Lynn and Shacham

properly at the time of the audit. The CS will

(BLS).

derive a response message from a function of the


stored data file and its verification metadata by

d) BLS Schema Details

executing GenProof. The TPA then verifies the

Let G1, G2, and GT be multiplicative cyclic group

response via VerifyProof.

of primary order and e: G1 X G2 GT . the BLS

Our frame work assumes the TPA is stateless,

map as introduced in mainly.

which is a desirable property achieved by our

Setup Phase: the cloud user runs KeyGen

proposed solution. It is easy to extend the

algorithm to generate the public and secrete

framework above to capture a stateful auditing

parameters. A then the user runs SigGen to

system, essentially by splitting the verification

compute authenticator for each block and name

metadata into two parts which are stored by the

is chosen by the user uniformity at random.

TPA and the cloud server respectively.


Audit Phase: the TPA first retrieves the file tag.
b) Basic Schemes

With respect to the mechanism in the setup

HLA-based solution. To effectively support


public auditability without having to retrieve the
data blocks themselves, the HLA technique can

phase, the TPA verifies the signature and quits


by emitting FALSE if the verification fails.
Otherwise, the TPA recovers name.

be user HLAs like MACs are also some

Now it comes to the core part of the auditing

unforgeable

that

process. To generate the challenge message for

authentication the integrity of a data block. The

the audit, the TPA picks a random element

difference is that HLAs can be aggregate.

subset of set [1, n]. then the TPA sends challenge

verification

c) Privacy-Preserving

metadata

Public

Auditing

Scheme

Upon receiving challenge, the server runs

To achieve privacy-preserving public auditing,


we propose to uniquely integrated the homophic
linear

authenticator

message to the server.

with random

masking

GenProof to generate a response proof of data


storage correctness.
b) Support for Batch Auditing

technique. On the other hand, the correctness


validation of the block- authenticator pairs can
still be carried out in a new way which will be
shown shortly, even with the presence of the
randomness. Our design makes use of public key
based HLA, to equip the auditing protocol with
public auditability. Moreover, we use the HLA
proposed, which is based on the short signature

With the establishment of privacy-preserving


public auditing the TPA concurrently handle
multiple upon different users delegation. The
individual auditing of these tasks for the TPA
can be tedious and very inefficient. It is more
advantageous for the TPA to batch these
multiple tasks together and audit at one time.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

116

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Keeping this in mind, we slightly modify our

stores, in Proc. of CCS07, Alexandria, VA,

protocol in a single user case, and achieve the

October 2007, pp. 598609.

aggregation of K verification equations into


[4] C.Wang, Q.Wang, K. Ren, andW. Lou,

single on.

Ensuring data storage security in cloud


III.

computing, in Proc. of IWQoS09, July

CONCLUSION

2009, pp. 19.


In this paper, we propose a privacy-preserving
public auditing system for data storage security

[5] R. C.Merkle, Protocols for public key

in cloud computing. We utilize the homomorphic

cryptosystems,

in

Proc.

of

IEEE

linear authenticator and random masking to

Symposium on Security and Privacy, Los

guarantee that TPA would not learn any

Alamitos, CA, USA, 1980.

knowledge about the data content stored on the


cloud server during efficient auditing process.

[6]

S. Yu, C. Wang, K. Ren, and W. Lou,


Achieving secure, scalable, and fine-

And we slight change our protocol, the TPA can

grained access control in cloud computing,

perform multiple audit sessions from different

in Proc. of IEEE INFOCOM10, San Diego,

users for their outsourced data files. Extensive

CA, USA, March 2010.

analysis shows that our schemes are provably


secure and highly efficient.

AUTHORS

REFERENCES
Mr.K
[1]Cong Wang, Qian Wang, Kui Ren, Wenjing

REDDY

Lou Privacy-Preserving Public Auditing for

received

Narayana

Secure Cloud Storage in Proc. of Compu-

the

Engineering

College, Nellore, B.Tech

ters, pp. 362-375, 2013.


[1] C. Erway, A. Kupcu, C. Papamanthou, and

ABHINANDAN

degree

in

computer

science & engineering from the Jawaharlal

Dynamic provable data

Nehru technological university Anantapur, in

possession, in Proc. of CCS09, 2009, pp.

2012, and received the Audisankara Institute of

213222.

Technology, Gudur M.Tech degree in computer

R. Tamassia,

science engineering from the Jawaharlal Nehru


[2] M. A. Shah, R. Swaminathan, and M. Baker,
Privacypreserving audit and extraction of
digital contents, Cryptology ePrint Archive,

technological university Anantapur in 2014,


respectively. He Participated National Level
Paper Symposiums in different Colleges.

Report 2008/186, 2008.


[3] G. Ateniese, R. Burns, R. Curtmola, J.
Herring, L. Kissner, Z. Peterson, and D.
Song, Provable data possession at untrusted

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

117

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Real Time event detection and alert system using sensors


V Anil Kumar1, V Sreenatha Sarma2
1

M.Tech, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India
2

Asst Professor, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India

I.

Abstract - Twitter a well-liked microbiologging


services, has received abundant attention recently.

INTRODUCTION

Twitter, a popular microblogging service, has

We investigate the $64000 time interaction of

received abundant attention recently. It is a web

events like earthquakes. And propose associate

social network employed by a lot of folks round

algorithm to watch tweets and to focus on event.

the world to remain connected to their friends,

To sight a target event, we device a classifier of

family members and fellow worker through their

tweets supported options like keyword in an

computers and mobile phones. Twitter was based

exceedingly tweet, the amount of words, and their

in March twenty one, 2006 within the headquarters

context. We have a tendency to manufacture a

of Sanfrancisco, California; USA. Twitter is

probabilistic spatiotemporal model for the target

classified

event that may notice the middle and therefore the

as

micro-blogging

services.

Microblogging is a style of blogging that permits

flight PF the event location. We have a tendency

users

to contemplate every twitter user as a device and

to

micromedia

apply kalman filtering and sensible filtering. The

send
like

.Microblogging

particle filter works higher than alternative

temporary
pictures

services

text
or

apart

updates
audio

or
clips

from Twitter

embrace Tumbler, plurk, Emote.in, Squeelr, Jaiku,

compared ways in estimating the middle of

identi.ca, and so on. They have their own

earthquakes and trajectories of typhoons. As

characteristics. Microblogging service of twitter

associate application, we have a tendency to

provides

construct associate earthquake coverage System in

immediacy

and

movableness.

Microblogging is slowly moving to the thought. In

Japan. Our System detects earthquakes promptly

USA, for example Obama microblogged from the

and sends e-mails and SMS alerts to registered

campaign path, to that BBC news gave link.

users .Notification is delivered abundant quicker


than the announcements that square measure
broadcast by JMA.

II. INVESTIGATION
Index terms - Tweets, social sensors, earthquake,
Earthquake is a disastrous event in which many

hand shaking.

people loss their life and property. Hence detection

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

118

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

and bringing awareness about the event is very

I happened to feel the shaking of the earth.

important to prevent hazardous environment.

These tweets are analyzed based on features such

Daily news paper, TV broadcast channels paves

as statically [2] [3]. They separate keywords and

way for it. But these systems are slower when it

words and confirm the event and sent to positive

comes for the real time occurrence as they are time

class. At time there may be tweets such as I read

consuming in sensing, reporting or publishing.

an article on earthquake or the tweets such as

Also in today s running world people does not pay

Earthquake before three days was heart stolen.

much attention all time in media .Later on SMS,

These tweets were analyzed and found that it is

calls facilitated the reporting system but due to the

not under the criteria and hence sent to the

ignorance of people under some situations such as

negative class

in the midst of night, expire of mobile phone


charges. This system also encountered many
failures when they happened to face the problem
of fakers. The mobile phone company was
helpless to stop the fake messages or confirm the
predicted information. Hence these calls and SMS
do not solve any problem but increases. At the
third stage social networks came to the play they

Fig.1 Twitter User Map

have lots of followers as Microblogging attraction


To category a tweet into a positive category or a

bloomed over them. These social networks served

negative class, we have a tendency to use a

to be both fun time and useful .It helped to bridge


the

gap

between

the

friends

of

support vector machine (SVM) [7] that may be a

different

wide used machine-learning algorithmic program.

community and get connected all over the world.

By making ready positive and negative examples


At the same time some of the facts and important

as a coaching set, we will manufacture a model to

messages were also known by the people by their

classify tweets mechanically into positive and

friends and followers updates, shares and likes.

negative classes. After confirmation of details

Their like gives rating to some of the ads or some

that\'s within the positive category the notification

blogs or groups. Considering these facts they

is distributed to the registered users victimization

decided to develop the reporting system for earth

particle filtering. Particle filtering is depended

quake. They choose twitter because of its

upon the rating for the location [6]. Rating is

popularity .They considers users as the social

depended informed the likes delivered to the zone.

sensors. These users posts report on earthquake

The intense disadvantage of existing system is that

such as I am attending Earthquake right now or

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

119

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

no confirmation of messages before the warning is

distribution for the situation estimation at time t,

distributed and also the notification is

selected because the belief Belxtp fx^(i ) t;


(_i^w)t g; i 1 . . . n. every x^i t may be a distinct
hypothesis associated with the thing location. The
w^it area unit area unit non negative weights,
referred to as importance factors.

Fig. 2 Number of tweets related to earthquakes


Delivered only to the registered users. To many
people who belong to the earth quake zone does
not receive these messages as because they did not
register.
III. PROPOSED SYSTEM
Fig. 3 System Architecture
In the projected system for the detection of
earthquake through twitter the user need n\'t

The

register for the earthquake awful tweets for the

algorithm is a Monte Carlo method that forms the

previous

basis for

info.

The

earthquake

is

detected

Sequential

Importance

Sampling

(SIS)

exploitation the keyword-based topic search.

particle filters. The SIS algorithm consists of

Tweet search API crawls the tweets containing the

recursive propagation of the weights and support

keywords like earthquake, disaster, damage,

points

shaking. Tweet crawler sends the keywords

sequentially.

containing sentence to mecab. In mecab keywords

Step1-Generation:

area unit keep and also the sentence containing

Generate and weight a particle set, which means N

keyword is distributed for linguistics analysis . A

discrete hypothesis evently on map.S0 s00 ; s10

particle filter may be a probabilistic approximation

; s20; . . . ; sN_1 0

algorithmic

program

implementing

as

each

measurement

is

received

mathematician filter, and a member of the family


Step 2: Resampling:

of consecutive Monte Carlo strategies. For


location estimation, it maintains a likelihood

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

120

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Resample Nparticles from a particle set

ISBN: 378 - 26 - 138420 - 5

using

Obama, cine stars and so on. Even some of the

weights of respective particles and allocate them

entertaining media like television and radio

on the map. (We allow resampling of more than

gathers information through the Twitter and spread

that of the same particles.).

it.
V.

Step 3: Prediction.
Predict the next state of a particle set

As we tend to mentioned during this paper,

from

Twitter user is employed as a device, and set the

Newtons motion equation.

matter as detection of an occurrence supported

Step 4:

sensory observations. Location estimation ways

Find the new document vector coordinates in this

like particle filtering are used

reduced 2-dimensional space.

locations

Step 5: Weighting
Re-calculate the weight of

Calculate the current object location ( ,

of

events.

As

to estimate the
associate

degree

application, we tend to developed associate degree

earthquake reportage system. Microblogging helps

Step 6: Measurement:
the average of ( ,

CONCLUSION

United States of America to unfold the news at

)by

quicker rates to the opposite subscribers of the

) .

page; it conjointly distinguishes it from different

Step 7: Iteration:

media within the sort of blogs and cooperative

Iterate Step 3, 4, 5 and 6 until convergence

bookmarks. The algorithmic program utilized in

IV. REPORTING SYSTEM

the projected system facilitates United States of

We developed an earthquake-reporting system

America to spot the keyword that\'s needed help

using the event detection algorithm. Earthquake

United States of America to assemble additional

information is much more valuable if it is received

info from varied places by SIS algorithmic

in real time. Given some amount of advanced

program that the folks are unaware of it. Finding

warning, any person would be able to turn off a

the folks of specific areas through particle filtering

stove or heater at home and then seek protection

and causation alarms to the folks is a brand new

under a desk or table if such a person were to have

conception of intimating folks which is able to pay

several seconds notice before an earthquake

some way for spreading the news at quicker rate

actually strikes an area. It goes without saying

instead of intimating the folks within the whole

that, for such a warning, earlier is better. It

world and spreading the news at slower rate.

provides advance announcements of the estimated


seismic intensities and expected arrival times. As
in the survey of social networking sites the twitter
have about 2 million users across the world
including the famous personalities like Barack

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

121

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

REFERENCES
[5] M. Ebner and M. Schiefner, Microblogging -

[1] S. Milstein, A. Chowdhury, G. Hochmuth, B.


Lorica, and R. Magoulas.Twitter and the micr-

More

messaging

Learning Conf., pp. 155-159, 2008.

revolution:

Communication,

than

Fun?Proc.

IADIS

Mobile

[6] L. Backstrom, J. Kleinberg, R. Kumar, and J.

connections, and immediacy.140 characters at

Novak, Spatial Variation in Search Engine

a time. OReilly Media, 2008.

Queries,Proc. 17th Intl Conf.World Wide

[2] V. Fox, J. Hightower, L. Liao, D. Schulz, and

Web (WWW 08), pp. 357-366, 2008.

G. Borriello, Bayesian Filtering for Location

[7] T. Joachims. Text categorization with support

Estimation, IEEE Pervasive Computing, vol.

vector machines. In Proc. ECML98, pages

2, no. 3, pp. 24-33, July-Sept. 2003.

137142, 1998.

[3]T. Sakaki, M. Okazaki, and Y. Matsuo,


Earthquake Shakes Twitter Users: Real-Time
Event Detection by Social Sensors, Proc.
19th Intl Conf. World Wide Web (WWW
10), pp. 851-860,2010.
[4] M. Cataldi, L. Di Caro, and C. Schifanella,
Emerging Topic Detection on Twitter Based
on Temporal and Social Terms Evaluation,Proc.10th Intl Workshop Multimedia
Data MiningMDMKDD10), pp. 1-10, 2010.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

122

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

An Efficient Secure Scheme for Multi User in Cloud


by Using Cryptography Technique
S.Santhosh1K.MadhuBabu 2
2

PG Scholar1, M.Tech, Assistant Professor


[email protected]
Department of Computer Science and Engineering,
Audisankara College of Engineering & Technology, Gudur.
A.P - India

ABSTRACT
The major aims of this technique a secure multi-owner knowledge sharing theme. It implies that any user within the cluster will
firmly share knowledge with others by the world organization trust worthy cloud. This theme is ready to support dynamic
teams. Efficiently, specifically, new granted users will directly rewrite knowledge files uploaded before their participation
whereas not contacting with knowledge owners. User revocation square measure just achieved through a very distinctive
revocation list whereas not modification the key. Keys of the remaining users the size and computation overhead of
cryptography square measure constant and freelance with the amount of revoked users. Weve a bent to gift a secure and
privacy-preserving access management to users that guarantee any member throughout a cluster to anonymously utilize the
cloud resource. We offer rigorous security analysis, and perform intensive simulations to demonstrate the potency of our theme
in terms of storage and computation overhead. To attain the dependable and adaptable in MONA, in this paper we are exhibiting
the new structure for MONA. In this strategy we are further exhibiting how we are dealing with the risks like failure of cluster
manager by increasing the number of backup cluster manager, hanging of gathering administrator on the off chance that number
of demands all the more by offering the workload in numerous gathering chiefs. Using this method states expected productivity,
scalability and a lot essentially dependability.
Keywords:- Multi owner, resource, cluster manager, revocation, Key Distribution
I.INTRODUCTION
CLOUD computing is recognized as another to ancient
information technology as a result of its resource -distribution
and low-maintenance distinctiveness. In cloud computing, the
cloud service suppliers (CSPs), like Amazon, unit able to
deliver various services to cloud users with the help of
powerful data centres. By migrating the local data
management systems into cloud servers, users can fancy highquality services and save vital investments on their native
infrastructures. One in every of the foremost basic services
offered by cloud suppliers is data storage. Permit US to place
confidence in a smart data application.
A company permits its staffs inside identical cluster or
department to store and share files inside the cloud. By
utilizing the cloud, the staffs square measure typically totally
discharged from the tough native data storage and
maintenance. However, it in addition poses a important risk to
the confidentiality of these hold on files. Specifically, the
cloud servers managed by cloud suppliers are not wholly
trustworthy by users whereas the data files hold on inside the
cloud may even be sensitive and confidential, like industry
plans. To conserve data privacy, a basic resolution is to jot
down in code data files, and so transfer the encrypted data into
the cloud [2].Sadly, planning degree economical and secure
data sharing theme for teams inside the cloud is not an easy
task as a result of the following issues.

However, the complexities of client participation and


revocation in these schemes square measure linearly increasing
with the amount of information homeowners and so the vary of
revoked users, severally. By setting a bunch with one attribute
level projected a secure origin theme supported the cipher textpolicy attribute-based cryptography technique, which allows
any member throughout a cluster to share data with others.
However, the matter of user revocation is not self-addressed in
their theme conferred a scalable and fine-grained data access
management theme in cloud computing supported the key
policy attribute-based cryptography (KP-ABE) technique
.Sadly, the one owner manner hinders the adoption of their
theme into the case, where any user is granted to store and
share data. Our contributions. To resolve the challenges
conferred higher than, we've an inclination to propose
Anglesey Island, a secure multi-owner data sharing theme for
dynamic groups inside the cloud.
To achieve secure data sharing for dynamic groups in the
cloud, we expect to combine the group signature and dynamic
broadcast encryption techniques. Specially, the group signature
scheme enables users to anonymously use the cloud resources,
and the dynamic broadcast encryption technique allows data
owners to securely share their data files with others including
new joining users.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

123

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

II. LITERATURE SURVEY

A.Plutus: Scalable Secure File Sharing on Untrusted


Storage
Plutus may be a scientific discipline storage system that
permits secure file sharing while not putting a lot of trust on
the file servers. Specifically, it makes novel use of scientific
discipline primitives to safeguard and share files. Plutus
options extremely scalable key management whereas
permitting individual users to retain direct management over
United Nations agency gets access to their files. We have a
tendency to make a case for the mechanisms in Plutus to scale
back the quantity of scientific discipline keys changed
between users by exploitation file groups, distinguish file scan
and write access, handle user revocation expeditiously, and
permit Associate in nursing untrusted server to authorize file
writes. Weve designed a epitome of Plutus on Open AFS.
Measurements of this epitome show that Plutus achieves
sturdy security with overhead such as systems that encrypt all
network traffic.
B. Sirius: Securing Remote Untrusted Storage
This paper presents SiRiUS, a secure classification system
designed to be superimposed over insecure network and P2P
file systems like NFS, CIFS, OceanStore, and Yahoo! case.
SiRiUS assumes the network storage is untrusted and provides
its acquire read-write cryptanalytic access management for file
level distribution. Key administration and revocation is easy
with tokenish out-of-band communication. Classification
system freshness guarantees square measure supported by
SiRiUS mistreatment hash tree constructions. SiRiUS contains
a completely unique technique of playing file random access
during a cryptanalytic classification system while not the
utilization of a block server. Extensions to SiRiUS embody
giant scale cluster sharing mistreatment the NNL key
revocation construction. Our implementation of SiRiUS
performs well relative to the underlying classification system
despite mistreatment cryptanalytic operations

Fig1. Existing Model

1. In the prevailing Systems, identity privacy is one among


the foremost vital obstacles for the wide preparation of cloud
computing. While not the guarantee of identity privacy, users
is also unwilling to affix in cloud computing systems as a
result of their real identities may be simply disclosed to cloud
suppliers and attackers. On the opposite hand, unconditional
identity privacy might incur the abuse of privacy. for instance,
a misbehaved employees will deceive others within the
company by sharing false files while not being traceable.
2. Only the cluster manager will store and modify knowledge
within the cloud.
3. The changes of membership create secure knowledge
sharing extraordinarily troublesome the difficulty of user
revocation isn't self-addressed.
4. On the other hand depending on consistency in addition to
scalability issue this method has to be workout further as if the
cluster mangers are amiss due to many requests caused by
various groups of owners, and then complete security
technique connected with MONA was unable lower.

IV. SYSTEM IMPLEMENTATION

.III. RELATED WORK


To preserve information privacy, a basic resolution is to
encipher information files, then transfer the encrypted
information into the cloud. sadly, coming up with associate
degree economical and secure information sharing theme for
teams within the cloud isn't a simple task.In the existing
System information homeowners store the encrypted
information files in untrusted storage and distribute the
corresponding cryptography keys solely to approved users.
Thus, unauthorized users moreover as storage servers cannot
learn the content of the information files as a result of they
need no data of the cryptography keys. However, the
complexities of user participation and revocation in these
schemes area unit linearly increasing with variety the amount
the quantity of knowledge homeowners and also the number
of revoked users, severally.

1. We have an inclination to propose a secure multi-owner


data sharing theme. It implies that any user inside the cluster
can firmly share data with others by the world organisation
sure cloud.
2. Our projected theme is during a place to support dynamic
teams expeditiously. particularly, new established users will
directly rewrite data files uploaded before their
participation
whereas not contacting with data
homeowners. User revocation is also merely achieved through
a completely unique revocation list whereas not amendment
the key keys of the remaining users. The size and computation
overhead of cryptography square measure constant and
freelance with the number of revoked users.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

124

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

3. We suggest secure and privacy-preserving access


management to users that guarantees any member throughout
a bunch to anonymously utilize the cloud resource. Moreover,
the vital identities of information homeowners are also
disclosed by the cluster manager once disputes occur.
4. We provide rigorous security analysis, and perform
intensive simulations to demonstrate the potency of our theme
in terms of storage and computation overhead.
.

ISBN: 378 - 26 - 138420 - 5

E. Group Signature :
A cluster signature theme permits any member of the group to
sign messages whereas keeping the identity secret from
verifiers. Besides, the selected cluster manager will reveal the
identity of the signatures mastermind once a dispute happens,
that is denoted as traceability.
F. User Revocation:

A.Cloud section:
In this module, we tend to produce an area Cloud and supply
priced torrential storage services. The users will transfer their
information within the cloud. we tend to develop this module,
wherever the cloud storage may be created secure. However,
the cloud isn't absolutely trusty by users since the CSPs are
terribly possible to be outside of the cloud users trusty
domain. kind of like we tend to assume that the cloud server is
honest however curious. That is, the cloud server won't
maliciously delete or modify user information owing to the
protection of information auditing schemes, however can try
and learn the content of the keep information and therefore the
identities of cloud users.
B. Group Manager :
Group manager takes charge of followings
1. System parameters generation,
2. User registration,
3. User revocation, and
4. Revealing the important identity of a dispute knowledge
owner.
Therefore, we tend to assume that the cluster manager is
absolutely trustworthy by the opposite parties. The cluster
manager is that the admin. The cluster manager has the logs of
every and each method within the cloud. The cluster manager
is accountable for user registration and additionally user
revocation too.
C. Group Member :
1.store their non-public knowledge into the cloud server and
2.Share them with others within the cluster.
Note that, the cluster membership is dynamically modified, as
a result of the workers resignation and new worker
participation within the company. The cluster member has the
possession of fixing the files within the cluster. Whoever
within the cluster will read the files that are uploaded in their
cluster and conjointly modify it.
D. File Security:
1. Encrypting the data file.
2. File stored in the cloud can be deleted by either the group
manager or the data owner. (i.e., the member who uploaded
the file into the server).

User revocation is performed by the cluster manager via a


public out there revocation list (RL), supported that cluster
members will cypher their information files and make sure the
confidentiality against the revoked users.
G. Enhanced work:

Fig 2.Proposed Model


In this strategy we are further showing how we are dealing
with the dangers like disappointment of gathering chief by
expanding the amount of reinforcement gathering supervisor,
hanging of gathering director on the off chance that number of
demands all the more by offering the workload in various
gathering supervisors.
This strategy cases obliged effectiveness, versatility and
above all dependability
To conquer the impediment of existing framework , in this
proposed Scheme of data security is if the gathering director
quit working because of expansive number of appeals
originating from diverse gatherings of holders, then
reinforcement bunch supervisor will stays accessible.
To attain the dependable and adaptable in MONA, in this
paper we are exhibiting the new structure for MONA. In this
strategy we are further exhibiting how we are dealing with the
risks like failure of cluster manager by increasing the number
of backup cluster manager, hanging of gathering administrator
on the off chance that number of demands all the more by
offering the workload in numerous gathering chiefs. Using
this method states expected productivity, scalability and a lot
essentially dependability.
In order to triumph over this downside connected with
existing system MONA, inside the recommended MONA will
be if the class manager stop working due to numerous asks
originating from diverse multiple proprietors, subsequently
back up class manager will probably remains offered.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

125

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

V. CONCLUSION
In this paper, we've an inclination to vogue a secure data
sharing theme, for dynamic groups in associate world
organisation trustworthy cloud. In Mona, a user is during a
position to share data with others at intervals the cluster
whereas not revealing identity privacy to the cloud.
Additionally, Angle Sea Island supports economical user
revocation and new user modification of integrity. lots of
specially, economical user revocation are achieved through a
public revocation list whereas not modification the private
keys of the remaining users, and new users can directly
decipher files keep at intervals the cloud before their
participation. Moreover, the storage overhead and additionally
the coding computation value are constant. Intensive analyses
show that our projected theme satisfies the specified security
requirements and guarantees efficiency additionally. Projected
a science storage system that allows secure file sharing on
world organisation trustworthy servers, named Plutus. By
dividing files into file groups and encrypting each file cluster
with a completely unique file-block key, the information
owner can share the file groups with others through delivering
the corresponding safe-deposit key, where the safe-deposit
key's wont to write the file-block keys. However, it brings a
few of significant key distribution overhead for large-scale file
sharing. In addition, the file-block key should be updated and
distributed yet again for a user revocation.

REFERENCES
[1]

M. Armbrust, A. Fox, R. Griffith, A.D. Joseph, R.H. Katz,


A.Konwinski,G. Lee, D.A. Patterson, A. Rabkin, I. Stoica, and M.
Zaharia, A View ofCloudComputing, Comm. ACM, vol. 53,no. 4, pp.
50-58, Apr. 2010.

[2]

S. Kamara and K. Lauter, Cryptographic Cloud Storage, Proc.Intl


Conf. Financial Cryptography and Data
Security (FC), pp. 136-149, Jan. 2010.
[3]
S. Yu, C. Wang, K. Ren, and W. Lou, Achieving Secure, Scalable,
and Fine-Grained Data Access Control in Cloud
Computing,
Proc. IEEE INFOCOM, pp. 534-542, 2010.
[4] M. Kallahalla, E. Riedel, R. Swaminathan, Q. Wang, and K. Fu,
Plutus: Scalable Secure File Sharing on Untrusted Storage,
Proc.USENIX Conf. File and Storage Technologies, pp. 29-42, 2003.
[5] E. Goh, H. Shacham, N. Modadugu, and D. Boneh, Sirius:Securing
Remote Untrusted Storage, Proc. Network and Distributed Systems
Security Symp. (NDSS), pp. 131-145, 2003.
[6] G. Ateniese, K. Fu, M. Green, and S. Hohenberger, Improved Proxy
Re-Encryption Schemes with Applications to Secure Distributed
Storage, Proc. Network and Distributed Systems Security
Symp. (NDSS), pp. 29-43, 2005.
[7] R. Lu, X. Lin, X. Liang, and X. Shen, Secure Provenance: The
Essential of Bread and Butter of Data Forensics in Cloud Computing,
Proc. ACM Symp. Information,
Computer and Comm.Security, pp. 282-292, 2010.
[8] B. Waters, Ciphertext-Policy Attribute-Based Encryption: An
Expressive, Efficient, and Provably Secure Realization, Proc. IntlConf.
Practice and Theory in Public Key
Cryptography
Conf.
Public
Key
Cryptography,
http://eprint.iacr.org/2008/290.pdf, 2008.
[9] V. Goyal, O. Pandey, A. Sahai, and B. Waters, AttributeBased Encryption for Fine-Grained Access Control of Encrypted Data,
Proc. ACM Conf. Computer and Comm.
Security (CCS), pp. 89-98, 2006.
[10] D. Naor, M. Naor, and J.B. Lotspiech, Revocation and Tracing Schemes
for Stateless Receivers, Proc. Ann. Intl Cryptology Conf. Advances in
Cryptology (CRYPTO), pp. 41-62, 2001

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

126

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Design and Implementation of secure cloud systems using


Meta cloud
Perumalla Gireesh
M.Tech 2nd year, Dept. of CSE, ASCET, Gudur, India
Email:[email protected]
_____________________________________________________________________________________

Abstract Cloud computing has recently emerged

way from existing schemes to solving this problem

as a new paradigm for hosting and delivering

effectively. But it does not consider the users data

services over the Internet. Cloud computing is

privacy in transforming Meta cloud. To address

attractive to business owners as it eliminates the

this problem, we introduce Business Continuity

requirement

Management (BCM). This is defined as a holistic

for

users

to

plan

ahead

for

provisioning, and allows enterprises to start from

management

the small and increase resources only when there

organization and reduces the impacts of data

is a rise in service demand. However, despite the

leakage issues.

fact

Index terms Meta cloud, Cloud Privacy, private

that

cloud

computing

offers

huge

opportunities to the IT industry, the development

process

that

identifies

to

an

clouds, security.

of cloud computing technology is currently at its

I. INTRODUCTION

infancy, with many issues still to be addressed. In


With the rapid development of processing and

this paper, we present a survey of cloud


computing,

highlighting

its

architectural

principles,

and

key

storage technologies and the success of the

concepts,

Internet,

state-of-the-art

computing

resources

have become

cheaper, more powerful and more ubiquitously

implementation as well as research challenges.

available than ever before. This technological


trend has enabled the realization of a new

Meta cloud based on a combination of existing

computing model called cloud computing, in

tools, concepts and provides the convenient to

which resources (e.g., CPU and storage) are

organize the private clouds. This can consider the

provided as general utilities that can be leased and

only vendor lock-in problem of different vendors

released by users through the Internet in an on-

in cloud. For that Meta cloud provides an abstract

demand fashion.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

127

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

The cloud computing paradigm has achieved

a) The Key Challenges

widespread adoption in recent years. Its

Being virtual in concept, the cloud environment

success is due largely to customers ability to

generates several questions in the minds of users

use services on demand with a pay-as-you go

with respect to confidentiality, integrity and

pricing model, which has proved convenient

availability. The key challenges for the adoption

in many respects. Low costs and high

of the cloud are as given below:

flexibility make migrating to the cloud

Assurance of the privacy and security

compelling. Despite its obvious advantages,

The cloud users are wary of the security and

however, many companies hesitate to move

privacy

to the cloud, mainly because of concerns

environment of the cloud is causing concerns

related to service availability, data lock-in,

amongst enterprises. As the same underlying

data security and legal uncertainties.

hardware may be used by other companies and

A previous study considers the data lock-in

competitors, it may lead to a breach of privacy.

problem and provides a convenient way to

Moreover, any data leakage or virus attack would

of

their

data.

The

multi-tenant

have a cascading effect on multiple organizations.

solve this using Meta cloud. The problem is


that once an application has been developed

Reliability and availability

based on one particular providers cloud

Instances of outages at the facilities of the cloud

services and using its specific API, that

service providers have raised concerns over the

application

provider;

reliability of the cloud solutions. Enterprises are

deploying it on another cloud would usually

recognizing that they would have to deal with

require completely redesigning and rewriting

some level of failures while using commodity-

is bound to that

based solutions. Also, the cloud providers cannot

it. Such vendor lock-in leads to strong

give an assurance on the uptime of their external

dependence on the cloud service operator.

internet connection, which cloud shut all access to

The Meta cloud framework contains the

the cloud.

following components: Meta cloud API, Meta


Data Security is Key Concern

cloud proxy, resource monitoring and so on.


But sometimes, transforming cloud as meta

There are a number of concerns surrounding the

cloud data security issues are raised which are not

adoption of the cloud especially because it is a

consider in the previous study.

relatively new concept. Assuring customers of


data security would be one of the biggest
challenges for cloud vendors to overcome. The

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

128

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

II.

Figure 1 shows the chart of the key barriers to


cloud adaptation.

ISBN: 378 - 26 - 138420 - 5

PROPOSED WORK

In this section we introduce a novel solution


Business continuity management (BCM) and
provide the overview of the Business Continuity
Management.
a) Business Continuity Management (BCM)
The BCMS will use to Plan Do Check Act
approach. The PDCA approach can be applied to
every element of the BCM lifecycle.
Business Continuity leads (BC leads)

Figure 1 Chart of the key barriers

Leads for business continuity management will be

To address this problem this paper introduce

appointed in each directorate, regional, area team

the Business Continuity Management (BCM)

and hosted bodies within the strategy.

is defined as a holistic management process that

BC leads will perform the following:

identifies to an organization and reduces the


Promote business continuity Management

impacts of data leakage issues. This contains


following stages Project initiation, understand the

Receive BC training

organization, BC strategies, develop Business


continuity planning, and Apply BCP. The

Facilitate the completion of BIAs

Business Continuity Planning is shown in the

Develop BCPs

following Figure 2.
Ensure that BCPs are available during
incident response
Ensure that incident responders receive
training appropriate to their role
Ensure that plans are tested, reviewed and
updated
Participate in the review and development
of the BCMS.
Figure 2 Business Continuity Management
Overview

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

129

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Business Continuity working Groups

Stage 1- Understanding the Organization

Working groups may be established to:

Understanding

Take control of resource allocation Set priorities


Set

continuity

organizations

strategies
objectives

in
and

line

ISBN: 378 - 26 - 138420 - 5

with

the business

is

essential

in

developing an appropriate BCM Programme. A

the

detailed understanding of what processes are

responsibilities

essential to ensure continuity of prioritized

Establish the measures that will be used to assure

activities to at least the minimum business

the BCMS remains current and relevant Report to

continuity objective level will be achieved by

top management on the performance of the

undertaking BIA. The BIA will incorporate

BCMS.

continuity requirements analysis which may

Emergency preparedness resilience and response

include

(EPRR)

qualifications required for prioritized activities.

The business continuity program will have close

BIAs will describe as follows:

links to EPRR because both desciplines aim to

the

staff

skills,

competencies

and

The prioritized activities of departments/

ensure the organization is resilient and able to

teams;

respond to threats and hazards. The BCMS


described in this strategy will ensure that the

The impact that the incidents will have on

organization is able to manage risks and incidents

prioritized activities

that directly impact on its ability to deliver

How long we could continue using the

business as usual.

emergency measures before we would have to

Assurance

restart our normal activities;

The National support centre will maintain an

A description of the emergency measures

overview of the BCMS. BC leads will be reuired

we have in place to deal with an incident;

to report on progress within their areas.


The threats to the continued delivery of
BCM Documentation

priority activate.

The National Support Centre will be given access


Stage 2 Determining BCM strategy

to related documentation by areas within the


scope, such as BCPs, training records, incident

BIAs

records and exercises to facilitate the sharing of

organizations

good practice throughout the organization. The

and business continuity risks. This information

Business

will be used to:

Continuity

management

has

the

will create a picture of the


dependencies,

vulnerabilities

following stages:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

130

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

To assist in deciding the scope of the

continuity plans will be based on different levels

BCM programme.

of response and escalation.

To provide the information from which

Business Continuity Plans

continuity options can be identified and

Various plans will continue to be developed to

evaluated.

identify the actions that are necessary and the


resources which are needed to enable business

To assist the preparation of detailed plans

continuity. Plans will be based upon the risks


Decisions that determine business continuity

identified, but will allow for flexibility.

strategies will be made at an appropriate level

Prioritized activity recovery plans (PARPs)

Recovery

Priority activities are those activities to which


priority must be given following an incident in

People

order to mitigate the impact. Activities of the


Premises

highest priority are those that if disrupted, impact


the organization to the greatest extent and in the

Technology and information

shortest possible time.


Suppliers and partners

Stage 4 Exercise, Audit, Marinating and


reviewing

Stage 3 Developing and implementing a BCM

Exercises

response

It is essential that regular BC exercises are carried

This stage considers the incident reporting

out to ensure that plans are tested and continue to

structure,

be effective and fit-for-purpose as operational

business

continuity

plans,

and

Prioritized activity recovery plans.

processes and technology configurations are


constantly changing. Exercise will rise awareness
of BCM procedures.

Incident Reporting Structure


There

are

various

sources

of

information

Audit

pertaining to business continuity threats such as

To

severe, flooding and soon.

organizations BCM polices and standards

The impact of all incidents will vary. It is

To

important that the response to an incident is

solutions

appropriate to the level of impact and remains

To validate the organizations BCM plans

flexible as the situation develops. Business

To verify that appropriate exercising and

validate

review

compliance

the

with

organizations

the

BCM

maintenance activities are taking place. To

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

131

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Business continuity exercises

highlight decencies and issues and ensure their


resolution

CONCLUSION

Management Review
An annual review of this strategy will be

In this paper we introduce a novel solution to

undertaken. However, events may prompt more

provide a convenient way to process to identify the

frequent re-examination, such as:

various security threats. This paper considers a

BIA

substantive

revision
changes

which
in

survey of Business continuity management (BCM)

identifies

processes

to avoid the security risks.

and

priorities;
A

REFERNCES

significant

assessment

and/or

change
risk

in

the

threat

appetite of

[1] ISO 22301 Societal Security - Business

the

Continuity

organization
New

Management

Systems

Requirements.
regulatory

or

legislative

[2] NHS England Core Standards for Emergency

requirements.

Preparedness,

Resilience

and

Response

(EPRR).

Embedding BCM in the Organizations


culture

[3]

J. Skene, D.D. Lamanna, and W. Emmerich,


Precise Service Level Agreements, Proc.

BCM must be an accepted management

26th Intl Conf. Software Eng. (ICSE 04),

process, full endorsed and actively promoted

IEEE CS Press, 2004, pp. 179188.

by directors. The communication of high-level


endorsement to all is essential. There are

[4] Q. Zhang, L. Cheng, and R. Boutaba, Cloud

various ways in which this be achieved:

Computing: State-of-the-Art and Research


Challenges,

Business continuity will be part of the

J.

Internet

Services

and

Applications, vol. 1, no. 1, 2010, pp. 718.

organizations induction for new starters

[5] The Route Map to Business Continuity

Participation in BIA and writing BCPs

Management: Meeting the Requirements of


Communication

of

risks,

alerts

ISO 22301.

and

incidents
Business continuity information will be
available on the staff intranet
Business continuity training

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

132

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

AUTHORS

Mr.P.Gireesh received the


Vaishnavi

Instutiate

Technology,

of

Tirupathi,

B.Tech degree in computer


science & engineering from
the Jawaharlal Nehru technological university
Anantapur, in 2011, and received the Audisankara
College of Engineering and Technology, Nellore
M.Tech degree in computer science engineering
from

the

Jawaharlal

Nehru

technological

university Anantapur in 2014, respectively. He


Participated National Level Paper Symposiums in
different
Networks,

Colleges.
Mobile

He

interests

Computer

Computing,

Network

Programming, and System Hardware. He is a


member of the IEEE.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

133

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Increasing network life span of MANET by using


cooperative MAC protocol
B Suman Kanth1, P Venkateswara Rao2
1

M.Tech , Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India.

Professor, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India.

Abstract Cooperative diversity has been shown

reducing the energy use for consistent data

to give important performance increases in

transmission becomes one of the most important

wireless

design concerns for such networks. As an

networks

where

communication

is

hindered by channel fading. In resource limit

promising

and

powerful

solution

that

can

networks, the advantages of cooperation can be

overcome the drawback of resource constraint

further exploited by optimally allocating the

wireless networks, cooperative communication has

energy and bandwidth resources among users in a

received major attention recently as one of the key

cross-layer way. In this paper, we examine the

candidates for meeting the stringent requirement

problem of transmission power minimization and

of the resource limited networks.

network lifetime maximization using cooperative


Cooperative communication is developed from the

diversity for wireless sensor networks, under the

traditional MIMO (multiple-input and multiple-

restraint of a target end-to-end transmission

output) techniques and the model of relay

consistency and a given transmission rate. By


using

cross-layer

optimization

channels. Though MIMO has been shown to be

scheme,

able to significantly enlarge the system throughput

distributive algorithms which mutually consider

and

routing, relay selection, and power allocation plans

reliability,

it

is

not

simple

to

be

straightforward to implement in the wireless

are proposed for the consistency constraint

sensor networks due to the control on the size and

wireless sensor networks.

hardware complexity of sensor nodes.

Index terms power minimization, cooperative


Cooperative communication, however, is able to

diversity, relay selection, cross-layer optimization.

achieve the same space diversity by forming a


I.

INTRODUCTION

virtual distributed antenna array where each


antenna belongs to a separate node. With

Wireless sensor networks are composed of nodes

cooperation, users that understanding a deep fade

powered by batteries, for which substitute or


recharging is very hard, if unfeasible. Therefore,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

134

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

in their link towards the destination can utilize

in MAC, and power allocation in PHY) for

quality of service (QoS).

cooperative networks is still an open issue.


Although there are some efforts concerned in

a variety of cooperative schemes have been

optimizing some metrics such as throughput,

developed so far. Cooperative beam forming


scheme

was

proposed for

delay, or power use under certain QoS conditions,

CDMA cellular

these efforts focus on either one-hop situation or

networks. Coded cooperation was proposed along

fading-free channels.

with analog-and-forward (AF) and decipher-andforward (DF) schemes. Diversity-multiplexing

Different from our work, we focus on various

tradeoff

the

optimization goal and conditions. We are planning

performance of diversive cooperative schemes

on energy minimization as well as network

such as fixed relaying, selection relaying, and

lifetime maximization. The QoS conditions in our

incremental relaying. Multi relay cooperative

work are end-to-end transfer reliability while the

protocol using space-time coding was proposed.

work is under the constraint of end-to-end

tools

were

used

to

analyze

transmission capacity. We believe in wireless


Opportunistic relaying (or selective cooperation)

sensor

was proposed with various relay selection policies.

network

and

the

network

lifetime

maximization, and the guarantee for end-to-end

Power allocation problem and SER performance

reliability

analysis in resource constraint networks. These

is

more

important

than

other

considerations. Moreover, we extend our work of

works are primarily focused on enhancinging the

cross-layer optimization for cooperative network.

link performance in the physical layer.

We decomposes the problem of minimizing

Cooperative communication presents at birth a

network power consumption into a routing sub-

cross-layer

communication

problem in the network layer and a combined relay

resources have to be carefully allocated among

selection and power allocation sub-problem in the

different nodes in the network. Therefore, the

physical layer.

problem,

since

integration and interaction with higher layers has

However, the decomposition method to solve this

become an active research area in recent years.

cross-layer problem is faulty for its complexity.

There have been a lot of efforts towards this

Since the algorithm projected is non-heuristic, it

consideration such as combining node cooperation

needs comprehensive iterations, with long meet

with ARQ in the link layer, or resource allocation

time and huge overhead for message exchanging,

in the MAC layer, and routing algorithm in the

thus inappropriate for sensor network application.

network layer.

On opposite side, in our work, we try to derive a

However, a complete cross-layer optimization

closed-form solution (though may not be optimal)

incorporating three layers (routing, relay selection

and propose our algorithm in a heuristic way; thus,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

135

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

it can be distributively employed in wireless

selection, and power allocation in arbitrarily

sensor

distributed wireless sensor networks.

networks.

In

the

cooperative

communication and data-aggregation techniques


are together implemented to reduce the energy

II. CROSS LAYER OPTIMIZATION PROBLEM IN


COOPERATIVE SENSOR NETWORK

spending in wireless sensor networks by reducing


the sum of data for transmission and better using

In this section, we prepare and analyze the min-

network

cooperative

power problem and the max-lifetime problem in

communication. There are also extra works in

the context of reliability conditioned cooperative

prolonging network lifetime using cooperative

networks and obtain relevant solutions, leading to

communication techniques.

our algorithms which will be explained in detail in

resources

through

the next section.

To address above stated problems, a cooperative


scheme is proposed join the maximum lifetime

Problem Formulation

power allocation and the energy aware routing to


Consider a multihop wireless sensor network

maximize the network lifetime. This method is not

containing of multiple arbitrarily disseminated

a cross-layer method since the energy saving


routing

is

formed

first,

and

sensor nodes and one sink. Each sensor node has a

cooperative

single

transmission is applied based on the build routes.


Therefore,

the

qualities

of

omnidirectional

antenna

and

can

dynamically adjust its transmitted power.

cooperative

transmission are not fully discovered since the best

We

cooperative route might be completely different

objectives. The first is called min-power problem:

from the noncooperative route.

given any source node, the goal is to find the route

consider

two

associated

optimization

that minimizes the total transmission power, while

A suboptimal algorithm for lifetime increasing is

fulfilling a required end-to-end broadcast success

developed based on the performance analysis for

probability and transmission rate.

MPSK modulation in the condition of adding


some special cooperative relays at the best places

The second is called max-lifetime problem: given

to prolong the network lifetime, while, in our

a set of source nodes, the goal is to offers an

work, the nodes are randomly distributed and no

information transferring scheme that increases the

additional relay nodes deployed.

network lifetime, defined as the lifetime of the


node whose battery consumes out first [1], while

In a word, we propose a fresh scheme to increase

fulfilling the same conditions as the min-power

the network lifetime by utilize the cooperative

problem.

diversity and together considering routing, relay

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

136

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

and the relay to confirm a accurate reception.


Otherwise, it sends a negative acknowledgment
(NACK) that allows the relay, if it received the
symbol properly, to transmit this symbol to the
receiver in the next time slot. 3) Multirelay
Cooperation

in

which

for

the

multirelay

collaboration mode, we use the opportunistic


relaying scheme according to this opportunistic
relaying scheme, for each frame, a node with the

Figure 1. Cooperative transmission (CT)

best immediate relaying channel condition among

and direct transmission (DT) modes as

several possible relays is selected to promote the

building blocks for any route.

packet in each frame.

From Figure 1, we can see that to promote data

Now consider the min-power problem described in

from node i to node j on link (i, j), either direct

the above. This is a convex problem and we can

transmission is used or a particular node r helps

solve it using Lagrangian multiplier techniques.

node i to forward data to node j using decipher-

To make best use of the lifetime of the entire

and-forward (DF) cooperative diversity.

network, the novel solution is that the lifetime of

The broadcast
transmission

rate and

(DT)

power

mode

and

for

each node in the route of flow is equal to a target

direct

lifetime.

cooperative

transmission (CT) mode classified into three cases

III.

COOPERATION-BASED CROSS-LAYER

such as 1) Direct Transmission in which node i


send the data to the node j.
Cooperation

in

which

for

SCHEMES

2) Single-Relay
the

In this section, we propose thorough minimum-

cooperative

transmission mode, the sender I sends its symbol

power

and

maximize-lifetime

in its time slot. Due to the transmit nature of the

algorithms, under the constraint of end-to-end

wireless medium, both the receiver and the relay

success probability and data rate, both in direct

receive noisy versions of the transmitted symbol.

mode and cooperative mode. We assume that each

We assume that the relay and the receiver decide

node transmits HELLO packet to its neighbors to

that the received symbol is properly received if the

update the local topology information. Our

received (SNR) is greater than a certain threshold.

algorithms are composed of two parts: routing

According to the incremental relaying scheme, if

(and relay selection) algorithm and power

the receiver deciphers the symbol suitably, then it

allocation algorithm. Algorithms are based on the

sends an acknowledgment (ACK) to the sender

conventional

Bellman-Ford

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

137

cross-layer

shortest

path

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

algorithm

which

can

be

distributively

ISBN: 378 - 26 - 138420 - 5

Algorithm 1 (Min-Power Cross-layer Scheme with


Direct Transmission (MPCS-DT)). Consider

implemented. In Bellman-Ford algorithm, each


node executes the iteration.

(1)

Notation

()
Qs
i)

except Cost (0) = 0 (node 0 represents the sink).

Table 1
Symbol and Explanation
Explanation
The effective distance between node
I and j
The latest estimate cost of the
shortest path from node j to the
destination
The set of neighboring nodes of node
i
Quality parameter

(2)

Every

distance

node

estimates

(3)

of

the

average

(4)

(5)

/
,

as

Cost(i)

min

(6)

optimization

( ) , and select node j as the next hop node.


If the required QoS parameter

is a

If not, the source will deliver a message

along the path about the

. Then each node in the

First we select the minimum-power route with the

route will adjust the transmit power.

least

(7)

correlated to

though the constructed route, informing all nodes

problem into two sub-problems.

the

/
,

will adjust the transmit power.

nothing to do with the QoS parameter. Thus we

Then,

( )

priori to the whole network, each node in the route

That means the formation of routing actually has

algorithm.

from

Every node updates its cost toward the

destination

We can see that to reduce is equal to minimizing.

conservative

SNR

Every node calculates the costs of its

outgoing links as

on

effective

periodically broadcasted HELLO message.

Transmission

based

the

of its outgoing links through the

measurement

Min-Power Cross-Layer Scheme with Direct

can decompose the cross-layer

Every node initiates its cost for routing as

Bellman-Ford

transmission

power
i)

Got to Step (2).

Min-Power Cross-Layer Scheme with Cooperative

for each node in the route is

Transmission

adjusted.
The

min-power

scheme

for

cooperative

Since the forward nodes in the route may not

communication is composed of two parts: single-

know

relay scheme and multirelay scheme. For single-

(if not a priori for the whole network),

the source node may need to transfer a message


containing

relay scheme, that to minimize

to the destination through the path

minimizing

is equal to

. Hence, the cross-layer best

to inform all the forward nodes. Thus, the cross-

strategy can be realized by three steps. First, the

layer scheme is as follows.

potential relay of each link is selected by


minimizing

. Then, the min-power route is

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

138

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

constructed as the route with the least

opportunistic

relaying

ISBN: 378 - 26 - 138420 - 5

multirelay

scenario)).

Consider

Finally, the transmission power of the nodes in the


route is adjusted. The algorithm as follows.

(1)

the similar as steps 1 & 2 Algorithm 1.

Algorithm 2 (Min-Power Cross-layer Scheme

(2)

the similar as steps 1 & 2 Algorithm 1.

Cooperative Transmission (MPCS-CT) (for single-

(3)

Every node sorts all its neighboring nodes

relay scenario)). Consider

ascending according to the value of

(1)

The similar as steps 1 & 2 in Algorithm 1.

(2)

The similar as steps 1 & 2 in Algorithm 1.

(3)

Every node calculates the costs of its

outgoing links as

= min

(, )

min

( , )

(4)

select node j as the next hop node.

Every node updates its cost toward the


+

Every node updates its cost toward the

(4)

of both I and j.

destination as Costi = min (


/

where N(i, j) denotes the set of neighboring node

of both i and j.

( )

Then it calculates the costs of its outgoing links as

where N(i, j) denotes the set of neighboring node

= min

and selects the first K nodes as potential relays.

, and select node k as the relay of this link,

destination as

), and

(5) (5) & (6) The similar as steps 5 & 6 in algorithm 1

and select node j as the next hop node.

except each path node and relay node in the route

(5)

adjust the transmit power.

& (6) The similar as steps 5 & 6 in

Algorithm 1 except each path node and relay node


ii)
in the route adjust the transmit power.

(6)

Go to (2).

(6)

Transmission

Max-Lifetime Cross-Layer Scheme with Direct

Go to step (2).

The remaining energy of each node is decreasing


For multirelay scheme (assume we need K relays

at unusual rates; the route may vary from time to

for each link), the difference with single-relay

time as the residual energy of the intermediate

scheme is that the K probable relays have to be

node in the route varies. The rate of the

selected and only the optimal relay is chosen from

recalculation of the route is determined by the rate

time-to-time in every transmission slot.

of HELLO message exchange. Thus, the algorithm

The best communicate is in charge of relaying the

is as follows.

packet, while further potential relays will turn off

Algorithm 4 (Max-Lifetime Cross-Layer Scheme

the radio and not receive the packet to save

with Direct Transmission (MLCS-DT)). Consider

energy. The algorithm is as follows.

(1)

& (2) The similar as steps 1 & 2 in

Algorithm 1.
Algorithm 3 (Min-Power Cross-Layer Scheme
Cooperative

Transmission

(MPCS-CT)

(for

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

139

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

(3)

Every node measures its residual energy

(5)

and its total power for the ongoing flows. Then it

and its total power

Every node updates its cost toward the


= min

destination as

calculate the cost of its outgoing links as

ISBN: 378 - 26 - 138420 - 5

()

and selects node j as the next hop node.

for the ongoing

flows.

(6)

(4) Every node updates its cost toward the

Algorithm 1 except each forward node and relay

destination as Costi = min

( )

& (7) The similar as steps 5 & 6 in

node in the route adjust the transmit power.

(8)

and select node j as the next hop node.

Go to (2).

(5) & (6) the similar as steps 5 & 6 in Algorithm 1

IV. EXPECTED RESULTS

except each node in the route adjust the transmit


power.

In this section we compare the Min-Power

(7) Go to (2).

Algorithms to demonstrate the effect of cross-layer

iii) Max-Lifetime

Cross-Layer

Scheme

design for communication; we implement two

with

min-power algorithms: MPCS-DT and MPCS-CT

Cooperative Transmission

in random networks. For better comparison, we


To find a route which can maximize, we should

also implement two other cooperation along the

find a path with the minimum, and the power of

shortest

nodes should be adjusted.

routing

(MPCR)

algorithm and cooperation along the shortest noncooperative path (CASNCP) algorithms.

Algorithm 5 (Max-Lifetime Cross-Layer Scheme


with Cooperative Transmission (MLCS-CT)).

And also we compare the Max-Life Algorithms,

Consider
(1)

non-cooperative

for that we consider three different schemes: (1)


max-lifetime cross-layer scheme with direct

& (2) the similar as steps 1 & 2 in

Algorithm 1.

communication (MLCS-DT)and (2) max-lifetime

(3) Every node calculates the effective distance of

cross-layer

its

communication (MLCS-CT) and (3) greedy power

outgoing
,

= min

( , )

links

as

with

cooperative

allocation and cost-based routing (GPA-CR).

), and select

node k as the relay of this link.


(4)

scheme

V. CONCLUSION

Every node measures its residual energy

In this paper, we build several cross-layer

and its total transmission power for the ongoing

algorithms for energy resourceful and consistent

flows. Then calculates the cost of its outgoing

data transfer in wireless sensor networks using

links.

cooperative diversity. We examine the problem of


how to minimize transmission power consumption

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

140

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

and

exploit

guaranteeing

the
the

network
end-to-end

lifetime

while

[4] L. Zheng and D. N. C. Tse, Diversity and

accomplishment

freedom: a fundamental tradeoff in multiple

probability.

antenna channels, in Proceedings of the IEEE


International

REFERENCES

Theory

and E. Cayirci, A survey on sensor

p.

476,

Lausanne,

Transactions on Information Theory, vol. 25,


no. 5, pp. 572584, 1979.

[2] A. Nosratinia, T. E. Hunter, and A. Hedayat,


in

wireless

[6] A. Sendonaris, E. Erkip, and B. Aazhang,

networks, IEEE Communications Magazine,

User cooperation diversity part I and part 2,

vol. 42, no. 10, pp. 7480, 2004.

IEEE Transactions on Communications, vol.


51, no. 11, pp. 19271938, 2003.

[3] Y.-W. Hong, W.-J. Huang, F.-H. Chiu, and C.C. J. Kuo, Cooperative communications in
wireless

'02),

Information

theorems for the relay channel, IEEE

vol. 40, no. 8, pp. 102105, 2002.

resource-constrained

(ISIT

on

[5] T. M. Cover and A. A. E. Gamal, Capacity

networks, IEEE Communications Magazine,

communication

Symposium

Switzerland, July 2002.

[1] F. Akyildiz, W. Su, Y. Sankarasubramaniam,

Cooperative

ISBN: 378 - 26 - 138420 - 5

[7] J. N. Laneman, D. N. C. Tse, and G. W.

networks,

Wornell, Cooperative diversity in wireless

IEEE Signal Processing Magazine, vol. 24, no.

networks: efficient protocols and outage

3, pp. 4757, 2007.

behavior, IEEE Transactions on Information


Theory, vol. 50, no. 12, pp. 30623080, 2004.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

141

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Scalable Mobile Presence Cloud with


Communication Security
N.Tejesh1 D.Prasanth2
1

PG Student, M.Tech. Computer Science and Engineering,


Audisankara College of Engineering & Technology, Gudur.A.P, India
[email protected]
2

Assistant Professor, Department of CSE,


Audisankara College of Engineering & Technology, Gudur.A.P, India
[email protected]

Abstract-

Within Portable existence services is often a


vital portion of a new social networking applications
on account of mobile users existence details such as
global setting method place, circle address, along
with online/offline rank are usually constantly
apprise to help users online close friends. A mobile
ubiquity companies can be an critical component
impair processing environments, for your reason that
maintains the up-to-date report on existence data
connected with mobile consumer. In the event that
existence improvements happen usually the amount
of announcements written by existence server may
cause scalability issue along with close friend number
look for issue with large-scale mobile existence
companies. For you to triumph over the particular
scalability issue recommended an efficient along with
ascendable server buildings termed existence impair.
It sets up the particular existence server straight into
quorum centered server-server buildings regarding
useful looking. If a mobile consumer ties together a
new circle or perhaps net, existence impair lookups
the particular existence data. Additionally, it
achieves modest regular look for latency from the led
look for protocol along with one-hop caching
approach.

keywordsSocial networks, mobile presence services, distributed


presence servers, cloud computing.

I. INTRODUCTION
The Presence [1] entitled applications such as Face-book,
Twitter etc., which is produced by mobile devices and cloud
computing [2] nature due to the prevalence of internet
[3].Way the members are engaged with their buddies on

internet are changed by the social network services [4]. In


order to interact with buddies across great distance
participants can dispense the live event immediately using
their mobile device. Mobile users presence information
details will be maintained by mobile presence service [5]. In
cloud computing environment mobile presence service is a
vital component of social network application. Because of the
ubiquity of the Internet, mobile devices and cloud computing
environments can provide presence-enabled applications, i.e.,
social network applications/services, worldwide. Face book,
Twitter Foursquare Google Latitude, buddy cloud and Mobile
Instant Messaging (MIM) are examples of presence-enabled
applications that have grown rapidly in the last decade. Social
network services are changing the ways in which participants
engage with their friends on the Internet. They exploit the
information about the status of participants including their
appearances and activities to interact with their friends.
Moreover, because of the wide availability of mobile devices
(e.g., Smartphones) that utilize wireless mobile network
technologies, social network services enable participants to
share live experiences instantly across great distances.
Presence information tells the detail about mobile users
availability, activity and machine capacity. Service does
binding of user id to his/her current presence information
details. Each individual mobile user has a buddy list which
includes details of whom he/she wants to interact with in
social network services. When a user does shipment from one
level to other, this change is instinctively transmitted to each
individual on the buddy list. Server cluster technology
increases the search speed and decrease the report time. For
example in social network application mobile user logs in
through his/her mobile device, the mobile presence services
searches and reveals each of them about users friend list such
as instant messaging system [6]. Potential of presence cloud
[5] [7] can be examined by using search cost and search
satisfaction without impaired neither of them. When a user
arrives presence server provoke a number of messages is
search cost. Time it takes to examine the arrival of users
buddy list is search satisfaction. To help the users who are

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

142

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

present worldwide, the services enhanced by Google [3] [8]


and Facebook [3] are proliferated among many servers.
Presence server used in large scale social network services to
ameliorate the coherence of mobile presence services. In this
section, examine the existing server architecture for buddy list
in large scale geographical information Centre. Overloading
of buddy search message on presence server leads to
scalability problem. Presence cloud disseminates many users
information details among many presence servers on the
internet, which is used as a building block of mobile presence
services. For efficient buddy list search there is no single point
collapse, since servers in presence cloud are organized in
quorum [9] based server to server architecture to gain small
search delay using directed buddy search algorithm.
Caching procedure is used to reduce buddy list search. The
potential of three architectures such as presence cloud, mesh
[10] based scheme and distributed hash table [11] are
examined in terms of search response time and friend
notification time. Presence information tells the detail about
mobile users availability, activity and machine capacity.
Service does binding of user id to his/her current presence
information details. Each individual mobile user has a buddy
list which includes details of whom he/she wants to interact
with in social network services. When a user does shipment
from one level to other, this change is instinctively transmitted
to each individual on the buddy list. Server cluster technology
increases the search speed and decrease the report time. For
example in social network application mobile user logs in
through his/her mobile device, the mobile presence services
searches and reveals each of them about users friend list such
as instant messaging system [6]. Potential of presence cloud
[5] [7] can be examined by using search cost and search
satisfaction without impaired neither of them. When a user
arrives presence server provoke a number of messages is
search cost. Time it takes to examine the arrival of users
buddy list is search satisfaction. To help the users who are
present worldwide, the services enhanced by Google [3] [8]
and Facebook [3] are proliferated among many servers.
Presence server used in large scale social network services to
ameliorate the coherence of mobile presence services. In this
section, examine the existing server architecture for buddy list
in large scale geographical information Centre. Overloading
of buddy search message on presence server leads to
scalability problem. Presence cloud disseminates many users
information details among many presence servers on the
internet, which is used as a building block of mobile presence
services. For efficient buddy list search there is no single point
collapse, since servers in presence cloud are organized in
quorum [9] based server to server architecture to gain small
search delay using directed buddy search algorithm. Caching
procedure is used to reduce buddy list search. The potential of
three architectures such as presence cloud, mesh [10] based
scheme and distributed hash table [11] are examined in terms
of search response time and friend notification time.

ISBN: 378 - 26 - 138420 - 5

II. RELEATED WORK


The rationale behind the design of Presence Cloud is to
distribute the information of millions of users among
thousands of presence servers on the Internet. To avoid single
point of failure, no single presence server is supposed to
maintain service-wide global information about all users.
Presence Cloud organizes presence servers into a quorum
based server-to-server architecture to facilitate efficient buddy
list searching. It also leverages the server overlay and a
directed buddy search algorithm to achieve small constant
search latency and employs an active caching strategy that
substantially reduces the number of messages generated by
each search for a list of buddies. We analyze the performance
complexity of Presence Cloud and two other architectures, a
Mesh based scheme and a Distributed Hash Table (DHT)based scheme. Through simulations, we also compare the
performance of the three approaches in terms servers on the
Internet. The design of Presence Cloud, a scalable server-toserver architecture that can be used as a building block for
mobile presence services. The rationale behind the design of
Presence Cloud is to distribute the information of millions of
users among thousands of presence servers on the Internet. To
avoid single point of failure, no single presence server is
supposed to maintain service-wide global information about
all users. Presence Cloud organizes presence servers into a
quorum-based server-to-server architecture to facilitate
efficient buddy list searching.

Fig1. Presence Cloud Architecture

Presence servers which are present in presence cloud, where


these presence servers are arranged in quorum based server to
server architecture and also load on servers are balance in
presence cloud sever overlay. All these presence server keeps
caches for buddies in order to increase query speed is one hop
caching approach. Finding small constant search delay results
in directed buddy search by decreasing network traffic using
one hop search strategy. Architecture of presence cloud which
is the proposed work is shown in Figure1, Using 3G or Wi-Fi
services mobile user access the internet and make a data link
to the presence cloud. Using secure hash algorithm mobile
users are intent to one of the presence servers. To transfer
presence information details, the mobile user is authenticated
to the mobile presence services and also opens a TCP link.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

143

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Once path is set up, the mobile user request for the friend list
to the presence server which is present in presence cloud. And
finally the request is responded by the presence cloud after
completing an efficient search of buddys presence
information. Analyze the performance complexity of Presence
Cloud and two other architectures, a Mesh based scheme and
a Distributed Hash Table (DHT)-based scheme. Through
simulations, we also compare the performance of the three
approaches in terms of the number of messages generated and
the search satisfaction which we use to denote the search
response time and the buddy notification time. The results
demonstrate that Presence- Cloud achieves major performance
gains in terms of reducing the number of messages without
sacrificing search satisfaction. Thus, Presence Cloud can
support a large-scale social network service distributed among
thousands of servers on the internet. Presence Cloud is among
the pioneering architecture for mobile presence services. To
the best of our knowledge, this is the first work that explicitly
designs a presence server architecture that significantly
outperforms those based distributed hash tables. Presence
Cloud can also be utilized by Internet social network
applications and services that need to replicate or search for
mutable and dynamic data among distributed presence servers.
The contribution is that analyzes the scalability problems of
distributed presence server architectures, and defines a new
problem called the buddy-list search problem. Through our
mathematical formulation, the scalability problem in the
distributed server architectures of mobile presence services is
analyzed. Finally, we analyze the performance complexity of
Presence Cloud and different designs of distributed
architectures, and evaluate them empirically to demonstrate
the advantages of Presence Cloud. Server architectures of
existing presence services, and introduce the buddy-list search
problem in distributed presence Architectures in large-scale
geographically data centers.

III. PRESENCE CLOUD


The past few years has seen a veritable frenzy of research
activity in Internet-scale object searching field, with many
designed protocols and proposed algorithms. Most of the
previous algorithms are used to address the fixed object
searching problem in distributed systems for different
intentions. However, people are nomadic, the mobile presence
information is more mutable and dynamic; anew design of
mobile presence services is needed to address the buddy-list
search problem, especially for the demand of mobile social
network applications. Presence Cloud is used to construct and
maintain distributed server architecture and can be used to
efficiently query the system for buddy list searches. Presence
Cloud consists of three main components that are run across a
set of presence servers. In the design of Presence Cloud, the
ideas of P2P systems and present a particular design for
mobile presence services has been refined. Presence Cloud
server overlay: It organizes presence servers based on the
concept of grid quorum system. So, the server overlay of

ISBN: 378 - 26 - 138420 - 5

Presence Cloud has a balanced load property and a two-hop


diameter node degrees, where n is the number of presence
servers. One-hop caching strategy: It is used to reduce the
number of transmitted messages and accelerate query speed.
All presence servers maintain caches for the buddies offered
by their immediate neighbors Directed buddy search: It is
based on the directed search strategy. Presence Cloud ensures
an one-hop search, it yields a small constant search latency on
average. The primary abstraction exported by our Presence
Cloud issued scalable server architecture for Mobile presence
services, and can be used to efficiently search the desired
buddy lists. We illustrated a simple overview of Presence
Cloud in Fig. 1. In the mobile Internet, a mobile user can
access the Internet and make a data connection to Presence
Cloud via 3G or Wifi services. After the mobile user joins and
authenticates himself/herself to the mobile presence service,
the mobile user is determinately directed to one of Presence
Servers in the Presence Cloud by using the Secure Hash
Algorithm, such as SHA-1. The mobile user opens a TCP
connection to the Presence Server (PSnode) for control
message transmission, particularly for the presence
information. After the control channel is
established, the mobile user sends a request to the connected
PSnode for his/her buddy list searching. Our Presence
Cloudshall do an efficient searching operation and return the
presence information of the desired buddies to the mobile
user.

IV. EVOLUTION
A cost analysis of the communication cost of Presence Cloud
in terms of the number of messages required to search the
buddy information of a mobile user. Note that how to reduce
the number of inter server communication messages is the
most important metric in mobile presence service issues. The
buddy-list search problem can be solved by a brute-force
search algorithm, which simply searches all the PS nodes in
the mobile presence service. In a simple mesh-based design,
the algorithm replicates all the presence information at each
PS node; hence its search cost, denote by QMesh, is only one
message. On the other hand, the system needs n _ 1 messages
to replicate a users presence information to all PS nodes,
where n is the number of PS nodes. The communication cost
of searching buddies and replicating presence information can
be formulated as Mcost = QMesh +RMesh, where RMesh is
the communication cost of replicating presence information to
all PS nodes. Accordingly, we have Mcost = O(n). In the
analysis of Presence Cloud, we assume that the mobile users
are distributed equally among all the PS nodes, which is the
worst case of the performance of Presence- Cloud. Here, the
search cost of Presence Cloud is denoted as Qp, which is
messages for both searching buddy lists and replicating
presence information. Because search message and replica
message can be combined into one single message, the
communication cost of replicating, Rp (0). It is straight
forward to know that the communication cost of searching

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

144

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

buddies and replicating presence information in Presence


Cloud is Pcost .However, in Presence Cloud, a PS node not
only searches a buddy list and replicates presence information,
but also notifies users in the buddy list about the new presence
event. Let b be the maximum number of buddies of a mobile
user. Thus, the worst case is when none of the buddies are
registered with the PS nodes reached by the search messages
and each user on the buddy list is located on different PS
nodes.

ISBN: 378 - 26 - 138420 - 5

[9]. Quorumb asedtechni queshttp:// www.en.wikip idiaorg/ wiki/Quoru


m_(distributed _computing).
[10]. Jianming zhao, Nianmin yao, Shaobin Cai, and Xiang li, Tree-Mesh
Based P2P Streaming data distribution schema,2012.
[11].
Distributed
hash
744/S07/lectures /16-dht.pdf.

tables

http://www.cs.cmu.edu/~dga/15-

[12]. Juuso aleksi lehtinen, Mobile peer-to-peer over session initiation


protocol, August 4,2008.
[13]. Kundan Singh and Henning Schulzrinne Department of Computer
Science, Columbia University {kns10,hgs}@
cs.columbia.edu, SIPPEER : A session iniiation protocol (SIP)- peer-topeer internet telephony cllient adaptor .
[14]. Michael Piatek, Tomas Isdal, Arvind Krishnamurthy , and Thomas
Anderson One hop Reputations for Peer to Peer File

V. CONCLUSION
In this paper In large scale social network services mobile
presence services is supported by the scalable server
architecture called as presence cloud.
A scalable server
architecture that supports mobile presence services in largescale social network services. Presence Cloud achieves low
search latency and enhances the performance of mobile
presence services. Total number of buddy search messages
increases substantially with the user arrival rate and the
number of presence servers. The growth of social network
applications and mobile device computing capacity to explore
the user satisfaction both on mobile presence services or
mobile devices. Presence Cloud could certificate the presence
server every time when the presence server joins to Presence
Cloud. The results of that Presence Cloud achieve
performance gains in the search cost without compromising
search satisfaction.

Sharing Workloads.
[15]. Brent Hecht, Jaime Teevan , Meredith Ringel Morris, and Dan Liebling,
SearchBuddies: Bringing Search Engines into the Conversation, 2012

REFERENCE
[1 ] R.B. Jennings, E.M. Nahum, D.P. Olshefski, D. Saha, Z.-Y. Shae, and C.
Waters, A Study of Internet Instant Messaging and Chat Protocols, IEEE
Network, vol. 20, no. 6, pp. 16-21, July/Aug. 2006.
[2] Z. Xiao, L. Guo, and J. Tracey, Understanding Instant Messaging
Traffic Characteristics, Proc. IEEE 27th Intl Conf. Distributed
Computing Systems (ICDCS), 2007.
[3] Chi, R. Hao, D. Wang, and Z.-Z. Cao, IMS Presence Server:aTraffic
Analysis and Performance Modelling, Proc. IEEE IntConf. Network
Protocols (ICNP), 2008.
[4] Instant Messaging and Presence Protocol IETF Working Group,http:
//www.ietf.org /html.charters/impp-charter.html, 2012.
[5] Extensible Messaging and Presence Protocol IETF Working Group,
http://www.ietf.org /html.charters/xmpp-chart er.html,2012.
[6] Open Mobile Alliance, OMA Instant Messaging and Presence Service,
2005.
[7] P. Saint-Andre, Interdomain Presence Scaling Analysis for the
Extensible Messaging and Presence Protocol (XMPP), IETF Internet
draft, 2008.
[8 ]X. Chen, S. Ren, H. Wang, and X. Zhang, SCOPE: Scalable Consistency
Maintenance in Structured P2P Systems, Proc. IEEE INFOCOM, 2005.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

145

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Adaptive and well-organized Mobile Video Streaming public


Networks in Cloud
1

B.Renuka, 2Ch.Madhu Babu


PG Scholor, Audisankara College of Engineering and Technology, Gudur.
Assoc Professor(CSE), Audisankara College of Engineering and Technology, Gudur.
Email: [email protected] , [email protected]

Abstract: Due to the high demands of video traffics over mobile networks, the wireless link
more fails to keep up the pace with the demand. There exits a gap between the request and the
link capacity which results in poor service quality of the video streaming over mobile networks
which includes disruptions and long buffering time. While demands on video traffic above
mobile webs have been float, the wireless link capacity cannot retain up alongside the traffic
request. The gap amid the traffic request and the link capacity, alongside time-varying link
conditions, by-product in poor ability quality of video streaming above mobile webs such as long
buffering period and intermittent confusion. Leveraging the cloud computing knowledge, we
advice a new mobile video streaming framework, dubbed AMES-Cloud that has two parts:
Adaptive Mobile Video Streaming (AMOV) and Efficient Communal Video Sharing (ESoV).
AMoV and ESoV craft a personal agent to furnish video streaming services effectually for every
single mobile user. For a given user, AMoV lets her private agent adaptively adjust her
streaming flow alongside a scalable video coding method established on the feedback of link
quality. In similar, ESoV monitors the common web contact amid mobile users, and their
confidential agents attempt to perfects video content in advance. We apply a prototype of the
AMES-Cloud structure to clarify its performance.

INDEX TERMS: Scalable Video Coding, Adaptive Video Streaming, Mobile Networks, Social
Video Sharing, Cloud Computing.

Trough cloud computing the user can

1. INTRODUCTION

decrease the cost and can use the resource at

Cloud computing is the lease of the

any time.

resources through which the users can use


the

resources

depending

upon

There are three types of cloud as shown in

the

fig1

requirement and pay based on the usage.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

146

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

i) Public cloud

deployment of applications without the cost

ii) Private cloud

and complexity of buying and managing the

iii) Hybrid cloud

underlying hardware and software and


provisioning hosting capabilities.

Public cloud: Public cloud or external cloud


is one in which the resources are leased on
self service basis over the internet, via web
applications/web services, from an off-site
third-party provider who shares resources
and bills on a fine-grained utility computing
basis.
Fig 1: Types of services

Private cloud: Private cloud is also called

Fig 2 shows the architecture of a

internal cloud; it is used to describe the


network.

typical cloud at a high level. An end user

Hybrid cloud is one

Bob connects to the cloud via a portal from

which contains multiple internal or external

his browser. Alternatively, a user Alice can

clouds. Means N number of internal and

choose to directly connect to the cloud

external clouds

manager via a command line interface

offerings

of

Hybrid cloud:

private

similar to that used in EC2. A cloud

AMES is based on platform as a service.

provides

Platform as a service (PaaS) is a category of

pool to store persistent user data. The users

(SaaS) and infrastructure as a service (IaaS),

will make the request and the cloud manager

it is a service model of cloud computing. In

will authenticate the user and he keep track

this model, the consumer creates the

of the users and their request and due to the

software using tools and/or libraries from

streaming techniques and AMoV will adjust

the provider. The consumer also controls

PaaS

offerings

the streaming flow with a video coding

configuration
facilitate

images can be run, and optionally a storage

service.[1] Along with software as a service

settings.

of resources:

a set of computer servers on which the VM

computing platform and a solution stack as a

and

types

collection of (VM) virtual machine images,

cloud computing services that provides a

software deployment

three

technique will adjust the flow and increase

the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

147

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

the

quality.

ISBN: 378 - 26 - 138420 - 5

into account for developing the proposed


system.

2.1. Scalability: Mobile video streaming


services should support a wide spectrum of
Fig 2: cloud architecture

mobile devices; they have different video


resolutions, different computing powers,
different wireless links (like 3G and LTE)

2. LITERATURE REVIEW

and so on. Also, the available link capacity


Several authors have developed the

of a mobile device may vary over time and

techniques related to storing the data and

space depending on its signal strength, other

also for maintaining the data and for security

users traffic in the same cell, and link

issues related to the cloud.

condition

The quality of service on mobile video is

versions (with different bit rates) of the

based on two factors: Literature survey is

same video content may incur high overhead

the

in terms of storage and communication.

most

important

step

in

software

variation.

Storing

multiple

development process. Before developing the


tool it is necessary to determine the time
factor, economy n company strength. Once

2.2.

Adaptability:

these things r satisfied, ten next steps are to

streaming

determine which operating system and

considering relatively stable traffic links

language can be used for developing the

between servers and users perform poorly in

tool. Once the programmers start building

mobile

the tool the programmers need lot of

fluctuating wireless link status should be

external support. This support can be

properly dealt with to provide tolerable

obtained from senior programmers, from

video streaming services. To address this

book or from websites. Before building the

issue, we have to adjust the video bit rate

system the above consideration are taken

adapting to the currently time-varying

techniques

environments

Traditional
designed

[11].

Thus

video
by

the

available link bandwidth of each mobile

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

148

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

user. Such adaptive streaming techniques

also developed client-side HTTP adaptive

can effectively reduce packet losses.

live streaming solutions.


2.2.2.

Mobile

Cloud

Computing

Techniques
The cloud computing

has

been well

positioned to provide video streaming


services, especially in the wired Internet
because of its scalability and capability. For
example, the quality-assured bandwidth
auto-scaling for VoD streaming based on the
cloud computing is proposed , and the
Fig. 3. An illustration of the AMES-

CALMS framework is a cloud-assisted live

Cloud framework
2.2.1.

Adaptive

media
Video

streaming

service

for

globally

distributed users. However, extending the

Streaming

cloud computing-based services to mobile

Techniques

environments requires more factors to


In the adaptive streaming, the video traffic

consider: wireless link dynamics, user

rate is adjusted on the fly so that a user can

mobility, the limited capability of mobile

experience the maximum possible video

devices. More recently, new designs for

quality based on his or her links time-

users on top of mobile cloud computing

varying bandwidth capacity. There are

environments

mainly two types of adaptive streaming

are

proposed,

which

virtualized private agents that are in charge

techniques, depending on whether the

of satisfying the requirements (e.g.QoS) of

adaptively is controlled by the client or the

individual users such as Cloudlets and

server. The Microsofts Smooth Streaming

Stratus.

is a live adaptive streaming service which


can switch among different bit rate segments

The Video usage and images plays a vital

encoded with configurable bit rates and

role in communication. The usage of

video resolutions at servers, while clients

traditional networking and service providers

dynamically request videos based on local

lacks to provide the quality centered and

monitoring of link quality. Adobe and Apple

reliable

service

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

149

to

the

mobile

users

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

concerning with the media data. The

Due to the fast development of the

problems that leads to the poor services

mobile communication technology, more

from the service providers would be low

people are getting addicted to video

bandwidth

efficient

streaming over phones. Over the few years,

transfer of video to the user, the disruption

video streaming is becoming challenging

of video streaming also occurs due to the

over wireless links than on wired links. The

low bandwidth. The buffer time of the video

increasing

over mobile devices which moves from

overwhelming the wireless link capacity.

place to place affects the smooth streaming

The

and also sharing of video from one user to

disruptions and very long buffering time

another user over social media. Our survey

while receiving video through networks like

shows the functioning of various methods

3G or 4G due to short bandwidth and link

and architecture which used cloud to provide

fluctuations. So, it is imperative to improve

effective solution for providing better

the services of video streaming over mobile

service to the users. AMES is cloud

networks. Scalability and adaptability are

architecture built specially to provide video

the two aspects in order to improve the

service to the user. The study has came up

quality of the video streaming over mobile

with a optimal solution, proposing with

networks.

which

affects

the

video cloud, which collects the video from

video

mobile

users

traffic

often

demands

suffer

are

from

Scalable video coding (SVC) and

video service providers and providing the

adaptive

reliable service to the user[1].The network

combined together to accomplish the best

providers

video

possible quality of the video streaming

downloads but it provides some delays due

services. So, that we can adjust the SVC

to network dynamics so this technique is

layers which depends on the current link

used to remove jitters and provide video on

status. The cloud computing technique is

demand[3].

streaming

ready to provide scalable resources to the

solutions for different mobile which shows

service providers and process offloading to

my realistic work relevant to streaming

the mobile users. So, cloud data centers can

methods with RTMP protocols family and

provision to large scale real time video

solutions for iPhone, Android, Smart mobile

services. In cloud more than one agent

YouTube

cloud

provide

centered

streaming

techniques

can

be

phones, Window and Blackberry phones etc.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

150

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

instances can be maintained dynamically

responsible to store the popular video clips.

and effectively due to mobile user demands.

tempVB is a video base which is temporary

The social network services (SNS)

and is utilized to cache new mobile users for

on the mobile networks is becoming

popular videos, while it counts the access

increasingly popular. In SNSs mobile users

frequency of each video.VC keeps on

might post, comment and share videos

executing a collector to look for videos

which can be viewed and by his/her friends.

which are popular already in video service

So,

the

provider (VSP), and it will re-encode the

relationship between the mobile users and

videos that are collected into scalable video

their SNS activities in order to perfects the

coding format and will save in tempVB.

first part of the video during streaming.

A sub video cloud (subVC) is dynamically

we

are

inspired

to

exploit

created if there is any ling of video demand


from the mobile user. A sub video base
(subVB) is present in subVC and it stores
segments of recently fetched video. The
subVC contains encoding functions, and if
the mobile users request a new video, which
is not in the subVB or the VB in VC, the
subVC will fetch, encode and move the
video. During the time of the streaming of
videos, the users of the mobile will report
the link conditions to the subVC and it will
offer adaptive streams. There is a temporary
storage in every mobile device which is
Fig.4.Context architecture

known as local video base (localVB), used


for perfecting and buffering.

3. CLOUD FRAMEWORK
As shown in the above figure, the video

4.

SOCIAL

streaming and storing system in the cloud is

PREFETCHING

called video cloud (VC). Within the video

In social network services, the mobile users

cloud, there is video base (VB), which is

can subscribe to their friends and content

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

151

AWARE

VIDEO

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

publishers, and there are numerous types of

about security, availability, and privacy of

activities socially. So it is required for us to

their data as it rest sin the cloud.

define different kinds of levels of strengths

In this technique we propose an adaptive

for those socially oriented activities to

mobile

indicate many different possibilities so that

framework, called AMES-Cloud, which

the videos that are shared by one mobile

efficiently stores videos in the clouds (VC),

user can be viewed by the receiver of his/her

and utilizes cloud computing to construct

sharing activities, so the sub video clouds

private agent (subVC) for each mobile user

may engage into pre fetching at subVB done

to try to offer non-terminating video

in background and may transfer to mobile

streaming adapting to the fluctuation of link

users local VB. Because after one shares a

quality based on the Scalable Video Coding

video, there can be a bit of delay that the

ability. Also AMES-Cloud can further seek

receiver will know the sharing, and starts to

to issue none buffering experience of

watch. So, advance pre fetching will not

video streaming by background pushing

affect the mobile users in most of the cases.

functions among the VB, subVBs and

But a mobile user may play the videos to

localVB of mobile users. We assess the

view without any delay due to buffering as

AMES-Cloud by prototype implementation

the first part or May even the entire video is

and shows that the cloud computing

locally pre fetched already.

technique brings significant improvement on

video

streaming

and

sharing

the adaptively of the mobile streaming. We


5. IMPLEMENTATION

disregard the cost of encoding workload in


the cloud while implementing the prototype.

Cloud computing promises lower costs,


rapid scaling, easier

maintenance, and

This method require three different steps

service availability anywhere, anytime, a key


1. Uploading and Rating videos:

challenge is how to ensure and build


confidence that the cloud can handle user

2. User details

data securely. A recent Microsoft survey


3. Rate videos

found that 58 percent of the public and 86


percent of business leaders are excited about

5.1. Uploading and rating Video: Here we

the possibilities of cloud computing. But

can upload the videos and also we can give

more than 90 percent of them are worried

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

152

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

rating to the videos depending upon the

adaptability and scalability of the mobile

priorities or the usage.

streaming, and the efficiency of intelligent


pre fetching.

5.2. User Details: In this we will maintain


the details of the users and also determine

7.REFERENCES

the usage of each user. And keep track of the

[1].A Survey on efficient video sharing and

videos the user is requesting and account

streaming in cloud environment using VC,

them.

M.Sona, D.Daniel, S.Vanitha.


[2]

.AMES-Cloud

Framework

of

5.3. Rate videos: This wills avoiding

Adaptive Mobile Video Streaming and

unexpected

videos

accept/reject

videos

from
then

users.

After

Efficient Social Video

only

users

[3].CloudStream:

delivering

high-quality

can/cannot view their own videos.

streaming videos through a cloud-based

6. CONCLUSION

SVC proxy ,Zixia Huang, ,ChaoMei,

In this paper we have discussed our

Li

Erran Li,Thomas

proposal of the cloud assisted adaptive

[4]. CISCO, Cisco Visual Networking

mobile video streaming and public websites

Index : Global Mobile Data Traffic Forecast

are fetching, which stores the videos

Update , 2011-2016, Tech. Rep., 2012.

efficiently in the clouds and constructs

[5] .Y. Li, Y. Zhang, and R. Yuan,

private agent (subVC) for active mobile

Measurement and Analysis

users in order to attempt to give non

of a Large Scale Commercial Mobile

terminating

Internet TV System, in ACM IMC,pp.

streaming

of

videos

by

adapting to the changes of quality of links

209224, 2011.

which depends on scalable video coding

[6]. T. Taleb and K. Hashimoto, MS2: A

technique, and to try to provide non

Novel

buffering video streaming occurrence by

Architecture, in IEEE Transaction on

background perfecting based on the tracking

Broadcasting,vol. 57, no. 3, pp. 662673,

of the interactions of mobile users in their

2011.

SNSs. We evaluated the framework by

[7] .X. Wang, S. Kim, T. Kwon, H. Kim, Y.

prototype

showed

Choi,Unveiling the BitTorrent Performance

successfully that the cloud computing

in Mobile WiMAX Networks, in Passive

method

andActive Measurement Conference, 2011.

implementation,

brings

and

improvement

to

the

Multi-Source

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

153

Mobile-Streaming

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[8] .A. Nafaa, T. Taleb, and L. Murphy,

Servicesand Applications, vol. 1, no. 1, pp.

Forward

Error

718, Apr. 2010.

Strategies

for

Correction

over

[13] .F. Wang, J. Liu, and M. Chen,

IEEE

CALMS : Cloud-Assisted Live Media

Communications Magazine, vol. 46, no. 1,

Streaming for Globalized Demands with

pp. 7279, 2008. Cloud-Based Mobile

Time / Region

Video Streaming Techniques By Saurabh

Diversities, in IEEE INFOCOM, 2012.

Goel

[14] .H. T. Dinh, C. Lee, D. Niyato, and P.

[9] .J. Fernandez, T. Taleb, M. Guizani, and

Wang,

N. Kato, Bandwidth Aggregation-aware

Computing : Architecture , Applications ,

Dynamic QoS Negotiation for Real-Time

and Approaches,

VideoApplications in Next-

in

Generation Wireless Networks, in IEEE

Communications and Mobile Computing,

Transaction on Multimedia, vol. 11, no. 6,

Oct. 2011.

pp. 10821093, 2009.

[15] .S. Chetan, G. Kumar, K. Dinesh, K.

[10]. T. Taleb, K. Kashibuchi, A. Leonardi,

Mathew, and M. A. Abhimanyu, Cloud

S. Palazzo, K. Hashimoto, N. Kato, and Y.

Computing for Mobile World, Tech. Rep.,

Nemoto, A Cross-layer Approach for

2010.

AnEfficient Delivery of TCP/RTP-based

[16] .N. Davies, The Case for VM-Based

Multimedia Applications in Heterogeneous

Cloudlets in Mobile Computing, in IEEE

Wireless Networks, in IEEE Transaction on

Pervasive Computing, vol. 8, no. 4, pp. 14

Vehicular Technology, vol. 57, no. 6, pp.

23, 2009.

38013814, 2008.

[17] .B. Aggarwal, N. Spring, and A.

[11].Y. Li, Y. Zhang, and R. Yuan,

Schulman,

Measurement and Analysis of a Large

Mobile

Scale Commercial Mobile Internet TV

Support, in ACM SIGCOMM DEMO,

System, in ACM IMC,pp. 209224, 2011.

2010.

Wireless

Media

Adaptation

Streaming

Networks,in

Survey

Wiley

of

Journal

Stratus

Communication

Mobile

of

Cloud

Wireless

Energy-Efficient
using

Cloud

[12].Q. Zhang, L. Cheng, and R. Boutaba,


Cloud Computing: State-of-the-art and
Research Challenges, in Journal of Internet

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

154

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Identifying And Preventing Resource Depletion Attack In


Mobile Sensor Network
V.Sucharitha
Associate Professor
[email protected]
Audisankara college of engineering and technology

M.Swapna M.Tech
[email protected]

ABSTRACT:
Ad-hoc low-power wireless networks are inspiring research direction in sense
and enveloping computing. In previous security work in this area has focused primarily
on inconsistency of communication at the routing or medium access control levels. This
paper explores resource depletion attacks at the navigation protocol layer, which
permanent disable networks by quickly draining nodes battery power. The Vampire
attacks are not specific protocol, but rather rely on the properties of many popular classes
of routing protocols. We find that all examined protocols are vulnerable to Vampire
attacks, which are demolish and difficult to detect, and easy to carry out using as few as
one malicious insider send only protocol compliant messages.

near future, such as omnipresent ondemand computing power, continuous


connectivity, and instantly-deployable
communication for military and first
responders. Such networks already
monitor
environmental
conditions,
factory
performance,
and
troop
deployment, to name a few applications.
As WSNs become more and more
crucial to the everyday functioning of
people and organizations, availability
faults become less tolerable lack of
availability can make the difference
between business as usual and lost
productivity,
power
outages,
environmental disasters, and even lost
lives; thus high availability of these

1.INTRODUCTION:
the last couple of years wireless
communication has become of such
fundamental importance that a world
without is no longer imaginable for
many of using. Beyond the establish
technologies such as mobile phones and
WLAN, new approaches to wireless
communication are emerging; one of
them are so called ad hoc and sensor
networks. Ad hoc and sensor networks
are formed by autonomous nodes
communicating via radio without any
additional backbone infrastructure. Adhoc wireless sensor networks (WSNs)
promise exciting new applications in the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

155

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

networks is a critical property, and


should hold even under malicious
conditions. Due to their ad-hoc
organization, wireless ad-hoc networks
are particularly vulnerable to denial of
service (DoS) attacks, and a great deal of
research has been done to enhance
survivability.
While these schemes can prevent
attacks on the short-term availability of a
network, they do not address attacks that
affect long-term available the most
permanent denial of service attack is to
entirely deplete nodes batteries. This is
an instance of a resource depletion
attack, with battery power as the
resource of interest. this paper we
consider how routing protocols, even
those designed to be secure, lack
protection from these attacks, which we
call Vampire attacks, since they drain
the life from networks nodes. These
attacks are distinct from previouslystudied DoS, reduction of quality (RoQ),
and routing infrastructure attacks as they
do not disrupt immediate availability,
but rather work over time to entirely
disable a network. While some of the
individual attacks are simple, and powerdraining and resource exhaustion attacks
have been discussed before, prior work
has been mostly confined to other levels
of the protocol stack, e.g. medium access
control (MAC) or application layers, and
to our knowledge there is little
discussion, and no thorough analysis or
mitigation, of routing-layer resource
exhaustion attacks.
Vampire attacks are not protocolspecific, in that they do not rely on
design properties or implementation
faults of particular routing protocols, but
rather exploit general properties of
protocol classes such as link-state,
distance-vector, source routing and
geographic and beacon routing. Neither

ISBN: 378 - 26 - 138420 - 5

do these attacks rely on flooding the


network with large amounts of data, but
rather try to transmit as little data as
possible to achieve the largest energy
drain, preventing a rate limiting solution.
Since Vampires use protocol-compliant
messages, these attacks are very difficult
to detect and prevent.
This paper makes three primary
contributions. First, we thoroughly
evaluate the vulnerabilities of existing
protocols to routing layer battery
depletion attacks. We observe that
security measures to prevent Vampire
attacks are orthogonal to those used to
protect routing infrastructure, and so
existing secure routing protocols such as
Ariadne, SAODV, and SEAD do not
protect against Vampire attacks. Existing
work on secure routing attempts to
ensure that adversaries cannot cause path
discovery to return an invalid network
path, but Vampires do not disrupt or
alter discovered paths, instead using
existing valid network paths and
protocol compliant messages. Protocols
that maximize power efficiency are also
inappropriate, since they rely on
cooperative node behavior and cannot
optimize out malicious action. Second,
we show simulation results quantifying
the
performance
of
several
representative protocols in the presence
of a single Vampire (insider adversary).
Third, we modify an existing sensor
network routing protocol to provably
bound the damage from Vampire attacks
during packet forwarding.
1.1.Wireless Adhoc Network:
An ad hoc wireless network is a
collection of wireless mobile nodes that
self-configure to form a network without
the aid of any established infrastructure,
as shown in without an inherent
infrastructure, the mobiles handle the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

156

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

necessary control and networking tasks


by themselves, generally through the use
of distributed control algorithms.
Multihop
connections,
whereby
intermediate nodes send the packets
toward their final destination, are
supported to allow for efficient wireless
communication between parties that are
relatively far apart. Ad hoc wireless
networks are highly appealing for many
reasons. They can be rapidly deployed
and reconfigured. They can be tailored
to specific applications, as implied by
Oxfords definition. They are also highly
robust due to their distributed nature,
node redundancy, and the lack of single
points of failure.

ISBN: 378 - 26 - 138420 - 5

necessary to determine the time factor,


economy n company strength. Once
these things r satisfied, ten next steps are
to determine which operating system and
language can be used for developing the
tool.

Once

the

programmers

start

building the tool the programmers need


lot of external support. This support can
be obtained from senior programmers,
from book or from websites. Before
building

the

system

the

above

consideration r taken into account for


developing the proposed system.
A wireless sensor network (WSN)
consists

of

spatially

distributed

autonomous sensors to monitor physical


Fig:Adhoc Network Structure

or environmental conditions, such as

Existing work on secure routing


attempts to ensure that adversaries
cannot cause path discovery to return an
invalid network path, but Vampires do
not disrupt or alter discovered paths,
instead using existing valid network
paths and protocol compliant messages.
Protocols
that
maximize
power
efficiency are also inappropriate, since
they rely on cooperative node behavior
and cannot optimize out malicious
action.

temperature, sound, pressure, etc. and to

2.LITERATURE REVIEW:

industrial and consumer applications,

Literature survey is the most important

such as industrial process monitoring

step in software development process.

and control, machine health monitoring,

Before

and so on.

developing

the

tool

it

cooperatively pass their data through the


network to a main location. The more
modern networks are bi-directional, also
enabling control of sensor activity. The
development of wireless sensor networks
was motivated by military applications
such as battlefield surveillance; today
such networks are used in

is

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

157

many

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The WSN is built of "nodes"

ISBN: 378 - 26 - 138420 - 5

3.IMPLIMENTATION:

from a few to several hundreds or even


thousands, where each node is connected

As

to one (or sometimes several) sensors.

cooperatively build a Chord overlay

Each such sensor network node has

network over the sensor network. Cloned

typically

node

several

parts:

radio

prerequisite,

may

not

all

participate

nodes

in

this

transceiver with an internal antenna or

procedure, but it does not give them any

connection to an external antenna, a

advantage of avoiding detection. The

microcontroller, an electronic circuit for

construction of the overlay network is

interfacing with the sensors and an

independent of node clone detection. As

energy source, usually a battery or an

a result, nodes possess the information

embedded form of energy harvesting. A

of their direct predecessor and successor

sensor node might vary in size from that

in the Chord ring. In addition, each node

of a shoebox down to the size of a grain

caches information of its g consecutive

of dust, although functioning "motes" of

successors in its successors table. Many

genuine microscopic dimensions have

Chord systems utilize this kind of cache

yet to be created. The cost of sensor

mechanism to reduce the communication

nodes is similarly variable, ranging from

cost and enhance systems robustness.

a few to hundreds of dollars, depending

More importantly in our protocol, the

on the complexity of the individual

facility

sensor nodes. Size and cost constraints

contributes to the economical selection

on sensor nodes result in corresponding

of inspectors. One detection round

constraints on resources such as energy,

consists of three stages.

memory,

and

Stage 1: Initialization

The

To activate all nodes starting a new

topology of the WSNs can vary from a

round of node clone detection, the

simple star network to an advanced

initiator uses a broadcast authentication

multi-hop wireless mesh network. The

scheme to release an action message

propagation technique between the hops

including a monotonously increasing

of the network can be routing or

nonce, a random round seed, and an

flooding.

action time. The nonce is intended to

computational

communications

speed

bandwidth.

of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

158

the

successors

table

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

prevent adversaries from launching a

period, during which nodes randomly

DoS attack by repeating broadcasting

pick up a transmission time for every

action messages. The action message is

claiming message.

dened by

Stage 3: Processing claiming messages


A claiming message will be forwarded to
its destination node via several Chord
intermediate nodes. Only those nodes in

Stage 2: Claiming neighbors information

the overlay network layer (i.e., the

Upon receiving an action message, a

source node, Chord intermediate nodes,

node veries if the message nonce is

and the destination node) need to process

greater than last nonce and if the

a message,

message signature is valid. If both pass,

whereas other nodes along the path

the node updates the nonce and stores

simply route the message to temporary

the seed. At the designated action time,

targets. Algorithm 1 for handling a

the node operates as an observer that

message is the kernel of our DHT-based

generates a claiming message for each

detection protocol. If the algorithm

neighbor (examinee) and transmits the

returns NIL, then the message has

message through the overlay network


with

respect

to

the

arrived at its destination. Otherwise, the

claiming

message will be subsequently forwarded

probability .The claiming message by


observer for examinee is

to the next node with the ID that is

constructed

returned by Algorithm 1.

by

Criteria

of

determining

inspectors:

During handling a message in Algorithm


1, the node acts as an inspector if one of
Where

, are locations of

,respectively.

Nodes

can

and

the following conditions is satised.

start

transmitting claiming messages at the


same time, but then huge trafc may
cause serious interference and degrade
the network capacity. To relieve this
problem, we may specify a sending

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

159

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

performance. By Algorithm 1, roughly

4.ALGORITHMS:

of all claiming messages related to a


same examinees ID will pass through
one

of

the

predecessors

of

the

destination. Thus, those nodes are much


more likely to be able to detect a clone
than randomly selected inspectors. As a
result, this criterion to decide inspectors
can increase the average number of
witnesses at a little extra memory cost.
We will theoretically quantify those
performance measurements later. In
Algorithm 1, to examine a message for
node clone detection, an inspector will
invoke Algorithm 2, which compares the
message

with

previous

inspected

messages that are buffered in the cache


table. Naturally, all records in the cache
table should have different examinee
IDs, as implied in Algorithm 2. If
detecting a clone, which means that
there exist two messages
satisfying

and

and

, the

1) This node is the destination node of

witness

the claiming message.

evidence to notify the whole network.

2) The destination node is one of the g

All integrity nodes verify the evidence

node

successors of the node. In other words,

then

broadcasts

message

the destination node will be reached in

the

and

stop communicating with the cloned

the next Chord hop. While the rst

nodes. To prevent cloned nodes from

criterion is intuitive, the second one is

joining the network in the future, a

subtle and critical for the protocol

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

160

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

revocation list of compromised nodes

expense

IDs may be maintained by nodes

probability. The RDE protocol shares the

individually. It is worth noting that

major merit with broadcasting detection:

messages

are

Every node only needs to know and

and,

buffer a neighbor-list containing all

respectively. Therefore, the witness does

neighbors IDs and locations. For both

not need to sign the evidence message. If

detection

a malicious node tries to launch a DoS

constructs a claiming message with

attack by broadcasting a bogus evidence

signed version of its neighbor-list, and

message,

node

then tries to deliver the message to

receiving it can immediately detect the

others which will compare with its own

wicked

the

neighbor-list to detect clone. For a dense

before

network, broadcasting will drive all

and

authenticated

by

the

next

behavior

signatures

observers

integrity

by

of

verifying

and

with

ISBN: 378 - 26 - 138420 - 5

adequate

procedures,

detection

every

node

neighbors of cloned nodes to nd the

forwarding to other nodes.

attack, but in fact one witness that


The DHT-based detection protocol

successfully catches the clone and then

can be applied to general sensor

noties the entire network would sufce

networks, and its security level is

for the detection purpose. To achieve

remarkable, as cloned nodes will be

that in a communicatively efcient way,

caught by one deterministic witness plus


several

probabilistic

we bring several

witnesses.

Chord

overlap

network

protocol. First, a claiming message

incurs

needs to provide maximal h op limit, and

considerable communication cost, which

initially it is sent to a random neighbor.

may not be desired for some sensor

Then,

networks that are extremely sensitive to

(RDE),

and

presents

optimal

subsequent

helps a message go through the network

which

as fast as possible from a locally optimal

tremendously reduces communication


cost

message

line. The line transmission property

challenge, we propose the randomly


exploration

the

transmission will roughly maintain a

energy consumption. To fulll this

directed

and

effectively construct a multicast routing

However, the message transmission over


a

mechanisms

perspective. In addition, we introduce

storage

border

determination

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

161

mechanism to

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

signicantly reduce communication cost.


We can do all of those because every
node is aware of its neighbors locations,
which is a basic assumption for all

where is time to live (a.k.a. message

witness-based detection protocols but

maximum hop). Since tt1 will be altered

rarely utilized by other protocols.

by

intermediate

transmission,

nodes

it

should

during
not

be

authenticated. The observer willdeliver


the claiming message r times. In each
time, the node transmits it to a random
neighbor as indicated. Note that can be a
real

number,

and

accordingly

an

observer transmits its claiming message


at least[r] ,up to ,[r] and on average r
times. When an intermediate node
receives a claiming message

4.1 Protocol Description:

it

launches , which is described by


pseudo code in Algorithm 3, to

One round of clone detection is still

process

the

message.

During

the

activated by the initiator. Subsequently,


processing, node

at the designated action time, each node

compares its own neighbor-list to the

creates its own neighbor-list including

neighbor-list in the message, checking if

the neighbors IDs and locations, which

there is a clone. Similarly, if detecting a

constitutes the sole storage consumption

clone,

of the protocol. Then, it, as an observer

the

claiming message containing its own ID,


and

its

claiming

message

neighb-list.
by

node

the

witness

node

will

broadcast an evidence messageto notify

for all its neighbors, starts to generate a

location,

, as an inspector,

whole

that

The

network

such

the

cloned

nodes are expelled from the sensor

is

network. To deal with routing, node

constructed by

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

162

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

decreases the messages by 1 and

can be directly discarded. In our

discards the message if reaches zero;

proposal for border local determination,


another parameter

4.4target range :
This is used along with ideal direction to
determine a target zone. When no
neighbor is found in this zone, the
current node will conclude that the
message has reached a border, and thus
throw it away.

Essentially, Algorithm 4 contains the


following three mechanisms.

4.2Deterministicdirected
transmission:
When node receives a claiming message
from previous node, the ideal direction
can be calculated. In order to achieve the
best effect of line transmission, the next
destination node should be node , which
is closest to the ideal direction.
Fig:Loose source routing performance

4.3Networkborder

compared to optimal, in a network with

determination:
This

takes

diameter slightly above 10. The dashed


into

trend line represents expected path

the

length when nodes store logN local state,

communication cost. In many sensor

and the solid trend line shows actual

network applications, there exist outside

observed performance.

consideration

network
to

shape
reduce

borders of network due to physical

5.CONCLUSION:

constrains. When reaching some border

We defined Vampire attacks, a new class

in the network, the claiming message

of resource consumption attacks that use

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

163

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

routing protocols to permanently disable

[5] J. Bellardo and S. Savage, 802.11

ad-hoc wireless sensor networks by

Denial-of-Service

depleting nodes battery power. These

Vulnerabilities and Practical Solutions,

attacks do not depend on particular

Proc. 12th Conf. USENIX Security,

protocols or implementations, but rather

2003.

expose vulnerabilities in a number of

[6] D. Bernstein and P. Schwabe, New

popular protocol classes. We showed a

AES Software Speed Records, Proc.

number

of proof-of-concept

Ninth Intl Conf. Cryptology in India:

against

representative

attacks

examples

of

Attacks:

Real

Progress in Cryptology (INDOCRYPT),

existing routing protocols using a small

2008.

number

[7] D.J. Bernstein, Syn Cookies,

of

weak

adversaries,

and

measured their attack success on a

http://cr.yp.to/syncookies.html, 1996.

randomly-generated topology of 30

[8] I.F. Blaked, G. Seroussi, and N.P.

nodes.

Smart, Elliptic Curves in cryptography,


vol. 265. Cambridge Univ. , 1999.

REFERENCES:

[9] J.W. Bos, D.A. Osvik, and D. Stefan,


Fast

[1] The Network Simulator - ns-2,

Various Platforms, Cryptology ePrint

http://www.isi.edu/nsnam/ns,2012.

Archive,

[2] I. Aad, J.-P. Hubaux, and E.W.

501,

and Privacy in Sensor Networks,

MobiCom, 2004.

Computer, vol. 36, no. 10, pp. 103-105,

[3] G. Acs, L. Buttyan, and I. Vajda,

Oct. 2003.

Provably Secure On-Demand Source

[11] J.-H. Chang and L. Tassiulas,

Routing in Mobile Ad Hoc Networks,

Maximum

IEEE Trans. Mobile Computing, vol. 5,

Lifetime

Routing

in

Wireless Sensor Networks, IEEE/ACM

no. 11, pp. 1533-1546, Nov. 2006.


Aura,

2009/

[10] H. Chan and A. Perrig, Security

in Ad Hoc Networks, Proc. ACM

T.

Report

http://eprint.iacr.org, 2009.

Knightly, Denial of Service Resilience

[4]

Implementations of AES on

Trans. Networking, vol. 12, no. 4, pp.

Dos-Resistant

609-619, Aug. 2004.

Authentication with Client Puzzles,


Proc. Intl Workshop Security Protocols,
2001.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

164

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[12] T.H. Clausen and P. Jacquet,


Optimized

Link

State

Routing

Protocol(OLSR), IETF RFC 3626, 2003.


[13] J. Deng, R. Han, and S. Mishra,
Defending against Path-Based DoS
Attacks in Wireless Sensor Networks,
Proc. ACM Workshop Security of Ad
Hoc and Sensor Networks, 2005.
[14] J. Deng, R. Han, and S. Mishra,
INSENS: Intrusion-Tolerant Routing
for

Wireless

Sensor

Networks,

Computer Comm., vol. 29, 1. 2, pp. 216230, 2006.


[15] S. Doshi, S. Bhandare, and T.X.
Brown, An On-Demand Minimum
Energy Routing Protocol for a Wireless
Ad Hoc Network, ACM SIGMOBILE
Mobile Computing and Comm. Rev.,
vol. 6, no. 3, pp. 50-66, 2002.
[16] J.R. Douceur, The Sybil Attack,
Proc.

Intl

Workshop

Peer-to-Peer

Systems, 2002.
[17] H. Eberle, A. Wander, N. Gura, C.S.

Sheueling,

and

V.

Gupta,

Architectural Extensions for Elliptic


Curve Cryptography over GF(2m) on 8bit Microprocessors, Proc. IEEE Intl
Conf Application- Specific Systems,
Architecture Processors (ASAP), 2005.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

165

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

EFFECTIVE FAULT TOERANT RESOURCE ALLOCATION WITH COST


REDUCTION FOR CLOUD
Dorababu Sudarsa M.Tech.,Ph.D,MISTE
Associate Professor
[email protected]
Audisankara college of engineering and technology

AllareddyAmulya M.Tech
[email protected]

Abstract:
In Cloud systems Virtual Machine technology being increasingly grown-up, compute resources which
can be partitioned in fine granularity and allocated them on require. In this paper we formulate a
deadline-driven resource allocation problem based on the Cloud environment that provides VM
resource isolation technology, and also propose an optimal solution with polynomial time, which
minimizes users payment in terms of their expected deadlines. We propose an fault-tolerant method
to guarantee tasks completion within its deadline. And then we validate its effectiveness over a real
VM-facilitated cluster environment under different levels of competition. To maximize utilization and
minimize total cost of the cloud computing infrastructure and running applications, efficient resources
need to be managed properly and virtual machines shall allocate proper host nodes . In this work, we
propose performance analysis based on resource allocation scheme for the efficient allocation of
virtual machines on the cloud infrastructure. Our experimental results shows that our work more
efficient for scheduling and allocation and improving the resource utilization.

Key words: fault torenant,resource allocation,cloud computing, cost reduction.


provides computation, software, data access,
and storage resources without requiring
1. INTRODUCTION:
Cloud Computing[1] is a model for enabling
cloud users to know the location and other
convenient, on-demand network access to a
details of the computing infrastructure.
shared pool of configurable and reliable
Cloud computing is transforming business
computing resources (e.g., networks,
by offering new options for businesses to
servers, storage, applications, services) that
increase efficiencies while reducing costs.
can be rapidly provisioned and released with
These problems include:
minimal consumer management effort or
a. High operational costs: typically
service
provider
interaction.
Cloud
associated with implementing and managing
computing is the delivery of computing as a
desktop and server infrastructures
service rather than a product, whereby
b. Low system utilization: often associated
shared resources, software, and information
with non-virtualized server workloads in
are provided to computers and other devices
enterprise environments
as a metered service over a network
(typically the Internet). Cloud computing

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

166

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

c. Inconsistent availability: due to the high


cost of providing hardware redundancy.
d. Poor agility: which makes it difficult for
businesses to meet evolving market
demands.
The reallocation in cloud computing is more
complex than in other distributed systems
like Grid computing platform. In a Grid
system [2], it is inappropriate to share the
compute resources among the multiple
applications at the same time running atop it
due to the unavoidable mutual performance
involvement among them. Whereas, cloud
systems usually do not providing physical
hosts directly to users, but leverage virtual
resources isolated by VM technology [3],
[4], [5]. Not only can such an elastic
resource usage way adapt to users specific
demand, but it can also maximize resource
utilization in fine granularity and isolate the
abnormal environments for safety purpose.
Some successful platforms or cloud
management tools leveraging VM resource
isolation technology include Amazon EC2
[6] and OpenNebula [7]. On the other hand,
with fast development of scientific research,
users may propose quite complicated
demands. For example, users may want to
minimize their payments when confirm their
service level such that their tasks can be
finished before deadlines. Such a deadline
ensure the reallocation with minimized
payment is rarely studied in literatures.
Moreover, inavoidable errors with an
anticipate the task workloads will definitely
make the problem harder. Based on the
elastic resource usage model, we aim to
design a reallocation algorithm with high
anticipate- error tolerance ability, also
minimizing users payments subject to their
expected deadlines. Since the idle physical
resources can be arbitrarily divide and
allocated to new tasks, the VM-based
divisible resource allocation could be very
flexible. This implies the feasibility of
finding the optimal solution through convex

ISBN: 378 - 26 - 138420 - 5

optimization strategies [8], unlike the


traditional Grid model that relies on the
indivisible resources like the number of
physical cores. However, we found that it is
in avoidable to directly solve the necessary
and sufficient condition to find the optimal
solution, a.k.a., Karush-Kuhn-Tucker (KKT)
conditions [9]. Our first contribution is
devising a new approach to solve the
problem.

2. RELATED WORKS:
A Static resource allocation based on peak
demand is not cost-effective because of poor
resource utilization during off-peak periods..
Resource provisioning for cloud computing,
an important issue is how resources may be
allocated to an application mix such that the
service level agreements (SLAs) of all
applications are met Heuristic algorithm that
determines a resource allocation strategy
(SA or DA) that results in the smallest
number of servers required to meet the SLA
of both classes; Comparative evaluation of
FCFS, head-of-the-line priority (HOL) and a
new scheduling discipline called probability
dependent priority (PDP). Scott et al[10]
proposed a finding the failure rate of a
system is a crucial step in high performance
computing systems analysis. Fault tolerant
mechanism, called checkpoint/ restart
technique, was introduced. Incremental
checkpoint model can reduce the waste time
more than it is reduced by the full
checkpoint model. Singh et al. presented a
slot-based provisioning model on grids to
provide scheduling according to the
availability and cost of resources.
2.1.Cloud Environment Infrastructure
Architecture:
Cloud users combine virtualization,
automated
software,
and
internet
connectivity [11] to provide their services. A
basic element of the cloud environment is
client, server, and network connectivity [13].

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

167

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

A hybrid computing model allows customer


to leverage both public and private
computing services to create a more flexible
and cost-effective computing utility. The
public cloud environment involves Web
based application, Data as a service (DaaS),
Infrastructure as a Service (IaaS), Software
as a service (SaaS), and Email as a service
(EaaS). A private cloud accesses the
resources from the public cloud organization
to provide services to its customers. In a
hybrid cloud environment, an organization
combines various services and data model
from various cloud environments to create
an automated cloud computing environment.

ISBN: 378 - 26 - 138420 - 5

taking rent of hardware resources based on


pay as you go basics. This process is
referred to as resource gathering. Resource
gathering by the hypervisor makes
virtualization possible, and virtualization
makes multiprocessing computing that leads
to an infrastructure shared by several users
with similar resources in regard to their
requirements.
2.3. Task Scheduling and Resource
Allocation :
To increase the flexibility, cloud allocates
the resources according to their demands.
Major problems in task scheduling
environment are load balancing, scalability,
reliability, performance, and re-allocation of
resources to the computing nodes
dynamically. In past days, there are various
methods and algorithms to solve the
problem of scheduling a resource in Preempt
able Job in cloud environment. In cloud
environment, resources are allocated to the
customers under the basics of pay per use on
demand. Algorithms used in the allocation
of the resources in cloud computing
environment differ according to schedule of
task in different environment under different
circumstances. Dynamic load balancing in
cloud allocates resource to computational
node
dynamical.
Task
Scheduling
algorithms aim at minimizing the execution
of tasks with maximizing resource usage
efficiently. Rescheduling is need only when
the customers request the same type of
resources. Each and every task is different
and autonomous their requirement of more
bandwidth,
response
time,
resource
expenses, and memory storage also differs.
Efficient scheduling algorithms maintain
load balancing of task in efficient manner.
Efficiency of cloud environment only
depends on the type of scheduling algorithm
used for task scheduling.

Fig 2.1: Cloud Environment Infrastructure


Architecture
2.2. Infrastructure as a service (IaaS) :
Infrastructure as a service (IaaS) controls
user and manage the systems. However, for
business IaaS takes an advantage in its
capacity. IT companies able to develop its
own software and implements that can able
to handles the ability to re-schedule
resources in an IaaS cloud. IaaS consists of a
combination of internal and external
resources. IaaS is low-level resource that
runs independent of an operating system
called a hypervisor and is responsible for

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

168

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

fewer then m jobs are active, they are


required to execute on the fastest processor
while the slowest are idled, and higher
priority jobs are executed on faster
processors. A formal verification which
guarantees all deadlines in a real-time
system would be the best. Then this
verification is called feasibility test.
Three different kinds of tests are available:1.Exact tests with long execution times or
simple models [11], [12], [13].
2. Fast sufficient tests which fail to accept
feasible task sets, especially those with high
utilizations [14], [15].
3. Approximations, which are allowing an
adjustment of performance and acceptance
rate [1],
Task migration cost might be very high. For
example, in loosely coupled system such as
cluster of workstation a migration is
performed so slowly that the overload
resulting from excessive migration may
prove unacceptable [3]. In this paper we are
presenting the new approach call the queue
set algorithm is used to reduce the efficent
time complexity.

3.IMPLEMENTATION
By using queue set scheduling for
scheduling the task we can obtain the high
task completion with in schedule. Whenever
the queue set scheduling event occurs the
task queue is searched for the process
closest to its deadline and is scheduled for
its execution.
In queue set scheduling , at every scheduling
point the task having the shortest deadline is
taken up for scheduling. The basic principle
of this algorithm is very sensitive and simple
to understand. If a new process arrives with
cpu burst time less than remaining time of
current executing process. Queue set
satisfies the condition that total processor
utilization (Ui) due to the task set is less
than 1. With scheduling periodic processes
that have deadlines equal to their periods,
queue set has a utilization bound of 100%.
For example let us Consider 3 periodic
processes scheduled using queue set
alogorithm, the following acceptance test
shows that all deadlines will be met.
Q2
Table1:Task Parameter
Process
P1
P2
P3

Execution Time=C
3
2
1

ISBN: 378 - 26 - 138420 - 5

Period=T
4
5
7

4.QUEUE
SET
SCHEDULING
ALGORITHM:
Let n denote the number of processing
nodes and m denote the number of Available
tasks in a uniform parallel real- time system.
C denotes the capacity vectore and D
denotes the deadline. In this section we are
presenting five steps of queue set scheduling
alogorithm.
obviously, each task which is picked for up
execution is not considered for execution by
other processors. Here we are giving
following methods for our new approach:
1. Perform a possible to check a specify the
task which has a chance to meet their
deadline and put them into a queue(2
) , Put the remaining tasks are also allocated
and assign that particular queue. We can

The utilization will be:


3/4+2/3+1/7=1.559=55.9%
The theoretical limit for any number of
processes is 100% and so the system is
schedulable. The queue set algorithm
chooses for execution at each instant in the
time currently active job(s) that have the
nearest
deadlines.
The
queue
set
implementation upon uniform parallel
machines is according to the following rules
[2], No Processor is idle while there are
active jobs waiting for execution, when

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

169

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

partition the task set by any existing


approach.
2. Sort the task queue set scheduling
according to their deadline in a nondescending order by using any of existing
sorting algorithms. Let k denote the
number of tasks in allocated in queue , i.e.
the number of tasks that have the
opportunity to meet their deadline.

ISBN: 378 - 26 - 138420 - 5

N1(corevalue0.2)+(cpuvalue0.5)+(me
mvalue 0.3);
DB.add(Ni );
end for

Node performance analysis algorithm


for each i N.size()
if !available(Ni , requestResource)
availableNodeList.add(Ni );
end if
end for Sort (availableNodeList );
while j availableNodeList.size()
if VM = empty creatVM(VM);
endifsuccess(NjVM)
vmtable.add(j,VM);
end if j++;
end while

3. For all processor j, (jmin(k,m)) check


whether a task which was last running on the jth
processor is among the first min(k,m) tasks of
set 1. If so assign it to the jth processor. At this
point there might be some processors to which
no task has been assigned yet.
4. For all j, (jmin(k,m)) if no task is assigned to
the jth processor , select the task with earliest
deadline from remaining tasks of set 1 and
assign it to the jth processor. If km, each
processor have a task to process and the
algorithm is finished.

Virtual machine scheduling algorithm After


getting the proper host, the scheduler will
return the host number to the virtual
machine manager for placement of virtual
machine on that host. Then the virtual
machine manager has all information about
the virtual machine and its location. It will
send a service activation message to the
client/user. After that, the client/user can
access the service for the duration specified.
And when the resources and the data are
ready, this tasks execution being.

5. If k<m, for all j, (k<jm) assign the task with


smallest deadline from B to the jth processor.
The last step is optional and all the tasks from
next set will miss their deadlines.

5. Resource allocation algorithm:


6. CONCLUSION:

Resource allocation is the process of


assigning available resources to the needed
cloud applications. Cloud resources consist
of physical and virtual resources. The user
request for virtualized resources is described
through a set of parameters detailing the
processing CPU, memory, disk, and so on.
For each I Node(Core,CPU,Mem)
StarttimeTimes();.
MemvalueInvertMatrix(Ni);
FinishtimeTimes();
CPUvalueFinishtimeStarttime;

Cloud Computing is a promising


technology to support IT organizations
in developing cost, time and resource
efective
products.
Since,
Cloud
computing is a pay-go-model, it is
necessary to reduce cost at the peak
hors inoredr to improve the business
performance of the cloud system. The
cost will be reduced and efficient
resource utilization also possible. to
have an effective error tolerant
approach that can efficiently allocate the
resources to reach the deadline.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

170

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

scheduling under time and resource


constraints, In: IEEE Transactions
on Computers C-36 (8) (1987)949
960.
10. Alex King Yeung Cheung and
Hans-Arno
Jacobsen, Green
Resource Allocation Algorithms for
Publish/Subscribe Systems, In: the
31th IEEE International Conference
on Distributed Computing Systems
(ICDCS), 2011.
11. Mrs.S.Selvarani1;
Dr.G.Sudha
Sadhasivam, Improved Cost Based
Algorithm
For
Task
Scheduling In Cloud Computing,
IEEE 2010.
12. S. Mohana Priya, B. Subramani, A
New Approach For Load Balancing
In
Cloud
Computing,
In:
International Journal Of Engineering
And Computer Science (IJECS2013) Vol 2, PP 1636-1640(2013).

7. REFERENCES:
1. Amazon Elastic Compute Cloud,
http://aws.amazon.com/ec2/, 2012.
2. D. Milojicic, I.M. Llorente, and R.S.
Montero, Opennebula: A Cloud
Management Tool, IEEE Internet
Computing, vol. 15, no. 2, pp. 11-14,
Mar./Apr. 2011.
3. S. Boyd and L. Vandenberghe,
Convex Optimization. Cambridge
Univ. Press, 2009.
4. E. Imamagic, B. Radic, and D.
Dobrenic, An Approach to Grid
Scheduling by Using Condor-G
Matchmaking Mechanism, Proc.
28th Intl Conf.
Information
Technology Interfaces, pp. 625-632,
2006.
5. Naksinehaboon N, Paun M, Nassar
R, Leangsuksun B, Scott S (2009)
High
performance
computing
systems with various checkpointing
schemes.
6. Ratan Mishra and Anant Jaiswal,
Ant colony Optimization: A
Solution of Load balancing in
Cloud, in: International Journal of
Web & Semantic Technology
(IJWesT-2012) Vol 3, PP 33-50
(2012). DOI: 15121/ijwest.2012.32
7. Chandrashekhar S. Pawar and
R.B.Wagh, A review of resource
allocation
policies
in
cloud
computing, IN: World Journal of
Science and Technology (WJST)
Vol 3, PP 165-167 (2012).
8. K C Gouda, Radhika T V, Akshatha
M,
"Priority
based
resource
allocation
model
for
cloud
computing", Volume 2, Issue 1,
January 2013, International Journal
of Science,
Engineering and
Technology Research (IJSETR).
9. W. Zhao, K.Ramamritham, and
J.A.Stankovic,
Preemptive

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

171

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

SCALABLE AND SECURE SHARING OF PERSONAL HEALTH


RECORDS IN CLOUD USING MULTI AUTHORITY
ATTRIBUTE BASED ENCRYPTION
PINISETTY CHANDRA SARIKA1
1

P.VENKATESWAR RAO2

PG Student, M.Tech, Computer Science and Engineering,

Audisankara College of Engineering & Technology, Gudur.A.P, India


[email protected]
2

Associate Professor, Department of CSE,

Audisankara College of Engineering & Technology, Gudur.A.P, India


[email protected]

ABSTRACT
This project is to design and implement personal health record (PHR) and provide security and
stored in cloud provider. This application allows user to access their lifetime health information, it
maintained in centralized server to maintain the patients personal and diagnosis information. The
person health records should maintain with high privacy and security. The patient data is protected
from the public access through these security techniques. To convince the patients access over to
their own PHRs, before outsourcing we have to encrypt the PHRs. To protect the personal health
data stored in a partly-trusted server, we have to adopt attribute-based encryption (ABE) as a main
encryption primitive. To obtain fine-grained and scalable information access control for PHRs and
to encrypt each patients PHR file we purchase attribute based encryption (ABE) techniques. In this
project we propose a new patient centric Architecture and access mechanism to control PHRs in
multitrusted servers. We get fine grained and scalable control of data access for personal health
records stored in multi-trusted servers, and attributes based encryption (ABE) technique to encrypt
the each patients medical record into a file. In this paper we are extend Multi Authority Attribute
Based Encryption (MA-ABE) for access control mechanism.
Keywords: Personal Health Records, Cloud Computing, Data Privacy, Fine-grained access
control, Multi Authority Attribute Based Encryption.

1. INTRODUCTION
In Recent years, personal health record is
maintained as a patient centric design of health
message exchange. It allows a patient to create
and control their medical data and it can be
maintained in a single place such as data
centres. High cost of building and managing
stream of data centres and many of PHR
services are outsourced to third party service
providers, for example Google Health,

Microsoft Health Vault. The PHR which is


stored in cloud computing have been proposed
in [1], [2]. When it is exciting to have
convenient PHR data passing for each one,
there are number of privacy and security risks
its a wide adoption. In the third party service
provider there is no security and privacy risk
for PHR. The maximum values of sensitive
Personal Health Information (PHI) the
unauthorized person storage service are often

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

172

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

to target the various malicious behaviours


which lead to exposure to the PHI.
The main concern is about privacy of
patients, personal health data and to find
which user could gain access to the medical
records stored in a cloud server. The famous
incident, department of Veterans Affairs
containing sensitive database PHI of 26.5
encryption pattern namely Attribute Based
Encryption (ABE). In this ABE it attributes
the users data to selects the access policies. It
allows a patient to share their own PHR
among a group of users by encrypting the file
under a set of attributes, need to know
complete information of users. The scope of
result the number of attributes is to determine
the complexities in the encryption technique,
security key generation, and decryption. The
Multi Authority Attribute Based Encryption
scheme is used to provide the multiple
authority based access mechanism. The aim
of this Patient Centric Privacy often disagrees
with the scalability and reliability in PHR
system. Only the authorized users can access
the PHR system for personal use or
professional purpose. We are referring about
two categories as personal and professional
users respectively.

ISBN: 378 - 26 - 138420 - 5

million military veterans, including their


health problems and social security numbers
was theft by an employee who take the data
home without authorization [13]. We ensure
the privacy control over their own PHRs, it is
essential to have fine-grained data access
control mechanisms that work with may
trusted servers. Then we can skip to a new
in the system. Not only may create a load
bottleneck but also it suffers from the security
key problems. The attribute management
tasks also include certification of all users
attributes and generating secret keys.

B. Revocable
Encryption:

Attributed

based

This scheme is a well-known challenging


problem to invoke users or attributes
efficiently and on-demand in ABE. Mainly
this technique is done by the authority
broadcasting periodic key updates to
unrevoked users frequently, and which does
not get complete forward/backward security
and efficiency is less.
In this paper for uniform security, we
are proposing framework of patients centric
sharing of PHR in several domain, several
authority PHR system with several users. It
captures framework application-level of
requirements for both public and personal use
of patients PHRs and it distributed users trust
on multiple authorities are better reflects
actuality.

2. RELATED WORK
This article is more related to operate in
cryptographically enforced control access for
outsourced data and attribute based on
encryption data. To improve the scalability of
the above result, one-to-many encryption
methods such as ABE can be utilized. The
basic property of ABE is preventing against
the user collusion.

3. IMPLEMENTATION
a. Requirements:
The most important task is to achieve patientcentric PHR sharing. That means, the patient
should contain the fundamental control over
their own health record. It also determines
which users should have access to their
medical data. The user control write/read
access and revocation are two main security
purposes for any type of electronic health
record system. The write access control is
controlled by the person to prevent in PHR

A.Trusted Authority:
Multiple operations used ABE to realize fine
grained access outsourced data control. Each
patient EHR files are encrypted by using a
broadcast variant of CP-ABE that allows
directly. Here several communication
drawbacks of the above mentioned
operations. Mainly they are usually
pretending the use of a single trusted authority

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

173

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

context entitles by the unauthorized users to


get access on the record and to modifying it.

ISBN: 378 - 26 - 138420 - 5

The below Fig.1 depicts the architecture of


proposed system for secure sharing of the
patient medical records. The system split the
users into two security domains namely,
public domains (PUDs) and personal domains
(PSDs) according to the different users data
access needs. The PUDs consist of users who
grant access based on their professional roles,
such as doctors, nurses, medical researchers
and insurance agents. For each PSD, the users
are personally associated with a data owner
(such as family members or close friends),
and they grant accesses to personal health
records based on access rights assigned by the
owner. Here we consider Data owner who
own the personal health record, data reader as
who can read the encrypted patient medical
record. In PSD, the owner uses key-policy
attributed based on encryption and generates
secret key for their PSD users and in PUD the
owner used multi-authority attribute based
encryption is preferred. Secret Key for PUD
users are produced by Multiple authority (For
this project we consider Specialization
Authority and Patient Medical Authority)
depending
on
both
profession
and
specialization.

b. Framework:
The purpose of our framework is to provide
security of patient-centric PHR access and
efficient key management at the same time. If
users attribute is not valid, then the user is
unable to access future PHR files using that
attribute. The PHR data should support users
from the personal domain as well as public
domain. The public domain may have more
number of users who may be in huge amount
and unpredictable, system should be highly
scalable in terms of complexity in key
management
system
communication,
computation and storage.
The owners
endeavour in managing users and keys should
be minimized to enjoy usability. By using
attribute based encryption we can encrypt
personal health records self-protective that is
they can access only authorized users even on
a semi trusted server.

Fig.1: Framework for PHR sharing


Fig.2: Architecture
record sharing

c. Architecture of implementation:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

174

of

patients

medical

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

authoritys domain and output secret key for


the user.

4. TECHNIQUES
a. Attribute Based Encryption

Central Key Generation: A central authority


can be used be run by a random algorithm. It
takes the master key as an input and a users
GID and outputs secret key for user.

The database security is provided by using


Attribute Based Encryption techniques. In this
the sensitive information is shared and stored
in the cloud provider; it is needed to encrypt
cipher text which is classified by set of
attributes. The private key is associated with
access make to control with cipher text a user
is able to decrypt. Here we are using Attribute
Based Encryption (ABE) as the principal
encryption primitive. By using ABE access
policies are declared based on the attributes of
user data, which make possible to selectively
share her/his PHR among a set of users to
encrypting the file under a set of attributes,
without a need of complete users. The
complexity per encryption, security key
generation, and decryption are only linear
with multi number of attributes are included.
When we integrate ABE into a large scale of
PHR system, the important dispute such as
dynamic policy updates, key management and
scalability, and an efficient on demand
revocation are non-retrieval to solve.

Encryption: This technique can be run by a


sender. Take a set of attributes as an input for
each authority, and the system public key.
The outputs are in the form of cipher text.
Decryption: This mechanism can be done by
a receiver. Takes input as a cipher text, which
was encrypted under a set of decryption keys
for attribute set.
By using this ABE and MA-ABE it
will increase the system scalability, there are
some restriction in building PHR system. The
ABE does not handle it efficiently. In that
scenario one may regard with the help of
attributes based broadcast encryption.

5. SECURITY MODEL OF THE


PROPOSED SYSTEM
i. Data confidentiality:

b. Multi-Authority ABE

This research plan reveals the data about each


user to access on the PHR among one another.
The different sets of documents are authorized
by the users to read the document.

A Multi-Authority ABE system is included


with k attribute authorities and one central
control. The value dk is assigned to every
attribute authority. In this proposed system
we can use the following algorithms:

ii.
User
Access
Confidentiality:

The random algorithm is run by the


central authority or some other trusted
security. It takes input as a security parameter
and outputs as a public key and secret key
pair for each of the attribute authorities and
also outputs as a system public key and
master secret key, which is used for central
authority.

Privilege

The system does not disclose the rights from


one person to another. This ensures the user to
access strong confidentiality. And also it
maintains both public domain and private
domain. Secure Sharing of Personal Health
Records System designer maintain Personal
Health Records with various user access
points. These data values are managed under a
third party cloud provider system. The cloud
provider will provide security for the data.
Multiple modules can be provided by this
system.

Attribute Key Generation: A random


algorithm is run by an attribute authority. The
secret key is to take as an input for security
authority and the authoritys value dk, a
users GID, and a set of attributes in the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

175

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Data owner is designed to manage the


patient details.
With multiple attribute
collections the PHR is maintained. Access
permission to different authorities can be
assigned by data header.

ISBN: 378 - 26 - 138420 - 5

[3] M. Li, S. Yu, K. Ren, and W. Lou,


Securing Personal Health Records in Cloud
Computing: Patient-Centric and Fine-Grained
Data Access Control in Multi-Owner
Settings,
Proc.
Sixth
Intl
ICSTConf.SecurityandPrivacyinComm.Netw
orks (SecureComm10), pp. 89-106, Sept.
2010.

Cloud provider module is used to store the


PHR values. The encrypted PHR is uploaded
by the data header to the cloud provider.
Patients can access data and also maintained
under the cloud provider.

[4] M. Chase and S.S. Chow, Improving


Privacy and Security in Multi-Authority
Attribute-Based Encryption, Proc. 16th
ACM Conf. Computer and Comm. Security
(CCS 09), pp. 121-130, 2009.

Key management is one of the main tasks


to plan and control key values for various
authorities. The owner of the data will update
the key values. This dynamic policy is based
on key management scheme.

[5] M. Li, S. Yu, N. Cao, and W. Lou,


Authorized Private Keyword Search over
Encrypted Personal Health Records in Cloud
Computing, Proc. 31st Intl Conf.
Distributed Computing Systems (ICDCS 11),
June 2011.

Patients are accessed by the client module.


This system uses the personal and
professional
access
pattern.
Access
classification is used to provide multiple
attributes. Clients access to log maintains to
the user request information to process
auditing.

[6] J. Benaloh, M. Chase, E. Horvitz, and K.


Lauter, Patient Controlled Encryption:
Ensuring Privacy of Electronic Medical
Records, Proc. ACM Workshop Cloud
Computing Security (CCSW 09), pp. 103114, 2009.

6. CONCLUSION
This PHR system fights against the security
attackers and hackers. The secure data sharing
is used to protect the information from
unauthorized user. We have proposed a novel
approach for existing PHR system providing
high security using Attribute Based
Encryption which plays main role, because
these are the unique competition, and it is
difficult to hack. The ABE model increases
and operates with MAABE.

[7] S. Yu, C. Wang, K. Ren, and W. Lou,


Achieving Secure, Scalable, and FineGrained Data Access Control in Cloud
Computing, Proc. IEEE INFOCOM 10,
2010.

7. REFERENCES

[8] V. Goyal, O. Pandey, A. Sahai, and B.


Waters, Attribute-Based Encryption for
Fine-Grained Access Control of Encrypted
Data, Proc. 13th ACM Conf. Computer and
Comm. Security (CCS 06), pp. 89-98, 2006.

[1] H. Lo hr, A.-R. Sadeghi, and M.


Winandy, Securing the E-Health Cloud,
Proc. First ACM Intl Health Informatics
Symp. (IHI 10), pp. 220-229, 2010.

[9] S. Narayan, M. Gagne, and R. SafaviNaini, Privacy preserving EHR system using
attribute-based infrastructure, ser. CCSW
10, 2010, pp. 4752.

[2] M. Li, S. Yu, N. Cao, and W. Lou,


Authorized Private Keyword Search over
Encrypted Personal Health Records in Cloud
Computing, Proc. 31st Intl Conf.
Distributed Computing Systems (ICDCS 11),
June 2011.

[10] J. Hur and D.K. Noh, Attribute-Based


Access Control with Efficient Revocation in
Data Outsourcing Systems, IEEE Trans.
Parallel and Distributed Systems, vol. 22, no.
7, pp. 1214-1221, July 2011.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

176

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[11] S. Jahid, P. Mittal, and N. Borisov,


Easier: Encryption-Based Access Control in
Social Networks with Efficient Revocation,
Proc. ACM Symp. Information, Computer
and Comm. Security (ASIACCS), Mar. 2011.
[12] S. Ruj, A. Nayak, and I. Stojmenovic,
DACC: Distributed Access Control in
Clouds, Proc. IEEE 10th Intl Conf. Trust,
Security and Privacy in Computing and
Comm. (TrustCom), 2011.
[13] At Risk of Exposure - in the Push for
Electronic Medical Records, Concern Is
Growing About How Well Privacy Can Be
Safeguarded,
http://articles.latimes.com/2006/jun/26/health/
he-privacy26, 2006.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

177

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

LATENT FINGERPRINT RECOGNITION AND MATCHING


USING STATISTICAL TEXTURE ANALYSIS
M.CHARAN KUMAR1, K.PHALGUNA RAO 2,
M.Tech 2nd year, Dept. of CSE, ASIT, Gudur, India
Professor, Dept. of CSE, ASIT, Gudur, India
[email protected]; 2 [email protected];

insisted on: distortion, dry, and wet ngerprints.

Abstract: Latents are the partial fingerprints that are


usually smudgy, with small area and containing large
distortion. Due to this characteristic, latents have a
significantly smaller number of minutiae points
compared to full (rolled or plain) fingerprints. The
small number of minutiae and noise of latents make
it extremely difficult to automatically match latents
their mated full prints that are stored in law
enforcement databases, although a number of
methods used for fingerprint recognition to extract
accurate results but is not up to level. The proposed
fingerprint recognition and matching using statistical
analysis gives efficient scheme of fingerprint
recognition for biometric identification of
individuals. Three statistical features are extracted to
represent in mathematical model. They are (1) an
entropy coefficient, for intensity histogram of the
image, (2) a correlation coefficient, for operation
between the original and filter image by using wiener
filter, and (3) an energy coefficient, obtaining image
in 5-level wavelet decomposition obtained after 5th
level decomposition. The approach can be easily used
to provide accurate recognition results.

Distortion of ngerprints seriously affects the


accuracy of matching. There are two main reasons
contributed to the ngerprint distortion. First, the
acquisition

of

ngerprint

is

three-

dimensional/two-dimensional warping process. The


ngerprint captured with different contact centers
usually results in different warping mode. Second,
distortion will be introduced to ngerprint by the no
orthogonal pressure people exert on the sensor. How
to cope with these nonlinear distortions in the
matching process is a challenging task. Several
ngerprint matching approaches have been proposed
in the literature. These include methods based on
point pattern matching, transform features and
structural matching. Many ngerprint recognition
algorithms are based on minutiae matching since it is
widely

believed that

the minutiae are most

discriminating and reliable features. Rather al.


Index Terms: - fingerprint recognition, entropy,

addressed a method based on point pattern matching.

correlation, wavelet energy.

The generalized Hough transform (GHT) is used to


recover the pose transformation between two

1. INTRODUCTION
SIGNIFICANT

impressions. Jain et al. proposed a novel later bank

improvements

recognition have been

in

ngerprint

based ngerprint feature representation method.

achieved in terms of

Jingled.

algorithms, but there are still many challenging tasks.

Addressed a method which relies on a

similarity measure dened between local structural

One of them is matching of nonlinear distorted

features, to align the two pat- terns and calculate a

nger-prints. According to Fingerprint Verication

matching score between two minutiae lists. Fantail.

Competition 2004 (FVC2004), they are particularly

Applied a set of geometric masks to record part of the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

178

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

rich information of the ridge structure. What et al.

not always well dened and, therefore, cannot be

Addressed a method using groups of minutiae to

correctly detected. A signicant number of spurious

dene local structural features. The matching is

minutiae may be created as a result. In order to

performed based on the pairs of corresponding

ensure that the performance of the minutiae

structural features that are identied between two

extraction algorithm will be robust with respect to the

ngerprint impressions. However, these methods do

quality of input ngerprint images, an enhancement

not solve the problem of nonlinear distortions.

algorithm which can improve the clarity of the ridge

Recently, some algorithms have been presented to

structures is necessary. However, for poor ngerprint

deal with the nonlinear distortion in ngerprints

image, some spurious minutiae may still exist after

explicitly in order

to improve the matching

ngerprint enhancement and post processing. It is

performance. Proposed a method to measure the

necessary to propose a method to deal with the

forces and torques on the scanner directly. This

spurious minutiae.

prevents capture with the aid of specialized hardware


when excessive force is applied to the scanner
Doraietal. Proposed a method to detect and estimate
distortion occurring in ngerprint videos, but those
two mentioned methods do not work with the
collected ngerprint images. Mao and Maltonietal.
Proposed a plastic distortion model to cope with the
nonlinear deformations characterizing nger- print
images taken from on-line acquisition sensors. This
model helps to understand the distortion process.
However, it is hard to automatically and reliably
estimate the parameter due to the insufficiency and
uncertainty of the information. Doggie Leeetal.
Addressed a minutiae-based ngerprints matching

Fig. 1. Feature set of a live-scan ngerprint image. (a)

algorithm using distance normalization and local

Original ngerprint image. (b) Thinned ridge image with


minutiae and sample points of (a).

alignment to deal with the problem of the nonlinear


distortion.

However,

rich

information

of

the
A method to judge whether an extracted minutia is a

ridge/valley structure is not used, and the matching

true one has been proposed in this paper. According

performance is moderate.

to our work, the distance between true minutiae is


However, in reality, approximately 10% [20] of

generally greater than threshold (three). While near

acquired ngerprint images are of poor quality due to

the spurious minutiae, there are usually other

variations

ridge

spurious minutiae. On the other hand, spurious

conguration, skin conditions, acquisition devices,

minutiae are usually detected at the border of

and non-cooperative attitude of subjects, etc. The

ngerprint image. Examples of spurious minutiae in

ridge structures in poor-quality ngerprint images are

poor quality ngerprint images are shown in Fig. 2.

in

impression

conditions,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

179

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Fingerprint

ISBN: 378 - 26 - 138420 - 5

recognition

(also

known

as

Dactyloscopy) is the process of comparing known


fingerprint against another or template fingerprint to
determine if the impressions are from the same finger
or not. It includes two sub-domains: one is fingerprint
verification

and

the

other

is

finger

print

identification. Verification specify an individual


fingerprint by comparing only one fingerprint
template stored in the database, while identification
specify comparing all the fingerprints stored in the
database. Verification is one to one matching and
Fig. 2. Examples of spurious minutiae in poor quality

identification is one to N (number of fingerprint

ngerprint images. The images have been cropped and

templates

scaled for view. (a) Original image. (b) Enhanced image of

available

in

database)

matching.

Verification is a fast process as compared to

(a), (c) original image, (d) enhanced image of (c). Many

identification.

spurious minutiae were detected in the process of minutiae


extraction. Near the spurious minutiae, there are usually
other spurious minutiae as indicated in ellipses, and
spurious minutiae are usually detected at the border of
ngerprint image as shown in rectangles.

2. EXISTING SYSTEM
A. Fingerprint Recognition
Fig.2. Fingerprint Recognition System

The existing algorithm uses a robust alignment

Fig.2 shows the basic fingerprint recognition system.

algorithm (descriptor-based Hough transform) to

First of all we take a fingerprint image. After taking

align fingerprints and measures similarity between

an input image we can apply fingerprint segmentation

fingerprints by considering both minutiae and

technique. Segmentation is separation of the input

orientation field information. To be consistent with

data into foreground (object of interest) and

the common practice in latent matching (i.e., only

background

minutiae are marked by latent examiners), the

ridges) from the background. This is very useful for

marked minutiae, it can be easily used in law

recovering false feature extraction. In some cases, a

enforcement applications. Experimental results on

correct segmentation is very difficult, especially in

two different latent databases show that the proposed


two

well

Before

to separate the fingerprint regions (presence of

Since the proposed algorithm relies only on manually

outperforms

information).

extracting the feature of a fingerprint it is important

orientation field is reconstructed from minutiae.

algorithm

(irrelevant

poor quality fingerprint image or noisy images.

optimized

Orientation

commercial fingerprint matchers.

field plays an

important role in

fingerprint recognition system. Orientation field

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

180

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

consist of four major steps (1) pre processing

techniques first of all we find minutiae points on

fingerprint image (2) determining the primary ridges

which we have to do mapping. However, there are

of fingerprint block (3) estimating block direction by

some difficulties when using this approach. It is

projective distance variance of such a ridge (4)

difficult to identify the minutiae points accurately

correcting the estimated orientation field. Image

when the fingerprint is of low quality.

enhancement is use to improve significantly the

Pattern-based (or image-based) matching:

image quality by applying some image enhancement

Pattern based technique compare the basic fingerprint

technique. The main purpose of such procedure is to

patterns (arch, whorl, and loop) between a previously

enhance the image by improving the clarity of ridge

stored template and a candidate fingerprint. This

structure or increasing the consistency of the fridge

requires that the images be aligned in the same

orientation. Fingerprint classification is used to check

orientation. In a pattern-based algorithm, the template

the fingerprint pattern type. After classification of

contains the type, size, and orientation of patterns

fingerprint. We can apply fingerprint ridge thinning

within the aligned fingerprint image.

which is also called block filtering; it is used to

The candidate fingerprint image is graphically

reduce the thickness of all ridges lines to a single

compared with the template to determine the degree

pixel width. Thinning does not change the location

to which they match.

and orientation of minutiae points compared to


original fingerprint which ensures accurate estimation

3. PROPOSED SYSTEM

of minutiae points. Then we can extract minutiae

A. entropy

points and generate data matrix. Finally we can use

Image entropy is an extent which is used to explain

minutiae matching to compare the input fingerprint

the business of an image, i.e. the amount of

data with the template data and give the result.

information which must be implicit for by a


compression algorithm. Low entropy Images, such as

B. Fingerprint Matching Techniques

those include a lot of black sky, have very little

There are many Fingerprint Matching Techniques.

difference and large runs of Pixels with the same or

Most widely used matching techniques are these:

parallel DN values. An Image that is entirely flat will


have entropy of Zero. Therefore, they can be

Correlation-based matching:

compressed to a relatively small size. On the other

In correlation based matching the two fingerprint

hand, high

images are matched through corresponding pixels

Entropy images such as an image of heavily formed

which is computed for different alignments and

areas on the moon have a great deal of

rotations. The main disadvantage of correlation based

Thing from one pixel to the next and accordingly

matching is its computational complexity.

cannot be compressed as much as low entropy


images. Image entropy as used in this paper is

Minutiae-based matching:

calculated with the same formula used by the Galileo

This is the most popular and widely used technique,

Imaging Team Entropy

for

fingerprint

comparison.

In

minutiae-based

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

181

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

In the above expression, P is the probability

That the difference between two adjacent pixels

Wiener Filtering:

The Wiener filtering is optimal in conditions of the

Is equal to i, and Log2i is the base 2 algorithm.

mean square error (MSE). In other words, it

Entropy successfully bounds the performance of


The strongest lossless compression feasible, which

minimizes the generally mean square error in the

can be realized in theory by using the

development of inverse to remove and noise

Distinctive set or in perform using Huffman.

smoothing. The Wiener filtering is a linear inference


of the new image. The advance is based on a
stochastic frame. The orthogonality principle implies

B. Correlation

that the Wiener filter in Fourier domain. To complete

Digital Image connection (correlation) and Tracking


(DIC/DDIT) is an optical method that employs

the

Wiener

filter

in

perform

we

have

to

tracking & image check techniques for accurate 2-D

approximation the power spectra of the original


image. For noise is remove the power spectrum is

and 3-D measurements of change in images. This is

equal to the variation of the noise. To estimate the

often used to measure deformation (engineering),

power range of the original image many methods can

displacement, strain, and visual flow, but it is widely

be used. A through estimate is the period gram

applied in many areas of science and engineering

estimate of the power spectral density (PSD).

calculations. The image is first subjected to a 2-D


wiener filter using a 3*3 mask. By using wiener filter
to remove the redundant noise or un-wanted pixels.

C. Energy

Digital image correlation (DIC) techniques have been

For calculating the energy coefficient, the image is


subjected to a wavelet decomposition using the

increasing in status, especially in micro- and neon-

Daubechies wavelet for up to 5 levels. The wavelet

scale mechanical testing applications due to its


relative ease of implementation and use. Advances in

decomposition involves the image with a low-pass

computer technology and digital cameras have been

filter for generating the approximation coefficients

the enabling technologies for this method and while

and a high pass filter for generating the detail

white-light optics has been the leading approach, DIC

coefficients, followed by a down-sampling. The data

can be and has been extensive to almost any imaging

image for each level is taken as the approximation

technology.

image for the previous level. Another related use is


in image transforms: for example, the DCT transform

The cross correlation coefficient is defined by

(basis of the JPEG compression method) transforms a

we

blocks of pixels (8x8 image) into a matrix of

represent as r, then we have

transformed coefficients; for distinctive images, it


results that, while the original 8x8 image has its
energy regularly distributed among the 64 pixels, the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

182

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

changed image has its energy determined in the


lower-upper "pixels, The decomposition operation
generates the approximation coefficientsA5 and
Where n is the number of variables, and Xi

detailed coefficients B5,B4,B3,B2,B1 as shown in

And Y is the values of the ith variable, at points X

below.

and Y correspondingly. i The Manhattan Distance is

Ea= (A5)/ (B5+ B5+B4+B3+B2+B1)

the distance between two points considered along


axes at right angles. The name alludes to the grid

The final feature vector is taken as the complex

explain of the streets of Manhattan, which cause the

formed of the above three components viz.

straight path a car could take between two points in

F= {Cc, En, Ea}

the city. For the 8-puzzle if xi(s) and y(s) are the x

Classification is done by mapping the feature vectors

and y coordinates of tile i in state s, and if upper line

of a training set and a testing set into appropriate

(xi) and upper-line (yii) are the x and y coordinates of

feature spaces and calculating differences using

tile i in the goal state, the heuristic is:

Manhattan distance.

ALGORITHM

FOR

CALCULATING

STATISTICAL TEXTURE FEATURES


Input: Query image for which statistical features has

been computed.
Output: feature vector
1. Calculate Entropy for query image (En) using -sum

Figure 3: Wavelet decomposition of an image

(p.*log2 (p)) formula


E. Manhattan Distance

2. Apply wiener filter for query image and then

The distance between two points in a grid base on a

calculate correlation coefficient (CC) for query image

firmly horizontal and/or vertical path (that is, along

and filtered image

the grid lines), as distinct to the diagonal or "as the

3. Apply 5 level decomposition for input query image

crow flies" distance. The Manhattan detachment is

and calculate energy for coefficients (Ea)

the plain sum of the horizontal and vertical works;

4. Calculate feature vector F for query image by

whereas the diagonal span might be computed by

using En, Ea, and CC.

apply the Pythagorean Theorem. The formula for this

Then compare feature vector F of query image with

distance between a point X=(X1, X2, etc.) and a point

the database image and if features are equal then the

Y= (Y1, Y2, etc.) is:

image is matched.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

183

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

0.420820.

4. EXPERIMENTAL RESULTS
The proposed algorithm has been participated in
FVC2004. In FVC2004, databases are more difficult
than in FVC2000/FVC2002 ones. In FVC2004, the
organizers have particularly insisted on: distortion,
dry, and wet ngerprints. Especially in ngerprints
database DB1 and DB3 of FVC2004, the distortion
between some ngerprints from the same nger is
large. Our work is to solve the problem of distorted
ngerprint matching, so the evaluation of the
proposed algorithm is mainly focused on DB1 and
DB3 of FVC2004. The proposed algorithm is also
compared with the one described by Luo et al. and
the

one

proposed

by

Bazen

et

al
Fig.5. Experimental results of the proposed algorithm on
103_2.tif and 103_4.tif in FVC2004 DB3. The images have
been scaled for view. (a) 103_2.tif. (b) Enhanced image of
103_2. (c) 103_4.tif. (d) Enhanced image of 103_4. The
similarity of these two ngerprints is 0.484 111.

5. CONCLUSION
This paper has proposed a quick and efficient
technique of fingerprint recognition using a set of
texture statistical based features. The features are
derived from a correlation coefficient, an entropy
coefficient and an energy coefficient. The features
can be calculated by using fingerprint miniature
points. Moreover such texture based by using color
finger print images. The fingerprint images may be

Fig.4. Experimental results of the proposed algorithm on

divided in to separation of red, green and blue

102_3.tif and 102_5.tif in FVC2004 DB1. The images have


been cropped and scaled for view. (a) 102_3.tif. (b)

components. And output part combine true color

Enhanced image of 102_3. (c)102_5.tif. (d) Enhanced

components. Future work would involve combining

image of 102_5. The similarity of these two ngerprints is

color and shape based techniques to study whether


these can be used to improve recognition rates.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

184

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

6. REFERENCES

IEEE Transactions on Image Processing, 9, 2000,

[1]. A. Lanitis, A Survey of the Effects of Aging on

pp. 846-853.

Biometric

Identity

Verification,

DOI:10.1504/IJBM.2010.030415
[2]. A. K. Jain, A. Ross and S. Prabhkar, An
Introduction

to

Biometric

Recognition,

AUTHORS

IEEE
K.Phalguna
Rao,
completed
M.Tech
information
technology
from Andhra University
presently Pursuing PhD.
Life member of ISTE. He
is working as Professor in
the Dept of CSE Published
several papers in the
International Journals and International and
national conferences. Attended several
International and national workshops.
Research Interest areas are Data Base
Systems,
Network
Security,
cloud
Computing, Bioinformatics.

Transactions on Circuits and Systems for Video


Technology, special issue on Image and Video
Based Biometrics, 14, 2004, pp. 4-20.
[3]. A. K. Jain, A. Ross and S. Pankanti, Biometrics:
A Tool for Information Security, IEEE Transactions
on Information Forensics and Security. 1, 2000.
[4]. A. K. Jain and A. Ross, Fingerprint Matching
Using Minutiae and Texture Features, Proceeding of
International

Conference on

Image Processing

(ICIP), 2001, pp. 282-285.


[5]. A. K. Jain, L. Hong, S. Pankanti and R. Bolle,
An

Identity-Authentication

System

using

Fingerprints, Proceeding of the IEEE. 85, 1997, pp.


M.Charan
Kumar
received sree kalahasthi
institute of technology
degree
in
computer
science engineering from
the Jawaharlal Nehru
technology
university
Anantapur, in 2010, and
received the Audisankara
institute of technology M.Tech degree in
computer science engineering from the
Jawaharlal Nehru technology Ananthapur in
2014, respectively. He published one
International journal and participated four
national conferences and participate One
International conference. He worked as
communication faculty for 3 years in Kerala
and Karnataka.

1365-1388.
[6]. D. Maltoni, D. Maio, A. K. Jain and S. Prabhkar,
Handbook of Fingerprint Recognition.
[7]. S. Chikkerur, S. Pankanti, A. Jea and R. Bolle,
Fingerprint Representation using Localized Texture
Features, The 18th

International Conference on

Pattern Recognition, 2006.


[8]. A. A. A. Yousiff, M. U. Chowdhury, S. Ray and
H. Y. Nafaa, Fingerprint Recognition System using
Hybrid Matching Techniques, 6th IEEE/ACIS
International

Conference

on

Computer

and

Information Science, 2007, pp. 234-240.


[9]. O. Zhengu, J. Feng, F. Su, A. Cai, Fingerprint
Matching

with

Rotation-Descriptor

Texture

Features,

The 8th International Conference on

Pattern Recognition, 2006, pp. 417-420.


[10]. A. K. Jain, S. Prabhkar, L. Hong and S.
Pankanti, Filterbank-Based Fingerprint Matching,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

185

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Rely on Administration with Multipath


Routing for Intrusion Threshold in
Heterogeneous WSNs.
CH. Sendhil Kumar
Department of CSE
ASCET, Gudur.
A.P, India
[email protected]

S. Dorababu
Department of CSE
ASCET, Gudur.
A.P, India
[email protected]

ABSTRACT
Here with in this report most of us recommend excess management connected with heterogeneous wireless sensor
networks (HWSNs), utilizing multipath aiming to help solution client enquiries near questionable in addition to
malevolent hubs. The main element thought of the repeating administration is usually to effort this buy and sell -off
concerning vitality utilization compared to this increase in trustworthiness, ease, In addition to security to help boost
this framework very helpful lifetime. All of us program this trade off being an enhancement matter intended for
alterably choosing the best repeating degree to use to help multipath aiming intended for disruption tolerance while
using the purpose that the inquiry reaction achievements likelihood is actually increased although dragging out this
very helpful lifetime. Other than, most of us consider this to be enhancement matter to the circumstance in which a
voting-based disseminated disruption finding calculations is actually linked with find in addition to oust malevolent
hubs in a HWSN. DSR seemed to be specifically designed for utilization in multi-hop wireless random
communities. Ad-hoc method permits the actual network to become entirely self-organizing in addition to selfconfiguring meaning you don't have on an active network structure or even management. To meet these trouble
when using the Slow Gradual cluster election Algorithm (GCA), it is reducing energy usage of community
groupings in addition to overall communities. The item elected the actual gradual cluster amid nodes that happen to
be practical for the actual gradual cluster head in addition to proved the power effectiveness inside network.
Key words Heterogeneous wireless sensor Networks; multipath routing; interruption location; dependability,
security; energy conversation
several completely new multipath direction-finding
protocols particularly intended for sensor networks.
Multipath direction-finding is an effective approach,
which often selects a number of trails to supply files
through supplier in order to desired destination in
wireless sensor systems. Multipath direction-finding
ensures a number of trails between your sourcedestination match. Within sole way completely new
route development procedure is actually begun,
which often boosts energy use. Node disappointment
additionally reasons packets to be lowered and may
cause a hold off in delivering your data to the
destroy, as a result the real-time demands of the
multimedia system apps will not be met. Multi-path
routing boosts the volume of feasible paths and also
via that that improves the heavy ness and also
throughput of the transmissions. Multipath directionfinding is actually commercial intended for major

I. INTRODUCTION
Wireless sensor Networks (WSN) is made of
spatially spread autonomous detectors in order to
check real as well as environmental conditions, for
instance temperature, noise , force, for example. and
to cooperatively go the files through the community
with a major position. Greater modern wireless
networks usually are bi-directional, additionally
which allows manage connected with sensor action.
The improvement connected with wireless sensor
networks has been encouraged simply by military
apps for instance battlefield surveillance; today like
wireless communities are widely-used in many
manufacturing and also client apps, for instance
manufacturing procedure monitoring and also
manage, appliance wellness monitoring, and so on.
Wireless sensor networks(WSNs) include resulted in

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

186

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

insert predicament when compared with sole way.


Multipath routing is actually mostly utilized often
intended for insert handling as well as intended for
dependability. Load handling is possible simply by
handling energy utilization among the nodes
increasing community life-time. The ability of your
community to provide services in order to decided on
community targeted visitors more than various
technology. The primary aim connected with QoS is
to offer priority including focused bandwidth,
controlled jitter and also latency need simply by
many real-time and also interactive targeted visitors,
and also increased reduction attributes. Intrusion
detection system (IDS) using the aim in order to
discover and also remove malevolent nodes. Within
sensor cpa networks, almost all adversaries would
targeted the direction-finding stratum, since which
allows the crooks to manage the data flowing inside
the community. In addition to, sensor networks
usually are mostly about exposure files back to the
bottom station, and also disrupting this would
produce a great episode an excellent 1. And so,
intended for like sensor networks, the most likely
structures on an IDS can be network based, compared
to host-based.

ISBN: 378 - 26 - 138420 - 5

neighbor nodes and the continuing energy level and


also one-hop neighbor details (GCA-ON), which
often elects group brains according to Im and the
comparable position details connected with sensor
nodes. This kind of mostly concentrates on the group
scalp choice. It focuses upon lowering energy
utilization of nearby groups and also overall wireless
systems. The idea decided the group scalp among
nodes that happen to be simple for the group scalp
and also turned out the energy productivity. The
group boss election will be based on the energy and
the boss can deliver the announcements in order to
inter group associates and the intra groups. And it
also locates the common energy of the overall
community and it re-elects the new group scalp.
II. LITERATURE SURVEY
D. Wagner et al., (2003)[16].yet none of them have
been outlined with security as an objective. We
propose security objectives for directing in sensor
systems, demonstrate how assaults against specially
I.R.Chen et ing.,[3] Suggested your data sensing are
used in endemic program in location including
security along with security supervising along with
order, control in struggle field. To produce adaptive
fault tolerant good quality regarding assistance
(QOS) control criteria depending on hop by-jump
files distribution using origin along with way
redundancy, having your objective to meet up with
program QOS requirements.

The network-based IDS works by using uncooked


community packets because the databases. The idea
listens upon the community and also catches and also
looks at person packets instantly. DSR can be a
reactive direction-finding process which often has the
ability to deal with the MANET without needing
recurrent table update announcements such as tabledriven direction-finding protocols complete. DSR has
been created intended for use within multi-hop
wireless ad hoc networks. Ad-hoc process makes it
possible for the community to be totally selforganizing and also self configuring which means
that there's no need on an present community national
infrastructure as well as supervision. Intended for
restricting the bandwidth, accomplishing this to get a
way is only carried out when a way is needed by way
of node (On- Demand-Routing).

Bao et ing.,[15] Suggested your scalable clusterbased hierarchical confidence operations standard
protocol with regard to WSN to help successfully
manage detrimental nodes.. Some sort of story
likelihood style summarize some sort of HWSN
composed a lot of sensor nodes having unique social
along with good quality regarding assistance (QOS)
behaviours. Anomaly structured breach recognition
used in both the recognition likelihood along with
false beneficial likelihood

Within DSR the sender (source, initiator) decides the


whole way on the supplier in order to the destination
node (Source-Routing) and also debris the deals with
of the second time beginners nodes of the route inside
the packets. When compared with other reactive
direction-finding protocols such as ABR as well as
SSA, DSR is actually beacon-less which means that
you will discover not any hello-messages utilized
between your nodes in order to notify the neighbours
about your ex occurrence. Gradual cluster election
Protocol (GCA) which often little by little elects
group brains in line with the proximity in order to

A.P.R da Silva (2005) [10] might be connected to


secure Wsns against a few sorts of assaults. Be that as
it may, there are a few assaults for which there is no
known aversion strategy. For these cases, it is
important to utilize some system of interruption
location. Other than keeping the gatecrasher from
bringing about harms to the system, the interruption
recognition framework (IDS) can get data identified
with the assault procedures, helping in the
improvement of avoidance frameworks.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

187

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

In this work we propose an IDS that fits the requests


and limitations of Wsns. Reenactment results uncover
that the proposed IDS is proficient and faultless in
discovering various types of reenacted assaults.. In
Traditional
Technique,
efficient
redundancy
supervision of your clustered HWSN to extend its
life-time procedure inside presence associated with
unreliable along with detrimental nodes. Most of us
tackle the actual trade-off concerning energy usage as
opposed to. QoS obtain throughout dependability,
timeliness along with protection with all the aim to
increase the actual lifetime of a new clustered HWSN
though gratifying app QoS requirements inside
wording associated with multipath course-plotting.
Far more exclusively, most of us assess the suitable
volume of redundancy where data are generally sent
with a remote control sink inside presence associated
with unreliable along with detrimental nodes, in order
that the issue achievement chance is actually
maximized though increasing the actual HWSN lifetime.
SYSTEM MODEL

ISBN: 378 - 26 - 138420 - 5

Because of constrained energy, a packet is sent


bounce by jump without utilizing affirmation or
retransmission. All sensors are liable to catch assaults
(inside aggressors). Since all sensors are haphazardly
placed in operational region catch rate applies to both
Chs and Sns thus traded off hubs are additionally
arbitrarily conveyed in operation territory. [19]
Compromised node performs two most energy
conserving attacks, bad-mouthing attacks, packet
dropping attacks. Environment conditions cause a
node to fail with certain probability include hardware
failure and transmission failure (due to noise and
interference). Hostility to HWSN is characterized by
as per node capture rate, determined based on
historical data and knowledge about the target
application environment.

WSN consists of sensors involving unique


functionality sorts of sensors are usually Cluster
Minds (CHs) in addition to Sensor Nodes (SNs). CHs
are usually better than SNs with power in addition to
computational means, stand for the original energy
level involving CHs in addition to SNs, in addition to
placed on almost any shape with the detailed region
[3]. CHs and SNs are distributed in the operational
area ensure coverage by deploying CHs and SNs
randomly and distributed according to homogeneous
spatial Poisson processes. Radio ranges of CH and
SN for transmission are initialized. Radio range and
transmission power of both CHs and SNs are
dynamically adjusted throughout system lifetime to
maintain connectivity between CHs and between
SNs. Multi-hop Routing is required for
communication between two nodes with distance
greater than single hop

Fig 2 Block diagram of WSN


A WSN should not just fulfil the application
particular Qos prerequisites, for example,
dependability,
opportunities
and
security,
additionally minimize vitality utilization to drag out
the framework valuable lifetime. It is regularly put
stock in the exploration group that bunching is a
powerful answer for accomplishing adaptability,
vitality preservation, and unwavering quality.
Utilizing heterogeneous hubs can further improve
execution and draw out the framework lifetime.In
existing system DSR is utilized to comprehend this
problem.dsr is a touchy directing convention which
can deal with a MANET without utilizing occasional
table-overhaul messages like table-driven steering
conventions do. DSR was particularly intended for
utilization in multi-jump remote impromptu systems.
Specially appointed convention permits the system to
be totally orchestrating toward oneself and
organizing toward oneself which implies that there is
no requirement for a current system base or
organization.

Fig1 : Cluster-based WSN Architecture.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

188

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

along with inside giving the user, self-assurance that


this brand new process will continue to work and
become powerful. The particular enactment level
entails careful setting up, research of the active
process along with its restrictions on enactment,
building involving ways to attain changeover along
with assessment involving changeover strategies.

Dynamic Source Routing Protocol


DSR is a receptive directing convention which can
deal with a MANET without utilizing occasional
table upgrade messages like table-driven steering
conventions do. DSR was particularly intended for
utilization in multi-jump remote specially appointed
systems. Impromptu convention permits the system
to be totally dealing with toward oneself and self
designing which implies that there is no requirement
for a current system base or organization.

SCHEMES:
1. Multi Course-plotting
2. Intrusion Patience
3. Energy Productive
4. Simulation Procedure

For limiting the data transfer capacity, the


methodology to discover a way is just executed when
a way is needed by a hub (On-Demand-Routing). In
DSR the sender (source, initiator) decides the entire
way from the source to the terminus hub (SourceRouting) and stores the locations of the moderate
hubs of the course in the bundles. Contrasted with
other responsive directing conventions like ABR or
SSA, DSR is guiding less which implies that there
are no welcome messages utilized between the hubs
to inform their neighbours about her vicinity. DSR
was produced for Manets with a little distance across
somewhere around 5 and 10 bounces and the hubs
ought to just move around at a moderate speed.dsr is
focused around the Link-State-Algorithms which
imply that every hub is proficient to spare the most
ideal path to an end. Additionally if a change shows
up in the system topology, then the entire system will
get this data by flooding.

1. Multi Course-plotting
Within this component, Multipath redirecting is
regarded a highly effective system regarding failing
along with attack tolerance to improve facts shipping
inside WSNs. Principle concept is actually that this
chance involving at least one particular journey
achieving this kitchen sink node or even base stop
increases as we convey more walkways carrying out
facts shipping. Alot of previous exploration dedicated
to using multipath redirecting to improve
consistency, a few attentionshas been paid out to be
able to using multipath redirecting to be able to
accept insider attacks. These reports, nonetheless,
generally pushed aside this tradeoff among QoS
attain versus. Power usage that may negatively
shorten the system life time.

Disadvantages Raises in traditional based protocols

2. Intrusion tolerance

Servicing standard protocol will not in the area mend


a new damaged web page link. Your damaged web
page link is merely divulged for you tothe actual
initiator. Your DSR standard protocol is merely
efficient in MANETs having below 190 nodes.
Difficulties glimpse simply by rapid going of much
more serves, so your nodes can solely move around
however using a reasonable swiftness. Inundating the
actual network might cause collusions relating to the
packets. Likewise you can a compact occasion hold
up at the get started of a brand-new relationship
because the initiator ought to first chose the approach
to the target.

Within these schemes, attack tolerance via multipath


redirecting, you will discover a pair of important
problems to fix[1]
(1) The amount of walkways to work with along with
(2) What exactly walkways to work with.
Towards ideal your knowledge, i am the very first to
cope with this how numerous walkways to be able
to use dilemma. With the what walkways to be able
to use dilemma, our strategy is actually distinct from
active function in this we all usually do not think
about distinct redirecting practices.

III. SYSTEM IMPLEMENTATION

3. Energy Productive
Within this component, you will discover a pair of
approaches by which power successful IDS might be
carried out inside WSNs. One particular strategy
particularly pertinent to be able to level WSNs is
designed for a second time beginners node to be able
to feedback maliciousness along with power

Rendering is the level of the venture when the


theoretical design and style is actually turned out in
to a functioning process. As a result it can be
regarded as being essentially the most vital level
inside obtaining an excellent brand new process

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

189

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

reputation involving its neighbour nodes for the


sender node (e. Gary the gadget guy., the origin or
even kitchen sink node) who is going to and then
operate the knowledge to be able to path packets in
order to avoid nodes with unsatisfactory
maliciousness or even power reputation. Another
strategy which usually we all take up with this
cardstock using regional host-based IDS regarding
power resource efficiency.

ISBN: 378 - 26 - 138420 - 5

Electing the Cluster And Cluster Heads:


In this module fundamentally concentrates on the
bunch head determination. The Gradual Cluster head
race Calculation (GCA) is utilized for choosing the
group and bunch heads. It chose the bunch head
among hubs which are feasible for the group head
and demonstrated the vitality proficiency.The group
pioneer decision will be focused around the vitality
and the pioneer will send the messages to bury bunch
parts and the intra groups .And it discovers the
normal vitality of the general system and it rechooses the new group head. the vitality utilization
will be less over than this element source steering
convention.

4 Simulation Procedures
Within this component, the cost of executing this
active redundancy managing protocol defined earlier
mentioned, as well as recurrent clustering, recurrent
attack diagnosis, along with query digesting via
multipath redirecting, with regard to power usage

Deploying Redundancy Management System:

IDS:
Intrusion Detection System (IDS)[9] is a device or
software application that monitors network or system
activities for malicious activities and produces
reports to a management station. Intrusion Detection
and prevention System (IDPS) are primarily focused
on identifying possible incidents, logging information
about them and reporting attempts.

The viable repetition administration framework was


conveyed .after the bunch and group head
determination the excess administration framework is
connected. Furthermore the Qos was enhanced, for
example, unwavering quality, and opportunities is
progressed. the information is exchanged from source
to sink. Through this repetition system the
information is steered from source to sink. So the
vitality of the hub will be expanded.

Gradual Cluster Head Election Algorithm (GCA):


In the proposed technique the way must be given and
a substitute must be picked by the directing method
and we execute the trust system to assess the
neighbour hubs and there is no requirement for
grouping system and the light weight IDS [20]. could
be an extra gimmick to catch the different assault
sorts and our adjustment predominantly concentrate
on the vitality utilization and the interruption
tolerance and our adjusted convention will act as
indicated by the trust usage it picks the main when
the trust quality accomplishes and there won't be any
steering overhead .And it will enhance the QOS and
the vitality utilization additionally will be less.

IV. PERFORMANCE EVALUATION


HWSN includes 3000 SN nodes and also 100 CH
nodes, started within a square area of 92(200m
200m).Nodes usually are dispersed in your
neighbourhood after having a Poisson process along
with denseness SN = 35 nodes/(20 20 m2) and also
CH = 1node/(2020 m2) in deployment time period.
Yourstereo varies rch and also rsh usually are
dynamically fine-tuned among 5m to be able to 25m
and also 25m to be able to 120m respectively to be
able to sustain multilevel online connectivity. The 1st
energy levels of SN and also CH nodes usually are
Esno= 0. 8 Joules and alsoEcho= 10 Joules so they
really wear out energy in concerning the same time
frame. The force dissipation to perform the actual
transmitter and also recipient circuitry is usually 50
nJ/bits. Your energy used by the actual monitor
amplifier to realize and also tolerable indicate to be
able to noises rate. Fig. 3&4 demonstrates a good
best combo (mp&ms) under low and also highget rate
for life time maximization.

The Gradual Cluster head decision Algorithm (GCA)


is utilized for choosing the group and bunch heads.
This technique chiefly concentrates on the group
head choice. It concentrates on decreasing vitality
utilization of near by bunches and general systems. It
chose the bunch head among hubs which are feasible
for the group head whats more demonstrated the
vitality effectiveness [20]. The bunch pioneer
decision will be focused around the vitality and the
pioneer will send the messages to entomb group parts
and the intra bunches .And it discovers the normal
vitality of the generally system and it re-chooses the
new group head.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

190

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

excess (mp) and source repetition (ms), and also the


best interruption identification settings regarding the
amount of voters (m) and the interruption conjuring
in term ( under which the lifetime of a heterogeneous
remote sensor system is boosted while fulfilling the
unwavering quality, convenience and security
necessities of question handling applications in the
vicinity of questionable remote correspondence and
noxious hubs. At last, we connected our investigation
results to the configuration of an element excess
administration calculation to distinguish and apply
the best plan parameter settings at runtime
accordingly to environment progressions to drag out
the framework lifetime.

Fig.3 Effect of T IDS on MTTF under low capture rate

The WSN mustn't just match the application


particular QoS demands for example stability,
timeliness as well as protection, but in addition limit
strength use to help prolong the machine helpful lifetime. It's generally thought in the analysis group in
which clustering is an efficient answer pertaining to
attaining scalability, strength resource efficiency, as
well as stability. Using heterogeneous nodes can
additional improve functionality as well as prolong
the actual program life-time. In current program DSR
is utilized to fix this concern. DSR is usually a
reactive direction-finding protocol which is capable
to take care of any MANET without needing regular
table-update messages similar to table-driven
direction-finding protocols perform. DSR has been
specifically designed pertaining to utilization in
multi-hop wireless ad hoc networks. Ad-hoc protocol
enables the actual circle to become absolutely selforganizing as well as self-configuring which means
that you don't have to have an current circle national
infrastructure or maybe current administration. To
fulfill the aforementioned problem utilizing the
Continuous Gradual cluster head election Algorithm,
it truly is cutting down strength usage of nearby
groupings as well as all round networks. This elects
the actual cluster mind among nodes which are
possible for the actual cluster mind as well as proved
the energy proficiency in the circle.

Fig.4 Effect of TIDS on MTTF under high capture rate

Energy Consumption

VI. REFERENCES

Iterations
Fig5Energy consumption in DSR & GCA method

[1]

V. CONCLUSION
In this paper we performed a trade-off investigation
of vitality utilization vs. Qos pick up in
dependability, opportunity, and security for excess
administration of grouped heterogeneous remote
sensor systems using multipath directing to answer
client inquiries. We created a novel likelihood model
to dissect the best excess level in wording of way

Hamid Al-Hamadi and Ing-Ray Chen (2013) Redundancy


Management of Multipath Routing for Intrusion Tolerance in
Heterogeneous Wireless Sensor Networks Vol 10.

[2] E. Felemban, L. Chang-Gun, and E. Ekici, "MMSPEED:


multipath Multi-SPEED protocol for QoS guarantee of
reliability and. Timeliness inwireless sensor networks," IEEE
Trans. Mobile Comput., vol. 5, no. 6, pp.738-754, 2006.
[3] I. R. Chen, A. P. Speer, and M. Eltoweissy, "Adaptive FaultTolerantQoS Control Algorithms for Maximizing System
Lifetime of Query-Based Wireless Sensor Networks," IEEE

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

191

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Trans. on Dependable andSecure Computing, vol. 8, no. 2, pp.


161-176, 2011.

ISBN: 378 - 26 - 138420 - 5

MANET,"
InternationalConference
on
Communication and Networking, 2008, pp.

[4] M. Yarvis, N. Kushalnagar, H. Singh, A. Rangarajan, Y. Liu,


and S.Singh, "Exploiting heterogeneity in sensor networks,"
24th Annu. JointConf.of the IEEE Computer and
Communications Societies (INFOCOM),2005, pp. 878-890
vol. 2.

Computing,

[19 ] Trust Management of Multipath Routing for


Intrusion Tolerance in Heterogeneous WSN
[20] Trust Based Voting Scheme and Optimal
Multipath Routing for Intrusion Tolerance in
Wireless Sensor Network

[5] H. M. Ammari and S. K. Das, "Promoting Heterogeneity,


Mobility, andEnergy-Aware Voronoi Diagram in Wireless
Sensor Networks," IEEETrans. Parallel Distrib. Syst., vol. 19,
no. 7, pp. 995-1008, 2008.
[6] X. Du and F. Lin, "Improving routing in sensor networks with
heterogeneous sensor nodes," IEEE 61st Vehicular
TechnologyConference, 2005, pp. 2528-2532.
[7] S. Bo, L. Osborne, X. Yang, and S. Guizani, "Intrusion
detectiontechniques in mobile ad hoc and wireless sensor
networks," IEEEWireless Commun., vol. 14, no. 5, pp. 56-63,
2007.
[8] I. Krontiris, T. Dimitriou, and F. C. Freiling, "Towards
intrusion detectionin wireless sensor networks," 13th
European Wireless Conference, Paris,France, 2007.
[9] J. Deng, R. Han and S. Mishra, INSENS: Intrusion tolerant
routing for wireless sensor networks, 2006.
[10] A. P. R. da Silva, M. H. T. Martins, B. P. S. Rocha, A. A. F.
Loureiro, L.B. Ruiz, and H. C. Wong, "Decentralized
intrusion detection in wirelesssensor networks," 1st ACM
Workshop on Quality of Service & Securityin Wireless and
Mobile Networks, Montreal, Quebec, Canada, 2005.
[11] Y. Zhou, Y. Fang, and Y. Zhang, "Securing wireless sensor
networks: asurvey," IEEE Communications Surveys &
Tutorials, vol. 10, no. 3, pp. 6[12] V. Bhuse and A. Gupta, "Anomaly intrusion detection in
wireless sensor networks," J. High Speed Netw., vol. 15, no.
1, pp.33-51, 2006.
[13] G. Bravos and A. G. Kanatas, "Energy consumption and
trade-offs on wireless sensor networks," 16th IEEE Int.
Symp.
OnPersonal,
Indoor
and
Mobile
Radio
Communications, pp. 1279-1283, 2005.
[14] A. P. R. da Silva, M. H. T. Martins, B. P. S. Rocha, A. A. F.
Loureiro, L. B. Ruiz, and H. C. Wong, "Decentralized
intrusion detection in wireless sensor networks," 1st ACM
Workshop on Quality of Service & Security in Wireless and
Mobile Networks, Montreal, Quebec, Canada, 2005.
[15] I.R. Chen, F. Bao, M. Chang, and J.H. Cho, "Trust
management for encounter-based routing in delay tolerant
networks" IEEEGlobecom 2010, Miami, FL, Dec. 2010.
16] C. Karlof and D. Wagner, "Secure routing in wireless sensor
networks: attacks and countermeasures," IEEE Int. Workshop
onSensor Network Protocols and Applications, pp. 113-127,
2003.
[17] Y. Lan, L. Lei, and G. Fuxiang, "A multipath secure routing
protocolbased on malicious node detection," Chinese Control
and DecisionConference, 2009, pp. 4323-4328.
[18] D. Somasundaram and R. Marimuthu, "A Multipath Reliable
Routing fordetection and isolation of malicious nodes in

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

192

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Electric power generation using piezoelectric crystal


P.BALAJI NAGA YASHWANTH, GAUTHAM KUMAR MOKA, DONEPUDI JASHWANTH

ANIL NEERUKONDA INSTITUTE OF TECHNOLOGY AND SCIENCES

ABSTRACT

Key words: energy generation ,piezoelectric


crystals

The usefulness of most high technology


devices such as cell phones, computers, and
sensors is limited by the storage capacity of
batteries. In the future, these limitations will
become more pronounced as the demand for
wireless
power
outpaces
battery
development which is already nearly
optimized. Thus, new power generation
techniques are required for the next
generation of wearable computers, wireless
sensors, and autonomous systems to be
feasible. Piezoelectric materials are excellent
power generation devices because of their
ability to couple mechanical and electrical
properties. For example, when an electric
field is applied to piezoelectric a strain is
generated and the material is deformed.
Consequently, when a piezoelectric is
strained it produces an electric field;
therefore, piezoelectric materials can convert
ambient vibration into electrical power.
Piezoelectric materials have long been used
as sensors and actuators; however their use
as electrical generators is less established. A
piezoelectric power generator has great
potential for some remote applications such
as in vivo sensors,embedded MEMS devices,
and distributed networking. Developing
piezoelectric generators is challenging
because of their poor source characteristics
(high voltage, low current, high impedance)
and relatively low power output. This paper
presents a theoretical analysis to increase
the piezoelectric power generation that is
verified with experimental results.

INTRODUCTION
Mechanical stresses applied to piezoelectric
materials distort internal dipole moments and
generate electrical potentials (voltages) in
direct proportion to the applied forces. These
same crystalline materials also
lengthen or shorten in direct proportion to
the magnitude and polarity of applied
electric fields. Because of these properties,
these materials have long been used as
sensors and actuators. One of the earliest
practical applications of piezoelectric
materials was the development of the first
SONAR system in 1917 by Langevin
who used quartz to transmit and receive
ultrasonic waves. In 1921, Cady first
proposed the use of quartz to control the
resonant frequency of oscillators. Today,
piezoelectric sensors (e.g., force, pressure,
acceleration) and actuators (e.g., ultrasonic,
micro positioning) are widely available. The
same properties that make these materials
useful for sensors can also be utilized to
generate electricity. Such materials are
capable of converting the mechanical energy
of compression into electrical energy, but
developing piezoelectric generators is
challenging because of their poor source
characteristics (high voltage, low current,
high impedance). This is especially true at
low frequencies and relatively low power
output.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

193

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

PIEZOELECTRIC CRYSTAL?
The phenomenon of generation of a voltage
under mechanical stress is referred to as the
direct piezoelectric effect, and the
mechanical strain produced in the crystal
under electric stress is called the converse
piezoelectric effect.

The effect is explained by the


displacement of ions in crystals that
have a nonsymmetrical unit cell
When the crystal is compressed, the
ions in each unit cell are displaced,
causing the electric polarization of
the unit cell.
Because of the regularity of
crystalline structure, these effects
accumulate, causing the appearance
of an electric potential difference
between certain faces of the crystal.
When an external electric field is
applied to the crystal, the ions in
each unit cell are displaced by
electrostatic forces, resulting in the
mechanical deformation of the whole
crystal.

P = d x stress and E = strain/d


Piezoelectricity, discovered by Curie
brothers in 1880, originated from the Greek
word piezenin , meaning, to press.
MAKING

The piezoelectric axis is then the axis


of polarization. If the polycrystalline
material is poled as it is cooled
through its curie point, the domains
in the crystals are aligned in the
direction of the strong electric field.
In this way, a piezoelectric material
of required size, shape and
piezoelectric qualities can be made
within limits.
In a given crystal, the axis of
polarization depends upon the type of
stress. There is no crystal class in
which the piezoelectric polarization
is confined to a single axis. In several
crystal classes, however, it is
confined to a plane. Hydrostatic
pressure produces a piezoelectric
polarization in the crystals of those
ten classes that show piezoelectricity,
in addition to piezoelectricity.

displacement of electrical charge due


to the deflection of the lattice in a
naturally piezoelectric quartz crystal
The larger circles represent silicon
atoms, while the smaller ones
represent oxygen.
Quartz crystals is one of the most
stable piezoelectric materials.

ARTIFICIAL MATERIALS polycrystalline, piezoceramics are


man made materials which are forced
to become piezoelectric by applying
large electric field.
high charge sensitivity

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

194

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

materials available which operate at


1000 F (540 C)
characteristics vary with temperature

ISBN: 378 - 26 - 138420 - 5

This help in reducing ripple in the


output waveform.
Here the output voltage produced is 2
to 3 volts.

CONFIGURATION:

Red indicates the crystal


Arrows indicate the direction of
applied force
the compression
design features
high rigidity, making it useful for
implementation in high frequency
pressure and force sensors.

FOOT
STEP
GENERATION:

The inverter circuit converts 12 v DC to 12 v


AC .
Now the obtained 12 v AC is connected to a
step up transformer.
Commercially available transformer to reach
is 12 v AC to 220 v AC.
Transformer step up 12 v AC to 220-230v
AC.
The obtained voltage can be used for many
applications.

POWER

When a piezoelectric material is put under


pressure it generates potential across the
other end.
This basic principal us used to produced
electrical supply.

By the ouput obtained a 12v battery


is charged .
Now the output of 12 v battery is
given to a inverter circuit.

The output obtained is a AC voltage.


This is the passed through a
rectifying circuit which converts AC
to DC voltage. The output is
connected to a rectifying circuit ,
which produce a pulsating dc.

APPLICATIONS:
The above method can be employed in many
ways like
1.railway station.
2. malls in cities.
3. escalators. Etc

To obtain pure DC we connect a


capacitor parallel to the load.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

195

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The other applications where pressure can be


employed are:
1. In vehicles tire
2. In runway
3. Industries
4. Ships
5. In drilling machines.

ISBN: 378 - 26 - 138420 - 5

[4]-Gaur and Gupta Engineering Physicscrystal structure.[58.8]

CONCLUSION:
As the results shows that by using double
actuators in parallel we can reduce the
charging time of the battery and increase the
power generated by the piezoelectric device.
In second research where a piezoelectric
generator was put to the test and generated
some 2,000 watt-hours of electricity. The
setup consists of a ten-meter strip of asphalt,
with generators lying underneath, and
batteries in the roads proximity. So that it is
clear by using parallel combination we can
overcome the problems like of impedance
matching and low power generation. The
results clearly show that piezoelectric
materials are the future of electric power
generation.

Reference:
[1]- Piezoelectric Plates and Buzzers, Oct.
17,
2009.
[Online].
Available:
http://jingfengele.en.alibaba.com/product/50
91700650159415/Piezoelectric_Buzzer_Plates.html
[Accessed: Oct.17, 2009].
[2]- Anil Kumar(2011) Electrical Power
Generation Using Piezoelectric CrystalInternational Journal of Scientific &
Engineering Research- Volume 2, Issue 5,
May-2011.
[3]- www.BEProjectReport.com

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

196

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

INTEGRATION OF DISTRIBUTED SOLAR POWER GENERATION


USING BATTERY ENERGY STORAGE SYSTEM
K.MOUNIKA

Sri.G.VEERANNA

M.E(Power Systems & Automation)

Asst.Professor

Department Electrical & Electronics Engineering,


S.R.K.R Engineering college,
Bhimavaram,
Andhra Pradesh

Abstract : This paper presents an overview of

sunlight into electricity using semiconductors that

the challenges of integrating solar power to the

exhibit the photovoltaic effect. Photovoltaic effect

electricity distribution system, a technical

involves the creation of voltage in a material upon

overview of battery energy storage systems, and

exposure to electromagnetic radiation. The solar cell

illustrates a variety of modes of operation for

is the elementary building block of the photovoltaic

battery energy storage systems in grid-tied solar

technology. Solar cells are made of semiconductor

applications. . Battery energy storage systems are

materials, such as silicon. One of the properties of

increasingly being used to help integrate solar

semiconductors that makes them most useful is that

power into the grid. These systems are capable of

their conductivity may easily be modified by

absorbing and delivering both real and reactive

introducing impurities into their crystal lattice.


The integration of significant amounts of

power with sub -second response times. With


these capabilities, battery energy storage systems

photovoltaic (PV) solar power generation to the

can mitigate such issues with solar power

electric grid poses a unique set of challenges to

generation as ramp rate, frequency, and voltage

utilities and system operators. Power from grid-

issues. Specifically, grid-tied solar power

connected solar PV units is generated in quantities

generation is a distributed resource whose output

from a few kilowatts to several MW, and is then

can change extremely rapidly, resulting in many

pushed out to power grids at the distribution level,

issues for the distribution system operator with a

where the systems were often designed for 1-way

large quantity of installed photovoltaic devices.

power flow from the substation to the customer. In


climates with plentiful sunshine, the widespread

Index Terms Battery energy storage systems,

adoption of solar PV means distributed generation on

photovoltaic, renewable, solar.

a scale never before seen on the grid.


Grid-connected

I.

INTRODUCTION

solar

PV

dramatically

changes the load pro-file of an electric utility

Photovoltaic is the field of technology and

customer. The expected widespread adoption of solar

research related to the devices which directly convert

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

197

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

generation by customers on the distribution system

counteract the change in generation. In small power

poses significant challenges to system operators both

systems, frequency can also be adversely affected by

in transient and steady state operation, from issues

sudden changes in PV generation. Battery energy

including voltage swings, sudden weather -induced

storage systems (BESS), whether centrally located at

changes in generation, and legacy protective devices

the substation or distributed along a feeder, can

designed with one-way power flow in mind. When

provide power quickly in such scenarios to minimize

there is plenty of sunshine during the day, local solar

customer interruptions. Grid-scale BESS can mitigate

generation can reduce the net demand on a

the above challenges while improving system

distribution feeder, possibly to the point that there is

reliability and improving the economics of the

a net power outflow to the grid. In addition, solar

renewable resource.

power is converted from dc to ac by power electronic

This paper describes the operation and

converters capable of delivering power to the grid.

control methodologies for

Due to market inefficiencies, the typical solar

designed to mitigate the negative impacts of PV

generator is often not financially rewarded for

integration,

providing reactive power support, so small inverters

distribution system efficiency and operation. The

are often operated such that they produce only real

fundamentals of solar PV integration and BESS

power while operating a lagging power factor,

technology are presented below, followed by specific

effectively taking in or absorbing reactive power, and

considerations in the control system design of solar

increasing the required current on the feeder for a

PV coupled BESS installations. The PV-coupled

given amount of real power. A radial distribution

BESS systems described in this paper utilize the XP-

feeder with significant solar PV generation has the

Dynamic Power Resource (XP-DPR).

potential to generate most of its own real power

II.

while

a grid-scale BESS

improving

overall

power

PHOTOVOLTAIC
INTEGRATION

during daylight hours, while drawing significant


reactive power.
Modest

Solar powers inherent intermittency poses

levels of solar PV generation

on

distribution circuits can be easily managed by the

challenges in terms of power quality and reliability.

distribution system operator (DSO). However, both

A weather event such as a thunderstorm has the

the DSO and the customers of electric retail service

potential to reduce solar generation from maximum

may soon feel the undesirable impacts on the grid as

output to negligible levels in a very short time. Wide-

PV penetration levels increase.

area weather related output fluctuations can be

A PV system consists of a number of

strongly correlated in a given geographical area,

interconnected components designed to accomplish a

which means that the set of solar PV generators on

desired task, which may be to feed electricity into the

feeders down-line of the same substation has the

main distribution grid. There are two main system

potential to drastically reduce its generation in the

configurations stand-alone and grid-connected. As

face of a mid-day weather event. The resulting output

its name implies, the stand-alone PV system operates

fluctuations can adversely affect the grid in the form

independently of any other power supply and it

of voltage sags if steps are not taken to quickly

usually supplies electricity to a dedicated load or

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

198

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

loads. It may include a storage facility (e.g. battery

= 0.01

ISBN: 378 - 26 - 138420 - 5

bank) to allow electricity to be provided during the

night or at times of poor sunlight levels. Stand-alone

(6)

systems are also often referred to as autonomous


systems since their operation is independent of other

Based on (6), it is evident that the power

power sources. By contrast, the grid-connected PV

delivered by the PV array is a function of insolation

system operates in parallel with the conventional

level at any given temperature.

electricity distribution system. It can be used to feed


electricity into the grid distribution system or to
power loads which can also be fed from the grid.
The PV array characteristic is described by
the following;
=

(2)

In (2), q is the unit charge, k the Boltzmans


constant, A is the p-n junction ideality factor, and Tc
the cell temperature. Current irs is the cell reverse
Fig. 1. Simplifi ed one-line diagram of a BESS in parallel with
a Solar PV fa-cility connected to the grid.

saturation current, which varies with temperature


according to
=

(3)

III.

BATTERY ENERGY STORAGE

In (3), Tref is the cell reference temperature,


the reverse saturation current at Tref. and EG

A. Battery Energy Storage Basics

the

band-gap energy of the cell. The PV current iph

A grid-scale BESS consists of a battery bank,

depends on the insolation level and the cell

control system, power electronics interface for ac - dc

temperature according to

power

conversion, protective circuitry, and a

transformer to convert the BESS output to the


= 0.01

(4)

transmission or distribution system voltage level. The

In (4), iscr is the cell short-circuit current at

one- line diagram of a simple BESS is shown in Fig.

the reference temperature and radiation, Kv a

1. A BESS is typically connected to the grid in

temperature coefficient, and the insolation level in

parallel with the source or loads it is providing

kW/m . The power delivered by the PV array is

benefits to, whereas tradi-tional uninterruptible

calculated by multiplying both sides of (2) by vpv.

power supplies (UPS) are installed in series with their

loads. The power conversion unit is typically a bi-

(5)

directional unit capable of four quadrant operation,

Substituting iph from (4) in (5), Ppv becomes

means that both real and reactive power can be


delivered or absorbed independently according to the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

199

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

needs of the power system, up to the rated apparent


power of the converter.

Most BESS control systems can be operated via


automatic generation control (AGC) signals much

The battery bank consists of many batteries


connected

in

combination

like a conventional utility generation asset, or it can

series-parallel

be operated in a solar-coupled mode where real and

configuration to provide the desired power and

reactive power commands for the converter will be

energy capabilities for the application. Units are

generated many times per second based on real -time

typically described with two numbers, the nameplate

PV output and power system data. In the case of the

power given in MW, and the maximum storage time

XP -DPR, three -phase measurements from potential

given in MWh. The BESS described in this paper is a

and current transducers (PTs and CTs) are taken in

1.5/1 unit, means it stores 1 MWh of energy, and can

real-time on an FPGA device, and once digitized

charge or discharge at a maximum power level of 1.5

these signals become the input for proprietary real

MW. In renewable energy applications, it is common

time control algorithms operating at kHz speeds.

to operate a BESS under what is known as partial

Various control algorithms have been used for PV

state of charge duty (PSOC), a practice that keeps the

applications, providing control of ramp rates,

batteries partially discharged at all times so that they

frequency support.

are capable of either absorbing from or discharging


power onto the grid as needed.

Fig.2.Configuration of the grid-connected hybrid PV /Battery generation system

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

200

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

generators. Frequency deviation is caused by a

B.Ramp Rate Control


Solar PV generation facilities have no

mismatch in generation and load, as given by the

inertial components, and the generated power can

swing equation for a Thevenin equivalent power

change very quickly when the sun becomes obscured

source driving the grid. The system inertia is

by passing cloud cover. On small power systems with

typically described using a normalized inertia

high penetrations of PV generation, this can cause

constant called the H constant, defined as

serious problems with power delivery, as traditional

thermal units struggle to maintain the balance of

power in the face of rapid changes. During solar -

H can be estimated by the frequency response of the

coupled operation, the BESS must counteract quick

system after a step-change such as a unit or load trip.

changes in output power to ensure that the facility

The equation can be re-written so that the system H is

delivers ramp rates deemed acceptable to the system

easily calculated from the change in frequency of the

operator. Allowable ramp rates are typically speci-

system after a generator of known size has tripped

fied by the utility in kilowatts per minute (kW/min),

off, according to

and are a common feature of new solar and wind

1
=2

power purchase agree-ments between utilities and

independent power producers. Here the ramp rate

1
2

refers only to real power, and that the reactive power


capabilities of the BESS can be dispatched simultane-

where the unit of H is seconds,

is system angular

ously and independently to achieve other power

speed,

system goals.

remaining generation online after the unit trip, and

Ramp Rate Control algorithm used in the XP-DPR

is the system frequency,

is the

is the size of the generator that has tripped.

continuously monitors the real power output of the


When frequency crosses a certain threshold, it is

solar generator, and commands the unit to charge or

desirable to command the BESS to charge in the case

discharge such that the total power output to the

of over-frequency events, typically caused by loss of

system is within the boundaries defined by the

load, or to discharge for under-frequency events,

requirements of the utility. The system ramp rate is

which often result when a generator has tripped

maintained to less than 50 kw/min, whereas the solar

offline. Using proportional control to deliver or

resource alone had a maximum second-to- second

absorb power in support of the grid frequency

ramp rate of over 4 MW/min.

stabilization is referred to as droop response, and this

C. Frequency Response

is common behavior in generator governors equipped

Even with ramp- rate control, there are still going

with a speed-droop or regulation characteristic.

to be occasional frequency deviations on the system.

Droop response in a governor is characterized as a

On small, low-voltage systems, it is common to see

proportional controller with a gain of 1/R, with R

frequency deviations of 13 Hz from the nominal 50

defined as

or 60 Hz frequency. Frequency deviation has adverse


effects on many types of loads as well as other

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

201

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Where

ISBN: 378 - 26 - 138420 - 5

is the grid frequency,

frequency dead band, and


Where

is steady-state speed at no load,

steady-state speed at full load, and

is the

is the power

rating of the BESS in KVA.

is

A set of droop characteristic curves for a 1 MW

is the nominal

BESS is depicted in Fig. 3.

or rated speed of the generator. This means that a 5%


droop response should result in a 100% change in
power output when frequency has changed by 5%, or
3 Hz on a 60 Hz power system.
Since the BESS uses a power electronics
interface, there is no inertia or speed in the system,
and we must approximate this desirable behavior
found in thermal generators. The straight forward
implementation is to digitally calculate an offset for
the BESS output power command as response
proportional to the frequency. The response has units
of kW and is determined as

Fig. 3. Frequency droop response curves for 5% response on a


1 MW BESS.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

202

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

IV.

ISBN: 378 - 26 - 138420 - 5

SIMULATION RESULTS

The photovoltaic and battery energy storage system


are combined and connected to the grid and is
simulated in Simulink /MATLAB R2009a.

Fig.4., Results For Solor Power Measured Over 24 Hours

Fig.5., Ramp Rate control to 50 kW/min for a 1 MW photovoltaic installation and a 1.5 MW/1 MWh BESS for a full day

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

203

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

204

ISBN: 378 - 26 - 138420 - 5

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig. 5., depicts the operation of an XP-DPR BESS

[3] C. Hill and D. Chen, Development of a real-

smoothing the volatile power output of a 1 MW solar

time testing environment for battery energy

farm. Here the system ramp rate is maintained to less

storage

than 50 kW/min, whereas the solar resource alone had a

applications, in Proc. IEEE Power Energy Soc.

maximum second-to- second ramp rate of over

Gen. Meeting, Detroit, MI, Jul. 2011.

4MW/min.

systems

in

renewable

energy

[4] A. Nourai and C. Schafer, Changing the


electricity game, IEEE Power Energy Mag.,
vol. 7, no. 4, pp. 4247, Jul./Aug. 2009.

V.

CONCLUSION

[5] R. H. Newnham, W. G. A Baldsing, and A.

Integration of energy storage systems into

Baldsing, Advanced man-agement strategies for

the grid to manage the real power variability of solar

remote-area power-supply systems, J. Power

by providing rate variation control can optimize the

Sources, vol. 133, pp. 141146, 2004.

benefits of solar PV. Using the BESS to provide

[6] C. D. Parker and J. Garche, Battery energy-

voltage stability through dynamic var support, and

storage systems for power supply networks, in

frequency regulation via droop control response

Valve-Regulated Lead Acid Batteries, D. A. J.

reduces integration challenges associated solar PV.

Rand, P. T. Mosely, J. Garche, and C. D. Parker,

Coupling solar PV and storage will drastically

Eds. , Amsterdam, The Netherlands: Elsevier,

increase reliability of the grid, enables more effective

2004, pp. 295326.

grid management, and creates a dispatchable power

[7] N. W. Miller, R. S. Zrebiec, R. W. Delmerico,

product from available resources. Battery energy

and G. Hunt, Design and commissioning of a 5

storage systems can also improve the economics of

MVA, 2.5 MWh battery energy storage, in

distributed solar power generation by reduced need

Proc. 1996 IEEE Power Eng. Soc. Transm.

for cycle traditional generation assets and increasing

Distrib. Conf., pp. 339345.

asset utilization of existing utility generation by

[8] Analysis of a valve-regulated lead-acid battery

allowing the coupled PV solar and BESS to provide

operating in utility en-ergy storage system for

frequency and voltage regulation services.

more than a decade, 2009.


[9] A. Nourai, R. Sastry, and T. Walker, A vision &
strategy for deploy-ment of energy storage in

VI.

REFERENCES

electric utilities, in Proc. IEEE Power En-ergy


Soc. Gen. Meet., Minneapolis, MN, Jul. 2010.

[1] F. Katiraei and J. R. Aguero, Solar PV

[10] P. Kundur, Power System Stability and Control.

Integration Challenges, IEEE Power Energy

New York: Mc-Graw-Hill, 1994, pp. 589594.

Mag., vol. 9, no. 3, pp. 6271, May/Jun. 2011.


[2] N. Miller, D. Manz, J. Roedel, P. Marken, and E.
Kronbeck, Utility scale battery energy storage
systems, in Proc. IEEE Power Energy Soc.
Gen. Meeting, Minneapolis, MN, Jul. 2010.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

205

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

ESTIMATION OF FREQUENCY FOR A SINGLE LINK-FLEXIBLE


MANIPULATOR USING ADAPTIVE CONTROLLER
KAZA JASMITHA
M.E (CONTROL SYSTEMS)

Dr K RAMA SUDHA
PROFESSOR

DEPARTMENT OF ELECTRICAL ENGINEERING


Andhra University,
Visakhapatnam,
Andhra Pradesh.
Abstract- In this paper, it is proposed an adaptive
control procedure for an uncertain flexible
robotic arm. It is employed with fast online
closed loop identification method combined with
an output feedback controller of Generalized
Proportional Integral (GPI).To identify the
unknown system parameter and to update the
designed certainty equivalence, GPI controller a
fast non asymptotic algebraic identification
method is used. In order to examine this method,
simulations are done and results shows the
robustness of the adaptive controller.
Index Terms- Adaptive control, algebraic
estimation,
flexible
robots,
generalized
proportional integral (GPI) control.

I.INTRODUCTION
FLEXIBLE arm manipulators mainly applied:
space robots, nuclear maintenance, microsurgery,
collision control, contouring control, pattern
recognition, and many others. Surveys of the
literature dealing with applications and challenging
problems related to flexible manipulators may be
found in [1] and [2].
The system, which is partial differential
equations (PDEs), is a distributed-parameter system
of infinite dimensions. It makes difficult to achieve
high- level performance for nonminimum phase
behavior. To deal with the control of flexible
manipulators and the modeling based on a truncated
(finite dimensional) model obtained from either the

finite-element method (FEM) or assumed modes


methods, linear control [3], optimal control [4].
Adaptive control [5], sliding-mode control [6],
neural networks [7], or fuzzy logic [8] are the
control techniques. To obtain an accurate trajectory
tracking these methods requires several sensors, for
all of these we need to know the system parameters
to design a proper controller. Here we propose a
new method, briefly explained in [9], an online
identification technique with a control scheme is
used to cancel the vibrations of the flexible beam,
motor angle obtained from an encoder and the
coupling torque obtained from a pair of strain
gauges are measured as done in the work in [10].
Coulomb friction torque requires a compensation
term as they are the nonlinearities effects of the
motor proposed in [11]. To minimize this effect
robust control schemes are used [12]. However, this
problem persists nowadays. Cicero et al. [13] used
neural network to compensate this friction effect.
In this paper, we proposed an output-feedback
control scheme with generalized proportional
integral (GPI) is found to be robust with respect to
the effects of the unknown friction torque. Hence,
compensation is not required for these friction
models. Marquez et al was first untaken this
controller. For asymptotically sable closed-loop
system which is internally unstable the velocity of a
DC motor should be controlled.For this we propose,
by further manipulation of the integral

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

206

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

reconstructor, an internally stable control scheme of


the form of a classical compensator for an angularposition trajectory task of an un-certain flexible
robot, which is found to be robust with respect to
nonlinearities in the motor. Output-feedback
controller is the control scheme here; velocity
measurements, errors and noise are produced which
are not required. Some specifications of the bar are
enough for the GPI control. Unknown parameters
should not affect the payload changes. Hence this
allows us to estimate these parameter uncertainties,
is necessary.
The objective of this paper is unknown
parameter of a flexible bar of fast online closedloop identification with GPI controller.
The author Fliess et al. [15] (see also [16])
feedback-control system s are reliable for state and
constant parameters estimation in a fast (see also
[17], [18]). As these are not asymptotic and do need
any statistical knowledge of the noise corrupting
data i.e. we dont require any assume Gaussian as
noise. This assumption is common in other methods
like maximum likelihood or minimum least squares.
Furthermore, for signal processing [19] and [20]
this methodology has successfully applied.

Fig. 1. Diagram of a single-link flexible arm

This paper is organized as follows. Section II


describes the flexible-manipulator model. Section
III is devoted to explain the GPI-controller design.
In Section IV, the algebraic estimator mathematical
development is explained. Section V describes the

ISBN: 378 - 26 - 138420 - 5

adaptive-control procedure. In Section VI,


simulations of the adaptive-control system are
shown. . Finally, Section VII is devoted to remark
the main conclusions.

II. MODEL DESCRIPTION


A. Flexible-Beam Dynamics
PDE describes the behavior of flexible slewing
beam which is considered as Euler-Bernoulli beam.
Infinite vibration modes involves in this dynamics.
So, reduced modes can be used where only the low
frequencies, usually more significant, are
considered. They are several approaches to reduce
model. Here we proposed:
1) Distributed parameters model where the infinite
dimension is truncated to a finite number of
vibration modes [3]
2) Lumped parameters models where a spatial
discretization leads to a finite-dimensional model.
In this sense, the spatial discretization can be done
by both a FEM [22] and a lumped-mass model [23].
As we developed in [23] a single-link flexible
manipulator with tip mass is modeled, it rotates Zaxis perpendicular to the paper, as shown in Fig. 1.
Gravitational effect and the axial deformation are
neglected. As the mass of the flexible beam is
floating over an air table itallows us to cancel the
gravitational effect and the friction with the surface
of the table. Stability margin of the system,
increases structural damping, a design without
considering damping may provide a valid but
conservative result [24].
In this paper the real structure is studied is made up
of carbon fiber, with high mechanical resistance
and very small density. We considered the studies
is under the hypothesis of small deformation as the
total mass is concentrated at the tip position
because the mass of the load is bigger than that of
the bar, then the mass of the beam can be

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

207

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

neglected. I.e., the flexible beam vibrates with the


fundamental mode; as the rest modes are very far
from the first so they can be neglected. Therefore
we can consider only one mode of vibration. Here
after the main characteristics of this model will
influence the load changes in very easy manner,
that make us to apply adaptive controller easily.
Based on these considerations, we propose the
following model for the flexible beam:
mL2t c( m t )
(1)

Motor servo amplifier system [in Newton meter


per volts] electromechanical constant is defined by

parameter k . m And m are the angular acceleration


of the motor [in radians per seconds squared] and
the angular velocity of the motor [in radians per
second], respectively. is the coupling torque
measured in the hub [in Newton meters], and n is
the reduction ratio of the motor gear. u is the motor
input voltage [in volts]. This is the control variable
of the system. It is given as the input to a servo
amplifier which controls the input current to the
motor by means of an internally PI current
controller [see Fig. 2(a)]. As it is faster than the
mechanical dynamics electrical dynamics are
rejected. Here the servo amplifier can be
considered as a constant relation ke between the

Where m is the unknown mass at the tip


position. L and c = (3EI/L) are the length of the
flexible arm and the stiffness of the bar,
respectively, assumed to be perfectly known. The
stiffness depends on the flexural rigidity EI and
on the length of the bar L. m is the angular
position of the motor gear.1tand t are the
unmeasured angular position and angular
acceleration of the tip, respectively.
B. DC-Motor Dynamics
In
many
control
systems
a
common
electromechanical actuator is constituted by the DC
motor [25]. A servo amplifier with a current inner
loop is supplied for a DC motor. The dynamics of
the system is given by Newtons law

ku jm vm c
(2)
n
Where J is the inertia of the motor [in kilograms
square meters], is the viscous friction coefficient
[in Newton meters seconds], and
c is the

voltage and current to the motor: im Vk e [see Fig.2


(b)], where imis the armature circuit current and
~
keincludes the gain of the amplifier k and R as the
input resistance of the amplifier circuit.

unknown Coulomb friction torque which affects the


motor dynamics [in Newton meters]. This
nonlinear-friction term is considered as a
perturbation, depending only on the sign of the
motor angular velocity. As a consequence,

Coulombs friction, when 0 , follow the model:


( 0)

coul m

c .sign (m )

coul (m 0)
(3)

And when 0
min(ku, coup )(u 0)
c .sign(u )

max(ku,coup )(u 0)

ISBN: 378 - 26 - 138420 - 5

For motor torque must exceed to begin the


movement coul is used as static friction value.

Fig.2.(a)Complete amplifier scheme. (b) Equivalent amplifier


scheme

The total torque given to the motor T is


directly proportional to the armature circuit in the
form T = kmim, where kmis the electromechanical
constant of the motor. Thus, theelectromechanical
constant of the motor servo amplifier system is k =
kekm.

(4)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

208

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

by

C. Complete-System Dynamics

ISBN: 378 - 26 - 138420 - 5

kuc Jm vm c

(10)
The controller to be designed will be robust with
respect to the unknown piecewise constant torque
disturbances affecting the motor dynamics. Then
the perturbation-free system to be considered is the
following:
kuc jm vm
(11)

The dynamics of the complete system, actuated


by a DC motor are formulated by given simplified
model:
mL2t c( m t )

ku Jm vm c
n
c( m t )

(5,6,7)
Equation (5) represents the dynamics of the
flexible beam; (6) expresses the dynamics of the dc
motor; and (7) stands for the coupling torque
measured in the hub and produced by the
translation of the flexible beam, which is directly
proportional to the stiffness of the beam and the
difference between the angles of the motor and the
tip position, restively.

Where K k / n .To specifies the developments,


let A K / J and B v / J .The transfer function of
a DC motor is written as
( s)
A
Gm( s) m

u c ( s) s( s B)

(12)
Fig.3 shows the compensation scheme of the
coupling torque measured in the hub.
The regulation of the load position t(t) to track a
given smooth reference trajectory * t (t ) is desired.

III. GPI CONTROLLER

To synthesis the feedback-control law, we will use


only the measured coupling motor position m and

The flexible-bar transfer function in Laplace


notation from (5) can be re written as
t (s)
2
Gb( s )

m (s) s 2 2
(8)
2 1/2
Where = (c/(mL )) is the unknown natural
frequency of the bar due to the lack of precise
knowledge of m. as done in [10], the coupling
torque can be canceled in the motor by means of
the a compensation term. In the case, the voltage
applied to the motor is of the form

u uc
k .n
(9)

coupling torque . One of the prevailing


restrictions throughout our treatment of the
problem is our desire of not to measure, or
compute on the basis samplings, angular velocities
of the motor shaft or of the load.

Fig. 4. Flexible-link dc-motor system controlled by a two-stage


GPI-controller design

A. Outer loop controller


Consider the model of the flexible link, given
in (1). This
subsystem is flat, with flat output given by
t t . This means that all variables of the

Fig.3 compensation of the coupling torque measured in the hub

Where uc is the voltage applied before the

unperturbed system may be written in terms of


the flat output and a finite number of its time
derivatives (See [9]). The parameterization of

compensation term. The system in (6) is then given

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

209

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

m t in terms of m t is given, in reduction

gear terms, by:

ISBN: 378 - 26 - 138420 - 5

Disregarding the constant error due to the


tracking error velocity initial conditions, the
estimated error velocity can be computed in the
following form:

mL2
t t
c

e t

(13)
System (9) is a second order system in which it
is desired to regulate the tip position of the
flexible bar t , towards a given smooth

c
mL2

(18)

e e d
m

The integral reconstructor neglects the possibly


nonzero initial condition e t 0 and, hence, it
exhibits a constant estimation error. When the
reconstructor is used in the derivative part of
the PID controller, the constant error is suitably
compensated thanks to the integral control
action of the PID controller. The use of the
integral reconstructor does not change the
closed loop features
of the proposed PID controller and, in fact, the
resulting characteristic polynomial obtained in
both cases is just the same. The design gains
k 0 , k1, k 2 need to be changed due to the use of

reference trajectory * t t with m acting as a


an auxiliary Control input. Clearly, if there
exists an auxiliary open loop control input,

* m t , that ideally achieves the tracking of


*t t for suitable initial conditions, it satisfies
then the second orderdynamics, in reduction
gear terms (10).

mL2 *
m t
t t *t t
c
*

(14)
Subtracting (10) from (9), we obtain an
expression in terms of the angular tracking
errors:
c
e t
e m e t
mL2
(15)

the integral reconstructor. Substituting the

Where e m m * m t , e t t * t t . For

s 0 *
*m 1
(19)
t t
s2
The tip angular position cannot be measured,
but it certainly can be computed from the
expression relating the tip position with the
motor position and the coupling torque. The
implementation may then be based on the use
of the coupling torque measurement. Denote
the coupling torque by it is known to be
given by:

integral reconstructor e t (14) byinto the PID


controller (12) and after some rearrangements
we obtain:

this part of the design, we view em as an


incremental
control
input
for
the
linksdynamics. Suppose for a moment we are
able to measure theangular position velocity
tracking error e t , then the outer loopfeedback
incremental controller could be proposed to be
the followingPID controller,

em et

mL2
K2et K1et K0 et d
c
0

(16)

c m t mL2 n coup

We proceed by integrating the expression (11)


once, to obtain:
c
e t t e t 0
mL2

(20)

Thus, the angular position is readily expressed


as,

e e d
m

1
t m
c

(21)

(17)
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

210

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

In Fig. 2 depicts the feedback control scheme


under which the outer loop controller would be
actually implemented in practice. The closed
outer loop system in Fig. 2 is asymptotically
exponentially stable. To specify the
parameters, 0 , 1 , 2 we can choose to locate

The open-loop control u * c (t ) that ideally


achieves the open-loop tracking of the inner loop
is given by
1
B
u * c (t ) * m (t ) m (t ) .
(27)
A
A
The inner loop system in Fig. 4 is
exponentially stable. We can choose to place all
the closed-loop poles in a desired location of the
left half of the complex plane to design the
parameters
{ 3 , 2 , 1 , 0 }. As done with the outer loop, all

the closed loop poles in the left half of the


complex plane. All three poles can be located
in the same point of the real line, s = a, using
the following polynomial equation,
(22)
s 3 3as 2 3 a 2 s a 3 0
Where the parameter a represents the desired
location of the poles. The characteristic
equation of the closed loop system is,

s 3 k 2 s 2 0 1 k1 s 0 k 2 k 0 0 (23)
Identifying each term of the expression (18)
with those of (19), the design parameters
2 , 1 , 0 can be uniquely specified.
2

poles can be located at the same real value,


3 , 2 , 1 , and 0 can be uniquely obtained by
equalizing the terms of the two following
polynomials:

B. Inner loop controller

(s p) 4 s 4 4 ps 3 6 p 2 s 2 4 p 3 s p 4 0
(28)
4
3
2
s ( 3 B ) s ( 3 B 2 A) s 1 As 0 A 0
(29)
Where the parameter p represents the common
location of all the closed-loop poles, this being
strictly positive.

The angular position m , generated as an auxiliary


control input in the previous controller design
step, is now regarded as a reference trajectory for
the motor controller. We denote this reference
trajectory by * mr .
The dynamics of the DC motor, including the
Coulomb friction term, is given by (10). The
design of the controller to be robust with respect
to this torque disturbance is desired.
The following feedback controller is proposed:

IV. IDENTIFICATION
As explained in the previous section, the
control performance depends on the knowledge of
the parameter . In order to do this task, in this
section, we analyze the identification issue, as
well as the reasons of choosing the algebraic
derivative method as estimator.
Identification
of
continuous-time
system
parameters has been studied from different points
of view. The surveys led by Young in [27] and
Unbehauen and Rao in [29] and [30], respectively,
describe most of the available techniques.
The different approaches are usually classified
into two categories:1) Indirect approaches: An
equivalent discrete-time model to fit the date is

k
e

k
e

k
3 m
2 m
1 e m ( ) d ( )
v
J
0

ev e m
t t

K
K
k 0 (e m ( 2 ))d ( 2 )d ( 1 )

0 0

(24)
the following integral reconstructor for the
angular-velocity error signal e is obtained:
m

K
v
e m ev ( ) d ( ) e m .
J 0
J

ISBN: 378 - 26 - 138420 - 5

rearrangements, the feedback control law is


obtained
s 2 1s 0
*
(u c u * c ) 2
( mr m ) . (26)
s( s 3 )

(25)

Replacing e in (25) into (24) and, after some


m

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

211

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

some arbitrarily small time t = >0, by means of


the
expression.
arbitary
for
t [0, )

2
oe ne (t )
(30)
t [,)
d e (t ) for

needed. After that, the estimated discrete-time


parameters are transferred to continuous time.2)
Direct approaches: In the continuous-time model,
the original continuous-time parameters from the
discrete-time
data
are
estimated
via
approximations for the signals and operators. In
the case of the indirect method, a classical wellknown theory is developed (see [31]).
Nevertheless, these approaches have several
disadvantages: 1) They require computationally
costly minimization algorithms without even
guaranteeing con-vergence. 2) The estimated
parameters may not be correlated with the
physical properties of the system. 3) At fast
sampling rates, poles and zeros cluster near the
1 point in the z-plane.
Therefore, many researchers are doing a big
effort following direct approaches (see [32][35],
among others). Unfortunately, identification of
robotic systems is generally focused on indirect
approaches (see [36], [37]), and as a consequence,
the references using direct approaches are scarce.
On the other hand, the existing identification
techniques, included in the direct approach, suffer
from poor speed performance.Additionally, it is
well known that the closed-loop identification is
more complicated than its open-loop counterpart
(see [31]). These reasons have motivated the
application of the algebraic derivative technique
previously presented in the introduction. In the
next point, algebraic manipulations will be
shownto develop an estimator which stems from
the differentialequations, analyzed in the model
description, incorporating the measured signals in
a suitable manner.
A. Algebraic Estimation of the Natural
Frequency.
In order to make more understandable the
equation deduction, we suppose that signals are
noise free. The main goal is to obtain an
estimation of 2 as fast as possible, which we
will denote by oe .

Where ne (t ) and de (t ) are the output of the


time-varying linear unstable filter
d e (t ) z 3
ne (t ) t 2 t (t ) z1

z1 z 2 4t t (t )
z2 2t (t )

z3 z 4
(31)

z4 t 2 ( m (t ) t (t ))

Proof: consider (5)


t 2 ( m t )

(32)

The Laplace transform of (32) is


s 2 t ( s ) s t (0) t (0) 2 ( m ( s ) t ( s )) (33)
Taking two derivatives with respect to the
complex variable s, the initial conditions are
cancelled
2
d 2 ( s 2 t )
d 2 ( t )
2 d ( m )

2
ds 2

ds 2
ds

(34)
Employing the chain rule, we obtain
s2

d 2 ( t )
d
d 2 ( m ) d 2 ( t )
4s t 2 t 2 (

)
2
ds
ds
ds 2
ds 2

(35)
Consequently, in order to avoid multiplications
by positive powers of s, which are translated as
undesirable time derivatives in the time domain,
we multiply the earlier expression by s 2 . After
some rearrangements, we obtain
d 2 ( t )
d
4 s 1 t 2 s 2 t
2
ds
2 ds
2
d ( m ) d 2 (t )
s 2 (

)
ds 2
ds 2
(36)
Let L denote the usual operational calculus
transform acting on exponentially bounded
signals with bounded left support (see[38]).
Recall that

L1 s() (d / dt )(), L1 (d v / ds v )() (1) v .t v (),


t

and L1(1 / s ) () ()( ) d .

taking

this

into

Proposition 4.1: The constant parameter 2 of


the noise free system described by (5)(7) can be
exactly computed, in a nonasymptotic fashion, at

account, we can translate(36) into the time


domain

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

212

www.iaetsd.in

INTERNATIONAL CONFERENCE
ON CURRENT
INNOVATIONS IN ENGINEERING AND TECHNOLOGY
t
t

[t t (t ) 4 t ( )d 2 t ( )dd
0

0 0

t t

t t
2

( )d d 2 t ( )dd

(37)
The time realization of (37) can be written via
time-variant linear (unstable) filters
d e (t ) z 3
ne (t ) t 2 t (t ) z1
0 0

0 0

z1 z 2 4t t (t )
z2 2t (t )

ISBN: 378 - 26 - 138420 - 5

Taking advantage of the estimator rational


form in (37), the quotient will not be affected
by the filters. This invariance is emphasized
with the use of the different notations in
frequency and time domain such as
n (t ) F (s )ne (t )
2 oe f

d f (t ) F ( s)d e (t )
(40)
Where n f (t ) and d f (t ) are the filtered

z3 z 4

z 4 t 2 ( m (t ) t (t ))

numerator and denominator, respectively, and


F(s) is the filter used. The choice of this filter
depends on the a priori available knowledge of
the system. Nevertheless, if such a knowledge
does not exist, pure integrations of the form ,

The natural frequency estimator of 2 is given by


arbitary
for
t [0, )

2 oe ne (t )
t [,)
d e (t ) for
(38)

1/ s p , p 1 may be utilized, where highfrequency noise has been assumed. This


hypothesis has been motivated by recent
developments in nonstandardanalysis toward a
new nonstochastic noise theory (more details
in [39]).
Finally, the parameter 2 is obtained by
arbitary
for
t [0, )

2
oe ne (t )
t [,)
d e (t ) for
(41)

Where ia an arbitrary small real number.


Note that, for the time t=o,ne(t) and de(t) are
both zero. Therefore, the quotient is undefined
for a small period of time. After a time t = >0,
the quotient is reliably computed. Note that t =
depends on the arithmetic processor precision
and on the data acquisition card. The unstable
nature of the linear systems in perturbed
Brunovskys form (38) is of no practical
consequence on the determination of the
unknown parameters since we have the following
reasons: 1) Resetting of the unstable timevarying systems and of the entire estimation
scheme is always possible and, specially, needed
when the unknown parameters are known to
undergo sudden changes to adopt new constant
values. 2) Once the parameter estimation is
reliably accomplished, after some time instant t =
>0, the whole estimation process may be
safely turned off. Note that we only need to
measure m and , since t is available according
to (21). Unfortunately, the available signals m
and are noisy. Thus, the estimation precision
yielded by the estimator in (30) and (31) will
depend on the signal-to-noise ratio3 (SNR).
B. Unstructured Noise
We assume that m and are perturbed by an
added noise with unknown statistical
properties. In order to enhance theSNR, we
simultaneously filter the numerator and
denominator by the same low-pass filter.

V.ADAPTIVE CONTROL PROCEDURE


Fig. 4 shows the adaptive-control system
implemented in practice in our laboratory. The
estimator is linked up, from time t0 0[ s] , to
the signals coming from the encoder m and the
pair of strain gauges . Thus, the estimator
begins to estimate when the closed loop begins
to work, and then, we can obtain immediately
the estimate of the parameter. When the natural
frequency of the system is estimated at time t1,
the switch s1 is switched on, and the control
system is updated with this new parameter
estimate. This is done in closed loop and at
real time in a very short period of time. The
updating of the control system is carried out by
substituting by the estimated parameter 0e
in (14) and (19). In fact, the feedforward term
which ideally controls in open loop the inner
loop subsystem u *c [see (27)] also depends on

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

213

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS


IN ENGINEERING AND TECHNOLOGY
*

from the knowledge of the system natural


frequency. Obviously, until the estimator
obtains the true value of the natural frequency,
the control system begins to work with an
initial arbitrary value which we select 0i.
Taking these considerations into account, the
adaptive controller canbe defined as follows.
For the outer loop, (14) is computed as
1
m* (t ) 2 t * (t ) t * (t )
x
(42)
Moreover,(23) is computed as
s 3 2s 2 x 2 (1 1 ) s x( 2 0 ) 0

(43)

For the inner loop only changes the


feedforward term in (27) which depends on the
bounded derivatives of the new *m (t ) in (42).
The variable x is defined as
x oi , t t1

x oe , t t1

ISBN: 378 - 26 - 138420 - 5

103 [N m s], and electromechanical


constant k = 0.21 [(N m)/V]. With these
parameters, A and B of the transfer function of
the dc motor in (12) can be computed as
follows: A = 61.14 [N/(V kg s)] and B=15.15
[(N s)/(kg m)]. The mass used to simulate
the flexible-beam behavior is m=0.03 [kg], the
length L=0.5 [m], and the flexural rigidity is EI
=0.264 [N m2]. According to these
parameters, the stiffness is c=1.584 [N m],
and the natural frequency is = 14.5 [rad/s].
Note that we consider that the stiffness of the
flexible beam is perfectly known; thus, the real
natural frequency of the beam will be estimated
as well as the tip position of the flexible bar.
Nevertheless, it may occur that the value of the
stiffness varies from the real value, and an
approximation is then included in the control
scheme. Such approximated value is denoted
by c0. In this case, we consider that the
computation of the stiffness fits in with the real
value, i.e., c0 c . A meticulous stability

the value because the variable m is obtained

(44, 45)

VI. SIMULATIONS
The major problems associated with the control
of flexible structures arise from the structure is
a distributed parameter system with many
modes, and there are likely to be many
actuators [40][44]. We propose to control a
flexible beam whose second mode is far away
from the first one, with the only actuator being
the motor and the only sensors as follows: an
encoder to measure the motor position and a
pair of strain gauges to estimate the tip
position. The problem is that the high modal
densities give rise to the well-known
phenomenon of spillover [45], where
contributions from the unmodeled modes affect
the control of the modes of interest.
Nevertheless, with the simulations as follows,
we demonstrate that the hypothesis proposed
before is valid, and the spillover effect is
negligible. In the simulations, we consider a
saturation in the motor input voltage in a range
of [10, 10] [in volts]. The parameters used in
the simulations are as follows: inertia J = 6.87
105 [kg m2], viscous friction = 1.041

analysis of the control system under variations


of the stiffness c is carried out in Appendix,
where a study of the error in the estimation of
the natural frequency is also achieved. The
sample time used in the simulations is 1 103
[s]. The value of 119.7 103 [N m]
taken in simulations is the true value estimated
in real experiments. In voltage terms is
c / k 0.57[V ] In order to design the gains of
the inner loop controller, the poles can be
located in a reasonable location of the negative
real axis. If closed-loop poles are located in, for
example, 95, the transfer function of the
controller from (26), that depends on the
location of the poles in closed loop of the inner
loop and the values of the motor parameters A
and B as shown in (28) and (29), respectively,
results in the following expression:

uc u *c 789s 2 5.6.104 s 1.3.106

(46) The
*m m
s( s 365)
feedforward term in (27), which depends on the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

214

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

values of the motor parameters, is computed in


accordance with
u *c 0.02*m (t ) 0.25 *m (t )
(47)

given by
m * (t )

with a natural frequency of the bar given by an


initial arbitrary estimate of oi 9 [rad/s]. The

RESULTS

(48)
1.2

The open-loop reference control input from


(14) in terms of the initial arbitrary estimate of
oi is given by
1 *
*
t (t ) t (t ) 12.3.10 3 * t (t ) *t (t )
2
oi

0.8

(49)

0.6

a n g le in ra d

m (t )

0.4

The desired reference trajectory used for the


tracking problem of the flexible arm is
specified as a Beziers eighth-order
polynomial. The online algebraic estimation of
the unknown parameter , in accordance with
(31), (40), and (41), is carried out in = 0.26 s
[see Fig. 5(a)]. At the end of this small time
interval, the controller is immediately replaced
or updated by switching on the interruptor s1
(see Fig. 4), with the accurate parameter
estimate, given by oe = 14.5 [rad/s]. When the

0.2

-0.2
0

0.5

1.5

2.5
time in sec

3.5

4.5

Without using estimation


1
input
output

0.9
0.8
0.7

controller is updated, s1 is switched off. Fig.


5(b) shows the trajectory tracking with the
adaptive controller. Note that the trajectory of
the tip and the reference are superimposed. The
tip position t tracks the desired trajectory *t

a n g le in r a d

0.6
0.5
0.4
0.3
0.2
0.1

with no steady-state error [see Fig. 5(c)]. In this

figure, the tracking error t * t is shown. The


corresponding transfer function of the new
updated controller is then found to be

m * m 0.3s 25.7

*
s 30
t t

1 *
t (t ) t * (t ) 4.7.103*t (t ) *t (t ) (51)
2
oi

The input control voltage to the dc motor is


shown in
Fig. 5(d), the coupling torque is shown in Fig.
5(e), and the Coulomb friction effect in Fig.
5(f). In Fig. 6, the motor angle m is shown

transfer function of the controller (19), which


depends on the location of the closedloop poles
of the outer loop, 10 in this case, and the
natural frequency of the bar as shown in (22)
and (23), respectively, is given by the
following expression.

m * m 2.7 s 17.7

*
s 30
t t

ISBN: 378 - 26 - 138420 - 5

from (14) in terms of the new estimate oe is

For

0.5

1.5

2.5
time in sec

3.5

4.5

14.5[rad / sec], A 61.14, B 15.15

(50)

The open-loop reference control input * m (t )

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

215

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

REFERENCES

-3

x 10

[1] S. K. Dwivedy and P. Eberhard, Dynamic


analysis of flexible manipulators, a literature
review, Mech.Mach. Theory, vol. 41, no. 7,
pp. 749777, Jul. 2006.
[2] V. Feliu, Robots flexibles: Hacia una
generacin de robots con nuevas prestaciones,
Revista Iberoamericana de Automtica e
Informtica Industrial, vol. 3, no. 3, pp. 2441,
2006.
[3] R. H. Canon and E. Schmitz, Initial
experiments on the end-point control
of a flexible robot, Int. J. Rob. Res., vol. 3, no.
3, pp. 6275, 1984.
[4] A. Arakawa, T. Fukuda, and F. Hara, H
control of a flexible robotics arm (effect of
parameter uncertainties on stability), in Proc.
IEEE/RSJ IROS, 1991, pp. 959964.
[5] T. Yang, J. Yang, and P. Kudva, Load
adaptive control of a single-link flexible
manipulator, IEEE Trans. Syst., Man,
Cybern., vol. 22, no. 1, pp. 8591, Jan./Feb.
1992.
[6] Y. P. Chen and H. T. Hsu, Regulation and
vibration control of an FEMbased single-link
flexible arm using sliding-mode theory, J.
Vib. Control, vol. 7, no. 5, pp. 741752, 2001.
[7] Z. Su and K. A. Khorasani, A neural
network-based controller for a single-link
flexible manipulator using the inverse
dynamics approach, IEEE Trans. Ind.
Electron., vol. 48, no. 6, pp. 10741086, Dec.
2001.
[8] V. G. Moudgal, W. A. Kwong, K. M.
Passino, and S. Yurkovich, Fuzzy learning
control for a flexible-link robot, IEEE Trans.
Fuzzy Syst., vol. 3, no. 2, pp. 199210, May
1995.
[9] J. Becedas, J. Trapero, H. Sira-Ramrez,
and V. Feliu, Fast identification method to
control a flexible manipulator with parameter
uncertainties, in Proc. ICRA, 2007, pp. 3445
3450.
[10] V. Feliu and F. Ramos, Strain gauge
based control of singlelink flexible very light

a n g le in ra d

-1

-2

-3
0

0.5

1.5

2.5
time in sec

3.5

4.5

ISBN: 378 - 26 - 138420 - 5

Error

CONCLUSION
A two-stage GPI-controller design scheme
is proposed in connection with a fast online
closed-loop continuous-time estimator of the
natural frequency of a flexible robot. This
methodology only requires the measurement of
the angular positionof the motor and the
coupling torque. Thus, the computation of
angular velocities and bounded derivatives,
which always introduces noise in the system
and makes necessary the use of suitable lowpass filters, is not required. Among the
advantages of this technique, we find the
following advantages: 1) a control robust with
respect to the Coulomb friction; 2) a direct
estimation of the parameters without an
undesired translation between discrete- and
continuous-time domains; and 3) independent
statistical hypothesis of the signal is not
required, so closedloop operation is easier to
implement. This methodology is well suited to
face the important problem of control
degradation in flexible arms as a consequence
of payload changes. Its versatility and easy
implementation make the controller suitable to
be applied in more than 1-DOF flexible beams
by applying the control law to each separated
dynamics which constitute the complete
system. The method proposed establishes the
basis of this original adaptive control to be
applied in more complex problems of flexible
robotics.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

216

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

weight robots robust to payload changes,


Mechatronics, vol. 15, no. 5, pp. 547571, Jun.
2005.
[11] H. Olsson, H. Amstrm, and C. C. de Wit,
Friction models and friction compensation,
Eur. J. Control, vol. 4, no. 3, pp. 176195,
1998.
[12] V. Feliu, K. S. Rattan, and H. B. Brown,
Control of flexible arms with friction in the
joints, IEEE Trans. Robot. Autom., vol. 9, no.
4, pp. 467475, Aug. 1993.
[13] S. Cicero, V. Santos, and B. de Carvahlo,
Active control to flexible manipulators,
IEEE/ASME Trans. Mechatronics, vol. 11, no.
1, pp. 7583, Feb. 2006.
[14] R. Marquez, E. Delaleau, and M. Fliess,
Commande par pid gnralis
dun moteur lectrique sans capteur
mecanique,
in
Premire
Confrence
Internationale Francophone dAutomatique,
2000, pp. 521526.
[15] M. Fliess and H. Sira-Ramrez, An
algebraic framework for linear identification,
ESAIM Control Optim. Calculus Variations,
vol. 9, pp. 151168, 2003.
[16] M. Fliess, M. Mboup, H. Mounier, and H.
Sira-Ramrez, Questioning Some Paradigms of
Signal Processing via Concrete Examples.
Mxico, Mxico: Editorial Lagares, 2003, ch.
1.
[17] J. Reger, H. Sira-Ramrez, andM. Fliess,
On non-asymptotic observation of nonlinear
systems, in Proc. 44th IEEE Conf. Decision
Control, Sevilla, Spain, 2005, pp. 42194224.
[18] H. Sira-Ramrez and M. Fliess, On the
output feedback control of a synchronous
generator, in Proc. 43rd IEEE Conf. Decision
Control, The Bahamas, 2004, pp. 44594464.
[19] J. R. Trapero, H. Sira-Ramrez, and V.
Feliu-Batlle, An algebraic frequency estimator
for a biased and noisy sinusoidal signal,
Signal Process., vol. 87, no. 6, pp. 11881201,
Jun. 2007.
[20] J. R. Trapero, H. Sira-Ramrez, and V.
Feliu-Batlle, A fast on-line frequency

ISBN: 378 - 26 - 138420 - 5

estimator of lightly damped vibrations in


flexible structures, J. Sound Vib., vol. 307, no.
1/2, pp. 365378, Oct. 2007.
[21] F. Bellezza, L. Lanari, and G. Ulivi,
Exact modeling of the flexible slewing link,
in Proc. IEEE Int. Conf. Robot. Autom., 1991,
vol. 1, pp. 734804.
[22] E. Bayo, A finite-element approach to
control the end-point motion of a single-link
flexible robot, J. Robot. Syst., vol. 4, no. 1, pp.
6375, Feb. 1987.
[23] V. Feliu, K. S. Rattan, and H. Brown,
Modeling and control of single-link flexible
arms with lumped masses, J. Dyn. Syst. Meas.
Control, vol. 114, no. 7, pp. 5969, 1992.
[24] W. Liu and Z. Hou, A new approach to
suppress spillover instability in structural
vibration, Struct. Control Health Monitoring,
vol. 11, no. 1, pp. 3753, Jan.Mar. 2004.
[25] R. D. Begamudre, Electro-Mechanical
Energy
ConversionWith
Dynamics
of
Machines. New York: Wiley, 1998.
[26] H. Sira-Ramrez and S. Agrawal,
Differentially Flat Systems. NewYork: Marcel
Dekker, 2004.
[27] P. C. Young, Parameter estimation for
continuous-time
modelsA
survey,
Automatica, vol. 17, no. 1, pp. 2329, Jan.
1981.
[28] H. Unbehauen and G. P. Rao,
Continuous-time approaches to system
identificationA survey, Automatica, vol. 26,
no. 1, pp. 2335, Jan. 1990.
[29] N. Sinha and G. Rao, Identification of
Continuous-Time Systems. Dordrecht, The
Netherlands: Kluwer, 1991.
[30] H. Unbehauen and G. Rao, Identification
of Continuous Systems. Amsterdam, The
Netherlands: North-Holland, 1987.
[31] L. Ljung, System Identification: Theory
for the User, 2nd ed. Englewood Cliffs, NJ:
Prentice-Hall, 1999.
[32] S. Moussaoui, D. Brie, and A. Richard,
Regularization aspects in continuous-time

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

217

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

model identification, Automatica, vol. 41, no.


2, pp. 197208, Feb. 2005.
[33] T. Sderstrm and M. Mossberg,
Performance evaluation of methods for
identifying continuous-time autoregressive
processes, Automatica, vol. 36, no. 1, pp. 53
59, Jan. 2000.
[34] E. K. Larsson and T. Sderstrm,
Identification
of
continuous-time
AR
processes from unevenly sampled data,
Automatica, vol. 38, no. 4, pp. 709718, Apr.
2002.
[35] K. Mahata and M. Fu, Modeling
continuous-time processes via input-tostate
filters, Automatica, vol. 42, no. 7, pp. 1073
1084, Jul. 2006.
[36] R. Johansson, A. Robertsson, K. Nilsson,
and M. Verhaegen, Statespace system
identification of robot manipulator dynamics,
Mechatronics, vol. 10, no. 3, pp. 403418, Apr.
2000.
[37] I. Eker, Open-loop and closed-loop
experimental on-line identification of a three
mass electromechanical system,Mechatronics,
vol. 14, no. 5, pp. 549565, Jun. 2004.
[38] J. Mikusinski and T. Boehme, Operational
Calculus, 2nd ed., vol. I.New York: Pergamon,
1987.
[39] M. Fliess, Analyse non standard du
bruit, C.R. Acad. Sci. Paris, 2006. p. Ser. I
342.
[40] T. nsay and A. Akay, Vibration
reduction of a flexible arm by time optimal
open-loop control, J. Sound Vib., vol. 147, no.
2, pp. 283300, 1991.
[41] D. Sun, J. Shan, Y. Su, H. H. T. Liu, and
C. Lam, Hybrid control of a rotational flexible
beam using enhanced pd feedback with a
nonlinear differentiator and PZT actuators,
Smart Mater. Struct., vol. 14, no. 1, pp. 6978,
Feb. 2005.
[42] J. Shan, H. T. Liu, and D. Shun, Slewing
and vibration control of a single link flexible
manipulator by positive position feedback

ISBN: 378 - 26 - 138420 - 5

(ppf), Mechatronics, vol. 15, no. 4, pp. 487


503, May 2005.
[43] C. M. A. Vasques and J. D. Rodrigues,
Active vibration control of smart piezoelectric
beams: Comparison of classical and optimal
feedback control strategies, Comput. Struct.,
vol. 84, no. 22/23, pp. 14021414, Sep. 2006.
[44] S. S. Han, S. B. Choi, and H. H. Kim,
Position control of a flexible gantry robot arm
using smart material actuators, J. Robot. Syst.,
vol. 16, no. 10, pp. 581595, Oct. 1999.
[45] M. J. Balas, Active control of flexible
systems, J. Optim. Theory Appl., vol. 25, no.
3, pp. 415436, Jul. 1978.
[46] E. G. Christoforou and C. Damaren, The
control of flexible link robots manipulating
large payloads: Theory and experiments, J.
Robot. Syst., vol. 17, no. 5, pp. 255271, May
2000.
[47] C. J. Damaren, Adaptive control of
flexible manipulators carrying large uncertain
payloads, J. Robot. Syst., vol. 13, no. 4, pp.
219288,Apr. 1996.
[48] V. Feliu, J. A. Somolinos, and A. Garca,
Inverse dynamics based control system for a
three-degree-of-freedom flexible arm, IEEE
Trans. Robot. Autom., vol. 19, no. 6, pp. 1007
1014, Dec. 2003.
[49] K. Ogata, Modern Control Engineering,
3rd ed. Englewood Cliffs, NJ: Prentice-Hall,
1996.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

218

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

MODELLING OF ONE LINK FLEXIBLE ARM MANIPULATOR USING


TWO STAGE GPI CONTROLLER
B Sudeep

Dr.K.Rama Sudha

ME Control systems

Professor

Department Electrical & Electronics Engineering,


Andhra University,
Visakhapatnam,
Andhra Pradesh
AbstractIn this article, a two stage Generalized

as Coulomb friction effect. With the controller proposed

Proportional Integral type (GPI), Controller is

no estimation of this nonlinear phenomena

designed for the control of an uncertain flexible robotic

is therefore required. This is a substantial improvement

arm with unknown internal parameters in the motor

over existing control schemes based on the on-line

dynamics. The GPI controller is designed using a two-

estimation of the friction parameters. Developments for

stage design procedure entitling an outer loop,

the controller are based on two concepts namely flatness

designed under the singular perturbation assumption

based exact feed forward linearization and Generalized

of no motor dynamics; and subsequently an inner loop

Proportional Integral (GPI) control. As a result using the

which forces the motor response to track the control

GPI control method, a classical compensating second

input position reference trajectory derived in the

order network with a roll-off zero bestowing a rather

previous design stage. For both, the inner and the

beneficial integral control action with respect to constant

outer loop, the GPI controller design method is easily

load perturbations is obtained. The control scheme

implemented to achieve the desired tracking control.

proposed in this article is truly an output feedback


controller since it uses only the position of the motor.

Key wordsFlexible Arm manipulator, trajectory

Velocity measurements, which always introduce errors in

tracking, generalized proportional integral (GPI)

the signals and noises and makes it necessary the use of

control.

suitable low pass filters, are not required. Furthermore, the


only measured variables are the motor and tip position.

I. INTRODUCTION

The goal of this work is to control a very fast system, only

In this paper, a two stage GPI controller is proposed for

knowing the stiffness of the bar, without knowing the rest

the regulation of an uncertain single-link flexible arm with

of the parameters in the system .A brief outline of this

unknown mass parameter at the tip, motor inertia, viscous

work is the following: Section II explains the system and

friction and electromechanical constant in the motor where

the

the tracking of a trajectory must be too precise. As in Feliu

proposed. An on-line closed loop algebraic identifier is

and Ramos [1] a two stage design procedure is used but

presented in the Section III, yielding expressions for DC

using now a GPI controller viewpoint the particular

motor and flexible bar parameters. This results will be

requirement of robustness with respect to unknown

verified via simulation in Section IV. Finally, the last

constant torque on the motor dynamics, this is known

section is devoted to main conclusions and further works.

generalized

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

219

proportional

integrator

controller

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

II. MODEL DESCRIPTION


Consider the following simplified model of a very

III. GENERALIZED PROPORTIONAL

lightight flexible link, with all its mass concentrated at the

INTEGRATOR CONTROLLER

tip, actuated by a DC motor, as shown in Fig. 1. The

In Laplace transforms notation, the flexible bar transfer

dynamics of the system is given by:

function, obtained from (1), can be written as follows,

mL2t c m t

(1)

ku J m V m T c T coup

T coup

Gb s t
2 0 2
m s s 0
2

(2)

c
m t
n

(4)

(3)
Where

c
mL

is

the

unknown

natural

where m and L are the mass in the tip position and the

frequency of the bar due to the lack of precise knowledge

length of the flexible arm, respectively, assumed to be

of m and L. It is assumed that the constant c is perfectly

unknown, and c is the stiffness of the bar, which is

known. As it was done in [1] the coupling torque can be

assumed to be perfectly known, J is the inertia of the

compensated in the motor by means of a feed-forward

motor,

term which allows to decouple the two dynamics, the

V the viscous friction coefficient, T c is the

unknown Coulomb friction torque, T

coup

flexible link dynamics and the bar dynamics, which allows

is the

an easier design task for the controller since the two

measured coupling torque between the motor and the link,

dynamics can be studied separately. In this case the

k is the known electromechanical constant of the motor,

voltage applied to the motor is of the form,

u is the motor input voltage,

stands for the acceleration

T coup
u uc
K

of the motor gear, m is the velocity of the motor gear.


The constant factor
gea.Thus,

n is the Reduction ratio of the motor

(5)

/n

where uc is the voltage applied before the feed-forward

is the angular position of

term. The system in (2) is then given by:

the motor and t is the unmeasured angular position of


the tip.

ku c J m V m T c

(6)

Where T c is a perturbation, depending only on the sign of


the angular velocity. It is produced by the Coulombs
friction phenomenon. The controller to be designed will be
robust with respect to unknown piecewise constant torque
disturbances affecting the motor dynamics. Then the
perturbation free system to be considered is the following:

ku c Jm Vm

(7)

Fig. 1. Diagram of a single link flexible arm

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

220

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

System (9) is a second order system in which it is desired

A K / J,B v/ J

To simplify the developments, let

ISBN: 378 - 26 - 138420 - 5

to regulate the tip position of the flexible bar

The DC motor transfer function is then written as:

a given smooth reference trajectory

m s
A

u c s s s B

proposed

* t t with m

acting as a an auxiliary Control input. Clearly, if there

(8)

exists an auxiliary open loop control input,


ideally achieves the tracking of

The

feed-forward

technique

has

t , towards

* m t , that

*t t for suitable initial

conditions, it satisfies then the second order dynamics, in

been

reduction gear terms (10).

successfully tested in previous experimental works where


it was implemented with direct driven motors [5], [6], and
in motors with reduction gears [1], [7], [8]. It is desired to
regulate the load position
reference trajectory

* m t

t t to track a given smooth

mL2 *
t t * t t
c

(10)

For the synthesis of the


Subtracting (10) from (9), obtain an expression in terms

feedback law only the measured motor position

m and

of the angular tracking errors:

the measured coupling torque T coup are used. Desired

e t

controller should be robust with respect to unknown


constant

torque

disturbances

affecting

the

motor

c
e m e t
mL2

(11)

dynamics. One of the prevailing restrictions throughout


*

t , e t

t * t t . For this

our treatment of the problem is our desire of not to

Where e m m

measure, or compute on the basis samplings, angular

part of the design, view e m as an incremental control

velocities neither of the motor shaft nor of the load.

input for the links dynamics. Suppose for a moment it is


possible to measure the angular position velocity tracking

A. Outer loop controller

error e t , then the outer loop feedback incremental


controller could be proposed to be the following PID

Consider the model of the flexible link, given in (1). This


Sub system is flat, with flat output given by

controller,

t t . This

means that all variables of the unperturbed system may be

em et

written in terms of the flat output and a finite number of its


time derivatives (See [9]). The parameterization of
in terms of

m t

m t is given, in reduction gear terms, by:

mL2
t t
c

mL2
K2et K1et K0 et d
c
0

(12)

Integrating the expression (11) once, to obtain:

e t t e t 0

(9)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

221

c
mL2

e e d
m

(13)

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Disregarding the constant error due to the tracking error

ISBN: 378 - 26 - 138420 - 5

1
t m
c

velocity initial conditions, the estimated error velocity can

(17)

be computed in the following form:


e t

c
mL2

t
m

In Fig. 2 depicts the feedback control scheme under which

(14)

e e d
t

the outer loop controller would be actually implemented in

The integral reconstructor neglects the possibly nonzero

practice. The closed outer loop system in Fig. 2 is

initial condition et 0 and, hence, it exhibits a constant

asymptotically exponentially stable. To specify the


parameters,

0 , 1 , 2 choose to locate the closed loop

poles in the left half of the complex plane. All three poles
estimation error. When the reconstructor is used in the

can be located in the same point of the real line, s = a,

derivative part of the PID controller, the constant error is


using the following polynomial equation,

suitably compensated thanks to the integral control action

s 3 3as 2 3a 2 s a 3 0

of the PID controller. The use of the integral reconstructor

(18)

does not change the closed loop features


of the proposed PID controller and, in fact, the resulting
characteristic polynomial obtained in both cases is just the

same. The design gains k 0 , k1, k 2

need to be changed

Where the parameter a represents the desired location of


the poles. The characteristic equation of the closed loop

due to the use of the integral reconstructor. Substituting

system is,

the integral reconstructor e t

s 3 k 2 s 2 0 1 k1 s 0 k 2 k 0 0
2

(14) by into the PID

controller (12) and after some rearrangements :

Identifying each term of the expression (18) with those of


(19), the design parameters

s 0 *
*m 1
t t
s2

(15)

2 , 1 , 0 can be uniquely

specified.

(15)

B. Inner loop controller


The tip angular position cannot be measured, but it
certainly can be computed from the expression relating the

The angular position

tip position with the motor position and the coupling

input in the previous controller design step, is now

torque. The implementation may then be based on the use

regarded as a reference trajectory for the motor controller.

of the coupling torque measurement. Denote the coupling

Denote this reference trajectory by

torque by it is known to be given by:

* mr .

The dynamics of the DC motor, including the Coulomb

c m t mL2 n coup

m generated as an auxiliary control

(16)

friction term, is given by (6). It is desired to design the


controller to be robust with respect to this torque

Thus, the angular position is readily expressed as,

disturbance. A controller for the system should then


include a double integral compensation action which is

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

222

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

capable of overcoming ramp tracking errors. The ramp


error is mainly due to the integral angular velocity
reconstructor, performed in the presence of constant,or
piece-wise constant, torque perturbations characteristicof
the Coulomb phenomenon before stabilization around zero
velocity. The integral reconstructor is hidden in the GPI
control scheme.
Fig. 2. Flexible link dc motor system controlled by a two

The following feedback controller is proposed.

stage GPI controller design


t

v
J
ev em k3em k2em k1 em ( )d( )
K
K
0
(20)

The closed inner loop system in Fig. 2 is asymptotically


exponentially

t t

stable.

To

design

the

parameters

3 , 2 , 1 , 0 choose to place the closed loop poles in

k0 (em ( 2 ))d( 2 )d(1)


00

a desired location of the left half of the complex plane. All

In order to avoid tracking error velocity measurements

poles can be located at the same real value, using the

again obtain an integral reconstructor for the angular

following polynomial equation,

velocity error signal

( s p ) 4 s 4 4 ps 3 6 p 2 s 2 4 p 3 s p 4 0 (24)
t

K
v
e m ev ( )d ( ) e m
J 0
J

(21)
Where the parameter p represents the common location of
all the closed loop poles. The characteristic equation of the

Replacing

e m (21)

into

(17)

and

after

closed loop system is,

some

rearrangements the feedback control law is obtained as:

s 4 (3 B)s3 (3 B 2 A)s 2 1 As 0 A 0 (25)

2 s 2 1 s 0
*
(uc u c )
( mr m )
s( s 3 )
*

(22)

Identifying the corresponding terms of the equations (24)


and (25), the parameters

3 , 2 , 1 , 0

may be

uniquely obtained.

The new controller clearly exhibits an integral action


which is characteristic of compensator networks that
robustly perform against unknown constant perturbation
inputs. The open loop control

that ideally achieves the

open loop tracking of the inner loop is given by

u *c (t )

1 *
B
m (t ) m (t )
A
A

(23)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

223

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

B. Outer loop

The parameter used for the flexible arm is c = 1.584 (Nm),


being unknown the mass (m) and the length (L). The poles
for the outer loop design are located at 35 in the real axis.
With a natural frequency of the bar given by an initial,
arbitrary, estimate of 0 9

(rad/sec), the transfer

function of the controller is given by the following


expression

m * m 2.7 s 17.7

s 30
*t t

Fig.3 Simulink file for GPI controller

The open loop reference control input

IV. SIMULATIONS

(28)

* m t in (10) is

given by:
A. Inner loop

* m t 0.123* t t * t t

(29)

The parameters used for the motor are given by an initial,


Arbitrary, estimate of:

C. Results

A = 61.14(N/(Vkgs)), B = 15.15((Ns)/(kgm))
The system should be as fast as possible, but taking care of

The desired reference trajectory used for the tracking

possible saturations of the motor which occur at 2 (V).

problem of the flexible arm is specified as a delayed

The poles can be located in a reasonable location of the

exponential function.The controlled arm response clearly

negative real axis. If closed loop poles are located in, say,

shows a highly oscillatory response. Nevertheless, the

60, the transfer function of the controller results in the

controller tries to track the trajectory and locate the arm in

following expression:

the required steady state position.

uc u * c 798s 2 56000s 1300000


(26)

ss 365
*m m
The feed-forward term in (23) is computed in accordance
with,

u * c 0.02* m 0.25 * m

(27)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

224

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Fig.6 Trajectory Tracking with GPI controller

Fig.4 Trajectory Tracking with integral controller

Fig.7 Comparison with different controllers

It can be noticed from Fig.7 that the reference trajectory


tracking error rapidly converges to zero by using a GPI
controller, and thus a quite precise tracking of the desired
Fig.5 Trajectory Tracking with PI controller

trajectory is achieved.

D. Some remarks

The Coulombs friction torque in our system is given by


.

sign( ) where is the Coulombs friction coefficient


m

and

is the motor velocity. The compensation voltage

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

225

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

that it would be required to compensate the friction torque,

REFERENCES

[1] V. Feliu and F. Ramos., Strain gauge based control

as it was proposed in [2] is of about 0.36sign( m ) Volts. In

of single-link flexible very lightweight robots robust to

our feedback control scheme there is no need of a


nonlinear

compensation

term

since

the

ISBN: 378 - 26 - 138420 - 5

payload changes. Mechatronics., vol. 15, pp. 547571,

controller

2004.

automatically takes care of the piecewise constant

[2] H. Olsson, K. Amstrom, and C. C. de Wit., Friction

perturbation arising from the Coulomb friction coefficient.

models and friction compensation, European Journal of


Control, vol. 4, pp. 176195, 1998.

V. CONCLUSIONS

[3]

M. Fliess and H. Sira-Ramrez, An algebraic

framework for linear identification, ESAIM Contr. Optim.


A two stage GPI controller design scheme has been

and Calc. of Variat., vol. 9, pp. 151168, 2003.

proposed and reference trajectory tracking of a single-link

[4]

flexible arm with unknown mass at the tip and parameters

M. Fliess, M. Mboup, H. Mounier, and H. Sira

Ramrez,

of the motor. The GPI control scheme here proposed only

Questioning

some

paradigms

of

signal

processing via concrete examples. Editorial Lagares,

requires the measurement of the angular position of the

Mexico City., 2003, ch. 1 in Algebraic methods in

motor and that of the tip. For this second needed

flatness, signal processing and state estimation, H. Sira-

measurement, a replacement can be carried out in terms of

Ramirez and G. Silva-Navarro (eds).

a linear combination of the motor position and the

[5] Fractional-order Systems and Controls Fundamentals

measured coupling torque provided by strain gauges

and Applications Concepcin A. Monje YangQuan Chen

conveniently located at the bottom of the flexible arm. The

Blas M. Vinagre Dingy Xue Vicente Feliu.

GPI feedback control scheme here proposed is quite robust

[6] , Modeling and control of single-link flexible

with respect to the torque produced by the friction term

arms with lumped

and its estimation is not really necessary. Finally, since the

masses. J Dyn Syst, Meas Control, vol. 114(3), pp. 59

parameters of the GPI controller depend on the natural

69, 1992.

frequency of the flexible bar and the parameters of the

[7] V. Feliu, J. A. Somolinos, C. Cerrada, and J. A.

motor A and B, this method is robust respect to zero mean

Cerrada, A new control

high frequency noises and yields good results, as seen

scheme of single-link flexible manipulators robust to

from digital computer based simulations. Motivated by the

payload changes, J.

encouraging simulation results presented in this article, is

Intell Robotic Sys, vol. 20, pp. 349373, 1997.

proposed that by use of an on-line, non asymptotic,

[8] V. Feliu, J. A. Somolinos, and A. Garca, Inverse

algebraic parameter estimator for this parameters will be

dynamics based control

the topic of a forthcoming publication.

system for a three-degree-of-freedom flexible arm, IEEE


Trans Robotics
Automat, vol. 19(6) pp. 1007-1014, 2003.
[9] H. Sira-Ramrez and S. Agrawal, Differentially flat
systems, Marcel
Dekker, 2004.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

226

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

DESIGN OF A ROBUST FUZZY LOGIC CONTROLLER FOR A SINGLELINK FLEXIBLE MANIPULATOR


K.RINI SOWHARIKA

Dr.R.VIJAYA SANTHI

PG Scholar

Asst.Professor
Department Electrical Engineering,
Andhra University,
Visakhapatnam,
Andhra Pradesh.

Abstract: In this paper, a single link flexible

truncated (finite dimensional) model obtained from

manipulator is designed and controlled using a

either the finite-element method (FEM) or assumed

fuzzy-PID controller. The controller is employed

modes method. These methods require several

to control an uncertain Flexible Robotic Arm and

sensors to be used in order to obtain an accurate

its internal parameters in the motor dynamics. In

trajectory tracking, and many of them also require the

order to verify this method, proposed controller is

knowledge of all the system parameters to design

being tested against the conventional Integral and

properly the controller. Here a new method is

PID controller. The simulation results show the

considered to cancel the vibration of the flexible

robustness of the proposed Fuzzy PID controller

beam

under motor dynamics.

technique with a control scheme in a suited manner,

which

gathers

an

online

identification

with the only measures of both the motor angle

I.

obtained from an encoder and the coupling torque

INTRODUCTION

obtained from a pair of strain gauges as done in the


work in. In that paper, the nonlinearities effect in the

FLEXIBLE arm manipulators span a wide range

motor dynamics, such as the Coulomb friction torque.

of applications: space robots, nuclear maintenance,

Robust control schemes minimized this effect.

microsurgery, collision control, contouring control,

Fuzzy logic is a form of many-valued logic;

pattern recognition, and many others. The system,

it deals with reasoning that is approximate rather than

described by partial differential equations (PDEs), is


a

distributed-parameter

system

of

fixed and exact. Compared to traditional binary sets

infinite

fuzzy logic variables may have a truth value that

dimensions. Its non minimum phase behavior makes

ranges in degree between 0 and 1. Fuzzy logic has

it difficult to achieve high-level performance. Control

been extended to handle the concept of partial truth,

techniques such as linear control, optimal control,


adaptive

control,

sliding-mode

control,

where the truth value may range between completely

neural

true

networks, or fuzzy logic deal with the control of

and

completely

false. Furthermore,

when

linguistic variables are used, these degrees may be

flexible manipulators and the modeling based on a

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

227

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

managed by specific functions. Irrationality can be

are considered. In order to reduce the model, several

described in terms of what is known as the

approaches were proposed: 1) distributed parameters

fuzzjective. Fuzzy Logic Control Technique has been

model where the infinite dimension is truncated to a

control

finite number of vibration modes; and 2) lumped

Techniques. Many researchers have suggested that

parameters models where a spatial discretization

these controllers have the potential for robust control

leads to a finite-dimensional model. In this sense, the

in

spatial discretization can be done by both a FEM and

good

the

replacement

face

of

for

system

conventional

parameter

and

load

uncertainties. It is realized that fuzzy logic control

a lumped-mass model.

can perform very effectively when the operating


conditions change rapidly. These features make up
very attractive for power system applications since
power system is a highly non-linear and chaotic
system.

Fig. 2. Solid model design of flexible joint arm

A single-link flexible manipulator with tip


mass is modeled, as developed in, that can rotate
about the Z-axis perpendicular to the paper, as shown
in Fig. 1. The axial deformation and the gravitational
effect are neglected, because the mass of the flexible
beam is floating over an air table which allows us to
cancel the gravitational effect and the friction with
Fig. 1. Diagram of a single-link flexible arm.

II.

the surface of the table. Since structural damping


always increases the stability margin of the system, a

MODEL DESCRIPTION

design without considering damping may provide a


valid but conservative result.

A. Flexible-Beam Dynamics

The real structure studied in this paper is


made of carbon fiber, with high mechanical

The flexible slewing beam studied in this

resistance and very small density. We study it under

paper is considered to be a EulerBernoulli beam

the hypothesis of small deformations with all its mass

whose behavior is described by a PDE. Its dynamics

concentrated at the tip position because the mass of

involves infinite vibration modes. As the frequency

the load is bigger than that of the bar, then the mass

of those modes increases, its amplitude decreases.

of the beam can be neglected. In other words, the

This means that reduced models can be used, where

flexible beam vibrates with the fundamental mode;

only the low frequencies, usually more significant,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

228

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

therefore, the rest of the modes are very far from the

The dynamics of the complete system,

first one and they can be neglected. Thus, we only

actuated by a dc motor, is given by the following

consider

simplified model:

one

mode

of

vibration.

The

main

= (

characteristic of this model is that the influence of the


load changes can be modeled in a very easy manner,

= (

Based on these considerations, we propose the

where

(3)
+

(4)
(5)

flexible beam;(4) expresses the dynamics of the dc

(1)

motor; and (5) stands for the coupling torque

is the unknown mass at the tip position.

and = (3

Equation (3) represents the dynamics of the

following model for the flexible beam:

= (

thus adaptive controller can be easily applied.

measured in the hub and produced by the translation

/ ) are the length of the flexible arm

of the flexible beam, which is directly proportional to

and the stiffness of the bar, respectively, assumed to

the stiffness of the beam and the difference between

be perfectly known. The stiffness depends on the

the angles of the motor and the tip position,

flexural rigidity

respectively.

and on the length of the bar .

is the angular position of the motor gear.

and

( )

( )=

( )

are the unmeasured angular position and angular


Where

acceleration of the tip, respectively.

=(

(6)

) is the unknown natural frequency

of the bar due to the lack of precise knowledge of .


The coupling torque can be canceled in the motor by
means of a compensation term. In this case, the
voltage applied to the motor is of the form;
=
Where

(7)

is the voltage applied before the

compensation term.
The system in (4) is then given by

Fig.3. compensation of coupling torque in a hub

(8)

B. DC-Motor Dynamics
The controller to be designed will be robust with
respect to the unknown piecewise constant torque

A common electromechanical actuator, in many


control systems, is constituted by the dc motor. The

disturbances affecting the motor dynamics

dc motor used is supplied by a servo amplifier with a

the perturbation-free system to be considered is the

current inner loop control. We can write the dynamic

following:
=

equation of the system by using Newtons second

. Then,

(9)

law.

(2)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

229

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

where = / . To simplify the developments, let


= /

ISBN: 378 - 26 - 138420 - 5

function is then written as

and = / . The dc-motor transfer


( )

( )=

( )

(10)

( ). Suppose for a moment, that we are

able to measure the angular-position tracking error


, then the outer loop feedback incremental

Fig. 3 shows the compensation scheme of

controller could be proposed to be the following PID

the coupling torque measured in the hub.

controller:

track a given smooth reference trajectory

( ) to

The regulation of the load position

( ) is

( )

(13)

desired. For the synthesis of the feedback-control


law, we are using only the measured motor position

In such a case, the closed-loop tracking error

and the measured coupling torque . One of the

evolves, governed by

prevailing restrictions throughout our treatment of the


problem is our desire of not to measure, or compute

( )

=0

(14)

on the basis samplings, angular velocities of the


motor shaft or of the load.

The design parameters { 2, 1, 0} are then


in terms of

The parameterization of

is

chosen so as to render the closed-loop characteristic

given, in reduction-gear terms, by

polynomial into a Hurwitz polynomial with desirable


roots.

(11)

III.

INNER LOOP
CONTROLLER

System (11) is a second-order system in


which to regulate the tip position of the flexible bar

The angular position

towards a given smooth reference trajectory,

( ) is desired, with

auxiliary control input in the previous controller

acting as an auxiliary

design step, is now regarded as a reference trajectory

control input. Clearly, if there exists an auxiliary

open loop control input


the tracking of

for the motor controller. We denote this reference

( )that ideally achieves

trajectory by

( )for suitable initial conditions, it

( ) +

The
()

. The design of the controller to be

desired.

reduction-gear terms.
( )=

robust with respect to this torque disturbance is

satisfies then the second order dynamics, in

, generated as an

(12)

following

feedback

controller

is

proposed:

Subtracting (12) from (11), an expression in

terms of the angular tracking errors is obtained

Where

(13)

( ) ( )

( ) ( ) ( )

(15)

( ),

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

230

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The following integral reconstructor for the angularvelocity error signal is obtained:

( ) ( )

Replacing

IV.

(16)

ISBN: 378 - 26 - 138420 - 5

FUZZY-PID CONTROLLER
Here, A classical FLS is represented as in

Fig-4. As shown, rules play a central role in the FLS

in (25) into (24) and, after

framework. Rules can be provided by experts or can

some rearrangements, the feedback control law is

be extracted from numerical data. The IF-part of a

obtained
(

)=

rule is an antecedent while the THEN-part of a rule is

) (17)

its consequent. Fuzzy sets are associated with terms


that appear both in the antecedent and in the

The open-loop control

) that ideally achieves the

consequent, while Membership Functions (MFs) are

open-loop tacking of the inner loop is given by

)=

( )+

( )

used to describe these fuzzy sets.

(18)

The inner loop system in Fig. 4 is


exponentially stable. We can choose to place all the
closed-loop poles in a desired location of the left half
of

the

complex
,

parameters {

plane
,

to

design

the

}. As done with the outer

loop, all poles can be located at the same real value,


,

and

equalizing

Fig.4., structure of the fuzzy logic controller

can be uniquely obtained by

the

terms

of

the

two

following
Table:1

polynomials:

Control rules of the fuzzy controller


( + ) =

+4

+6

+4

=0

(19)

Input2

+(
=0

+ )

+(

Input1
N
Z
P

N
P
N
N

Z
N
P
N

P
N
P
N

(20)

Where the parameter

Fuzzification: Translates inputs (real values)

represents the common

to fuzzy values. Inference System: Applies a fuzzy

location of all the closed-loops poles, this being

reasoning mechanism to obtain a fuzzy output. Type

strictly positive.

Defuzzificator/Reductor:

The

defuzzificator

transduces one output to precise values; the type


redactor transforms a fuzzy set into a Type-1 fuzzy
set.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

231

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Knowledge Base: Contains a set of fuzzy

and input2, are first fuzzified by two interval t fuzzy

rules, and a membership functions set known as the

sets i.e., positive and negative represented by


(

database. The two normalized input variables, input1

1) and

2) respectively.

BLOCK DIAGRAM OF THE SINGLE LINK FLEXIBLE MANIPULATOR USING FUZZY PID
CONTROL

Fig.5., Block Diagram Of The Single Link Flexible Manipulator Using Fuzzy PID Control

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

232

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

V.

ISBN: 378 - 26 - 138420 - 5

They are;

SIMULATION & RESULTS


The major problems associated with the

1.

Integral controller

control of flexible structures arise from the structure

2.

Conventional PID controller

is a distributed parameter system with many modes,

3.

Fuzzy PID controller

The simulation results of the above controllers and

and there are likely to be many actuators.

given bellow

There are three types of controllers used to


control the system which are simulated and compared
for better results.

I)

Using Integral Controller

Fig.6, the simulation result of integral controller

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

233

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

II)

ISBN: 378 - 26 - 138420 - 5

Using Conventional PID Controller

Fig.7, the simulation result of the conventional pid controller

III)

Using Fuzzy PID Controller

Fig.8, the simulation result of the fuzzy PID controller

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

234

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

IV)

ISBN: 378 - 26 - 138420 - 5

The Combined Figure Of The Output Simulated

Fig.9. The Combined Figure Of The Output Simulated

Therefore, the above results prove that the

proposed controller, as it is being compared with the

system controlled by the fuzzy PID controller has

conventional integral and PID controller.

better n accurate results when compared to other


controllers

VI.

VII.
CONCLUSION:

REFERENCES:

[1] S. K. Dwivedy and P. Eberhard, Dynamic


analysis of flexible manipulators, a literature review,

Here, fuzzy PID controller is designed to control

Mech.Mach. Theory, vol. 41, no. 7, pp. 749777, Jul.

the single link flexible manipulator . The proposed

2006.

controller

the

[2] R. H. Canon and E. Schmitz, Initial experiments

conventional PID controller. The proposed Fuzzy

on the end-point control of a flexible robot, Int. J.

PID controller is tested for different motor dynamics.

Rob. Res., vol. 3, no. 3, pp. 6275, 1984.

Simulation results show the efficiency of the

[3] Y. P. Chen and H. T. Hsu, Regulation and

is

being

designed

based

on

vibration control of an FEM based single-link

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

235

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

flexible arm using sliding-mode theory, J. Vib.

[11] V. Feliu, K. S. Rattan, and H. Brown, Modeling

Control, vol. 7, no. 5, pp. 741752, 2001.

and control of single-link flexible arms with lumped

[4] Z. Su and K. A. Khorasani, A neural-network-

masses, J. Dyn. Syst. Meas. Control, vol. 114, no. 7,

based controller for a single-link flexible manipulator

pp. 5969, 1992.

using the inverse dynamics approach, IEEE Trans.

[12] W. Liu and Z. Hou, A new approach to

Ind. Electron., vol. 48, no. 6, pp. 10741086,

suppress spillover instability in structural vibration,

Dec.2001.

Struct. Control Health Monitoring, vol. 11, no. 1, pp.

[5] V. G. Moudgal, W. A. Kwong, K. M. Passino,

3753, Jan.Mar. 2004.

and S. Yurkovich, Fuzzy learning control for a

[13] R. D. Begamudre, Electro-Mechanical Energy

flexible-link robot, IEEE Trans. Fuzzy Syst., vol. 3,

ConversionWith Dynamics of Machines. New York:

no. 2, pp. 199210, May 1995.

Wiley, 1998.

[6] J. Becedas, J. Trapero, H. Sira-Ramrez, and

[14] H. Sira-Ramrez and S. Agrawal, Differentially

V.Feliu, Fast identification method to control a

Flat Systems. NewYork: Marcel Dekker, 2004.

flexible manipulator with parameter uncertainties, in

[15] P. C. Young, Parameter estimation for

Proc. ICRA, 2007, pp. 34453450.

continuous-time modelsA survey, Automatica,

[7] V. Feliu and F. Ramos, Strain gauge based

vol. 17, no. 1, pp. 2329, Jan. 1981.

control of single link flexible very light weight robots

[16] H. Unbehauen and G. P. Rao, Continuous-time

robust to payload changes, Mechatronics, vol. 15,

approaches to system identificationA survey,

no. 5, pp. 547571, Jun. 2005.

Automatica, vol. 26, no. 1, pp. 2335, Jan. 1990.

[8] H. Olsson, H. Amstrm, and C. C. de Wit,

[17] N. Sinha and G. Rao, Identification of

Friction models and friction compensation, Eur. J.

Continuous-Time

Control, vol. 4, no. 3, pp. 176195, 1998.

Netherlands: Kluwer, 1991.

[9] S. Cicero, V. Santos, and B. de Carvahlo, Active

[18] H. Unbehauen and G. Rao, Identification of

control to flexible manipulators, IEEE/ASME Trans.

Continuous Systems. Amsterdam, The Netherlands:

Mechatronics, vol. 11, no. 1, pp. 7583, Feb. 2006.

North-Holland, 1987.

Systems.

Dordrecht,

The

[10] E. Bayo, A finite-element approach to control


the end-point motion of a single-link flexible robot,
J. Robot. Syst., vol. 4, no. 1, pp. 6375, Feb. 1987.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

236

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

FPGA Implementation Of RF Technology And Biometric Authentication


Based ATM Security

Author-1: K.MOHAN

Author-2: S.VAMSEE KRISHNA

Email.id: [email protected]

Email id: [email protected]

MTECH in VLSI SYSTEM DESIGN

Assistant Professor In Dept Of ECE

Siddarth institute of engineering and technology

Siddarth institute of engineering and technology

puttur.

Puttur.

ABSTRACT
Three-factor authentication method was introduced as
advancement to two-factor authentication schemes in
remote authentication. The three factors used in
authentication are a smart card, password and a
biometric. The authentication is based on the
characteristics of these three factors. To improve the
security in the remote authentication, biometric was
introduced. Due to the uniqueness and the
characteristics of biometrics, they are quite suitable for
user authentication and also reduce the drawbacks
inherited from passwords and smart cards.

something you are is taken into thought


multifactor.
Most early authentication mechanisms area unit
entirely supports secret. whereas such protocols
area unit comparatively simple to implement,
passwords (and human generated passwords in
particular) have several vulnerabilities. As
associate degree example, human generated and
unforgettable passwords area unit typically short
strings of characters and (sometimes) poorly
designated. By exploiting these vulnerabilities,
straightforward wordbook attacks will crack
passwords during a short time [1]. Due to these
issues, hardware authentication tokens area unit
introduced to strengthen the protection in user
authentication, and smart-card-based secret
authentication has become one amongst the
foremost common authentication mechanisms.

Instead of Smart cards we are designed the RF


Technology to identify the account person details.
A face recognition system is a computer automatically
application for identifying or verifying a person from
a digital image from a video. One of the ways to do this
is by comparing selected facial features from the image
and a facial database. With the help of Camera to
detect whether that account person is authorized or
unauthorized. If the authorized person only to access
the account. If unauthorized person we will give certain
intimation given to owners mobile by using the MMS
Modem. The above process will be done by FPGA and
Mat lab.

Smart-card-based Arcanum authentication provides


two-factor authentication, particularly a in login
needs the shopper to own a legitimate smart-card
and an accurate Arcanum.

Keywords: RF Technology, Face Recognition


Method, VLSI.

An authentication issue could be a piece of


knowledge and method wont to demonstrate or
verify the identity of someone or different entity
requesting access below security constraints.
Multifactor authentication (MFA) could be a
system wherever in 2 or a lot of various factors
area unit employed in conjunction to demonstrate.
victimization over one issue is typically referred to
as strong authentication. the method that solicits
multiple answers to challenge queries in addition
as retrieves something you have or something
you are is taken into account multifactor. True
multifactor authentication needs the employment of
resolution from 2 or a lot of the 3 classes of things.

INTRODUCTION
An authentication issue can be a bit of data and
technique accustomed certify or verify the identity
of a personal or completely different entity
requesting access below security constraints. 3
authentication could be a system wherever in 2 or a
lot of various factors area unit wont to demonstrate
the persons. victimization higher than one issue is
typically referred to as strong authentication. the
method of multiple answers to challenge queries
equally as retrieves something you have or

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

237

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

victimization multiple solutions from constant


class wouldn't represent multifactor authentication
. Two-factors or multi-factor authentication is
precisely what it feels like. rather than
victimization just one kind of authentication issue,
like solely things a user is aware of (Login Ids,
passwords, secret pictures, shared secrets, invited
personnel info, etc), two-factor authentication
needs the addition of a second issue, the addition of
one thing the user HAS or one thing the user IS.
Two-factor authentication isn't a replacement idea
particularly within the banking business. Twofactor authentication is employed whenever a bank
client visits their native ATM. One authentication
issue is that the physical ATM cards the client
slides into the machine. The second issue is that
the PIN they enter. while not each, authentication
cannot occur.

ISBN: 378 - 26 - 138420 - 5

countersign mustn't be a word that may be found in


an exceedingly lexicon. A lexicon attack uses a
info of words like a lexicon, attempting all the
words within the info for a match. Its price stating
the plain here[md]attackers have access to
dictionaries in different languages. In different
words, a countersign employing a word from
another language is as straightforward to crack as a
countersign employed in your language. The
common method that tokens area unit used for
authentication is with websites. The user varieties
within the variety displayed within the token on an
internet page. If the user varieties within the same
variety well-known by the server at that point, the
user is echt. Its common to use multifactor
authentication with token-based authentication.
additionally to coming into the quantity displayed
within the token, the user is commonly needed to
enter a username and countersign. This proves they
need one thing (the token), and that they grasp one
thing (their password).

Three issue is that the commonest issue used and


might be a countersign or a straightforward
personal number (PIN). However, it's conjointly
the best to beat. once victimization passwords, its
necessary to use sturdy passwords. a robust
countersign encompasses a mixture of upper-case
letter, lower case, numbers, and special characters.
within the past, security professionals suggested
that passwords ought to be a minimum of eight
characters long. However, with the increasing
strength of countersign bats, its common to listen
to professionals recommending longer passwords.
as an example, several organizations need that
administrator passwords be a minimum of fifteen
characters long.

AUTHENTICATION METHODS:
Token Based Authentication:
The Token primarily based technique class is once
more because the name suggests authentication
supported a TOKEN such as: a key, a magnetic
card, a wise card, a badge and a passport. even as
once someone loses a key, he wouldn't be ready to
open the lock, a user agency loses his token
wouldn't be ready to login, per se the token
primarily based authentication class is kind of
liable to fraud, thieving or loss of the token itself.

Longer passwords area unit more durable to


recollect unless theyre place into some kind of
purposeful order. as an example, a phrase like
Security breeds success will become a
countersign of S3curityBr33d$Succ3$. Notice
that every word starts with a capital, every
minuscule s is modified to a $, every minuscule
e is modified to a three, and also the areas area
unit removed. The countersign is less complicated
to recollect, nonetheless is extremely advanced.
However, if a user is needed to recollect an
extended countersign with none which means, like
1kqd9% lu@7cpw#, they're way more probably
to write down the countersign down, weakening
the protection.

Knowledge Based Authentication


The thought of data based mostly Authentication is
just the utilization of typical passwords, pins or
pictures to achieve access into most laptop systems
and networks. matter (alphabetical) and graphical
user authentications area unit 2 strategies that area
unit presently used. True matter authentication that
uses a username and watchword has inherent
weaknesses and disadvantages which can be
mentioned within the following section.
Inherit Based Authentication
The Inherent primarily based Authentication class
that is additionally called identification, because

Passwords mustn't embody personal knowledge


sort of a users name or username. in addition a

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

238

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

the name suggests, is that the machine-driven


method/s
of
biometric
identification
or
identification supported measurable physiological
or behavioural characteristics like fingerprints,
palm prints, hand pure mathematics, face
recognition, voice recognition and such different
similar strategies. Biometric characteristics square
measure neither duplicatable nor transferable.
they're constant and immutable . so it's close to not
possible to change such characteristics or faux
them. what is more such characteristics cannot be
transferred to different users nor be taken as
happens with tokens, keys and cards. not like the
protection of a users secret, biometric
characteristics, for example the users fingerprint
or iris pattern, aren't any secret. thus there's no
danger of an opening in security.

In proposed algorithm 1st level authentication is


provided by a smart card by using RF Transmitter
and RF Receiver. Whenever authorized frequency
occurs then it automatically opens the camera then
2 Level authentication will be started this will be
done by Face recognition using PCA algorithm
implemented in Mat Lab then if person was
authorized then he forwarded to text based
Password i.e. 3rd level else it automatically send a
MMS to the owner when an unauthorized was
detected and door lock and buzzer will be
activated. If owner replies with a secret code then
the person can access the account.
The second step is that the capturing of a face
image. this may commonly be done employing a
still or video camera. The face image is passed to
the popularity computer code for recognition
(identification or verification). this may commonly
involve variety of steps like normalizing the face
image then making a template of print to be
compared to those within the information. The
match will either be a real match which might
cause investigatory action or it'd be a false
positive which suggests the popularity algorithmic
rule created a blunder and also the alarm would be
off. every component of the system is set at totally
different locations at intervals a network, creating
it simple for one operator to retort to a spread of
systems.

PROPOSED AUTHENTICATION TECHNIQUE:

The information age is quickly revolutionizing the


method transactions square measure completed.
Everyday actions square measure more and more
being handled electronically, rather than with
pencil and paper or face
to
face.
This
Advancement in electronic transactions has
resulted in a very bigger demand for quick and
correct user identification and authentication.
Access codes for buildings, banks accounts and
laptop systems typically use PIN's for identification
and security clearances. exploitation the right PIN
gains access, the booming transactions will occur,
however the user of the PIN isn't verified. once
sensible cards square measure lost or taken, AN
unauthorized user will typically come back up
with the correct personal codes. This paper
describes however face recognition technology will
facilitate to the $64000 world banking machines.

RFID
Recieve

Door
Lock

Webcam

FPGA

PC

ALARM

MMS
Modem

ISBN: 378 - 26 - 138420 - 5

FPGA
An FPGA could be a device that contains a matrix
of reconfigurable gate array logic electronic
equipment. once a FPGA is organized, the inner
electronic equipment is connected in a very means
that makes a hardware implementation of the
software package application. in contrast to
processors, FPGAs use dedicated hardware for
process logic associate degree don't have an
software. FPGAs are actually parallel in nature
therefore totally different process operations don't
need to contend for identical resources. As a result,
the performance of 1 a part of the appliance isn't
affected once further process is additional. Also,
multiple management loops will run on one FPGA
device at totally different rates. FPGA-based
management systems will enforce essential
interlock logic and may be designed to forestall I/O
forcing by associate degree operator. However, in
contrast to hard-wired computer circuit board
(PCB) styles that have fastened hardware

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

239

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

the info out pins of the HT-12D, D0,D1,D2 and


D3. The HT-12D receives the 12-bit word and
interprets the primary 8-bits as address and also the
last 4-bits as information. Pins 1-8 of the HT-12E
area unit the address pins. victimisation the address
pins of the HT-12E, we are able to choose
completely different addresses for up to 256
receivers. The address is set by setting pins 1-8 on
the HT-12E to ground, or simply feat them open.
The address designated on the HT-12E circuit
should match the address designated on the HT12D circuit (exactly), or the knowledge are going
to be unnoticed by the receiving circuit.

resources, FPGA-based systems will virtually wire


their internal electronic equipment to permit
reconfiguration when the system is deployed to the
sphere. FPGA devices deliver the performance and
responsibility of dedicated hardware electronic
equipment.
A single FPGA will replace thousands of separate
parts by incorporating legion logic gates in a very
single computer circuit (IC) chip. the inner
resources of associate degree FPGA chip include a
matrix of configurable logic blocks (CLBs)
enclosed by a boundary of I/O blocks . Signals ar
routed among the FPGA matrix by programmable
interconnect switches and wire routes

When the received addresses from the encoder


matches the decoders, the Valid Transmission pin17 of the HT-12D can go HIGH to point that a
sound transmission has been received and also the
4-bits of information area unit barred to the info
output pins, 10-13. The electronic transistor circuit
shown within the schematic can use the American
state, or valid transmission pin to light-weight the
light-emitting diode. once the American state pin
goes HIGH it activates the 2N2222 electronic
transistor that successively delivers power to the
light-emitting diode providing a visible indication
of a sound transmission reception.

FPGAs contain programmable logic parts referred


to as "logic circuits", and a hierarchy of
reconfigurable interconnects that permit the blocks
to be "wired together" somewhat like several
(changeable) logic gates which will be inter-wired
in (many) totally different configurations. Logic
blocks will be organized to perform advanced
combinatory functions, or just easy logic gates like
AND and XOR. In most FPGAs, the logic blocks
conjointly embody memory components, which
can be easy flip-flops or additional complete blocks
of memory.

Controlling the Project with a FPGA


Using these RF transmitter &amp; receiver circuits
with a FPGA would be easy. we are able to merely
replace the switches used for choosing knowledge
on the HT-12E with the output pins of the FPGA.
conjointly we are able to use another output pin to
pick out TE, or transmit change on the HT-12E.
By taking pin-14 LOW we tend to cause the
transmitter section to transmit the info on pins 1013.

RF ENCODER AND DECODER:


General Encoder and Decoder Operations
The Holtek HT-12E IC encodes 12-bits of {data of
knowledge} and serially transmits this data on
receipt of a Transmit change, or a coffee signal on
pin-14 /TE. Pin-17 the D_OUT pin of the HT-12E
serially transmits no matter information is out there
on pins ten,11,12 and 13, or D0,D1,D2 and D3.
information is transmitted at a frequency
designated by the external generator electrical
device.

To receive info merely attach the HT-12D output


pins to the FPGA. The VT, or valid transmission
pin of the HT-12D might signal the FPGA to grab
the 4-bits of knowledge from the info output pins.
If you're employing a FPGA with interrupt
capabilities, use the Green Mountain State pin to
cause a jump to associate interrupt vector and
method the received knowledge.

By victimisation the switches connected to the info


pins on the HT-12E, as shown within the
schematic, we are able to choose the knowledge in
binary format to send to the receiver. The receiver
section consists of the Ming dynasty RE-99 and
also the HT-12D decoder IC. The DATA_IN pin14 of the HT-12D reads the 12-bit binary info sent
by the HT-12E then places this information on its
output pins. Pins 10, 11,12 and thirteen area unit

The HT-12D knowledge output pins can LATCH


and stay during this state till another valid
transmission is received. NOTE: you may notice
that in each schematics every of the Holtek chips
have resistors hooked up to pins fifteen and
sixteen. These resistors should be the precise
values shown within the schematic. These resistors

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

240

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

set the inner oscillators of the HT-12E/HT-12D. it's


counseled that you simply opt for a tenth electrical
device for every of those resistors to make sure the
right circuit oscillation.
Range of Operation
Figure 4: RF-434 Pin Diagram

The normal operating range using (only)


the LOOP TRACE ANTENNA on the transmitter
board is about 50 feet. By connecting a quarter
wave antenna using 9.36 inches of 22 gauge wire
to both circuits, you can extend this range to
several hundred feet. Your actual range may vary
due to your finished circuit design and
environmental conditions. The transistors and
diodes can be substituted with any common
equivalent type. These will normally depend on
the types and capacities of the particular loads you
want to control and should be selected accordingly
for your intended application.

MMS Modems
A GSM electronic equipment is AN external
electronic equipment device, like the Wavecom
FASTRACK electronic equipment. Insert a GSM
SIM card into this electronic equipment, And
connect the electronic equipment to an offered port
on your laptop. A GSM electronic equipment is a
laptop Card put in an exceedingly pc, like the
Nokia Card Phone.
A GSM electronic equipment may even be a
typical GSM mobile with the suitable cable and
code driver to attach to a port on your laptop.
Phones like the Nokia 7110 with a DLR-3 cable, or
varied Ericsson phones, square measure typically
used for this purpose.

RF DETAILS
The TWS-434 and RWS-434 are
extremely small, and are excellent for applications
requiring short-range RF remote controls. The
transmitter module is only 1/3 the size of a
standard postage stamp, and can easily be placed
inside a small plastic enclosure. TWS-434: The
transmitter output is up to 8mW at 433.92MHz
with a range of approximately 400 foot (open area)
outdoors. Indoors, the range is approximately 200
foot, and will go through most walls.....

A dedicated GSM electronic equipment (external


or laptop Card) is typically preferred to a GSM
mobile. this is often attributable to some
compatibility problems that may exist with mobile
phones. for instance, if you want to be ready to
receive inward MMS messages along with your
entree, and you're employing a mobile as your
electronic equipment, you want to utilize a mobile
that doesn't support WAP push or MMS. this is
often as a result of the mobile mechanically
processes these messages, while not forwarding
them via the electronic equipment interface.
equally some mobile phones won't permit you to
properly receive SMS text messages longer than
one hundred sixty bytes (known as concatenated
SMS or long SMS). this is often as a result of
these long messages are literally sent as separate
SMS messages, and therefore the phone tries to
piece the message before forwarding via the
electronic equipment interface. (Weve ascertained
this latter downside utilizing the Ericsson R380,
whereas it doesn't seem to be a tangle with several
different Ericsson models.)When you install your
GSM electronic equipment, or connect your GSM
mobile to the pc, make certain to put in the suitable
Windows electronic equipment driver from the
device manufacturer. To modify configuration, the
currently SMS/MMS entree can communicate with
the device via this driver. a further advantage of
utilizing this driver is that you simply will use
Windows medical specialty to make sure that the

RF 434 MHz Transmitters. Modulation: ASK


The TWS-434 transmitter accepts both
linear and digital inputs, can operate from 1.5 to 12
Volts-DC, and makes building a miniature handheld RF transmitter very easy. The TWS-434 is
approximately the size of a standard postage stamp.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

241

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

electronic equipment is communication properly


with the pc. The currently SMS/MMS entree will at
the same time support multiple modems, only if
your component has the offered communications
port resources.

Figure 5: RTL view of VHDL Code

GSM smart modem

Previous Analysis:

Preserving security and privacy is a


challenging issue in distributed systems.
This paper makes a step forward in solving
this issue by proposing a generic
framework for three-factor authentication
to protect services and resources from
unauthorized use. The authentication is
based on password, smart-card and
biometrics. Our framework not only
demonstrates how to obtain secure threefactor authentication from two-factor
authentication, but also addresses several
prominent
issues
of
biometric
authentication in distributed systems (e.g.,
client privacy and error tolerance). The
analysis shows that the framework
satisfies all security requirements on threefactor authentication and has several other
practice-friendly properties (e.g., keyagreement, forward security and mutual
authentication). The future work is to fully
identify the practical threats on threefactor authentication and develop concrete
three-factor authentication protocols with
better performances.

Figure 6: Technological Schematic


Whenever the face was detected true then it
automatically opens a login form shown below

ATM Sequrity Page

Results:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

242

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

After Recognition

Unauthorized Person
Whether person is checking as authorized or
unauthorized

8. CONCLUSION:
There are several schemes that manage three-factor
authentication method. however it's a really
troublesome task to get both client aspect and
server aspect security. They additionally tried to
produce privacy of the user biometric. Even though
the theme achieved privacy protection, it couldnt
face up to positive identification attack.
additionally server aspect attack is another crucial
issue in such remote authentication schemes. Face
recognition technologies have been associated
generally with very costly top secure applications.
Hence, our projected theme in all probability
addresses the concerns of user privacy, example
protection and trust problems and gives advantage
of protective data from the user except the
specified identity.

Authorized Person

REFERENCES
[1] D.V. Klein, Foiling the Cracker: A Survey of,
and Improvements to, Password Security, Proc.
Second USENIX Workshop Security, 1990.
[2] A.K. Jain, R. Bole, and S. Pankanti, Eds.,
Biometrics: Personal Identification in Networked
Society, Norwell, MA: Kluwer, 1999.
[3] D. Malone, D. Maio, A. K. Jain, and S.
Prabhakar, Handbook of Fingerprint Recognition
ACM SIGOPS Operating Syst. Rev., vol. 38, no. 4,
pp. 91-96, Oct. 2004
[4] Ed. Dawson, J. Lopez, J. A. Montenegro, and
E. Okamoto, BAAI: Biometric Authentication and
Authorization Infrastructure, Proc. IEEE Intern.
Conference on Information Technology: Research
and Education (ITRE03), pp. 274-278, 2004.
[5] J.K. Lee, S.R. Ryu, and K.Y. Yoo, Fingerprint
Based Remote User Authentication Scheme Using
Smart Cards, Electron. Lett., vol. 38, no. 12, pp.
554-555, Jun. 2002.

Bank Login Page

Remaining Balance

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

243

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[6] C.C. Chang and I.C. Lin, Remarks on


Fingerprint-Based Remote User Authentication
Scheme Using Smart Cards,
[7] C.H. Lin and Y.Y. Lai, A Flexible Biometrics
Remote User Authentication Scheme, Compute.
Standards Interfaces, vol. 27, no. 1, pp. 19-23,
Nov. 2004
[8]3D Face Tracking and Expression Interference
from a 2D sequence Using Manifold Learning:
WeikaiLiao and GerardMedioni,
[9] A. Elgammal. Learning to track: Conceptual
manifoldmapforclosedformtracking.CVPR2005,pp.724730.1
[10] A.Elgammal and.-S. Lee. Inferring
3dbodyposefromsilhouettes
using
activity
manifold learning.CVPR2004,pp.681688
[11]L.GuandT.Kanade.3dalignment of face
single image. CVPR 2006,pp.13051312.

in a

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

244

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Design and Simulation of High Speed CMOS Full Adder

Author 1: V.VINAY KUMAR

Author2: V.VISWANATH
Email id: [email protected]

Email.id: [email protected]
M.TECH in VLSI System Design,

Assistant prof. In Department of ECE,

Siddarth Institute of Engineering and Technology,

Siddarth Institute of Engineering and Technology,

Puttur, Chittor(Dist).

Puttur, Chittor(Dist).

faster operation speed and more regular circuit


structure
The arithmetic circuit is the important core in
electronic systems. If the arithmetic circuit has good
characteristics, the overall performance of electronic
systems will be improved dramatically. Obviously,
the performance of the arithmetic circuit directly
determines whether the electronic system in market is
competitive. It is well known that full adder is the
crucial building block used to design multiplier,
microprocessor, digital signal processor (DSP), and
other arithmetic related circuits. In addition, the full
adder is also dominant in fast adder design.
Therefore, to effectively design a full adder with
smaller chip area, low power consumption, fast
operation speed and regular circuit structure, are the
common required for IC designers.

ABSTRACT: Adder are the basic building blocks of


any computing system.. These Arithmetic operations
are widely used in most digital computer systems.
Addition will be the basic component in arithmetic
operation and is the base for arithmetic operations
such as multiplication and the basic adder cell can
be modified to function as subtractor by adding
another xor gate and can be used for division.
Therefore, 1-bit Full Adder cell is the most
important and basic block of an arithmetic unit of a
system. In this paper we analysis the 1-bit full adder
using SR-CPL style of full adder design and
Transmission gate style of design.
Keywords: SR-Cpl, Transmission Gate Full Adder,
Leakage, T-Spice

1. Introduction:
Due to rapid advances in electronic technology,
electronics market is becoming
More competitive, which results in consumer
electronic products requiring even more stringently
high quality. The design of consumer electronic
products requires not only light weight and slim size,
but also low power and fast time-to-market.
Therefore, the integrated circuit (IC) designers have
to consider more important issues such as chip area,
power consumption, operation speed, circuit
regularity, and so on. Due to these design issues
relevant to the key competitive factors of electronic
systems, IC designers and electronic design
automation (EDA) vendor are very concerned about
the development of effective methodologies to fetch
smaller chip area design, lower power consumption,

Since full adder plays an extremely important role


in arithmetic related designs, many IC designers puts
a lot of efforts on full adder circuit research.
Consequently, there are many different types of full
adders have been developed for a variety of different
applications. These different types of full adders have
different circuit structures and performance. Full
adder designs have to make tradeoff among many
features including lower power consumption, faster
operating speed, reduced transistor count, full-swing
output voltage and the output driving capability,
depending on their applications to meet the needs of
electronic systems.
One important kind of full adder designs focus on
adopting minimum transistor count to save chip area
[1, 2, 3, 4, 5]. These full adder designs with fewer

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

245

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

transistors to save chip area does have excellent


performance, however, due to MOS transistors
reduced, these full adders have threshold voltage loss
problem and poor output driving capability. Some
full adders are designed to emphasize making up for
threshold voltage loss to improve circuit performance
[6, 7, 8]. These full-swing full adder designs insist on
using fewer MOS transistors to reduce circuit
complexity to go along with reduced power
consumption and delay time. However, the full-swing
full adders have no output driver in design leading to
signal attenuation problems when they are connected
in series to construct multi-bit adders. Therefore,
many studies focus on gathering many features such
as full-swing voltage, powerful output driving
capability and good power delay product [9, 10, 11,
12, 13] in the meantime to boost the performance of
full adder circuit design as a whole. However, the
penalties have to pay for taking too many design
issues into consideration are increased circuit
complexity, larger chip area, difficult layout design,
and increased transistor count. Therefore, how to
design a full adder circuit with better performance
and simpler structure is the main goal of full adder
design field. In order to design a full adder with low
circuit complexity, good circuit performance and the
modularized structures, a multiplexer-based full
adder is proposed in this study. The multiplexerbased full adder has not only regularly modularized
structure, but also superior circuit performance.

ISBN: 378 - 26 - 138420 - 5

improve the total performance.

Fig. 2: Full adder logic architecture with four


modules.

The traditional full adder logic architecture can be


divided into three modules [6, 7, 8], and the logic
architecture block diagram is shown in Fig. 1. NewHPSC full adder [9] and Hybrid-CMOS full adder
[10] also belong to this category. These two full
adders achieve logic functions of three modules by
using pass transistor logic (PTL) and static
complementary metal-oxide-semiconductor (CMOS)
circuit design techniques.
Fig. 2 schematically show another logic
architecture block diagram of a full adder in which
logic architecture is divided into four modules.
DPLFA full adder [13] and SR-CPL full adder [13]
also belong to this category. DPLFA full adder and
SR-CPL full adder achieve logic functions of full
adder modules by using double pass transistor logic
(DPL) and swing restored complementary passtransistor logic (SR-CPL) circuit design techniques,
respectively.

The rest of this paper is organized as follows: In


section 2, some previous works on full adder design
are discussed. A novel multiplexer-based full adder
design is presented in section 3. In section 4, we
show the experimental results and make a discussion.
Finally, a brief conclusion is given in section 5.
2. Previous Works on Full Adder Design
The full adder function is to sum two binary operands
A, B and a carry input Ci, and then generate a sum
output (S) and a carry output (Co). There are two
factors affecting the performance of a full adder
design: one is the full adder logic architecture, and
the other is the circuit design techniques to perform
the logic architecture function. Therefore, the full
adder design approach requires using different types
of logic architecture and circuit design technique to

2. Design Considerations
A. Impact of Logic Style
The logic style used in logic gates basically
influences the speed, size, power dissipation, and the
wiring complexity of a circuit. The circuit delay is
determined by the number of inversion levels, the
number of transistors in series, transistor sizes [3]
(i.e., channel widths), and intra- and inter-cell wiring

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

246

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

capacitances. Circuit size depends on the number of


transistors and their sizes and on the wiring
complexity. Power dissipation is determined by the
switching activity and the node capacitances (made
up of gate, diffusion, and wire capacitances), the
latter of which in turn is a function of the same
parameters that also control circuit size. Finally, the
wiring complexity is determined by the number of
connections and their lengths and by whether Singlerail or dual-rail logic is used. All these characteristics
may vary considerably from one logic style to
another and thus make the proper choice of logic
style crucial for circuit performance. As far as cellbased design techniques (e.g., standard-cells) and
logic synthesis are concerned, ease-of-use and
generality of logic gates is of importance as well.
Robustness with respect to voltage and transistor
scaling as well as varying process and working
conditions, and compatibility with surrounding
circuitries are important aspects influenced by the
implemented logic style.

ISBN: 378 - 26 - 138420 - 5

should be kept minimal. Another source for


capacitance reduction is found at the layout level,
which, however, is not discussed in this paper.
Transistor downsizing is an effective way to reduce
switched capacitance of logic gates on noncritical
signal paths. For that purpose, a logic style should be
robust against transistor downsizing, i.e., correct
functioning of logic gates with minimal or nearminimal transistor sizes must be guaranteed (ratioless
logic).
Supply voltage reduction: The supply voltage and
the choice of logic style are indirectly related through
delay-driven voltage scaling. That is, a logic style
providing fast logic gates to speed up critical signal
paths allows a reduction of the supply voltage in
order to achieve a given throughput. For that purpose,
a logic style must be robust against supply voltage
reduction, i.e., performance and correct functioning
of gates must be guaranteed at low voltages aswell.
This becomes a severe problem at very low voltage
of around 1 V and lower, where noise margins
become critical.

B. Logic Style Requirements for Low Power


According to the formula_

Switching activity reduction:


Switching activity of a circuit is predominantly
controlled at the architectural and registers transfer
level (RTL). At the circuit level, large differences are
primarily observed between static and dynamic logic
styles. On the other hand, only minor transition
activity variations are observed among different static
logic styles and among logic gates of different
complexity, also if glitching is concerned.

The dynamic power dissipation of a digital CMOS


circuit depends on the supply voltage, the clock
frequency, the node switching activities, the node
capacitances, the node short circuit currents and the
number of nodes. A reduction of each of these
parameters results in a reduction of dissipated power.
However, clock frequency reduction is only feasible
at the architecture level, whereas at the circuit level
frequency is usually regarded as constant in order to
fulfill some given throughput requirement. All the
other parameters are influenced to some degree by
the logic style applied. Thus, some general logic style
requirements for low-power circuit implementation
can be stated at this point.

Short-circuit current reduction: Short-circuit


currents (also called dynamic leakage currents or
overlap currents) may vary by a considerable amount
between different logic styles. They also strongly
depend on input signal slopes (i.e., steep and
balanced signal slopes are better) and thus on
transistor sizing. Their contribution to the overall
power consumption is rather limited but still not
negligible (1030%), except for very low voltages
where the short-circuit currents disappear. A lowpower logic style should have minimal short-circuit
currents and, of course, no static currents besides the
inherent CMOS leakage currents.

Switched capacitance reduction: Capacitive load,


originating from transistor capacitances (gate and
diffusion) and interconnect wiring, is to be
minimized. This is achieved by having as few
transistors and circuit nodes as possible, and by
reducing transistor sizes to a minimum. In particular,
the number of (high capacitive) inter-cell connections
and their length (influenced by the circuit size)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

247

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

C. Logic Style Requirements for Ease-of-Use For


ease-of-use and generality of gates, a logic style
should be highly robust and have friendly electrical
characteristics, that is, decoupling of gate inputs and
outputs (i.e., at least one inverter stage per gate) as
well as good driving capabilities and full signal
swings at the gate outputs, so that logic gates can be
cascaded arbitrarily and work reliably in any circuit
configuration. These properties are prerequisites for
cell-based design and logic synthesis, and they also
allow for efficient gate modeling and gate-level
simulation. Furthermore, a logic style should allow
the efficient implementation of arbitrary logic
functions and provide some regularity with respect to
circuit and layout realization. Both low-power and
high-speed versions of logic cells (e.g., by way of
transistor sizing) should be supported in order to
allow flexible power-delay tuning by the designer or
the synthesis tool.
D. Static versus Dynamic Logic Styles
A major distinction, also with respect to power
dissipation, must be made between static and
dynamic logic styles. As opposed to static gates,
dynamic gates are clocked and work in two phases, a
precharge and an evaluation phase. The logic
function is realized in a single NMOS pull-down or
PMOS pull-up network, resulting in small input
capacitances and fast evaluation times. This makes
dynamic logic attractive for high-speed applications.
However, the large clock loads and the high signal
transition activities due to the recharging mechanism
result in excessive high power dissipation. Also, the
usage of dynamic gates is not as straight forward and
universal as it is for static gates, and robustness is
considerably degraded. With the exception of some
very special circuit applications, dynamic logic is no
viable candidate for low-power circuit design

ISBN: 378 - 26 - 138420 - 5

The XORXNOR module performs XOR and XNOR


logic operations on inputs A and B, and then
generates the outputs H and H. Subsequently, H and
H both are applied to the sum and the carry modules
for generation of sum output so and carry output Co.
1) SR-CPL Full Adder Design:
A SR-CPL is designed using a combination of pass
transistor logic (PTL) and static CMOS design
techniques to provide high energy efficiency and
improve driving capability. An energy efficient
CMOS FA is implemented using swing restored
complementary pass-transistor logic (SR-CPL) and
PTL techniques to optimize its PDP.

Fig4: SR-CPL Full Adder Design


2) Transmission gate CMOS (TG) uses
transmission gate logic to realize complex logic
functions using a small number of complementary
transistors. It solves the problem of low logic level
swing by using pMOS as well as nMOS

3. Design Technologies:
There are many sorts of techniques that intend to
solve the problems mentioned above
Full Adder Design:
So = HCi + HCo (1)
Co = HCi + HA (2)
Where H = A Xor B and H = A Xnor B. A Full
Adder is made up of an XORXNOR module, a sum
module and a carry module.

Fig1: Transmission Gate Full Adder

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

248

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

4. Simulation Analysis:
Full Adder was designed with different logic using
Tanner Tools and simulated

Fig6: Transmission Gate Full Adder simulation

Fig11: SR-CPL full Adder Design

Tabulation:
Circuit
SR-CPL

Power Dissipation
2.309128e-005 watts

Transmission Gate

4.876595e-008 watts

The above circuits of full adder are simulated using


Tanner Tools using TSMC018 and its delay and
power of individual circuits are tabulated
5. Conclusion:
In this paper we show the low power adder design
full adder with less number transistors. CMOS design
styles make the full adder to design in less no of
transistors as wells as with lower power dissipation.
For performance validation, Tanner simulations were
conducted on FAs implemented with TSMC 018
CMOS process technology in aspects of power
consumption, delay time. In contrast to other types of
FAs with drivability, an SR-CPL-FA is superior to
the other ones and can be applied to design related
adder-based portable electronic products in practical
applications in todays competitive markets.

Fig12: SR-CPL full Adder simulation

References:
[1] Neil H. E. Weste & David Harris, CMOS VLSI
Design- A circuit and Systems Perspective, 4th
edition, Addison Wesley, 2010
[2] C. N.Marimuthu, Dr. P. Thangaraj, Aswathy
Ramesan, Low power shift and add multiplier
design", International Journal of Computer
Science and Information Technology, June 2010,
Vol. 2, Number 3.
[3] Marc Hunger, Daniel Marienfeld, New SelfChecking Booth Multipliers, International Journal of
Applied Mathematics Computer Sci.,

Fig5: Transmission Gate Full Adder Design

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

249

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

2008, Vol. 18, No. 3, 319328


[4] C. Jaya Kumar, R. Saravanan, VLSI Design for
Low Power Multiplier using Full Adder, European
Journal of Scientific Research,
ISSN 1450-216X Vol.72 No.1 (2012), pp. 5-16
[5] Ravi Nirlakalla, Thota Subba Rao, Talari
Jayachandra Prasad, Performance Evaluation of
High Speed Compressors for High Speed
Multipliers, Serbian Journal of Electrical
Engineering, Vol. 8, No. 3, November 2011, 293-306
[7] G. E. Sobelman and D. L. Raatz, Low-Power
multiplier
designusing
delayed
evaluation.
Proceedings of the International Symposium on
Circuits and Systems (1995), pp. 1564 1567.
[8] T. Sakuta, W. Lee, and P. T. Balsara, Delay
balanced multipliers for low power/low voltage DSP
core. Proceedings of IEEE Symposium on Low
Power Electronics (1995), pp. 3637.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

250

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

FPGA Implementation Of Various Security Based Tollgate system Using ANPR


Technique.
Author 1: S.SHASHIDHAR

Author2: S.VAMSEE KRISHNA


Email id: [email protected]

Email.id: [email protected]
M.TECH in Vlsi System Design,

Assistant prof. In Department of ECE,

Siddarth Institute of Engineering And Technology,

Siddarth Institute Of Engineering And Technology,

Puttur, Chittor(Dist).

Puttur, Chittor(Dist).

ABSTRACT:
This paper presents the look and development of
vehicle plate recognition for automated toll
collection. Vehicle plate recognition (LPR) is that
the extraction of car plate information from a
picture since it's simpler and faster than the
normal token based ticket system, it's all the
potential to interchange the prevailing system.
Moreover, it saves users valuable time by
reducing the queue length before of the toll
counter. Its accustomed pay the quantity
automatically and open &amp; close the toll gate
automatically. A vehicle plate is installed on each
vehicle. A recognition device at the gate reads
this knowledge from the vehicle and compares it
with the info within the on-line database and
permits the access consequently by gap the gate.
This data is employed to print a daily or monthly
bill for toll collection from the vehicles. This
model has low complexity and takes fewer times
in terms of car plate segmentation and character
recognition. We aim to scale back the time
consumed to pay the toll gate amount and also to
assist the RTO, local department to trace the
vehicle, just in case} if it absolutely was stolen or
used for any illegal activities. Yet as we are
reaching to increase the protection features
within the toll gate because now a days toll gate
are the doorway to the most cities. If we increase
the protection within the toll gate section
automatically the protection within the city are
also increased. The proposed open-end credit
has been designed using very (high-speed
integrated
circuit)
hardware
description
language (VHDL) and simulated. Finally, it's
downloaded in a very field programmable gate
array (FPGA) chip and tested on some given
scenarios. The FPGA implementation is
administrated in one among the applying area
automatic toll assortment.

rating and automatic toll assortment. Attributable to


totally different operating environments, LPR
techniques vary from application to application.
Point able cameras produce dynamic scenes after
they move. A dynamic scene image could contain
multiple car places or no license plate the least bit.
Moreover, after they do seem in a picture, license
plates could have impulsive sizes, orientations and
positions. And, if complicated backgrounds area
unit concerned, detective work license plates will
become quite a challenge. Typically, Associate in
Nursing LPR method consists of 2 main stages (1)
locating license plates and (2) distinctive license
numbers. Within the 1st stage, car place
candidates area unit determined supported the
options of license plates. Options ordinarily used
are derived from the car place format and therefore
the alphanumeric characters constituting license
numbers. The options concerning car place format
embody form, symmetry height-to dimension
magnitude relation colour texture of achromatic
colour abstraction frequency and variance of
intensity values Character options embody line
blob the sign transition of gradient magnitudes, the
ratio of characters the distribution of intervals
between characters and therefore the alignment of
characters. In reality, a tiny low set of sturdy,
reliable, and easy-to-detect object options would be
adequate.
The car place candidates determined within the
locating stage area unit examined within the
identification number identification stage. There
are unit 2 major tasks concerned within the
identification stage, variety separation and variety
recognition. Variety separation has within the past
been accomplished by such techniques as
projection morphology relaxation labelling,
connected elements and blob colouring. Since the
projection methodology assumes the orientation of
a car place is thought and therefore the morphology
methodology needs knowing the sizes of
characters. A hybrid of connected elements and
blob colouring techniques is taken into account for
character separation. For this, we tend to develop
our own character recognition technique that is
predicated on the disciplines of each artificial
neural networks and mechanics.

Keywords: License plate recognition,


INTRODUCTION:
AUTOMATIC car place recognition (LPR)
plays a crucial role in various applications like
unattended parking heaps security management of
restricted areas traffic enforcement congestion

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

251

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

License Plate Recognition:


Most of the quantity plate detection algorithms fall
in additional than one class supported totally
different techniques. To sight vehicle variety plate
following factors ought to be considered:

ISBN: 378 - 26 - 138420 - 5

character into an identical size. Each example scans


the character column by column to calculate the
normalized cross correlation. The example with the
most price is that the foremost similar one.
Example matching is useful for recognizing singlefont, rotated, broken, and fixed-size characters. If a
personality is completely totally different from the
example, the example matching produces incorrect
recognition result. Among the disadvantage of
recognizing inclined characters is solved by storing
several templates of an identical character with
totally different inclination angles.

(1). Plate size: a plate will be of various sizes in a


very vehicle image.
(2). Plate location: a plate will be settled anyplace
within the vehicle.
(3). Plate background: A plate will have totally
different background colours supported vehicle
sort. As an example a government vehicle variety
plate may need totally different background than
different public vehicles.
(4). Screw: A plate might have screw which may be
thought-about as a personality.
A number plate will be extracted by mistreatment
image segmentation technique. There ar various
image segmentation strategies accessible in
numerous literatures. In most of the strategies
image binarization is employed. Some authors use
Otsus technique for image binarization to convert
colour image to grey scale image. Some plate
segmentation algorithms are supported colour
segmentation. A study of car place location
supported colour segmentation is mentioned.
Within the following sections common variety
plate extraction strategies are explained, that is
followed by elaborate discussion of image
segmentation techniques adopted in numerous
literature of ANPR or LPR.

Fig2: Input Images

Fig3: Different Template Images

Fig1: License Plate recognition


The templates that square measure almost like the
registration number plate character is known by the
comparison method with the character hold on
within the information. Extracted registration
number plate characters could have some noise or
they'll be broken. The extracted characters could in
addition be inclined. Example matching may be a
straightforward and straightforward technique in
recognition. The similarity between character and
therefore the example is measured. Example
matching is performed once resizing the extracted

Database Image

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

252

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

quantity of vehicles. The quantity on the idea of


weight &amp; the count of vehicles is additionally
displayed on the screen. The quantity to be paid is
mechanically deduced from the various checking
accounts.

Block Diagram for Toll Gate System:

If a vehicle carries any reasonably gas that


shouldnt be carried, the gas sensing element
detects the gas within the vehicle. Just in case if
there's any reasonably gas that's detected, the RF
transmitter is employed to alert the close police
headquarters associated an alarm is enabled to alert
the encircling areas. Afterwards motor one is
employed to shut the gate and at the same time
motor two is employed to drag up the spikes so as
to puncture the vehicle.
PRESSURE SENSOR:
A piezoelectric sensor as shown in figure
5 is a device that uses the piezoelectric effect to
measure pressure, acceleration, strain or force by
converting them to an electrical charge. Here a
simple pressure sensor is used to protect door or
window. It generates a loud beep when somebody
tries to break the door or window. The alarm stops
automatically after three minutes. The circuit uses a
piezo element as the pressure sensor.

Fig: 4New Toll Gate system


The planned system makes certain that the traffic at
the toll gates is efficient and security is additionally
gift. The tax that is collected is predicated on the
load carried by the vehicle. Through this technique
we will additionally determine taken vehicles. The
reading the knowledge from vehicle plate
recognition system, computer compares the
information within the info and permits the access
consequently by opening/closing the gate. This
knowledge is employed to print a daily or monthly
bill for toll assortment from the vehicles. This
fashion even taken vehicles is known.

Piezo buzzer exploits the piezoelectric


property of the piezo electric crystals. The
piezoelectric effect may be direct piezoelectric
effect in which the electric charge develops as a
result of the mechanical stressor reverse or indirect
piezoelectric effect (Converse piezoelectric effect)
in which a mechanical force such as pressure
develops due to the application of an electric field.

The pressure of the vehicle is obtained


victimization the pressure sensing element and
consequently the pressure of the vehicle is showed
on the display. A counter is employed to count the
quantity of vehicles. The quantity on the idea of
weight &amp; the count of vehicles is additionally
displayed on the screen. The quantity to be paid is
mechanically deduced from the various checking
accounts.
If a vehicle carries any reasonably gas that
shouldnt be carried, the gas sensing element
detects the gas within the vehicle. Just in case if
there's any reasonably gas that's detected, the RF
transmitter is employed to alert the close police
headquarters associated an alarm is enabled to alert
the encircling areas. Afterwards motor one is
employed to shut the gate and at the same time
motor two is employed to drag up the spikes so as
to puncture the vehicle.

Fig5: Piezo electric sensor


A typical example of direct piezoelectric effect is
the generation of measurable amount of
piezoelectricity when the Lead Zircon ate Titan ate
crystals are deformed by mechanical or heat stress.
The Lead Zircon ate Titan ate crystals also shows
indirect piezoelectric effect by showing pressure
when an electric potential is applied.

The pressure of the vehicle is obtained


victimization the pressure sensing element and
consequently the pressure of the vehicle is showed
on the display. A counter is employed to count the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

253

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

OPERATION:

STRUCTURE AND CONFIGURATION:

Operation of pressure sensor is very


simple. Here we have two plates that is one is input
plate and the other is output plate, whenever
pressure is applied as shown in figure 6 then these
two plates come into contact we get voltage as a
output then this output is send to the FPGA in turn
it shows the weight of the vehicles is shown in the
display as per our project. Accordingly toll tax is
calculated. It is made up of a piezoelectric crystal.
Depending on how a piezoelectric material is cut,
three main modes of operation can be distinguished
into transverse,
longitudinal, and
shear.

Structure and configuration of MQ-2 gas


sensor is shown as figure, sensor composed by
micro AL2O3 ceramic tube, Tin Dioxide (SnO2)
sensitive layer, measuring electrode and heater are
fixed into a crust made by plastic and stainless steel
net. The heater provides necessary work conditions
for work of sensitive components. The enveloped
MQ-2 has 6 pin, four of them are used to fetch
signals, and other two are used for providing
heating current.
WIRELESS COMMUNICATION MODULE:
A general RF communication block
diagram is shown in figure.8 since most of the
encoders/decoders/microcontrollers
are
TTL
compatible and mostly inputs by the user will be
given in TTL logic level. Thus, this TTL input is to
be converted into serial data input using an encoder
or a microcontroller. This serial data can be directly
read using the RF Transmitter, which then performs
ASK (in some cases FSK) modulation on it and
transmit the data through the antenna. In the
receiver side, the RF Receiver receives the
modulated signal through the antenna, performs all
kinds of processing, filtering, demodulation, etc
and gives out a serial data. This serial data is then
converted to a TTL level logic data, which is the
same data that the user has input.

Figure 6 Pressure sensor operations


GAS SENSOR:
The Flammable Gas and Smoke sensors can detect
the presence of combustible gas and smoke at
concentrations from 300 to 10,000 ppm. Owing to
its simple analogue voltage interface, the sensor
requires one analogue input pin from the FPGA.

Fig8: RF communication block diagram


Results:

Fig7: Gas Sensor (type-MQ2)


The product can detect the pressure of the
smoke and send the output in the form of analogue
signals. Our range can function at temperature
ranging from -20 to 50 degree Celsius and consume
less than 150 mA at 5V.
Sensitive material of MQ-2 gas sensor in
the figure is SnO2 (Tin dioxide), which with lower
conductivity in clean air. When the target
combustible gas exist, the sensors conductivity is
higher along with the gas concentration rising.
MQ-2 gas sensor has high sensitivity to
LPG, Propane and Hydrogen, also could be used to
Methane.

Fig9: Before Segmentation Number Plate

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

254

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

lighting variations, headlight dazzle, and partial


vehicle obscuration. The next stage of development
will involve increasing the vehicle coverage of the
reference database and further trials at sites
including parking facilities and public highways.
FUTURE WORK:
The future analysis of ANPR ought to focus on
multi style plate recognition, video-based ANPR
mistreatment temporal information, multi plates
process, high definition plate image process,
ambiguous-character recognition, and so on. In four
important factors were planned to identify the multi
style license plate problem: license plate rotational
angle, character line number, the alphanumerical
varieties used and character formats

Fig10: After Segmentation Number Plate

REFERENCES
[1]
Kin Seong Leong, Mun Leng Ng, Member,
IEEE, Alfio R. Grasso, Peter H. Cole,
"Synchronization of RFID Readers for Dense RFID
Reader Environments", International Symposium
on Applications and the Internet Workshops
(SAINTW06), 2005
[2]
Manish Buhptani, Shahram Moradpour,
"RFID Field Guide - Developing Radio Frequency
Identification Systems", Prentice Hall, 2005, pp 79, 16-225, 160, 231

Fig11: Number Plate Recognition

[3]
Raj Bridgelall, Senior Member, IEEE, "
Introducing a Micro-wireless Architecture for
Business Activity Sensing ", IEEE International
Conference RFID, April 16-17,2008
[4]
Sewon Oh, Joosang Park, Yongioon Lee,
"RFID-based Middleware System for Automatic
Identification", IEEE International Conference on
Service Operations and Logistics, and Information,
2005
[5]
Shi-Cho Cha Kuan-Ju Huang Hsiang-Meng
Chang, " An Efficient and Flexible Way to Protect
Privacy in RFID Environment with Licenses ", IEEE
International Conference RFID, April 16-17,2008

Fig12: After Matching


CONCLUSION:
Motive of this projected system is to detect the
images of the vehicle plate accurately. The
automated real-time vehicle plate recognition
system has been demonstrated to provide correct
identification of vehicles captured images from the
camera. The system has been shown to be robust to

[6]
Urachada Ketprom, Chaichana Mitrpant,
Putchapan Lowjun, Closing Digital Gap on RFID
Usage for Better Farm Management, PICMET
2007, 5-9 August 07

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

255

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

MINIMIZATION OF VOLTAGE SAGS AND SWELLS USING DVR


N.VISWANATH

Dr. K. RAMA SUDHA

PG Scholar

Professor

Department Of Electrical Engineering,


Andhra University,
Visakhapatnam,
Andhra Pradesh
most concerning disturbance

quality is voltage sag. Voltage sag is a sudden

ABSTRACT: - Power quality problem is an

drop in the Root Mean Square (RMS) [1] voltage

occurrence of non-standard voltage, current or

and is usually characterized by the retained voltage.

frequency that results in a failure or a

The major source of voltage sag is short- circuits

disoperation of end user equipments. Utility

on the utility lines. Faults from the disturbed

distribution networks, sensitive industrial loads

process will generate a momentary voltage sag

and critical commercial operations suffer from


various

types

interruptions

of
which

outages
can

and

cost

[2][3] in the electrical environment to the end user.

service

The Dynamic Voltage Restorer (DVR) is an

significant

effective Custom Power device which is used to

financial losses. With the restructuring of power


systems and

affecting power

mitigate the impacts of voltage sags on sensitive

with shifting trend towards

loads in distribution systems. DVR is used for

distributed and dispersed generation, the issue

balancing the load voltage due to harmonics and

of power quality is going to take newer

unbalancing at the source end, in order to eliminate

dimensions. The present work is to identify the

switching transients. DVR has to inject voltages

prominent concerns in this area and hence the

with large

measures that can enhance the quality of the

magnitude,

which is completely

undesirable. By varying load voltage angle, if the

power are recommended. This work describes

required nominal voltage is injected at the system

the techniques of correcting the supply voltage

frequency, the control operation will be efficient.

sag, swell and interruption in a distributed

To realize this, a method for estimating the

system. At present, a wide range of very flexible

frequency from the sampled injected voltage signal

controllers, which capitalize on newly available

has been presented.

power electronics components, are emerging for


custom power applications. Among these, the

DVR consists of energy storage device,

distribution static compensator and the dynamic

pulse width modulation inverter, LC filter and

voltage restorer are most effective devices, both

series transformer. Pulse Width Modulated (PWM)

of them based on the VSC principle.

control technique is applied for inverter switching


to produce a three phase 50 Hz sinusoidal voltages

KEY WORDS: Dynamic voltage restorer,

at the load terminals. The PWM scheme which is

Voltage Sag and swell, PWM Generator.

used to synthesize the injected voltage generates


switching frequency harmonics must be prevented

I INTRODUCTION

from entering into the utility and customer system.


The quality of output power delivered

A low-pass filter is introduced to accomplish this

from the utilities has become a major concern. The

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

256

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

function. Literature shows that a number of

voltage level for mitigation of power quality

techniques are available for improving power

phenomenon, known as Custom Power Devices,

quality problems and frequency estimation to

able to deliver customized solution to power

measure the signals which are available in distorted

quality problems.

form. Least mean square, Kalman filtering,

mitigating devices are normally connected between

Discrete Fourier transform, Smart discrete Fourier

the supply and the load.

Voltage sag and interruption

Transform and Newton method are some of the


Dynamic voltage restorer [5] is a series

techniques shown in literature. Faults in the

connected device designed to maintain a constant

distribution system may cause voltage sag or swell

RMS voltage across a sensitive load. The structure

in the large parts of the system. Voltage sag and

of DVR is shown in Fig. I. The DVR consists of:

swell can cause sensitive equipment to fail and


create a large current unbalance that trips the

Voltage Source Inverters: Voltage Source Inverters

circuit breakers. These effects can be very

converts the dc voltage from the energy storage

expensive for the customer, to avoid equipment

unit to a controllable three phase ac voltage. The

damage. There are many different methods to

inverter switches are normally fired using a

mitigate voltage sags and swells, but the use of a

sinusoidal Pulse Width Modulation scheme.

DVR is considered to be the most cost efficient


method. DVR with PI controller has a simple

Injection

structure and offers a satisfactory performance

used in the DVR plays a crucial role in

over a wide range of operating conditions. The

ensuring

main problem of Conventional Controllers [3][4] is

effectiveness of the restoration scheme. It is

the correct tuning of the controller gains. When

connected in series with the distribution feeder.

there are variations in the system parameters and

transformers: Injection

the

maximum

transformers

reliability

and

Passive Filters: Passive Filters are placed at the

operating conditions, the controller may not

high voltage side of the DVR to filter the

provide the required control performance with fixed

harmonics. These filters are placed at the high

gains.

voltage side as placing the filters at the inverter


side introduces phase angle shift which can disrupt

Power Quality problem is the main

the control algorithm.

concern in electricity industry. Power Quality


includes a wide range of disturbances such as

Energy storage devices: Examples of energy

voltage sags/swells, flicker, harmonics distortion,

storage devices are dc capacitors, batteries, super-

impulse transient, and interruptions. And the

capacitors,

majority of power quality problems are due to

superconducting

magnetic

energy

Storage and flywheels. The capacity of energy

different fault conditions. These conditions cause

storage

voltage sag. Voltage sag can occur at any instant of

device

compensation

time, with amplitude ranging from 10-90% and a

has

capability

big
of

impact
the

on

the

system.

Compensation of real power is essential when large

duration lasting for half a cycle to one minute. It is

voltage sag occurs.

generally caused by faults in the power system


and characterized by its magnitude and duration.
The duration of voltage sag depends on clearance
of fault by using protective devices. Power
Electronics based devices installed at medium

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

257

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

DVR is connected between a terminal bus


and load bus. The control technique to be adopted
depends on the type of load as some loads are
sensitive to only magnitude change whereas some
other loads are sensitive to both magnitude and
phase angle shift. Control techniques that utilize
real

Fig 1: Structure of Dynamic Voltage Restorer

and

reactive

power

compensation

are

generally classified as pre-sag compensation, inII COMPENSATION OF VOLTAGE SAG

phase compensation and energy optimization

USING DVR

technique. The single line diagram of DVR


connected in the distribution system

The single line diagram of test system is


shown in Fig.2. The voltage source is connected
to a feeder with an impedance of

Rs + jXS

(1)

The load is balanced and the impedance of the


load is given by

RL + jXL

(2)

Fig 3 shows the test system with 3phase fault.


Fig.3. single line diagram of dynamic voltage restorer

VL is the source voltage in volts


s

connected to distribution system

vt is voltage at point of common coupling in volts.


t

When the source

Rs+ jX s is impedance of the feeder in ohms

voltage drops

or

increases, the dynamic voltage restorer injects a


series

is the load voltage in voltage

voltage through the injection transformer

so that the desired load [11] voltage magnitude


can be maintained. The series injected voltage of

RL+ jX L is the load impedance in ohms.

the DVR, Vk can be written as:

Is is the source current and IL is the load current

Vk = Vt + Vl

(3)

Vk is the series injected voltage in the distribution


system such that it mitigates the voltage sag and
regulates the load bus voltage, Vl to a reference
value Vl*. It is pre specified value. The reference
voltage of the DVR can be written as
Vk* = Vt + Vl*

(4)

Fig.2. Single line diagram of test system

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

258

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

III Function of DVR:

ISBN: 378 - 26 - 138420 - 5

contribute to the losses. The DVR will be most of


the time in this mode. In boost mode (VDVR>0),

The main function of a DVR is the


protection

of

sensitive

loads

from

the DVR is injecting a compensation voltage

voltage

through the booster transformer due to a detection

sags/swells coming from the network. Therefore as

of a supply voltage disturbance.

shown in Figure, the DVR is located on approach


of sensitive loads. If a fault occurs on other lines,
DVR

inserts

series

voltage [6]

VDVR

and

compensates load voltage to pre fault value. The


momentary amplitudes of the three injected phase
voltages are controlled such as to eliminate any
detrimental effects of a bus fault to the load voltage
VL. This means that any differential voltages caused
by transient disturbances in the ac feeder will be
compensated by an equivalent voltage generated by
the converter and injected on the medium voltage
Fig.4.Equivalent Circuit of DVR

level through the booster transformer.


Figure 4 shows the equivalent circuit of
The DVR works independently of the type

the DVR, when the source voltage is drop or

of fault or any event that happens in the system,

increase, the DVR injects a series voltage Vinj

provided that the whole system remains connected

through the injection transformer [9][10] so that the

to the supply grid, i.e. the line breaker does not trip.

desired load voltage magnitude VL can be

For most practical cases, a more economical design

maintained.

can be achieved by only compensating the positive


The series injected voltage of the DVR

and negative sequence [7] components of the

can be written as

voltage disturbance seen at the input of the DVR.


This option is Reasonable because for a typical

Vinj = VL + VS

distribution bus configuration, the zero sequence


part of a disturbance will not pass through the step

(5)

Where;

down transformer because of infinite impedance for


VL is the desired load voltage magnitude

this component.

VS is the source voltage during sags/swells

The DVR has two modes of operation

condition.

which are: standby mode and boost mode. In


standby mode (VDVR=0), the booster transformers

The load current ILoad is given by,

low voltage winding is shorted through the


converter. No switching of semiconductors occurs

in this mode of operation, because the individual

(6)

converter legs [8] are triggered such as to establish


a short-circuit path for the transformer connection.
Therefore, only the comparatively low conduction
losses of the semiconductors in this current loop

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

259

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Pulse-width modulation (PWM):


PWM also works well with digital

Pulse-width modulation (PWM), or pulse-duration

controls, which, because of their on/off nature, can

modulation (PDM), is a modulation technique that

easily set the needed duty cycle. PWM has also

controls the width of the pulse, formally the pulse

been used in certain communication systems where

duration, based on modulator signal information.

its duty cycle has been used to convey information

Although this modulation technique can be used to

over a communications channel.

encode information for transmission, its main use is


to allow the control of the power supplied to
electrical devices, especially to inertial loads such
as motors. In addition, PWM is one of the two
principal algorithms used in photovoltaic solar
battery chargers, the other being MPPT.
The average value of voltage (and current)
fed to the load is controlled by turning the switch
between supply and load on and off at a fast pace.
The longer the switch is on compared to the off
periods, the higher the power supplied to the load.
The PWM switching frequency has to be
much higher than what would affect the load (the
device that uses the power), which is to say that the
resultant waveform perceived by the load must be
as smooth as possible. Typically switching has to
be done several times a minute in an electric stove,
120 Hz in a lamp dimmer, from few kilohertz
(kHz) to tens of kHz for a motor drive and well into
the tens or hundreds of kHz in audio amplifiers and

Fig 5: An Example of PWM in an Ac Motor Driver

computer power supplies.


The

term duty

cycle describes

the

proportion of 'on' time to the regular interval or


'period' of time; a low duty cycle corresponds to
low power, because the power is off for most of the
time. Duty cycle is expressed in percent, 100%
being fully on.
The main advantage of PWM is that
power loss in the switching devices is very low.
When a switch is off there is practically no current,
and when it is on and power is being transferred to
the load, there is almost no voltage drop across the
switch. Power loss, being the product of voltage
and current, is thus in both cases close to zero.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

260

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

IV SIMULATION FIGURES

Fig.6. Main Block Diagram of DVR

Fig.7. Control System of the DVR

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

261

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

V SIMULATION RESULTS
The faults in the three phase source can be eliminated by calculating phase angle . But the calculation of
becomes complex some times. So by using PWM generator the calculation of the phase angle can be found
easily from the magnitude part only.
The figure8 shows the three phase waveform where fault occur at phase A. By using DVR with PWM
generator the fault is eliminated and the output waveform is shown in figure9.

Fig.8.The simulation of the input fault

Fig.9. Simulation result of the output

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

262

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[11] S. S. Choi, B. H. Li, and D. M. Vilathgamuwa,Dynamic

VI CONCLUSION

voltage restoration with minimum energy injection, IEEE Trans.

In the simulation study, MATLAB Simulink is used to

Power. Syst., vol. 15, no. 1, pp. 5157, Feb. 2000.


[12] Y. W. Li, D. M. Vilathgamuwa, P. C. Loh, and F. Blaabjerg, A

simulate the model of dual dynamic voltage restorer is an

dual- functional medium voltage level DVR to limit downstream

effective custom power device for voltage sags and swells

fault cur- rents, IEEE Trans. Power. Electron., vol. 22, no. 4, pp.

mitigation.

13301340, Jul. 2007.

The Dual DVR controls under different faults

without any difficulties and injects the appropriate voltage

[13] Y. W. Li, D. M. Vilathgamuwa, F. Blaabjerg, and P. C. Loh,


A Ro- bust control scheme for medium-voltage-level DVR

component to correct rapidly any abnormally in the supply

implementation, IEEE Trans. Ind. Electron., vol. 54, no. 4, pp. 2249

voltage to keep the load voltage balanced and constant at the

2261, Aug. 2007.

nominal value.

[14] S. S. Choi, T. X. Wang, and D. M. Vilathgamuwa, A series


compen- sator with fault current limiting function, IEEE Trans.
Power Del., vol. 20, no. 3, pp. 22482256, Jul. 2005.
[15] B. Delfino, F. Fornari, and R. Procopio, An effective SSC

VII REFERENCES

control scheme for voltage sag compensation, IEEE Trans. Power


Del., vol.

[1]

Fault Current Interruption by the Dynamic Voltage Restorer

20, no. 3, pp. 21002107, Jul. 2005.

Firouz Badrkhani Ajaei, StudentMember, IEEE, Shahrokh Farhangi, and Reza

[16] C. Zhan, V. K. Ramachandaramurthy, A. Arulampalam, C.

Iravani, Fellow, IEEE

Fitzer, S. Kromlidis, M. Barnes, and N. Jenkins, Dynamic voltage

[2]

Bollen M. H. J,(2000) :Understanding Power Quality Problems;

Voltage

Sags

and

interruptions,

ser.

IEEE

Press

Series

restorer based on voltage-space-vector PWM control, IEEE Trans.

on

Ind. Appl., vol. 37, no. 6, pp. 18551863, Nov./Dec. 2001.

PowerEngineering, Piscataway, NJ.

[17] D. M. Vilathgamuwa, P. C. Loh, and Y. Li, Protection of

[3]

microgrids during utility voltage sags, IEEE Trans. Ind. Electron.,

Choi, S.S.; Li, B.H.; Vilathgamuwa, D.M.(2000) : A Comparative

Study Of Inverter- And Line-Side Filtering Scheme In Dynamics Voltage

vol. 53, no. 5, pp. 14271436, Oct. 2006.

Restorer, Power Engineering Society Winter Meeting,. IEEE. Vol. 4, pp.

[18] F. Badrkhani Ajaei, S. Afsharnia, A. Kahrobaeian, and S.

2967-2972.

Farhangi, A fast and effective control scheme for the dynamic

[4] N. G. Hingorani, Introducing custom power, IEEE Spectr., vol. 32, no.

voltage restorer, IEEE Trans. Power Del., vol. 26, no. 4, pp. 2398

6, pp. 4148, Jun. 1995.

2406, Oct. 2011.

[5] J. G. Nielsen, F. Blaabjerg, and N. Mohan, Control strategies for dy-

[19] M. S. Sachdev and M. A. Barlbeau, A new algorithm for

namic voltage restorer compensating voltage sags with phase jump, in

digital impedance relays, IEEE Trans. Power App., Syst., vol.

Proc. IEEE APEC, 2001, pp. 12671273.

PAS-98, no.6, pp. 22322240, Nov./Dec. 1979.

[6]

Ghosh A ; Ledwich G.( 2001) : Structures and control of a dynamic

voltage regulator (DVR), in Proc. IEEE Power Eng. Soc. Winter Meeting,
Columbus, OH.
[7]

Ghosh A ; Ledwich G.( 2002) : Power Quality Enhancement Using

Custom Power Devices, Norwell, MA: Kluwer.


[8] G. J. Li, X. P. Zhang, S. S. Choi, T. T. Lie, and Y. Z. Sun, Control strategy for dynamic voltage restorers to achieve minimum power
injection without introducing sudden phase shift, Inst. Eng. Technol. Gen.
Transm. Distrib., vol. 1, no. 5, pp. 847853, 2007.
[9] S. S. Choi, B. H. Li, and D. M. Vilathgamuwa, Design and analysis of
the inverter-side filter used in the dynamic voltage restorer, IEEE Trans.
Power Del., vol. 17, no. 3, pp. 857864, Jul. 2002.
[10] B. H. Li, S. S. Choi, and D. M. Vilathgamuwa, Design considerations on
the line-side filter used in the dynamic voltage restorer, Proc. Inst. Elect. Eng.,
Gen. Transm. Distrib., vol. 148, no. 1, pp. 17, Jan. 2001.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

263

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

POWER-QUALITY IMPROVEMENT OF GRID INTERCONNECTED


WIND ENERGY SOURCE AT THE DISTRIBUTION LEVEL
BHARTH KUMAR.D
Dr.R.VIJAYA SANTHI
PG Scholar
Asst.Professor
DEPARTMENT OF ELECTRICAL ENGINEERING
Andhra University,
Visakhapatnam,
Andhra Pradesh.
ABSTARCT: Utility of power electronic
converters are increasingly connected in
distribution systems using renewable energy
resources. But due to their intermittent nature,
they may pose a threat to network in terms of
stability, voltage regulation and power quality
issues. The present work explains the concept of
grid interconnection of Wind Farm using grid
interfacing inverter along with the facility of
power quality improvement. In this paper it
presents control strategy for achieving
maximum benefits from grid-interfacing
inverter when interconnected in 3-phase 4-wire
distribution system is installed. For controlling
the performance as a multi-function device
which is active power filter functionality
inverters are used, using hysteresis current
control technique. With such a control, the
combination of grid-interfacing inverter and the
3-phase 4-wirelinear/non-linear unbalanced load
at point of common coupling appears as
balanced linear load to the grid. This new
control concept is demonstrated with extensive
MATLAB/Simulink simulation study.

energy solution. Since the past decade,


there has been an enormous interest in
many countries on renewable energy for
power
generation.
The
market
liberalization and governments incentives
have further accelerated the renewable
energy sector growth.
A renewable resource is a natural
resource which can replenish with the
passage of time, either through biological
reproduction or other naturally recurring
processes. Renewable resources are a part
of Earth's natural environment and the
largest components of its ecosphere.
Renewable resources may be the source
of power for renewable energy.
Renewable energy is generally
defined as energy that comes from
resources which are naturally replenished
on a human timescale such as sunlight,
wind, rain, tides, waves and heat.
Renewable energy replaces conventional
fuels such as coal, nuclear, natural gas etc.
in three distinct areas. The extensive use of
power electronics based equipment and
non-linear loads at PCC generate harmonic
currents, which may deteriorate the quality
of power [1], [2]. In [3] an inverter
operates as active inductor at a certain
frequency to absorb the harmonic current.
A similar approach in which a shunt active
filter acts as active conductance to damp
out the harmonics in distribution network
is proposed in [4]. A [5] control strategy

Index TermsActive power filter (APF),


distributed generation (DG), distribution
system, grid interconnection, power quality
(PQ), renewable energy.

I INTRODUCTION
Electric utilities and end users of
electric power are becoming increasingly
concerned about meeting the growing
energy demand. Seventy five percent of
total global energy demand is supplied by
the burning of fossil fuels. But increasing
air pollution, global warming concerns,
diminishing fossil fuels and their
increasing cost have made it necessary to
look towards renewable sources as a future

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

264

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

for renewable interfacing inverter based on


p-q theory is proposed.

ISBN: 378 - 26 - 138420 - 5

(small hydro, modern biomass, wind,


solar, geothermal, and bio-fuels) account
for another 3% and are growing
rapidly. At the national level, at least 30
nations around the world already have
renewable energy contributing more than
20% of energy supply. National renewable
energy markets are projected to continue
to grow strongly in the coming decade and
beyond. Wind power, for example, is
growing at the rate of 30% annually, with
a worldwide installed capacity of 282,482
megawatts (MW) at the end of 2012.

Power generation: Renewable energy


provides 19% of electricity generation
worldwide. Renewable power generators
are spread across many countries, and
wind power alone already provides a
significant share of electricity in some
areas: for example, 14% in the U.S. state
of Iowa, 40% in the northern German state
of Schleswig-Holstein, and 49% in
Denmark. Some countries get most of their
power from renewable, including Iceland
(100%), Norway (98%), Brazil (86%),
Austria (62%) and Sweden (54%).

Renewable
energy
resources
mainly can be used for power generation,
heating and transportation purposes. In
case of power generation, the generation
can take place either as a separate unit to
feed a particular area or the renewable
energy resources can be interconnected to
the grid at the transmission level, subtransmission level and distribution level to
enhance load supplying capability of the
grid. Large wind farms, concentrated solar
power photovoltaic system, bio-power,
hydro power, geothermal power are
interconnected at the transmission and subtransmission levels. Photovoltaic system,
small wind farm, hydro power and fuel
cells are interconnected at the distribution
level. The resources are connected to grid
using grid interfacing inverter by suitable
controlling of the inverter switches. But
their highly intermittent nature may result
in instability and power quality problems;
hence an appropriate control circuitry is
required.

Heating: Solar hot water makes an


important
contribution
to renewable
heat in many countries, most notably in
China, which now has 70% of the global
total (180 GWth). Most of these systems
are installed on multi-family apartment
buildings and meet a portion of the hot
water needs of an estimated 5060 million
households in China. Worldwide, total
installed solar water heating systems meet
a portion of the water heating needs of
over 70 million households. The use of
biomass for heating continues to grow as
well. In Sweden, national use of biomass
energy has surpassed that of oil.
Transport
fuels:
Renewable biofuels have contributed to a significant
decline in oil consumption in the United
States since 2006. The 93 billion liters of
bio-fuels produced worldwide in 2009
displaced the equivalent of an estimated 68
billion liters of gasoline, equal to about 5%
of world gasoline production.

II SYSTEM DESCRIPTION:
The proposed system consists of RES
connected to the dc-link of a gridinterfacing inverter as shown in Fig. 1. The
voltage source inverter is a key element of
a DG system as it interfaces the renewable
energy source to the grid and delivers the

About 16% of global final energy


consumption
presently
comes
from renewable resources, with 10% of all
energy from traditional biomass, mainly
used
for heating,
and
3.4%
from hydroelectricity. New renewable

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

265

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

generated power. The RES may be a DC


source or an AC source with rectifier
coupled to dc-link. Usually, the fuel cell
and photovoltaic energy sources generate
power at variable low dc voltage, while the
variable speed wind turbines generate
power at variable ac voltage. Thus, the
power generated from these renewable
sources needs power conditioning (i.e.,
dc/dc or ac/dc) before connecting on dclink [6]-[8]. The dc-capacitor decouples
the RES from grid and also allows
independent control of converters on either
side of dc-link.

ISBN: 378 - 26 - 138420 - 5

grid via the dc-link. The current injected


by renewable into dc-link at voltage level
can be given as

Fig. 2. DC-Link equivalent diagram.

1=
Where
RES.

. (1)
is the power generated from

The current flow on the other side of dclink can be represented as,

. (2)

Where
,
and
are total power
available at grid-interfacing inverter side,
active power supplied to the grid and
inverter losses, respectively. If inverter
losses are negligible then
=
B. Control of Grid Interfacing Inverter:
Fig.1. Schematic of proposed renewable
based distributed generation system
A. DC-Link Voltage and Power Control
Operation:
Due to the intermittent nature of RES,
the generated power is of variable nature.
The dc-link plays an important role in
transferring this variable power from
renewable energy source to the grid. RES
are represented as current sources
connected to the dc-link of a gridinterfacing inverter. Fig. 2 shows the
systematic representation of power transfer
from the renewable energy resources to the

Fig.3. Block diagram representation of


grid-interfacing inverter control.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

266

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

The control diagram of grid- interfacing


inverter for a 3-phase 4-wire system is
shown in Fig. 3. The fourth leg of inverter
is used to compensate the neutral current
of load. The main aim of proposed
approach is to regulate the power at PCC
during: 1)
= 0; 2) PRES< total load
power (PL); and 3) PRES > PL. While
performing the power management
operation, the inverter is actively
controlled in such a way that it always
draws/ supplies fundamental active power
from/ to the grid. If the load connected to
the PCC is non-linear or unbalanced or the
combination of both, the given control
approach also compensates the harmonics,
unbalance, and neutral current. The duty
ratio of inverter switches are varied in a
power cycle such that the combination of
load and inverter injected power appears
as balanced resistive load to the grid. The
regulation of dc-link voltage carries the
information regarding the exchange of
active power in between renewable source
and grid. Thus the output of dc-link
voltage regulator results in an active
current Im. The multiplication of active
current component (Im) with unity grid
voltage vector templates (Ua, Ub and Uc)
generates the reference grid currents (Ia*,
Ib* and Ic*). The reference grid neutral
current (In*) is set to zero, being the
instantaneous sum of balanced grid
currents. The grid synchronizing angle ()
obtained from phase locked loop (PLL) is
used to generate unity vector template.

( )

. (4)

= sin +

. (5)

( )

( ) ..

(6)

The output of discrete-PI regulator at nth


sampling instant is expressed as

. (7)
Where
and
are proportional and integral gains of dcvoltage regulator. The instantaneous
values of reference three phase grid
currents are computed as
. (8)
.. (9)
. (10)
The neutral current, present if any, due to
the loads connected to the neutral
conductor should be compensated by forth
leg of grid-interfacing inverter and thus
should not be drawn from the grid. In other
words, the reference current for the grid
neutral current is considered as zero and
can be expressed as

= sin(). (3)
= sin

ISBN: 378 - 26 - 138420 - 5

and in the generated reference current


signals. The difference of this filtered dclink voltage and reference dc-link voltage
(Vdc*) is given to a discrete- PI regulator to
maintain a constant dc-link voltage under
varying generation and load conditions.
The dc-link voltage error (Vdcerr (n)) at nth
sampling instant is given as:

.. (11)
The reference grid currents (Ia* , Ib* ,Ic*
and In*)are compared with actual grid
currents
(Ia* , Ib* ,Ic* and In*) to compute the
current errors as

The actual dc-link voltage (Vdc) is sensed


and passed through a second-order low
pass filter (LPF) to eliminate the presence
of switching ripples on the dc-link voltage

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

267

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

.... (12)

ISBN: 378 - 26 - 138420 - 5

(23)

. (13)
.. (24)

.. (14)
Similarly
. (15)

the

charging

currents

, and
on dc
bus due to the each leg of inverter can be
expressed as

These current errors are given to hysteresis


current
controller.
The
hysteresis
controller then generates the switching
pulses (P1 to Pg ) for the gate drives of
grid-interfacing inverter. The average
model of 4-leg inverter can be obtained by
the following state space equations

.. (25)
.. (26)
(27)

... (16)

... (28)

... (17)

The switching pattern of each IGBT inside


inverter can be formulated On the basis of
error between actual and reference current
of inverter, which can be explained as:

.. (18)

If
switch

, then upper
will be OFF

lower switch will be ON


phase
a
leg
of

. (19)

and
in the
inverter.
If

, then upper switch


will be ON

(20)

will be OFF
a leg of inverter.

Where
, and
are
the three-phase ac switching voltages
generated on the output terminal of
inverter. These inverter output voltages
can be modeled in terms of instantaneous
dc bus voltage and switching pulses of the
inverter as

and lower switch


in the phase

Where hb is the width of hysteresis band.


On the same principle, the switching
pulses for the other remaining three legs
can be derived.

.. (21)

(22)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

268

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

III SIMULATION RESULTS:

Fig:7 Simulation result of PQ-Grid

Fig:4 Simulation result of Grid Voltages

Fig:8 Simulation result of PQ-Load

Fig:5 Simulation result of Grid Current

Fig:9 Simulation result of Grid Voltages

Fig:6 Simulation result of unbalanced


load currents

Fig:10 Simulation result of dc-link


Voltages

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

269

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

IV CONCLUSION

ISBN: 378 - 26 - 138420 - 5

[4] U. Borup, F. Blaabjerg, and P. N.


Enjeti, Sharing of nonlinear load in
parallel-connected
three-phase
converters, IEEE Trans. Ind. Appl., vol.
37, no. 6, pp. 18171823, Nov./Dec. 2001.

Copious literature survey has been


done with fruitful discussions about
controlling of grid interfacing inverter to
improve power quality. From literature
survey it is found that hysteresis current
control technique is well suited to meet the
requirements. The proposed scheme is
validated by connecting the controller to
the 3-phase 4-wire system. This approach
eliminates the need for additional power
conditioning equipment to improve the
quality of power at PCC.

[5] P. Jintakosonwit, H. Fujita, H. Akagi,


and S. Ogasawara, Implementation and
performance of cooperative control of
shunt active filters for harmonic damping
throughout a power distribution system,
IEEE Trans. Ind. Appl., vol. 39, no. 2, pp.
556564, Mar./Apr. 2003.

Simulations have been done to


validate the performance of the proposed
system
in
MATLAB/Simulink
environment. The THD in grid current
with grid interfacing inverter comes out to
be 1.64% for unbalance non-linear load
and 1.67% for unbalanced linear load
which is well within of 5 percent limits
laid down in IEEE std.

[6] J. P. Pinto, R. Pregitzer, L. F. C.


Monteiro, and J. L. Afonso, 3-phase 4wire shunt active power filter with
renewable energy interface, presented at
the Conf. IEEE Renewable Energy &
Power Quality, Seville, Spain, 2007.
[7] F. Blaabjerg, R. Teodorescu, M.
Liserre, and A. V. Timbus, Overview of
control and grid synchronization for
distributed power generation systems,
IEEE Trans. Ind. Electron., vol. 53, no. 5,
pp. 13981409, Oct. 2006.

V REFERENCES
[1] M. Singh , K. vinod , A. Chandra, and
Rajiv K.varma. Grid interconnection of
renewable
energy sources at the distribution level
with
power-quality
improvement
features, IEEE Transactions on power
delivery, vol. 26, no. 1, pp. 307-315,
January 2011

[8] J. M. Carrasco, L. G. Franquelo, J. T.


Bialasiewicz, E. Galvn, R. C. P. Guisado,
M. . M. Prats, J. I. Len, and N. M.
Alfonso, Power electronic systems for the
grid integration of renewable energy
sources: A survey, IEEE Trans. Ind.
Electron., vol. 53, no. 4, pp. 10021016,

[2] J. M. Guerrero, L. G. de Vicuna, J.


Matas, M. Castilla, and J. Miret, A
wireless controller to enhance dynamic
performance of parallel inverters in
distributed generation systems, IEEE
Trans. Power Electron., vol. 19, no. 5, pp.
12051213, Sep. 2004.

Aug. 2006.
[9] B. Renders, K. De Gusseme, W. R.
Ryckaert, K. Stockman, L. Vandevelde,
and M. H. J. Bollen, Distributed
generation for mitigating voltage dips in
low-voltage distribution grids, IEEE
Trans. Power. Del., vol. 23, no. 3, pp.
15811588, Jul. 2008.

[3] J. H. R. Enslin and P. J. M. Heskes,


Harmonic interaction between a large
number of distributed power inverters and
the distribution network, IEEE Trans.
Power Electron., vol. 19, no. 6, pp. 1586
1593, Nov. 2004.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

270

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

An Enhancement for Content Sharing Over

Smartphone-Based Delay Tolerant Networks


P. Guru Tejaswini, K. Phalguna Rao
Student, Audisankara institute of technology,
Gudur, Nellore, Andhra Pradesh, India.
[email protected]
HOD, Audisankara institute of technology,
Gudur, Nellore, Andhra Pradesh, India.

Abstract: From the last few years, the Smartphone users has swiftly increased so peer-to-peer ad hoc content sharing is liable to crop up frequently.
As the usual data delivery schemes are not proficient for content sharing because of random connectivity amid Smartphones latest content sharing
mechanisms should be developed. To achieve data delivery in such exigent environments, researchers have anticipated the use of encounter-based
routing or store-carry-forward protocols, in this a node stores a message may be a note and carries it until a forwarding chance arises through an
encounter with other node. Earlier studies in this field focused on whether two nodes would come across each other and the place and time of
encounter. This paper proposes discover-predict-deliver as proficient content sharing scheme.Here we make use of a hidden markov model in order to
predict the future mobility information of individual's .The existing system approximately consequences in a 2 percent CPU overhead , diminish the
Smartphone battery lifetime by 15 percent .So to minimize energy consumption we propose the use of sensor scheduling schemes in an opportunistic
context.

Key words: Encounter Based Routing, Content Sharing, sensor scheduling Schemes, hidden markov model

1.Introduction
The Smartphone users have been rapidly increasing
day-by day[1]. A Smartphone consists of more advanced computing
capability and connectivity than basic phones. As interfaces of
Smartphone are more handy and accessible users can share any type
of contents like images, videos such multimedia content. But content
sharing is bothersome. It involves numerous user activities. To
minimize users burden we can depend upon an ad hoc technique of
peer-to-peer content sharing. Mobile ad hoc network is characterized
as multi-hop wireless communications between mobile device.
Smartphone's consists of many network interfaces like Bluetooth and
Wi-Fi so ad hoc networks can be easily constructed with them. The
Connectivity among Smartphone's is likely to be alternating because
of movement patterns of carriers and the signal transmission
phenomena. A wide variety of Store-carry-forward protocols have
been anticipated by researchers.

Delay Tolerant Network (DTN) routing protocols attain


enhanced performance than usual ad hoc routing protocols. Over
the anticipated DTN routing protocols, Epidemic routing is an
vital DTN routing solution. In Epidemic routing by vahdat et al[2],
messages are forwarded to each encountered node that does not have
a replica of the same message. This solution exhibits the finest
performance in terms of delivery pace and latency, but it involves
abundant resources, such as storage, bandwidth, and energy.
This paper spotlight mainly on efficiency of content
discovery and its delivery to the targeted destination. Here we
suggest recommendation based discover-predict-deliver(DPD) as
efficient and effective content sharing scheme for smart phone
based DTNs. DPD suppose that smart phones can hook up when
they are in close proximity that is where the Smartphone users reside
for a longer period. Earlier studies have shown that Smartphone
users stay indoors for a longer period where GPS cannot be
accessed.

The objective of our work is to discover solutions to the


problems in content sharing and to minimize the energy
Routing in delay-tolerant networking concerns itself consumption using sensor scheduling schemes.
with the ability to route, data from a source to a destination, which is
a vital ability of all communication networks must have. In these
2. Related work
exigent environments, mostly used or familiar ad hoc routing
A delay tolerant network (DTN) is a mobile network
protocols fail to launch routes. This is because , these protocols first
where a existing source-destination path may not exist amid a pair of
try to establish a complete route and then, once the route has been
nodes and messages are forwarded in a store-carry-forward routing
established forwards the actual data. Still, when immediate end-tohypothesis [6].
end paths are complicated or unfeasible to institute, routing protocols
The objective of our work is to discover the
should take to a "store and then forward" method or approach, where
content sharing problem in Smartphone based DTNs involves
data or a message is moved and stored incrementally all over the
minimizing energy consumption using sensor scheduling schemes.
network in hops that it will finally arrive at its destination. A general
technique used to maximize the likelihood of a message being
effectively transferred is to duplicate many copies of the message in
Content sharing in DTN'S involves the following problems:
hops that one will be successful in reaching its destination.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

271

www.iaetsd.in

limit and
query lifetime limit. We employ
the- 26controlled
INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING
AND split
TECHNOLOGY
ISBN: 378
- 138420 - 5
replication-based [9] routing scheme that performs a single-copy
scheme. This single-copy scheme turn both query lifetime and
distance limits into random walk, and the scheme is not powerful as
In this segment we examine the problem of soon as content-carrier nodes (i.e., destinations) are not eminent. By
content sharing in delay tolerant networks and depict substitute distinguishing, the controlled replication-based scheme dispenses a
solutions. As specified in the introduction, we spotlight on mobile set of message replicas and evade the excessive spread of messages.
opportunistic networking scenarios where the nodes will be
communicating using the DTN bundle protocol. A few devices in the 2.1.2. Content delivery
network store content which they are ready to share with others. All
When the
query matching content is discovered, the
nodes are willing to assist and provide a restricted amount of their
local system resources (bandwidth, storage, and dispensation power) content carrying node should transmit only a subset of results. This
to aid other nodes. Our objective is to permit users to issue queries constraint is needed to limit the amount of resources utilized both
for content that is stored on the other nodes everywhere in the locally and globally for sending and storing the responses, and to
network and consider the possibility of such a node to acquire the eliminate potential copies . The query originator sets a limit for both
required information. To ease searching, we suppose that nodes are the number of replications or duplicates and the amount of content
capable to carry out searches on their local storage and uncover the that should be produced. When nodes require to forward a query
appropriate results for a given query. The content sharing process is message,
characterized into two stages: the content discovery phase and the the limits incorporated in the query message are used to make the
content delivery phase. In content discovery phase, the user inputs forwarding decision. If the amount of the content go beyond the
or enters in a content sharing application requests for the content. response limit, the node wants to select which ones to forward.
The application initially searches the content it its own or individual
database and if not found, the application then creates a query that is
forwarded based on the users request. The content delivery phase is 2.2. Mobility Prediction
commenced, only when the content is found and the content is then
Numerous studies have largely specified
another
forwarded to the query originator.
problem of content sharing: mobility learning and prediction.
Beacon Print discover meaningful places by constantly determining
constant scans for a time period. Place Sense senses the arrival and
exit from a place by utilizing invasive RF-beacons. The system uses
a radio beacons retort rates to attain vigorous beacon conclusion.
EnTracked is a position tracking system for GPS-enabled devices.
The system is configurable to recognize different tradeoffs amid
energy consumption and heftiness.
Mobility prediction has been extensively
studied in and out of the delay-tolerant networking area. Markovbased schemes, make the problem as a Hidden Markov or semiMarkov model and probabilistic prediction of human mobility. In
contrast, neural network based schemes try to match the observed
user behaviour with earlier observed behaviour and estimate the
prospect based on the experimental patterns.
2.1. Content sharing

Figure: Processing of incoming query.


2.1.1 Content discovery
In content discovery, mainly systems spotlight on how to
formulate queries, that depends on assumptions about the format or
layout of the content to be discovered. A common protocol should
sustain various forms of queries and content, but we summarize from
the actual similar or matching process in order to spotlight on
discovering content in the network. The easiest strategy to discover
and deliver the contents is Epidemic routing. But, due to resource
limits, Epidemic routing is regularly extravagant, so we have to
consider methods that limits the system resources used up on both
content discovery and delivery. Preferably, a query should only be
forwarded to neighbours that hold on the matching contents or those
are on the pathway to other nodes having matching content .
Different nodes should return no overlapping responses to the
requester. As total knowledge or active coordination is not an
alternative in our state, one node can merely make autonomous
forwarding decisions. These autonomous forwarding decisions
should attain a fine trade off amid discovery efficiency and
necessary resources. Analogous limitations pertain to content
delivery. A few methods anticipated by Pitkanen et al. may be used
for restraining the distribution of queries. Additionally, we study two
substitutes for restraining the distribution of queries:a query distance

Markov based schemes are suitable for resource- restricted devices,


like smartphones, owing to their low computation overhead and
reserved storage requirements. In our work, we have to develop a
mobility learning and prediction method. This method has been
built to offer coarse-grained mobility information with a less
computation overhead. When the difficulty of mobility learning and
prediction scheme can be mistreated, the schemes specified in can
be worn to offer fine-grained mobility information.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

272

www.iaetsd.in

influence
evaluation of encounter opportunity
two
nodes.- 5
INTERNATIONAL
CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING
ANDthe
TECHNOLOGY
ISBN: amid
378 - 26
- 138420
3.Problem Definition
In the existing system, the energy consumption is more so
the battery lifetime will be reduced. By using sensor scheduling
mechanisms energy consumption can be reduced and can increase
the lifespan of the batteries.

For example, if two distinct places are recognized as identical ones,


we may improperly estimate that two nodes will encounter each
other when they visit two distinct places. Also, the accurate
computation of and rely on the geographical location information of
the nodes.

4. Proposed System

Figure: Mean energy consumption in a day

Figure: learning accuracy

Fig shows the daily energy utilization outline, which is calculated


using the composed stationary and movement time from seven
different users in four weeks . The scrutiny does not comprise the
energy utilization of content swap as it mainly rely on the volume
and communication pace of the nodes. The typical energy
consumptions of GPS, Wi-Fi, and the accelerometer varies. The
accelerometer do have the maximum energy consumption as it is
used endlessly over 24 hours. Wi-Fi energy utilization is observed
owed to the scanning of neighbour APs for place recognition. GPS
has a huge discrepancy in energy consumption as this may not be
available in all places.
In order to minimize energy consumption or utilization we use
sensor scheduling schemes or mechanisms[10]. Sensor systems
have an wide-ranging diversity of prospective, functional and
important applications. In any case, there are questions that have to
be inclined for prolific procedure of sensor system frameworks in
right applications. Energy sparing is one fundamental issue for
sensor systems as most of the sensors are furnished with no
rechargeable batteries that have constrained lifetime. To enhance the
lifetime of a sensor set up, one vital methodology is to attentively
schedule sensors' work sleep cycles (or obligation cycles). In
addition, in cluster based systems, grouping heads are usually
selected in a way that minimizes or reduces the aggregate energy
utilization and they may axle among the sensors to fine-tune energy
utilization. As a rule, these energy productive scheduling
components or mechanisms (furthermore called topology
arrangement components) required to accomplish certain application
requirements while sparing energy. In sensor arranges that have
various outline requirements than those in conventional remote
systems. Distinctive instruments may make characteristic suspicions
about their sensors together with identification model, sense zone,
transmission scope, dissatisfaction or disappointment model, time
management, furthermore the capability to get area and parting data.

5.Results

5.2.Discovery Efficiency
The ratio of discovered contents to the generated queries within a
specified period or time is discovery ratio or efficiency. DPDs
discovery performance is skewed to the two forwarding. In Epidemic
routing , queries are forwarded to each and every node. In hops-10
and hops-5, a query message is then forwarded till its hop count
achieve 10 and 5, correspondingly. When a query matching content
is accessible only on a small number of nodes, the discovery
methods illustrates a low discovery speed. With an rising query
lifespan, both DPD and Epidemic demonstrate a high discovery ratio
since with a longer time, each query is forwarded to more number of
nodes.

Figure: Discovery Efficiency

5.3.Prediction accuracy

Mobility prediction is a main aspect in the estimation of utility


function. Here, we estimate our prediction process according to
trajectory deviation, prediction accuracy, as shown in the figure
5.1.Learning accuracy
below. Trajectory deviation specify the abnormality of a users
Learning accuracy demonstrates how capably and mobility. For this assessment, we mutate the existing mobility
exactly the places were identified. The accuracy of place learning information with noise data. Thus, 10, 20, and 30 % of the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

273

www.iaetsd.in

meaningful places
are chosen
random INNOVATIONS
locations for IN
trajectory
[7] J. Wu,
Lu, and F. Li, Utility-Based Opportunistic
Routing
INTERNATIONAL
CONFERENCE
ONatCURRENT
ENGINEERING
ANDM.
TECHNOLOGY
ISBN: 378 - 26
- 138420in- 5
deviations of 0.1, 0.2, and 0.3, correspondingly. So as the trajectory
deviation raise, the prediction accuracy drop off. Prediction accuracy
is calculated as the ratio of accurately predicted locations to the
overall predicted locations.

MultiHop Wireless Networks, Proc. *28th Intl Conf. Distributed


Computing Systems (ICDCS 08), pp. 470-477,+ 2008.
[8] T. Spyropoulos, K. Psounis, and C.S. Raghavendra, Spray and
Wait: An Efficient Routing Scheme for Intermittently Connected Mobile Networks, Proc. ACM SIGCOMM Workshop DelayTolerant Networking (WDTN 05), pp.. 252-259, 2005.
[9] T. Spyropoulos, K. Psounis, and C. Raghavendra, *Efficient
Routing in Intermittently Connected Mobile Networks: The Single
Copy Case, IEEE/ACM Trans. Networking,/. vol. 16, no. 1, pp. 6376, Feb. 2008.
[10] Ling Shi, Michael Epstein, Bruno Sinopoli and
Richard.M.Murray," Effective Sensor Scheduling Schemes
Employing Feedback in the
Communication Loop"

Figure: Prediction accuracy

6.Conclusion
In this paper we have proposed a proficient content
sharing scheme for Smartphone based DTNs. In this we have
anticipated discover-predict-deliver as a effectual content sharing
method which is capable of discovering the content and delivers it
to the appropriate destination. The scheme also present the mobility
information of individuals. We made an try to make use of the
availability and communication technology of current Smartphone.
We have also compared our proposed scheme with traditional
schemes.
In this paper we have also proposed sensor scheduling
schemes to enhance the lifespan of a battery. By the effectiveness of
the sensing in sensor scheduling we can reduce energy consumption
of the smartphones.
Finally, our system has still has room for improvement
by considering the privacy issues.

6. References
[1] T3IGroupLLc , Http:://www.telecomweb.com, 2010.
[2] A. Vahdat and D. Becker, Epidemic Routing for Partially
Connected Ad-Hoc Networks, technical reports, Dept. of Computer
Science and engineering, Duke Univ.., Sept. 2000.
[3] A. Balasubramanian, B.N. Levine, and A. Venkataramani.,
DTN Routing as a .Resource Allocation Problem, Proc. ACM
SIGCOMM.., pp. 373-384, 2007.
[4] R.C. Shah, S. Roy, S. Jain, and W. Brunette, Data MulesModeling a Three-Tier Architecture for Sparse Sensor Networks,
Elsevier Ad Hoc Networks J...., vol. 1, pp. 215-233, Sept. 2003.
[5] A. Lindgren, A. Doria, and O. Schelen, Probabilistic -Routing in
Intermittently Connected Networks, SIGMOBILE Mobile
Computer Comm....Rev., vol. 7, no. 3, pp. 19-20, 2003.
[6] C. Liu and J. Wu, An Optimal Probabilistic Forwarding
Protocol in Delay Tolerant Networks, Proc. ACM MobiHoc, pp. 14,
2009.
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

274

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

EFFICIENT RETRIEVAL OF FACE IMAGE FROM


LARGE SCALE DATABASE USING SPARSE
CODING AND RERANKING
P. Greeshma1, K. Palguna rao2
1

Student, Audisankara institute of technology,


Gudur, Nellore, Andhra Pradesh, India.
[email protected]

HOD, Dept of CSE, Audisankara institute of technology,


Gudur, Nellore, Andhra Pradesh, India.

Abstract: Due to largely mounting of photo sharing in social network services, there is a strong need for large scale content-based face image
retrieval which is enabling device for many emerging applications. It is very exigent to find a human face image from large scale database which might
hold huge amount of face images of people. Existing methods regulate the location and illumination differences between the faces and abolish
background contents while retrieving face images, which leads to decline the important context information. This paper intends to utilize automatically
discovered human attributes that contain semantic signs of the face photos to recover large scale content-based face image retrieval. In this paper two
methods are used to develop image retrieval in both offline and online stages, they are attribute-enhanced sparse coding (ASC) and attribute-embedded
inverted indexing (AEI). Reranking is further used with these two methods to attain significant retrieval outcome. Proposed system exploits
automatically discovered human attributes which balance the information loss and attain good retrieval performance compared to existing system.

Keywords: Content-based image retrieval, face image, human attributes, sparse coding, reranking.
.

1. Introduction
The unpredictable expansion of image data escorts to the need of
explore and enlargement of Image Retrieval. Image retrieval is the
field of study concerned with searching and retrieving digital images
from a collection of database. However, Image retrieval survey
moves from keyword, to low level features and then to semantic
features. Compel towards semantic features is due to the problem of
the keywords/text which can be much distorted and time consuming
while low level features cannot always describe high level notions in
the users mind.

Figure 1: Two different people face images might be similar in low


level feature space due to lack of semantic description.

To deal with this problem, two methods are proposed


named attribute-enhanced sparse coding (ASC) and attributeembedded inverted indexing (AEI). In this paper, low level features
are integrated with high level attributes which provide semantic
descriptions. Reranking is further used with these two methods to
Large scale image search has recently fascinated significant
discard forged images and retrieve specific image results.
attentiveness due to easy accessibility of huge amount of data. Since
databases enclose even billions of samples, such large-scale search
insists extremely efficient and precise retrieval methods. CBIR has
many applications in different areas. For example, in forensics, it can 2. Related Work
help with crime investigation. The objective of face image retrieval
is to resolve the ranking result from most to least related face images This thesis is interrelated to different research fields, including
in a face image database for a specified query. For large scale CBIR, automatic discovery of human attributes and sparse coding.
datasets, it is essential for an image search application to rank the Content-based image retrieval (CBIR) has concerned significant
images such that the most relevant images are sited at the top.
deliberation over the past decade. Instead of taking query words as
input, CBIR techniques directly take an image as query and seek to
However, all present CBIR systems tolerate deficient return similar images from a large scale database. Before CBIR, the
generalization performance and accuracy as they are not capable to conventional image retrieval is typically based on text or keywords.
produce a flexible relation between image features and high-level Keyword-based image retrieval has some boundaries, they are:
concepts. Earlier schemes employ low level features (e.g., texture, Language and civilization variations always cause problems, the
color, shape) to retrieve image, but low level features doesnt afford same image is usually text out by many different ways, Mistakes
semantic descriptions of face and human face images generally such as spelling error or spell difference escort to totally different
consists of high level features (e.g., expression, posing). Therefore, results.
retrieval results are intolerable as shown in figure 1.
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

275

www.iaetsd.in

In order
to conquerONthese
restrictions,
CBIR IN
was
first
INTERNATIONAL
CONFERENCE
CURRENT
INNOVATIONS
ENGINEERING
AND TECHNOLOGY
introduced by Kato [1]. The term, CBIR, is widely used for
retrieving desired images from a large collection, which is based on
extracting the features from images themselves. In general, the
purpose of CBIR is to present an image conceptually, with a set of
low-level optical features such as color, texture, and shape [2].

Large scale
database

ISBN: 378 - 26 - 138420 - 5

Query image

Preprocessing

One major difficulty when creating CBIR system is to make


a system general-purpose. CBIR for common function image
databases is a highly exigent problem because of the enormous
volume of the databases, the obscurity of accepting images both by
people and computers, and concern of evaluating results properly.
All these methods endure from low recall problems due to semantic
gap. Semantic image presentations were initiated to bridge this
semantic gap. In this paper, automatic discovery of human attributes
are used to construct sparse codewords for face image retrieval
operation, instead of using identity information which necessitate
manual interpretation.

Face
detection

Aligned face

Facial landmark
detection

Face alignment

Attribute detection

Local patches
Patch-level
LBP features

3. Problem Definition
Several schemes were introduced for face image retrieval, but all
these techniques endure with some limitations. Previous mechanisms
for face image retrieval stabilize the location and lighting differences
between the faces and prohibit background contents. Such common
approaches give up the important context information. Using
automatic discovery of human attributes it is possible to balance
such information loss. Existing methods uses low level features for
face retrieval that have lack of semantic descriptions about face
which escorts to poor results. Also all current retrieval systems
struggle with low recall problems which diminishes the performance
of the system. By concerning all these factors, this paper proposes
two methods (ASC, AEI) which detect the human attributes
automatically to resolve recall problems. Reranking method is united
with these two methods to boost up system performance.

Attribute enhanced
sparse coding

Patch-level
sparse codewords

Reranking image
results (no forged
images)

Attribute embedded
inverted indexing

Reranking

Image results
(with forged
face images)

Figure 2: proposed system framework

4. Proposed System
This paper intends to employ automatic discovery of human
attributes which contains semantic signs of face image for efficient
large scale face image retrieval. To improve CBIR two methods are
used named ASC and AEI that combines low level features with
high level concepts to afford semantic descriptions of face. Another
technique named reranking is used with these two methods to
improve the performance of retrieval system.

4.1 Attribute Enhanced Sparse Coding (ASC)

The attribute-enhanced sparse coding is used in offline stage which


describes the automatic detection of human attribute from the image
and also generates the codewords to the image in the database by
combining the low level features with high-level attributes to give
semantic description about the face. By incorporating low level and
high level attribute it is possible to get promising result to retrieve
For each image in the database, initially apply Viola-Jones
similar faces from large scale database.
face detector [3] to discover the pose of faces. Active shape model
[4] is applied to trace facial landmarks on the image. By using these 4.2 Attribute Embedded Inverted Indexing (AEI)
facial landmarks, next align every face in the image with the face
mean shape [5] using barycentric coordinate based mapping process. In on-line image retrieval, the user can submit a query image to the
For each identified facial part, take out grids, each grid is retrieval system to search for desired images. Retrieval is performed
represented as a square patch [6]. Extract an image patch from each by applying an indexing scheme to afford an efficient way of
grid and examine LBP feature descriptor as a local feature. After searching the image database. In the end, the system indicates the
attaining local feature descriptors, quantize all descriptors into search results and then returns the results that are related to the query
codewords by attribute enhanced sparse coding.
image. Attribute embedded inverted indexing collects the sparse
codewords from the attribute- enhanced sparse coding and check the
Attribute embedded inverted index is then construct for codewords with the online feature database and retrieve the related
efficient retrieval. When a query image appears, it will experience images similar to the query image.
the same procedure to attain sparse codewords and human attributes,
and these codewords are used with binary attribute signature to
retrieve images from large scale database, if those retrieving images
contain forged images then apply reranking method to evade those
forged images. Figure 2 demonstrate the framework of our system.
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

276

www.iaetsd.in

4.3 Reranking CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING


automatic
of human attributes for
content-based
face- 5
INTERNATIONAL
AND discovery
TECHNOLOGY
ISBN:
378 - 26 - 138420
The fundamental proposal of reranking is to evade forged images
present in database. A human face image employs very light and
compulsive global signatures, so using reranking we can improve the
accuracy but without trailing the scalability. Human images contain
discrepancies provoked by changes in pose, expression, and
illumination, taking all these intra-class variations into account, a
technique called reranking is used to elude such discrepancies.
Reranking is applied to set of retrieved human face images to evade
forged images. Reranking technique is vibrant to erroneous images.
Below figure shows how forged images are rejected after applying
reranking method.

5. Experimental Results
we estimate and visualize results using real examples. Figure 3
illustrates the results of ranking images. Red boxes in the figure
specify forged images. After using the reranking method majority of
images are correct. In figure 4 the graph demonstrates how efficient
the image retrieval is in the proposed method. In proposed system
individual attribute detection can be done within few milliseconds,
because here we are using automatic attribute detection.

image retrieval. Attribute-enhanced sparse coding is used in the


offline stage and present semantic descriptions of face. Attributeembedded inverted used in the online stage and ensures image
retrieval. Finally, reranking method is used to discard forged images
and obtain accurate image results.

References
[1] M. Lew, N. Sebe, C. Djeraba and R. Jain, Content-based
Multimedia Information Retrieval: State of the Art and Challenges,
Transactions on Multimedia Computing, Communications, and
Applications, 2006.
[2] N. Krishnan, and C. Christiana, Content-based Image Retrieval
using Dominant Color Identification Based on Foreground Objects,
Fourth International Conference on Natural Computation, 2008.
[3] Rapid object detection using boosted cascade of simple features,
P. Viola and Jones, IEEE Conference on Computer Vision and
Pattern Recognition, 2001.
[4] Locating facial features using extensive active shape model,
European Conference on Computer Vision, S. Milborrow and F.
Nicolls, 2008.
[5] Face image retrieval and matching with biometrics, IEEE
Transactions on Information Forensics and Security, U. Park and A.
K. Jain, 2010.
[6] Scalable face image retrieval using identity based quantization
and multi reference relevance ranking, Z. Wu, Q. Ke, and H.Y.
Shum, IEEE Conference on Computer Vision and Pattern
Recognition, 2010.

Figure 3: Reranking Image Results

Figure 4: Query processing time per image

6. Conclusion
All current image retrieval systems standardize the position and
illumination variations between the faces and abolish background
contents while retrieving face images, which escorts to refuse the
vital context information. In this paper, two schemes are projected
and united to extend automatically detected human attributes to
extensively develop content-based face image retrieval. This is the
primary idea ofASSOCIATION
uniting low-level
features, high&level
attributes and
INTERNATIONAL
OF ENGINEERING
TECHNOLOGY
FOR SKILL DEVELOPMENT

277

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

ELIMINATING HIDDEN DATA FROM AN IMAGE


USING
MULTI CARRIER-ITERATIVE GENERALISED
LEAST SQUARES
Author1:Ch.AnushaM.tech,Guide2:V.sireeshaM.tech (PhD)
1

Student, Computer science and engineering.


Audisankara institute of technology
Nellore, Andhra Pradesh, India
[email protected]

Associate professor, Computer Science and Engineering


Audisankara institute of technology
Nellore, Andhra Pradesh, India
[email protected]

ABSTRACT:Data hiding and extraction schemes are growing in todays communication world suitable to rapid growth of data tracking and tampering
attacks. Data hiding, a type of steganography, embeds information into digital media for the reason of identification, annotation, and copyright. In this
narrative techniques are used for addressing the data-hiding method and estimate these techniques in glow of three applications: copyright protection,
tamper proofing, and augmentation data embedding. Thus we necessitate a proficient and vigorous data hiding schemes to defend from these attacks.
In this project the blindly extraction method is measured. Blindly extraction means the novel host and the embedding carriers are not necessitate
to be recognized. Here, the hidden data embedded to the host signal, via multicarrier SS embedding. The hidden data is extracted from the digital media
like audio, video or image. The extraction algorithm used to extract the hidden data from digital media is Multicarrier Iterative Generalized Least Squares
(M-IGLS).

KEY WORDS:Data hiding, Blind Extraction, Data tracking, Tampering attacks, Steganography.
distorted or not. Equally the data extraction schemes also offer a
good recovery of hidden data.This is the purpose of the protected
communication.

1. INTRODUCTION
Data hiding, while comparable to compression, is divergent from
encryption. Its objective is not to limit or standardize admittance to
the host signal, other than slightly to guarantee that embedded data
continue inviolate and recoverable. Two significant uses of data
hiding in digital media are to afford evidence of the copyright, and
assertion of content integrity. Data tracking and tampering are
hastily rising in all over the place like online tracking,Mobile
tracking etc. hence we require a tenable communication scheme for
transmitting the data. Forthat, we are having lots of data hiding
schemes and extraction schemes. Data hiding schemes are
primarily used in military communication systems similar to
encrypted message, for finding the sender and receiver or its
extremely subsistence. Originally the data hiding schemes are used
for the copy write purpose.[1]Breakable watermarks are used for
the certification purpose, i.e. to find whether the data has been

2. RELATED WORK
The techniques used for data hiding contrast depending on the
magnitude of data being hidden and the mandatory invariance of
those data to exploitation. Since that no one method is proficient of
achieving each and every one these goals, a group of processes is
considered necessary to extent the variety of likely applications.
The procedural challenges of data hiding are terrible. There are
numerous data hiding and data extraction schemes are comes into
existence. The key data hiding procedure is steganography. It is
fluctuate from cryptography in the means of data hiding. The target
of steganography is to conceal the data from a third party where the
purpose of cryptography is to create data incomprehensible by a
third party.

1
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

278

www.iaetsd.in

INTERNATIONAL
ON CURRENT
INNOVATIONS
IN ENGINEERING
AND TECHNOLOGY
In this [2] CONFERENCE
steganalysis process
is used.
The ambition
of
substitute
the "intricate

ISBN:
378of
- 26
- 138420
5
areas" on the bit
planes
the
vessel- image
with the secret data. The most significant feature of Steganography
is that the embedding capability is incredibly huge. For a 'normal'
image, approximately 50% of the data might be disposable with
secret data earlier than image damage becomes perceptible.

[3] steganalysis is to decide if an image or additional carrier


contains an embed message. To enhance the protection and
payload speed the embedder will acquire multicarrier embedding
model in the [4] spread spectrum communication is explained.
Here a contracted band signal is transmitted above a lot better
bandwidth such that the signal force nearby in any particular
frequency is unnoticeable. Correspondingly in [5] SS embedding
scheme, the secret data is extend over many samples of host signal
by adding a low energy Gaussian noise progression. In[6] the
Generalized Gaussian Distribution (GGD) has-been used to form
the statistical performance of the DCT coefficients. In[7] there are
many extraction measuresto search for the hidden data. But it is
havingsome drawback. IterativeLeast Square Estimation (ILSE) is
unaffordabledifficult even for judicious values. Pseudo-ILS (ILSP)
algorithm is not definite to congregate in universal and also it
afforddemonstrablybadoutcome.So, these two algorithms united
and so called Decoupled weighted ILSP(DW-ILSP).But at this
juncture also have an drawback like ,it cannot be applicable for
huge N.

Fig: steganographic model


3.2 Multi-Carrier Spread Spectrum Embedding

3. PROPOSED SYSTEM

The procedure of spread spectrum may possibly permit partially to


fulfill the above requirements. The embedding technique is
intended to assure the perceptual limit and advance the perceive
capability as well as the embedding charge. As a substitute of the
pixel rate, the histogram can be customized to embed the data. If
we observe distinctive histograms of DCT coefficients we will
locate some trial include high amplitudes that the widespread
Gaussian technique cannot effectively established. We will believe
the DCT coefficients whose amplitude is beneath a confident
threshold importance. In this embedding proposal, the hidden data
is widen over various test of host signal or image by totaling the
DCT coefficient as the carrier.
Advantages of spread spectrum
procedures are broadly well-known: Invulnerability against multipath alteration, no necessitate for frequency preparation, high
elasticity and uneven data rate transmission. The propensity of
diminishing multiple access interference in direct-sequence codedivision-multiple-access system is specified by the crosscorrelation properties of spreading codes. In the case of multi-path
transmission the ability of distinctive one section from others in the
complex received signal is obtainable by the auto-correlation
properties of the scattering codes. The following figures show
entered data, transform data using DCT, embedded image
respectively.

The proposed method employs blind resurgence of data and it


utilizes the DCT transform as a carrier for insert the data in digital
media. Insert is achieved by using multicarrier SS embedding
procedure. It uses M-IGLS algorithm for the removal of the
concealed data. It is a low convolution algorithm and offer tough
improvement performance. It achieves equal prospect of fault
recovery to identified host and embedding carriers. It is used as a
performance study tool for the data thrashing scheme.
The proposed system includes 4 techniques:
1. Steganography
2. Multicarrier spread spectrum embedding
3. Image encryption and watermarking
4. Image decryption and extraction

3.1 Steganography
Steganography can be used to hide a message deliberate for
afterward reclamation by a definite person or collection. In this
case the intent is to avoid the message being perceived by any
other revelry.Steganography includes the cover up of information
inside computer files..The other major area of steganography is
copyright marking, where the message to be included is used to
declare patent over a article. This can be further divided into
watermarking and fingerprinting. In digital steganography,
electronic communications may include steganographic coding
inside of a transportlayer, such as a document file, image file,
program or protocol
Digital steganography can conceal top secret data (i.e. secret
files) extremely strongly by embedding them into some media data
known as "vessel data." The vessel data is also referred to as
"carrier, cover up, or replica data". In Steganography images used
for vessel data. The embedding action put into practice is to

Fig: Data entered

2
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

279

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


The host image

ISBN: 378 -grey


26 - 138420
5
is an 8-bit or privileged
height- image
which has to perfectly be the similar dimension as the basictext
image or else resized consequently with the same magnitude.

Pre-conditioning the cipher and the complication practice


are take on using a Discrete Cosine Transform (DCT).The output
will comprise negative hovering point numbers ahead pleasing the
real constituent of a intricate array. The array must be correct by
totaling the biggest negative rate in the output array to the
equivalent array prior to normalization. For color host images, the
twofoldcoded text can be included into single or entire of the RGB
components. The binary plaintext image should include
homogeneous margins to minimize the special effects of buzzing
due to edge effects when dealing out the data using Cosine
transform. The following figure shows the embedding of

Fig: DCT transformation

watermark and detecting watermark.

Fig: Watermarked Image


3.4 Image decryption and extraction
Decryption is exactly the reverse procedure of encryption. When
the receiver obtains encrypted image, extraction of the data from
random values and flag values are to be done. This extraction of
data from image is considered as the highlighting factor.
The steps to perform M-IGLS algorithm for extracting
data from an image is as follows:
Initialize B^ irrationally and swap step wise stepamong
(1) and (2) to accomplish at every pace conditionallyindiscriminate
least squares rough of one matrix bound particular the further.
The equations used above for computation are
V^GLS=argVRLxR||Rz-1/2(Y-VB)||2F
=YBT(BBT)-1
(1)
binary
KxM
B^
min ||Rz-1/2(Y-VB)||2F
GLS=argB{1}
sgn {(VTRy-1V)-1VTRy-1Y}
(2)

Fig: Embedded Image

3.3 Image encryption and watermarking

Encryption is the method of converting the information for its


protection. Many image substance encryption algorithms have
been projected. To create the data safe from a variety of assault and
for the reliability of data we should encrypt the data prior to it is
transmitted or accumulated. Government, military, financial
institution, hospitals and private business covenant with
confidential images about their patient (in Hospitals), geographical
areas (in research ), enemy positions (in defense), product,
financial status.

End when convergence is accomplished. Observe that (2)


stimulateunderstanding of the autocorrelation matrix Ry, which
can be conservative by figure averaging over the expected data
interpretation,
R^y = 1/MMm=1y(m)y(m)T.
The M-IGLS extraction algorithm is review in Table I.
Superscripts signify iteration index. The computational density of
every iteration of the M-IGLS algorithm is
O(2K3+2LMK+K2(3L+M)+L2K)
and, experimentally, the number of stepsis accomplished between
20 and 50 in broad-spectrum

Imperceptible digital watermarks are a innovative


technology which could solve the trouble of make compulsory
the patent of content transmitted across shared networks. They
allow a patent holder to insert a concealed message (invisible
watermark) within images, moving pictures, sound files, and even
raw text. moreover, the author can supervise traffic on the shared
network for the occurrence of his or her watermark via network
system. Because this method obscure both at ease of the message
(cryptography) and the occurrence of the message (steganography)
an imperceptible watermark is very hard to eradicate.

Table-1
Multi-carrier iterative generalized least squares
Algorithm

3
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

280

www.iaetsd.in

CONFERENCE
ON CURRENT
INNOVATIONS IN ENGINEERING
AND TECHNOLOGY
ISBN: 378 - 26
- 138420 - which
5
1) d INTERNATIONAL
:= 0; initialize B^(0)
2 {1}KM
arbitrarily.
system
is to afford a good quality extraction
technique
2) d := d + 1;
measured the blindly improvement of data. This technique uses the
V^(d) := Y(B^(d-1))T[B^(d-1)(B^(d-1))T]-1;
M-IGLS algorithm for the extraction. The data is entrenched via
B^(d) := sgn{(V^(d))TRy^-1 (V^(d)))-1(V^(d))T Ry^-1 Y}
DCT transform by multicarrier SS embedding. This extraction
3) Repeat Step 2 until B^(d) = B^(d1).
procedure will afford high signal to noise fraction and it will
achieve the possibility of fault improvement equals to notorious
host and embedding carriers. This method is improved by using
harmony search algorithm where it offers small time utilization and
RESULTS
high assault confrontation.

The proposed technique is to remove the concealed data from the


digital media. Here blindly improvement of data is measured. That
is the original host end embedding carrier is not necessitating to be
known. This technique uses multicarrier embedding and DCT
transformation for the embedding the data into the host image. The
M-IGLS algorithm is used for the extraction purpose. The
following figure shows extracted data and graph for existing and
proposed.

REFERENCES
[1]

[2]

[3]

[4]

[5]

[6]

[7]

F.A.P.Petitcolas, R.J.Anderson, and M.G.Kuhn.Information hiding. A


survey,,Proc. IEEE, Special Issue on Identification and Protection
ofMultimedia Information, vol. 87, no. 7, pp. 10621078, Jul. 1999.
S. Lyu and H. Farid, Steganalysis using higher-order image
statistics,IEEE Trans. Inf. Forensics Security, vol. 1, no. 1, pp. 111
119, Mar.2006.
G. Gul and F. Kurugollu, SVD-based universal spatial domain image
steganalysis, IEEE Trans. Inf. Forensics Security, vol. 5, no. 2, pp.349
353, Jun.2010.
. M. Gkizeli, D. A. Pados, and M. J. Medley, Optimal signature design
for spread-spectrum steganography, IEEE Trans. Image Process.,
vol.16, no.2, pp. 391405, Feb. 2007
. C. Fei, D. Kundur, and R. H. Kwong, Analysis and design of
watermarking algorithms for improved resistance to compression, IEEE
Trans. ImageProcess., vol. 13, no. 2, pp. 126144, Feb. 2004.
C. Qiang and T. S. Huang, An additive approach to transformdomaininformation hiding and optimum detection structure, IEEE
Trans. Multimedia, vol. 3, no. 3, pp. 273284, Sep. 2001.
T. Li andN.D. Sidiropoulos, Blind digital signal separation using
successive interference cancellation iterative least squares, IEEE Trans
Signal Process., vol. 48, no. 11, pp. 31463152, Nov. 2000

Fig: Extracted data

Fig: Graph for extracted data

CONCLUSION AND FUTURE WORK


Data tracking and tampering is speedily growing in
communication. So we have to lock the data from the trackers
.Hence we require a vigorous and protected data hiding and
extraction format. The most important accord of the proposed

4
INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

281

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Implementation of Context Features using ContextAware Information Filters in OSN


A Venkateswarlu1, B Sudhakar2.
1

M.Tech , Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India.

Assoc Professor, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P

Abstract Information filtering has becoming a

which specifies the filtering rules to filter the

key technology for current information systems.

unwanted messages.

The goal of information filter is to filter the


unwanted messages from user walls in online
social networks. In order to deal with huge
amount of messages, the information filtering is
very crucial. In the current system we study the
filter walls a rule-based information filtering
system in which we implement the filtered rules
and discuss the blacklist rules in online social
network such as facebook. But the current
system not considers the context features of a
message. To address this problem, in this paper
we propose context aware information Filtering
(CAIF) using the EAGER Algorithm.
Index terms Information Filtering, Contextaware information filtering, AGILE Algorithm,

But these methods or filtered rules are inefficient


in identifying context updates of the message
and message routing. To address this problem,
this paper presents Context-aware Information
Filters (CAIF). In contrast to the current
information filters, a CAIF has two input
streams: (i) an order of messages that need to be
routed and (ii) a stream of context information
updates. This way, a CIF provides a undefined
solution to information delivery for the routing
of

messages

and

to

handle the

context

information.
The problem of implementing Context-aware
Information Filters is that the two goals to route
messages and record context updates efficiently

indexing.

are crucial. The current approaches to identifying


I.

INTRODUCTION

profiles are very efficient in routing messages,

Information filter systems manage continuous

but are inefficient when it comes to processing

streams of messages that must be routed

context updates.

according to rules or so-called profiles. In order


to

implement

information

filters,

several

methods have been proposed in the past. The


focus of all that work was on the development of
scalable rule-based system for filtering system

To close this gap, this paper presents AGILE.


AGILE is generic way to extend existing
structures in order to make them consistency to
context updates and achieve a huge amount of
message throughput at the same time.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

282

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

A) Usecases for CAIF

ISBN: 378 - 26 - 138420 - 5

The main issue for context-based information

To implement better understanding of what a


context is and what the benefits of a contextware information filter are a few usecases are as

filters can be summarized as follows: Given a


large set of profiles, high message rates and
varying rates of context updates, provide the best
possible throughput of messages. No message

follows:

must be dropped or sent to the wrong user


Massage Broker with State

because a change in context has not yet been

A message broker routes messages to a specific

considered by the filter. This constraint rules out

application and location. Each message can

methods

change the state of the receivers and affects

periodically. In the following, we define the

future routing decisions dynamically.

terms context, profile and message.

Generalized Location-Based Services

i)

With an increased availability of mobile, yet

A context is set of attributes associated with an

network-connected devices, the possibilities for

entity the values of those attributes can change at

personalized

varying rates.

information

delivery

have

multiplied. So far, those services mostly use very


little context information, such as the location of
the device. A more solution is to extend those
systems to a more elaborate context.

that

update

the

context

only

Context

Gathering context information is outside the


scope of this paper and has been addressed, e.g.,
in the work on sensor networks, data cooking,
context/sensor fusion, or the Berkeley HiFi

Stock Brokering

project.

Financial information systems require sending

ii) Messages

only the relevant market updates to specific


applications or brokers.

A message is a set of attributes associated to


values.

To sum up, some contexts have a high update

iii) Profiles

rate, others have a low update rate, but many


have

varying,

bursty

update rates.

All

examples involve a high message rate and a


large number of profiles.

A profile is a continuous query specifying the


information

interests

of

subscriber.

Expressions in profiles can refer to a static


condition or a dynamic context. Static conditions

Skipping updates in order to reduce update rates


has to be avoided because it leads to costly errors
in information filtering.

change relatively seldom, since they specify


abstract

interests.

In

contrast,

context

information can change frequently. We define


profiles as proposed in pub/sub-systems, using

II. PROBLEM STATEMENT

the disjunctive normal form (DNF) of atomic


comparisons. This definition allows the use of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

283

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

context information in profiles in multiple ways:

Find all profiles that match the message m,

Message attributes can be compared with

considering the current state. Return this set of

constants and with context attributes. The latter

profiles.

is the novel aspect of this work and a significant

2. update_context(Context x, Attribute a, Value

extension to the way profiles are defined in

v):

traditional pub/sub systems and information


filters. This extension enables easier modeling,
richer profiles and, as we will see, several

III. CONTEXT AWARE INFORMATION FILTERS


This section introduces a processing model and
architecture

for

v, i.e., c.a:=v. All profiles referencing this


context must consider this new value.

opportunities for optimization.

reference

Set the attribute a of context c to the new value

Context-Aware

Information Filters (CIF).

b) CIF Architecture
As shown also in Figure, a CIF has four main
components:

(a)

context

management,

(b)

indexes, (c) merge, (d) postifier. A similar


architecture without context management has
also been devised for information filters.
i)
The

Context management
first

component

manages

context

information. It stores the values of static


attributes and values of context attributes which
are used in predicates of profiles. Any context
change is recorded by this component.
Fig. 1 Architecture of a CIF
a) CIF Processing Model

This component interacts heavily with indexes


and postfiltering, which both consume this
information. Both indexes and postfiltering

Figure 1 shows the processing of a CIF. The CIF

require values of constants and context attributes

keeps profiles of subscribers and context

in order to evaluate predicates of profiles for

information. The CIF receives two input streams:

incoming messages.

a message stream and a context update stream.


These two streams are serialized so that at each

ii) Indexes

point in time either one message or one update is

Given a message, the information filter must find

processed.

all profiles that match. This filtering can be

In order to deal with the two input streams, a CIF


must support the following methods:

accelerated by indexing the profiles or predicates


of the profiles. The most important method
supported by an index is probe, which is invoked

1. Handle_message(Message m):

by the CIFs handle message method. Probe


takes a message as input and returns a set of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

284

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

profiles that potentially match that message.


Furthermore, an index provides insert and delete
methods in order to register new profiles or
delete existing profiles.

i)

ISBN: 378 - 26 - 138420 - 5

General Idea

The key idea of AGILE is to dynamically reduce


the accuracy and scope of an index if context
updates are frequent and to increase the accuracy

iii) Merge

and scope of an index if context updates are

Profiles are conjunctions and disjunctions of


predicates. Since it is not efficient to use a highdimensional index to cover all conjunctions and
disjunctions, an individual index typically only
covers one type of predicate. Therefore, several
potential indexes are probed in order to process a

seldom and handle message calls are frequent.


This way, AGILE tries to act in the same way as
the NOINDEX approach for update context
operations and like the EAGER approach for
handle message operations, thereby combining
the advantages of both approaches. In order to do
so,

message.

AGILE

generalizes

techniques

from

PARTIAL and GBU.


Several optimization techniques for this merge
operation exist. One idea is to optimize the order
in which intersection and union operations are
applied. Another class of optimizations involves
the algorithms and data structures used to

The operation to reduce the accuracy is called


escalation; it is triggered by context updates in
order to make future context updates cheaper.
The operation that increases the accuracy of an
index is called de-escalation; it is triggered by

implement the intersect and union operations.

handle message events in order to make future


iv) Postfilter

message

processing

more

efficient.

Both

The last step of processing a message eliminates

operations are carried out in the granularity of

false positives. This step is necessary if

individual index entries. This way, the index

inaccurate indexes are used or if the merge

remains accurate for profiles that are associated

operation does not involve all kinds of

with contexts that are rarely updated and the

predicates. The Postfilter operation takes a set of

index moves profiles that are associated with

profiles as input and checks which profiles

contexts that are frequently updated out of scope.

match

the

As a result, AGILE only escalates and

predicates of the profiles based on the current

deescalates as much as necessary and can get the

state of the context.

best level of accuracy for all profiles. AGILE is

the

message

by

reevaluating

IV. ADAPTIVE INDEXING : AGILE

generic and can be applied to any index


structure; in particular, it can be used for the

This section presents the AGILE (Adaptive


Generic Indexing
algorithm.

with

Local

Escalations)

index

structures

devised

specifically

for

information filtering. It works particularly well


for hierarchical index structures.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

285

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Escalation
This operation is triggered by increasing the
stock of Warehouse by one; i.e., a context update

ISBN: 378 - 26 - 138420 - 5

result of a handle message operation in order to


reduce the number of false positives for future
calls to handle message.

from two to three. Rather than carrying out an

The advantages of AGILE are that it effectively

insert and delete on the binary tree, the

combines the advantages of the NOINDEX and

escalation moves the Identifier up to the left set

EAGER approaches in a fine-grained way. It can

of the parent node.

deal with workloads in which certain contexts

De-escalation

are frequently updated by escalating the entries


for those contexts: in the extreme case to the root

This operation is triggered if the handle message


operation is called several times for orders of,
say, three or four items and Warehouse A was
returned by the index as a potential candidate

of the data structure or even outside of the scope


of the index. Likewise, AGILE is suitable for
workloads in which context updates are seldom
and many messages need to be processed.

and had to be filtered out by the Postfilter step. If


this happens, AGILE decides to deescalate the
index entry for Warehouse in order to improve
the performance of future calls to the handle
message operation.

The basic idea of AGILE is not new. In some


sense, R-Trees and other indexes for spatial data
apply a similar approach. In R-Trees, identifiers
are associated with bounding boxes which can
also be interpreted as sets of keys.

ii) Properties of AGILE Indexes


The difference is that AGILE uses escalatations
The escalate and deescalate operations can be
implemented on any index structure, and thus
AGILE can be used to extend any index
structure for context-aware information filtering.
The insert and delete operations of an index are
not modified and are the same as in the basic
(non AGILE) version of the index. However,
AGILE allows efficient implementations of the
update operation that assigns a new value to an
identifier.

and de-escalations in order to control the


accuracy of the index, whereas an R-Tree does
not adjust its accuracy depending on the
update/probe workload.
iii) AGILE Algorithm
Based on the escalate and deescalate operations,
we are now ready to present the algorithms for
the

handle

message

and

update

context

operations. The algorithm for the handle


message operation is almost the same as the

What are escalations and de-escalations in this

algorithm for the EAGER approach. The only

framework? Both operations re-assign a new set

difference is that a special implementation of the

of keys to an identifier. Escalations are carried

Postfilter operation is used (Lines 2 to 9). If a

out as a result of update context operations in

profile is detected to be a false positive (i.e., test

order to make future calls to update context

in Line 3 fails for one of the predicates of the

cheaper. De-escalations are carried out as a

profile), then a DEpolicy function (Line 5)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

286

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

determines whether to deescalate the index for

(4)

that profile. The function that tests the profile

(5) EndFor

(test) returns 0 if the profile matches the

(6) DataStore[c].a := v;

ISBN: 378 - 26 - 138420 - 5

EndIf

message. If not, test returns a reference to the


index for the predicate that failed; this index
CONCLUSION

returned the profile as a false positive and


therefore is a candidate for de-escalation (Line

Information filtering has matured to a key

6). Since de-escalation is an expensive operation,

information processing technology. Various

it is not necessarily a good idea to carry it out

optimization techniques and index structures

whenever a false positive occurs.

have been proposed for various kinds of

The

algorithm

for

update

context

is

straightforward. In the first step (Lines 2 to 5), it


checks whether an escalation is necessary. In the
second step (Line 6), the context is updated, just
as in every other CIF approach.

applications, data formats and workloads. This


paper picked up this challenge by providing
simple extensions to existing index structures for
information filtering systems. We called this
approach AGILE for Adaptive Generic Indexing
with Local Escalations. The proposed extensions

Function AGILE.handle message


Input: Message m

any index structure. The key idea is to adapt the

Output: Set of matching profiles RES


(1) RES := merge(AGILEindex[1].probe(m),...
AGILEindex[N].probe(m))
(2)

forEach p in RES

(3)

f := test(p, m)

(4)

if (f > 0)

(5)

RES := RES \ {p}

(6)

if (DEpolicy(p))

(7)

AGILEindex[f].deescalate(p)

(8)
(9)
(10)

are universal and can, in theory, be applied to

endIf

accuracy and scope of an index to the workload


of a context-aware information filter or more
generally, to information filtering and message
routing with state.
REFERENCES
[1] S. Abiteboul. Querying Semi-Structured
Data. In ICDT, 1997.
[2] N. Beckmann, H. Kriegel, R. Schneider, and
B. Seeger. The R*-Tree: An Efficient and

endIf

Robust Access Method for Points and

EndFor

Rectangles. In SIGMOD, 1990.

(11) return RES


Procedure AGILE.update context

[3] R. R. Brooks and S. Iyengar. Multi-Sensor

Input: Context c, Attribute a, Value v

Fusion: Fundamentals and Applications in

(1) For (i:=1 to Att)

Software. Prentice Hall, 1997.

(2)

If (AGILEindex[i] indexes a ^
(c.a AGILEindex[i].probe(v))

(3)

AGILEindex[i].escalate(c, v)

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

287

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

[4] H.-J. Cho, J.-K. Min, and C.-W. Chung. An

ISBN: 378 - 26 - 138420 - 5

[8] P. T. Eugster, P. A. Felber, R. Guerraoui, and

Adaptive Indexing Technique Using Spatio-

A.-M.

Kermarrec.

Temporal Query Workloads. Information

Publish/Subscribe.

and Software Technology, 46(4):229241,

35(2):114131, 2003.

The

Many

ACM

Faces

Comput.

of

Surv.,

2004.
[9] F. Fabret, H. A. Jacobsen, F. Llirbat, J.
[5] O. Cooper, A. Edakkunni, M. J. Franklin, W.

Pereira, K. A. Ross, and D. Shasha. Filtering

Hong, S. R. Jeffery, S. Krishnamurthy, F.

Algorithms and Implementation for Very

Reiss, S. Rizvi, and E. Wu. HiFi: A Unified

Fast

Architecture for High Fan-in Systems. In

SIGMOD, 2001.

Publish/Subscribe

Systems.

In

VLDB, 2004.
[6] A. K. Dey. Understanding and Using
Context.

Personal

and

Ubiquitous

Computing Journal, 5(1):47, 2001.


[7] Y. Diao, M. Altinel, M. J. Franklin, H.
Zhang, and P. Fischer. Path Sharing and
Predicate Evaluation for High-Performance
XML

Filtering.

TODS,

28(4):467516,

2003.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

288

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

ENHANCEMENT OF FACE RETRIVAL DESIGEND FOR


MANAGING HUMAN ASPECTS
Devarapalli Lakshmi Sowmya 1, V.Sreenatha Sarma2
Student, Dept of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India
2Associate Professor, Dept of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India
1 M.Tech

ABSTRACT:
Traditional methods of content-based image retrieval make use of image content like color,
texture as well as gradient to symbolize images. By combine low-level characteristics with
high level human features, we are capable to discover enhanced feature representations and
attain improved retrieval results. The recent effort explains automatic attribute recognition
has sufficient quality on numerous different human attributes. Content-based face image
retrieval is strongly associated to the problems of face recognition however they focus on
finding appropriate feature representations in support of scalable indexing systems. To
leverage capable human attributes automatically identified by attribute detectors in support of
getting better content-based face image retrieval, we put forward two orthogonal systems
named attribute enhanced sparse coding as well as attribute-embedded inverted indexing.
Attribute-embedded inverted indexing believes human attributes of chosen query image in a
binary signature as well as make available resourceful recovery in online stage. Attributeenhanced sparse coding make use of global structure and employ quite a lot of human
attributes to build semantic-aware code words in offline stage. The projected indexing system
can be effort less integrated into inverted index, consequently maintaining a scalable
structure.
Keywords: Content-based image retrieval, Attribute recognition, Feature representations,
Binary signature, Semantic-aware code words.
substantial computation outlay in support
of dealing with high dimensional features
as well as generate explicit models of
classification, it is non-trivial to directly
pertain it towards tasks of face retrieval.
Even though images obviously have
extremely high dimensional representation,
those within similar class generally lie on a
low dimensional subspace [1]. Sparse
coding can make use of semantics of
information and attain capable results in
numerous different applications for
instance image classification as well as
face recognition. Even though these works
accomplish significant performance on
keyword based face image recovery as
well as face recognition, we put forward to
make use of effectual ways to merge lowlevel features and automatically noticed
facial attributes in support of scalable face
image retrieval [11]. Human attributes are
high level semantic description concerning
an individual. The recent effort explains

1. INTRODUCTION:
In the recent times, human attributes of
automatically detected have been revealed
capable in various applications. To get
better the value of attributes, relative
attributes were applied. Multi-attribute
space was introduced to standardize
assurance scores from various attributes
[4]. By means of automatically detected
human attributes, excellent performance
was achieved on retrieval of keyword
based face image as well as face
verification detectors in support of search
of similar attribute. Due to increase of
photo sharing or social network services,
there increase tough needs for extensive
content-based retrieval of face image.
Content-based face image retrieval is
strongly associated to the problems of face
recognition however they focus on finding
appropriate feature representations in
support of scalable indexing systems [8].
As face recognition generally necessitate

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

289

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

automatic attribute recognition has


sufficient quality on numerous different
human attributes. Using human attributes,
numerous researchers have attained
capable results in various applications for
instance face verification, identification of
face, keyword based face image recovery
as well as similar attribute search [3]. Even
though human attributes have been
revealed
practical
on
applications
associated towards face images, it is not
trivial towards concerning in retrieval task
of content-based face image due to quite a
lot of reasons. Human attributes simply
enclose limited dimensions. When there
are more over numerous people in dataset,
it loses discrimin ability as assured people
may have comparable attributes [14].
Human attributes are represented as vector
concerning floating points. It does not
effort well with increasing extensive
indexing methods, and consequently it
suffers from slow reply and scalability
concern when the data size is enormous.
2. METHODOLOGY:
Traditional methods of content-based
image retrieval make use of image content
like color, texture as well as gradient to
symbolize images [13]. Traditional
methods in support of face image retrieval
typically employ low-level features to
correspond to faces but low-level
characteristics are be short of of semantic
meanings as well as face images typically
include high intra class variance thus the
recovery results are unacceptable [9].
When specified a face image query,
retrieval of content-based face image effort
to discover comparable face images from
huge image database. It is an enabling
knowledge for numerous applications
includes automatic face annotation,
investigation of crime. By combining lowlevel characteristics with high-level human
features, we are proficient to discover
enhanced feature representations and attain
improved retrieval results [7]. To deal with
extensive information, mainly two types of
indexing systems are employed. Numerous
studies have leveraged inverted indexing

ISBN: 378 - 26 - 138420 - 5

or else hashbased indexing pooled with


bag-of-word representation as well as
local features to attain well-organized
similarity search. Even though these
methods can attain high precision on rigid
object recovery, they go through from low
recall difficulty due to semantic gap [2].
The significance as well as sheer amount
of
human
face
photos
makes
manipulations of extensive human face
images actually significant research
difficulty and facilitate numerous real
world applications. In recent times, some
researchers have fixed on bridging
semantic gap by discovery of semantic
image representations to get better
performance of content-based image
retrieval [16]. To leverage capable human
attributes automatically identified by
attribute detectors in support of getting
better content-based face image retrieval,
we put forward two orthogonal systems
namedchosen query image in a binary
signature as well as make available
resourceful recovery in online stage [12].
Attribute-enhanced sparse coding make
use of global structure and employ quite a
lot of human attributes to build semanticaware code words in offline stage. By
incorporating these methods, we put up an
extensive content-based face image
retrieval scheme by taking benefits of lowlevel features as well as high-level
semantic [5]. When a query image arrives,
it will experience same process to get hold
of sparse code words as well as human
attributes, and make use of these code
words by binary attribute signature to get
back images in index system. By means of
sparse coding, a mark is a linear grouping
of column vectors of dictionary [15]. As
learning dictionary with a huge vocabulary
is lengthy we can presently make use of
randomly sampled image patch as
dictionary and pass over prolonged
dictionary learning measure. To believe
human attributes in sparse illustration, we
initially put forward to make use of
dictionary selection to compelte images
with various attribute values to enclose

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

290

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

various codewords [10]. For a single


human aspect, we separate dictionary
centroids into two various subsets, images
all the way through positive attribute
scores will make use of subset as well as
images by negative attribute scores will
employ the other [6]. For cases of
numerous attributes, we separate sparse
representation into numerous segments
based on number of features, and every
section of sparse representation is
produced depending on distinct aspect.

ISBN: 378 - 26 - 138420 - 5

indexing system can be effortlessly


integrated
into
inverted
index,
consequently maintaining a scalable
structure. Certain informative attributes
were discovered in support of face
retrieval across various datasets and these
aspects are capable for other applications.
Attribute-enhanced sparse codewords
would additionally get better accurateness
of retrieval of content-based face image.
4. CONCLUSION:
Even though human attributes have been
revealed
practical
on
applications
associated towards face images, it is not
trivial towards concerning it in retrieval
task of content based face image due to
quite a lot of reasons. Traditional methods
in support of face image retrieval typically
employ low level features to correspond to
faces but low-level characteristics are be
short of semantic meanings as well as face
images typically include high intra-class
variance thus the recovery results are
unacceptable. Using human attributes,
numerous researchers have attained
capable results in various applications for
instance face verification, identification of
face, keyword based face image recovery
as well as similar attribute search.
Numerous studies have leveraged inverted
indexing or else hash based indexing
pooled with bag-of word representation as
well as local features to attain wellorganized similarity search. To believe
human attributes in sparse illustration, we
initially put forward to make use of
dictionary selection to compel images with
various attribute values to enclose various
code words. The significance as well as
sheer amount of human face photos makes
manipulations of extensive human face
images actually significant research
difficulty and facilitate numerous real
world applications. In recent times, some
researchers have fixed on bridging
semantic gap by discovery of semantic
image representations to get better
performance of content-based image
retrieval. The experimental result illustrate
that by means of codewords generated by

Fig1: An overview of structure of proposed system

3. RESULTS:
Attribute-enhanced sparse coding make
use of global structure and employ quite a
lot of human attributes to build semanticaware code words in offline stage.
Attribute embedded inverted indexing
additionally believe local attribute
signature concerning attribute-enhanced
sparse coding as well as attributeembedded inverted indexing as shown in
fig1.
Attribute-embedded
inverted
indexing believes human attributes of
query image and still make sure proficient
recovery in online stage. The experimental
result illustrate that by means of code
words generated by projected coding
system, we can decrease the quantization
error as well as attain salient gains in face
recovery on public datasets. The projected

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

291

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

projected coding system, we can decrease


the quantization error as well as attain
salient gains in face recovery on public
datasets.
REFERENCES:
[1] J. Wright, A. Yang, A. Ganesh, S.
Sastry, and Y. Ma, Robust face
recognition via sparse representation,
IEEE Transactions on Pattern Analysis and
Machine Intelligence (PAMI), 2009
[2] Scalable Face Image Retrieval using
Attribute-Enhanced Sparse Code words
Bor-Chun Chen, Yan-Ying Chen, Yin-Hsi
Kuo, Winston H. Hsu,2013
[3] G. B. Huang, M. Ramesh, T. Berg, and
E. Learned-Miller, Labeled faces in the
wild: A database for studying face
recognition in unconstraine denvironments
, University of Massachusetts, Amherst,
Tech.Rep. 07-49, October 2007.
[4] Y.-H. Kuo, H.-T. Lin, W.-H. Cheng,
Y.-H. Yang, and W. H.
Hsu
,Unsupervised auxiliary visual words
discovery for large scale image object
retrieval, IEEE Conference on Computer
Vision and Pattern Recognition, 2011
[5] J. Yang, K. Yu, Y. Gong, and T.
Huang, Linear spatial pyramid matching
using
sparse
coding
for
image
classification, IEEE Conference on
Computer Vision and Pattern Recognition,
2009
[6] N. Kumar, A. C. Berg, P. N.
Belhumeur,
and
S.
K.
Nayar,
Describablevisual attributes for face
verification and image search, in IEEE
Transactionson Pattern Analysis and
Machine
Intelligence
(PAMI),
SpecialIssue
on
Real-World
Face
Recognition, Oct 2011
[7] O. Chum, J. Philbin, J. Sivic, M. Isard
and
A.
Zisserman,
Total
Recall:Automatic Query Expansion with a
Generative
Feature
Model
for
ObjectRetrieval,
IEEE
International
Conference on Computer Vision, 2007
[8] A. Torralba, K. P. Murphy, W. T.
Freeman, and M. A. Rubin, Context base
dvision system for place and object

ISBN: 378 - 26 - 138420 - 5

recognition, International Conference on


Computer Vision, 2003.
[9] J. Wang, J. Yang, K. Yu, F. Lv, T.
Huang, and Y. Gong, Locality
constrained linear coding for image
classification, IEEE Conference on
Computer Vision and Pattern Recognition,
2010.
[10] Z. Wu, Q. Ke, J. Sun, and H.-Y.
Shum, Scalable face image retrieval with
identity-based quantization and multireference reranking, IEEE Conference on
Computer Vision and Pattern Recognition,
2010.
[11] W. Scheirer and N. Kumar and P.
Belhumeur and T. Boult, Multi-Attribute
Spaces: Calibration for Attribute Fusion
and Similarity Search, IEEE Conference
on Computer Vision and Pattern
Recognition, 2012.
[12] J. Wang, S. Kumar, and S.-F. Chang,
Semi-supervised hashing for scalable
image retrieval, IEEE Conference on
Computer Vision andPattern Recognition,
2010.
[13] M. Douze and A. Ramisa and C.
Schmid, Combining Attributes and Fisher
Vectors for Efficient Image Retrieval,
IEEE Conference on Computer Vision and
Pattern Recognition, 2011
[14] S. Lazebnik, C. Schmid, and J. Ponce,
Beyond bags of features :Spatial pyramid
matching for recognizing natural scene
categories,IEEE Conference on Computer
Vision and Pattern Recognition, 2006
[15] H. Jegou, M. Douze, and C. Schmid,
Hamming
embedding
and
weak
geometric consistency for large scale
image search, European Conference on
Computer Vision, 2008
[16] W. Scheirer, N. Kumar, K. Ricanek,
T. E. Boult, and P. N. Belhumeur, Fusing
with context: a bayesian approach to
combining
descriptive
attributes,
International
Joint
Conference
on
Biometrics, 2011.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

292

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Design and Implementation of secure cloud systems using


Meta cloud
Perumalla Gireesh
M.Tech 2nd year, Dept. of CSE, ASCET, Gudur, India
Email:[email protected]
_____________________________________________________________________________________

Abstract Cloud computing has recently emerged

way from existing schemes to solving this problem

as a new paradigm for hosting and delivering

effectively. But it does not consider the users data

services over the Internet. Cloud computing is

privacy in transforming Meta cloud. To address

attractive to business owners as it eliminates the

this problem, we introduce Business Continuity

requirement

Management (BCM). This is defined as a holistic

for

users

to

plan

ahead

for

provisioning, and allows enterprises to start from

management

the small and increase resources only when there

organization and reduces the impacts of data

is a rise in service demand. However, despite the

leakage issues.

fact

Index terms Meta cloud, Cloud Privacy, private

that

cloud

computing

offers

huge

opportunities to the IT industry, the development

process

that

identifies

to

an

clouds, security.

of cloud computing technology is currently at its

I. INTRODUCTION

infancy, with many issues still to be addressed. In


With the rapid development of processing and

this paper, we present a survey of cloud


computing,

highlighting

its

architectural

principles,

and

key

storage technologies and the success of the

concepts,

Internet, computing resources

state-of-the-art

have

become

cheaper, more powerful and more ubiquitously

implementation as well as research challenges.

available than ever before. This technological


trend has enabled the realization of a new

Meta cloud based on a combination of existing

computing model called cloud computing, in

tools, concepts and provides the convenient to

which resources (e.g., CPU and storage) are

organize the private clouds. This can consider the

provided as general utilities that can be leased and

only vendor lock-in problem of different vendors

released by users through the Internet in an on-

in cloud. For that Meta cloud provides an abstract

demand fashion.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

293

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

The cloud computing paradigm has achieved

a) The Key Challenges

widespread adoption in recent years. Its

Being virtual in concept, the cloud environment

success is due largely to customers ability to

generates several questions in the minds of users

use services on demand with a pay-as-you go

with respect to confidentiality, integrity and

pricing model, which has proved convenient

availability. The key challenges for the adoption

in many respects. Low costs and high

of the cloud are as given below:

flexibility make migrating to the cloud

Assurance of the privacy and security

compelling. Despite its obvious advantages,

The cloud users are wary of the security and

however, many companies hesitate to move

privacy

to the cloud, mainly because of concerns

environment of the cloud is causing concerns

related to service availability, data lock-in,

amongst enterprises. As the same underlying

data security and legal uncertainties.

hardware may be used by other companies and

A previous study considers the data lock-in

competitors, it may lead to a breach of privacy.

problem and provides a convenient way to

Moreover, any data leakage or virus attack would

of

their

data.

The

multi-tenant

have a cascading effect on multiple organizations.

solve this using Meta cloud. The problem is


that once an application has been developed

Reliability and availability

based on one particular providers cloud

Instances of outages at the facilities of the cloud

services and using its specific API, that

service providers have raised concerns over the

application is bound to that provider;

reliability of the cloud solutions. Enterprises are

deploying it on another cloud would usually

recognizing that they would have to deal with

require completely redesigning and rewriting

some level of failures while using commoditybased solutions. Also, the cloud providers cannot

it. Such vendor lock-in leads to strong

give an assurance on the uptime of their external

dependence on the cloud service operator.

internet connection, which cloud shut all access to

The Meta cloud framework contains the

the cloud.

following components: Meta cloud API, Meta


Data Security is Key Concern

cloud proxy, resource monitoring and so on.


But sometimes, transforming cloud as meta

There are a number of concerns surrounding the

cloud data security issues are raised which are not

adoption of the cloud especially because it is a

consider in the previous study.

relatively new concept. Assuring customers of


data security would be one of the biggest
challenges for cloud vendors to overcome. The

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

294

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

II.

Figure 1 shows the chart of the key barriers to


cloud adaptation.

ISBN: 378 - 26 - 138420 - 5

PROPOSED WORK

In this section we introduce a novel solution


Business continuity management (BCM) and
provide the overview of the Business Continuity
Management.
a) Business Continuity Management (BCM)
The BCMS will use to Plan Do Check Act
approach. The PDCA approach can be applied to
every element of the BCM lifecycle.
Business Continuity leads (BC leads)

Figure 1 Chart of the key barriers

Leads for business continuity management will be

To address this problem this paper introduce

appointed in each directorate, regional, area team

the Business Continuity Management (BCM)

and hosted bodies within the strategy.

is defined as a holistic management process that

BC leads will perform the following:

identifies to an organization and reduces the


Promote business continuity Management

impacts of data leakage issues. This contains


following stages Project initiation, understand the

Receive BC training

organization, BC strategies, develop Business


continuity planning, and Apply BCP. The

Facilitate the completion of BIAs

Business Continuity Planning is shown in the

Develop BCPs

following Figure 2.
Ensure that BCPs are available during
incident response
Ensure that incident responders receive
training appropriate to their role
Ensure that plans are tested, reviewed and
updated
Participate in the review and development
of the BCMS.
Figure 2 Business Continuity Management
Overview

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

295

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Business Continuity working Groups

Stage 1- Understanding the Organization

Working groups may be established to:

Understanding the

Take control of resource allocation Set priorities


Set

continuity strategies

organizations

objectives

in line
and

ISBN: 378 - 26 - 138420 - 5

business

is

essential

in

developing an appropriate BCM Programme. A

with the

detailed understanding of what processes are

responsibilities

essential to ensure continuity of prioritized

Establish the measures that will be used to assure

activities to at least the minimum business

the BCMS remains current and relevant Report to

continuity objective level will be achieved by

top management on the performance of the

undertaking BIA. The BIA will incorporate

BCMS.

continuity requirements analysis which may

Emergency preparedness resilience and response

include

(EPRR)

qualifications required for prioritized activities.

The business continuity program will have close

BIAs will describe as follows:

links to EPRR because both desciplines aim to

the

staff

skills,

competencies

and

The prioritized activities of departments/

ensure the organization is resilient and able to

teams;

respond to threats and hazards. The BCMS


described in this strategy will ensure that the

The impact that the incidents will have on

organization is able to manage risks and incidents

prioritized activities

that directly impact on its ability to deliver

How long we could continue using the

business as usual.

emergency measures before we would have to

Assurance

restart our normal activities;

The National support centre will maintain an

A description of the emergency measures

overview of the BCMS. BC leads will be reuired

we have in place to deal with an incident;

to report on progress within their areas.


The threats to the continued delivery of
BCM Documentation

priority activate.

The National Support Centre will be given access


Stage 2 Determining BCM strategy

to related documentation by areas within the


scope, such as BCPs, training records, incident

BIAs will create a picture of the

records and exercises to facilitate the sharing of

organizations

good practice throughout the organization. The

and business continuity risks. This information

Business

will be used to:

Continuity

management

has

the

dependencies,

vulnerabilities

following stages:

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

296

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

To assist in deciding the scope of the

continuity plans will be based on different levels

BCM programme.

of response and escalation.

To provide the information from which

Business Continuity Plans

continuity options can be identified and

Various plans will continue to be developed to

evaluated.

identify the actions that are necessary and the


resources which are needed to enable business

To assist the preparation of detailed plans

continuity. Plans will be based upon the risks


Decisions that determine business continuity

identified, but will allow for flexibility.

strategies will be made at an appropriate level

Prioritized activity recovery plans (PARPs)

Recovery

Priority activities are those activities to which

People

priority must be given following an incident in


order to mitigate the impact. Activities of the

Premises

highest priority are those that if disrupted, impact


the organization to the greatest extent and in the

Technology and information

shortest possible time.


Suppliers and partners

Stage 4 Exercise, Audit, Marinating and


reviewing

Stage 3 Developing and implementing a BCM

Exercises

response

It is essential that regular BC exercises are carried

This stage considers the incident reporting

out to ensure that plans are tested and continue to

structure,

be effective and fit-for-purpose as operational

business

continuity

plans,

and

Prioritized activity recovery plans.

processes and technology configurations are


constantly changing. Exercise will rise awareness
of BCM procedures.

Incident Reporting Structure


There

are

various

sources

of

information

Audit

pertaining to business continuity threats such as

To

severe, flooding and soon.

organizations BCM polices and standards

The impact of all incidents will vary. It is

To

important that the response to an incident is

solutions

appropriate to the level of impact and remains

To validate the organizations BCM plans

flexible as the situation develops. Business

To verify that appropriate exercising and

validate

review

compliance

the

with

organizations

the

BCM

maintenance activities are taking place. To

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

297

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Business continuity exercises

highlight decencies and issues and ensure their


resolution

CONCLUSION

Management Review
An annual review of this strategy will be

In this paper we introduce a novel solution to

undertaken. However, events may prompt more

provide a convenient way to process to identify the

frequent re-examination, such as:

various security threats. This paper considers a

BIA

substantive

revision
changes

which
in

survey of Business continuity management (BCM)

identifies

processes

to avoid the security risks.

and

priorities;
A

REFERNCES

significant

assessment

and/or

change

in

the

risk

appetite

threat
of

[1] ISO 22301 Societal Security - Business

the

Continuity

organization
New

Management

Systems

Requirements.
regulatory

or

legislative

[2] NHS England Core Standards for Emergency

requirements.

Preparedness,

Resilience

and

Response

(EPRR).

Embedding BCM in the Organizations


culture

[3] J. Skene, D.D. Lamanna, and W. Emmerich,


Precise Service Level Agreements, Proc.

BCM must be an accepted management

26th Intl Conf. Software Eng. (ICSE 04),

process, full endorsed and actively promoted

IEEE CS Press, 2004, pp. 179188.

by directors. The communication of high-level

[4] Q. Zhang, L. Cheng, and R. Boutaba, Cloud

endorsement to all is essential. There are


various ways in which this be achieved:

Computing: State-of-the-Art and Research


Challenges,

Business continuity will be part of the

J.

Internet

Services

and

Applications, vol. 1, no. 1, 2010, pp. 718.

organizations induction for new starters

[5] The Route Map to Business Continuity

Participation in BIA and writing BCPs

Management: Meeting the Requirements of


Communication

of

risks,

alerts

and

ISO 22301.

incidents
Business continuity information will be
available on the staff intranet
Business continuity training

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

298

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

AUTHORS

Mr.P.Gireesh received the


Vaishnavi

Instutiate

Technology,

of

Tirupathi,

B.Tech degree in computer


science & engineering from
the Jawaharlal Nehru technological university
Anantapur, in 2011, and received the Audisankara
College of Engineering and Technology, Nellore
M.Tech degree in computer science engineering
from

the

Jawaharlal

Nehru

technological

university Anantapur in 2014, respectively. He


Participated National Level Paper Symposiums in
different
Networks,

Colleges.
Mobile

He

interests

Computer

Computing,

Network

Programming, and System Hardware. He is a


member of the IEEE.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

299

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


1

ISBN: 378 - 26 - 138420 - 5

ASYNCHRONOUS DATA TRANSACTIONS ON SoC USING FIFO


BETWEEN ADVANCED EXTENSIBLE INTERFACE 4.0 AND ADVANCED
PERIPHERAL BUS 4.0
A VIJAY KUMAR*, T VINAYSIMHA REDDY**,M SANTHOSHI***.
*

ECE DEPARTMENT, MRCET, INDIA.


ECE DEPARMENT, MRCET, INDIA.
***
ECE DEPARMENT, CVRCC, INDIA.
**

ABSTRACT : Recently, VLSI technology has improved


significantly and more transistors can be integrated into a chip.
This makes the ideal of system-on-a-chip (SoC) more of an
achievable goal than an abstract dream. The on-chip-bus (OCB)
which connects silicon intellectual property (SIP) in a system-ona-chip (SoC) plays a key role in affecting the system
performance. Advanced Microcontroller Bus Architecture
(AMBA) bus protocol has been proposed by ARM community to
justify the uneven demand of integrity. Recently, a new
generation of packet-based OCB protocol called Advance
extensible Interface 4.0 (AXI4.0) has been proposed. AMBA
AXI4.0 protocol system supports 16 masters and 16 slaves
interfacing. It supports for removal of locked transactions. . This
paper presents a proj e ct aimed to do data transactions on SoC
to low speed APB4.0 from the AXI4.0 using asynchronous
FIFO.A asynchronous FIFO has been considered to avoid the
complex hand shaking mechanism. By setting write pointer, read
pointer, empty flags & full flags for the read operation and write
operation the data will be transmitted between them. This paper
modelled in a Verilog hardware description language (HDL) and
simulation results for read and write operation of data and
address are shown in ISE simulator.
.Keywords - AMBA Bus Protocol, APB 4.0, AXI4.0,
Asynchronous FIFO, FPGA Vertex 3, SoC , VerilogHDL, and
ISE Simulator.
1.INTRODUCTION
Recently due to the miniaturization of semiconductor process
technology and computation for survival in the current market
conditions constant customization is required. VLSI technology
has improved significantly and more transistors can be integrated
into a chip. This makes the ideal of system-on-a-chip (SoC) It
may consists all intellectual property blocks on a single chip
substrate. IP is an inevitable choice with size constraint. A SoC
platform usually consists of various design components
dedicated to specified application domains SoC buses are used to
interconnect an Intellectual Property (IP) core to the other
intellectual property core. They are reside in Field Programmable
Gate Array (FPGA). The AMBA (Advanced Microcontroller Bus
Architecture) on-chip interconnect system is an established open
specification that details a strategy on the interconnection and
management of functional blocks that makes up a SoC. AMBA
was introduced in the year 1996 by ARM limited.
AMBA has four generations of buses. The details of the
generations and their interfaces are as given below:
AMBA specification

Advanced System Bus (ASB)


Advanced Peripheral Bus (APB)

AMBA 2.0
Advanced System Bus (ASB)
Advanced Peripheral Bus (APB2 or APB)

AMBA 3.0
Advanced eXtensible Interface ( AXI v1.0)
Advanced High-performance Bus Lite

(AHB-Lite v1.0)
Advanced Peripheral Bus (APB3 v1.0)
Advanced Trace Bus (ATB v1.0)

AMBA 4.0
AXI Coherency Extensions (ACE)
AXI Coherency Extensions Lite (ACE-Lite)
Advanced eXtensible Interface 4 (AXI4)
Advanced eXtensible Interface 4 Lite (AXI4Lite)
Advanced eXtensible Interface 4 Stream (AXI4-Stream v1.0)
Advanced Trace Bus (ATB v1.1)
Advanced Peripheral Bus (APB4 v2.0)
AMBA defines both a bus specification and a technologyindependent methodology for designing, high-integration
embedded controllers. Advanced eXtensible Interconnect
(AXI4.0) was introduced in AMBA 3.0 as the successor on-chip
bus protocol of the AHB in AMBA 2.0. The AXI4.0 protocol is a
high-performance and high bandwidth bus includes a number of
features that make it suitable for high-speed submicrons
interconnect.
1.1 AMBA axi4.0 Architecture
AMBA AXI4 supports burst and unaligned data transfers. In
AMBA AXI4.0 system i n t e r fa c e 16 masters to slaves. Each
master and slave has their own 4 bit identification tags. .
AMBA AXI4 system c o n s i s t s o f m a s t e r , s l a v e a n d
b u s i n t e r c o n n e c t s . The AXI4.0 consists of five channels
namely write address, write data, read data, read address, a n d
w r i t e r e s p o n s e ch an n el s.
The A X I 4 . 0 protocol supports the following mechanisms:
Burst and unaligned data transfers and up-dated write
r e s p on s e a ckn owl e d g m en t .
A burst data bits wide 8, 16, 32, 64, 128, 256,512 or
1024 bits.
Updated AWCACHE and ARCACHE signalling
details.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

300

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


2

The AMBA AXI4.0


specification satisfies four key
requirements:
To design high speed embedded micro controller
products with one or more CPUs signal processors
System macro-cells and highly reusable peripheral can
be migrated across a diverse range IC processes and be
appropriate for full-custom ,standard cell and gate array
technologies
To improve independence of processor ,providing a
development road-map for advanced cached CPU cores
and the development of peripheral libraries
For success efficient on-chip and off-chip
communication
To encourage modular system design to improve processor
independence, providing a development road- The AMBA 4.0
specification defines five buses/interfaces.
Advanced eXtensible Interface (AXI)
Advanced High-performance Bus (AHB)
Advanced System Bus (ASB)
Advanced Peripheral Bus (APB)
Advanced Trace Bus (ATB)

ISBN: 378 - 26 - 138420 - 5

arid, arcache, arlock, arprot, arburst]. The addresses of read


and write operations are validated by VALID signals and sent to
interface unit. The acknowledgement signals of slave will
become incoming signals to master.

Fig.3 Write address and data burst


The write operation process starts when the master sends an
address and control information on the write address channel
as shown in figure.3. The master then sends each item of write
data over the write data channel. The VALID signal is low by
master until the write data available in write data channel .The
WLAST signal goes HIGH for the master last data. Slave
drives a write response signal BRESP[1:0] back to the master
to indicate that the write transaction is complete When the it
has accepted all the data items, . The completion of the
write transaction indicated by this signal. OKAY, EXOKAY,
SLVERR, and DECERR are the allowable responses by the
slave to the master.

Fig.1 Block diagram of AMBA AXI4.0 bus interconnect.


Fig.4 Block diagram of AXI4.0 slave
The data transfer occurs on the read data channel when the read
address appears on the address bus hence the data transfer occurs
on the read data channel as shown in figure 5. The VALID signal
LOW by the slave until the read data is present in read data
channel. For the last data transfer of the burst signal, the slave
provides the RLAST signal to show that the last data item is
being moved. The status of the read transfer indicated by the
RRESP[1:0] signal ,OKAY, EXOKAY, SLVERR, and DECERR
are the allowable responses..
Fig.2 Block diagram of AXI 4.0 master
To perform write address and data operation the
transaction is initiated with concatenated input of [awaddr, awid,
awcache, awlock, awprot, awburst]. On the same lines for read
address and data operations the concatenated input is [araddr,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

301

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


3

ISBN: 378 - 26 - 138420 - 5

Fig.7 Finite state machine of axi4.0 read and write operation.

Fig.5: Read address and data burst

1 . 2 AMBA AXI4.0 Interconnect


The interconnect block consists of decoder and arbiter. When
more than one master initiates a transaction simultaneously,
the priority is given by the arbiter to access the bus. The
address sent by any of the master is decoded by the decoder
and the control goes to one slave out of 16 and also address
goes by one of the salve out of 16. The AMBA AXI4.0
interface decoder is centralized block.

2. PROPOSED WORK
The AXI4.0 to APB4.0 Bridge provides an interface between the
high-performance AXI domain and In this proposed work data
transactions can be done between high speed AXI4.0 bus to low
power APB4.0 bus using asynchronous FIFO.Which act as a
interface between them. Read and write transfers on the AXI4.0
bus are converted into corresponding transfers on the APB4.0.
2.1 Asynchronous FIFO
Asynchronous FIFOs are widely used in the computer
networking industry to receive data at a particular clock and
transmit them at another clock. An asynchronous FIFO has two
different clocks one for read and one for write. There are issues
that arise when passing data over asynchronous clock values.
Data could be over-written and hence lost when the write clock is
faster than the read clock, . In order to overcome these problems,
control signals like write pointer, read pointer, empty and full
flags are required.

Fig 6: Signals used to design AMBA AXI4.0 modules

1.3 Finite state machine of AXI 4.0

Fig.8 Asynchronous FIFO state machine

Write and read operations between the channels can be


explained by using below finite state machine.

2.1.1 Functional description of A Asynchronous


FIFO
Functionally the FIFO works as follows: At reset, the write and
read pointers are both at null value. This is the empty condition
of the FIFO, and empty is pulled high (we use the active high
convention) and full is low. At null or empty, read operations are
stopped and so the only write operations are possible in to the
FIFO. Location 0 of the RAM nothing but the array loaded by
Write and increments the write pointer to. This results the empty
flag to go LOW. Consider that there are subsequent cycles only
write to the FIFO and no reads, there will be a certain time when

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

302

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


4

the write pointer will equal array _size -1. This means that the
last location in the array is the next location that will be written
to. At this condition, the write pointer to become 0 due to write,
and set full. Note that in this condition the write pointer and read
pointers are equal, and not empty or null but the FIFO said to be
full. This implies that the FIFO full/empty decision depends on
the write pointer and read pointers when they will become equal
due to the read or write operation between them but not based on
the pointer values alone. When the pointer equality is a reset or a
read, the FIFO said to be empty; if the cause is a write, the FIFO
said to be full.
Now consider that we begin a multiple reads. The read pointer is
incremented by each read operation till the point where the read
pointer equals array_ size -1. At this point, the data from this
location of RAM or array size is available on the output bus of
FIFO.Succeeding logic reads this data and provides a read signal
(active for one clock). And hence the read pointer to become
equal to the write pointer again (after both pointers have
completed one cycle through the array). However, empty is set
since pointers are equal due to read operation.

ISBN: 378 - 26 - 138420 - 5

2.4 Mechanisms of AXI4.0 & APB4.0


2.4.1 Flow chart representation
A diagram of the sequence of movements or actions of people or
things involved in a complex system or activity. A graphical
representation of a computer program in relation to its sequence
of functions (as distinct from the data it processes).It can be
applicable for all the engineering branches.
The flow of moving data from FIFO between protocols is as
follows
If the FIFO is empty then empty flag indicates to the
write pointer to write data into the FIFO and then write
pointer incremented.
If the FIFO is full then full flag indicates to the read
pointer to readout data from the FIFO to the APB
master.

Fig.10 Flowchart of general FIFO

2.4.2 Finite state machine representation of


Asynchronous FIFO

Fig.9 Circuit for write and read pointers of FIFO

2.2 AXI4.0 Slave Interface


The AXI4.0 Slave Interface module provides a bi-directional
slave interface to the AXI. The AXI address and data bus widths
are always fixed to 32-bits and 1024bits.When both write and
read transfers are simultaneously requested on AXI4.0, more
priority is given for read request and less priority is given for the
write request. This block also contains the data phase time out
logic for generating OK response on AXI interface when APB4.0
does not respond.

2.3 APB4.0 Master Interface


The APB4.0 module provides the APB master interface on the
APB. This interface can be APB2 or APB3 or APB4, these are
chosen by setting the generic C_M_APB_PROTOCOL. When
C_M_APB_PROTOCOL=apb4.
At the APB Interface
M_APB_PSTRB and M_APB_PPROT signals are driven. The
APB address bus and data bus widths are fixed to 32-bits.

FSM is a mathematical tool used to design system programs and


digital circuits. It is a behaviour model composed of a finite
number of states, transitions among all states, similar to a flow
graph in which one can inspect the way logic runs when certain
conditions are met. It is considered as an abstract machine, which
is in only one state at a time. The current state is a state which is
present at particular time .By initiated a triggering event or
condition, It can change from one state to another state is called a
transition for that state. A finite state machine is a mathematical
abstraction model .According AXI4.0 specification, the read
address channel, write address channel, read data channel and
write data channel are completely independent. A read and a
write requests may be issued simultaneously from AXI4, the
AXI4.0 to APB Bridge of asynchronous will give more priority
to the read request and less priority to the write request. That is,
the write request is initiated on APB4.0 after the read is
requested on APB4.0 when both write and read requests are
valid.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

303

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


5

ISBN: 378 - 26 - 138420 - 5

Fig.11 FSM of Asynchronous FIFO

2.5 Advanced Peripheral Bus (APB4.0) Protocol


The Advanced Peripheral Bus (APB) is a part of AMBA
Architecture protocol family. It t is optimized for minimal power
consumption and reduced interface complexity. It defines a lowcost interface. To enable the integration of APB4.0 peripherals
easily into any design flow all signal transitions are only related
to the rising edge of the clock. Every transfer takes At least two
cycles are required for the every data transfer by APB4.0.The
APB can interface with the AMBA Advanced High-performance
Bus Lite (AHB-Lite) and AMBA AXI4.0 Advanced Extensible
bus.

2.6 Handshake Mechanism of AXI4.0 & APB4.0


In AXI 4.0 specification, VALID and READY signals are
present in each and every channel for hand shaking mechanism.
When the control information or data is available then source
asserts VALID and when destination can accept the control
information or data then it asserts READ signal. When both the
VALID and READY signals are asserted then only transfer takes
place. Note that when source asserts VALID, the corresponding
control information or data must also be available, at the same
time at the positive edge of clock transfer takes place. Therefore,
the source and destination need register inputs to sample the
READY and VALID signals. Therefore source and destination
should use combinational circuit as output. In short, AXI4.0
protocol is suitable input register and combinational output
circuit.
The APB4.0 Bridge buffers address, control and data from
AXI4.0, drives the APB4.0 peripherals and returns data and
response signal to the AXI4.0.By using internal address map it
decodes the address to select the peripheral. The bridge is
designed to operate when the APB4.0 and AXI4.0have
independent clock frequency and phase by using asynchronous
FIFO. For every AXI4.0 channel, invalid commands are not
forwarded and an error response generated. the APB Bridge will
generate DE CERR response through the response channel (read
or write) if peripheral accessed does not exist. And it asserts
PSLVERR if the target peripheral exists, but, it will give a
SLVERR response.

Fig.12 Block diagram of signal connections between AXI4.0 and


APB4.0

3. SIMULATION AND SYNTHESIS RESULT

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

304

Fig.13 Channels in AXI 4.0

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


6

ISBN: 378 - 26 - 138420 - 5

Fig.14 Burst data transactions between AXI4.0 master and slave


Fig.18 AXI 4.0 to APB4.0 RTL schematic view

Fig.15 AXI4.0 RTL Schematic view


Fig.19 Top module RTL schematic view of AXI4.0 to APB4.0

Fig.16 AXI 4.0 Technology schematic view

Fig.20 AXI 4.0 to APB4.0 tech schematic view

3. CONCLUSION

Fig.17 Data transaction between AXI4.0 and APB4.0

The implementation asynchronous data transactions on


SoC using FIFO between Advanced eXtensible
Interface and Advanced peripheral Bus is designed
.Verilog HDL has been used for implementing bridge
between them.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

305

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY


7

The bridge defines low cost interface optimized for


minimal power consumption and reduced interface
complexity.

4. FEATURES

32-bit AXI4.0 slave and APB4.0 master interfaces.


PCLK clock domain completely independent of clock
domain.
Support up to 16 APB4.0 peripherals
Burst length is 32 bits
PREADY signal is supported, which translates to wait
states on AXI.
An error on any transfer results in SLVERR as the
AXI4.0 read/write response.

5. FUTURE SCOPE

The design will be extended by developing a total


system around it.
This work provides an ideal platform for enhancement
or further development of the Bus Bridge Design
between AXI4.0 and APB4.0 protocols.
When read request and write request are simultaneously
occurs the bridge gives the high priority to read request.
this condition creates the race condition in APB4.0
Bridge.
For date item from the FIFO and now the read date
FIFO was empty. Again read and write requests are
example FIFO has only one date item to read operation.
If read and write requests are simultaneously occurs the
bridge first execute the read request and it read the
single simultaneously occurs again it execute read
request but there is no data item in read data FIFO so
transaction will fail. This situation is called race
condition.

ISBN: 378 - 26 - 138420 - 5

[5]
http://www.arm.com/products/system-ip/amba/ambaopenspecifications.php
[6]
ARM, "AMBA Protocol
Specification 4.0",
www.arm.com, 2010 ARM,AMBA Specification
[7] LogiCORE IP AXI to APB Bridge (v1.00a) DS788 June
22, 2011 Product Specification.
[8] Simulation and Synthesis Techniques for Asynchronous
FIFO Design Clifford E.Cummings, Sunburst
Design, Inc. SNUG San Jose 2002 Rev 1.2.,FIFO
Architecture, Functions, and Applications SCAA042A
November 1999.
[9]
Lahir, K., Raghunathan A., Lakshminarayana G.,
LOTTERYBUS: a new high-performance
communication architecture for system-on- chip deisgns,
in Proceedings of Design Automation
Conference, 2001.
[10] Ying-Ze Liao, "System Design and Implementation of AXI
Bus", National Chiao Tung University,
October 2007.
AUTHOR BIOGRAPHY

A VIJAY KUMAR received B. Tech degree in Electronics and


Communication Engineering from JNTU affiliated college in
2009 and Pursuing M. Tech in VLSI & Embedded systems from
JNTU affiliated college.

6. ACKNOWLEDGMENT
I A VIJAY KUMAR would like to thank Assistant Prof T.
Vinaysimha Reddy, who had been guiding through out to
complete the work effectively and successfully, and would also
like to thank the Prof.P Sanjeeva Reddy HOD, ECE Department
and other Professors for extending their help & support in giving
technical ideas about the paper and motivating to complete the
work effectively & successfully.

7. REFERENCES
ARM, AMBA Specifications (Rev2.0). [Online]. Available
at http://www.arm.com, 1999
[2] ARM, AMBA AXI Protocol Specification (Rev 2.0),
Available at http://www.arm.com, March 2010
[3] Shaila S Math, Manjula R B, Survey of system on chip
buses based on industry standards, Conference on
Evolutionary Trends in Information Technology(CETIT),
Bekgaum,Karnataka, India, pp. 52, May 2011

T VINAYSIMHA REDDY received M.Tech in VLSI System


Design from JNTU affiliated college in 2010 from JNTU
affiliated college. And his interested areas are in Embedded
systems and VLSI Design .Now he is working as an Asst. Prof.
in JNTU affiliated college, Hyderabad, INDIA.

[1]

[4]
Design and Implementation of APB Bridge based on
AMBA 4.0 (IEEE 2011), ARM Limited.

M. Santhoshi received M.Tech in VLSI System Design from


JNTU affiliated college in 2012 from JNTU affiliated college.
And her interested areas are in Image and video processing .Now
she is working as an Asst. Prof. in JNTU affiliated college,
Hyderabad, INDIA.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

306

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Appliances of Harmonizing Model in Cloud


Computing Environment
Abstract
Introducing a system that uses virtualization
technology to assign data center resources
dynamically based on request demands and
support green computing by optimizing the
number of servers in use. We introduce the
concept of skewness to measure the
unevenness in the multidimensional store
utilization of a server. Minimizing skewness,
we can combine different types of
workloads nicely and improve the overall
utilization of server resources. Developing a
set of heuristics that prevent overload in the
system effectively while saving energy used.
Trace driven simulation and experiment
results demonstrate that our algorithm
achieves good performance. This article
introduces a better load balance model for
the public cloud based on the cloud partition
concept with a switch mechanism to choose
different strategies for different situations.
The algorithm applies the game theory to the
load balancing strategy to improve the
efficiency in the public cloud environment.

inside the VMs. The algorithm can capture


the rising trend of resource usage patterns
and help reduce the placement churn
significantly.Since the job arrival pattern is
not predictable and the capacities of each
node in the cloud differ, for load balancing
problem, workload control is crucial to
improve system presentation and maintain
stability. Load balancing schemes depending
on whether the system dynamics are
important can be either static and dynamic .
Static schemes do not use the system
information and are less complex while
dynamic schemes will bring additional costs
for the system but can change as the system
status changes. A dynamic scheme is used
here for its exibility. The model has a main
controller and balancers to gather and
analyze the information. Thus, the dynamic
control has little inuence on the other
working nodes. The system status then
provides a basis for choosing the right load
opposite strategy. The load balancing model
given in this article is aimed at the public
cloud which has numerous nodes with
dispersed computing resources in many
different geographic locations. Thus, this
model divides the public cloud into several
cloud partitions. When the environment is
very large and complex, these divisions
simplify the load balancing. The cloud has a
main manager that chooses the suitable
partition for arriving jobs while the balancer
for each cloud partition chooses the best
load balancing strategy.

1. Introduction
We present the design and execution of an
automated resource management system that
achieves a good balance between the two
goals. We make the following contributions:
We develop a resource allocation system
that can avoid excess in the system
effectively while minimizing the number of
servers used. We introduce the concept of
skewness to compute the uneven
utilization of a server. By minimizing
skewness, we can improve the overall
utilization of servers in the face of
multidimensional resource constraints. We
design a load prediction algorithm that can
capture the future resource usages of
applications accurately without looking

2. Related Work
There have been many studies of load
balancing for the cloud environment. Load
balancing in cloud computing was describe
in a white paper written by Adler who

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

307

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

introduced the tools and techniques usually


used for load balancing in the cloud.
However, load balancing in the cloud is still
a new problem that needs new architectures
to adapt to many changes. Chaczko et al.
described the role that load balancing plays
in improving the performance and
maintaining stability. There are many load
balancing algorithms, such as Round Robin,
Equally
Spread
Current
Execution
Algorithm, and Ant Colony algorithm.
Nishant et al.used the ant colony
optimization method in nodes load
balancing. Randles et al.gave a compared
analysis of some algorithms in cloud
computing by checking the performance
time and cost. They concluded that the
ESCE algorithm and throttled algorithm are
better than the Round Robin algorithm.
Some of the classical load balancing
methods are similar to the allocation method
in the in service system, for example, the
Round Robin algorithm and the First Come
First Served (FCFS) rules. The Round Robin
algorithm is used here because it is fairly
simple.

ISBN: 378 - 26 - 138420 - 5

details of our algorithm. Analysis of the


algorithm is obtainable in the additional file,
which can be

Comparison of SPAR and FUSD.

found on the ComputerSociety Digital


Libraryhttp://doi.ieeecomputersociety.org/10
.1109/TPDS.2012.283.

3.1 Hot and Cold Spots


Our algorithm executes sometimes to
evaluate the resource allocation status based
on the predicted future resource demands of
VMs. We define a server as a hot spot if the
operation of any of its resources is above a
hot threshold. This indicates that the server
is overloaded and hence some VMs running
on it should be migrated away. We define
the temperature of a hot spot p as the square
sum of its resource utilization beyond the
hot threshold:

3. The Skewness Algorithm


We introduce the concept of skewness to
quantify the unevenness in the utilization of
multiple resources on a server. Let n be the
number of resources we consider and be the
utilization of the i th resource. We define the
resource skewness of a server p as

where R is the set of overloaded resources in


server p and r is the hot threshold for
resource r. (Note that only overloaded
resources are considered in the calculation.)
The temperature of a hot spot reflects its
degree of overload. If a server is not a hot
spot, its temperature is zero.We define a
server as a cold spot if the utilizations of all
its resources are below a cold threshold.
This indicates that the server is mostly idle
and a probable candidate to turn off to save
energy. However, we do so only when the

where
is the average use of all resources
for server p. In practice, not all types of
resources are performance dangerous and
hence we only need to consider bottleneck
resources in the above calculation. By
minimizing the skewness, we can combine
different types of workloads nicely and
improve the overall utilization of server
resources. In the following, we describe the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

308

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

average resource utilization of all actively


used servers (i.e., APMs) in the system is
below a green computing threshold. A server
is actively used if it has at least one VM
running. Otherwise ,it is inactive. Finally,
we define the warm threshold to be a level
of resource utilization that is adequately
high to justify having the server running but
not so high as to risk becoming a hot spot in
the face of temporary fluctuation of
application resource demands. Different
types of resources can have different
thresholds. For example, we can define the
hot thresholds for CPU and memory
resources to be 90 and 80 percent,
respectively. Thus a server is a hot spot if
either its CPU usage is above 90 percent or
its memory usage is above 80 percent.
3.2 Hot Spot Mitigation

ISBN: 378 - 26 - 138420 - 5

we can find a destination server for any of


its VMs, we consider this run of the
algorithm a success and then move onto the
next hot spot. Note that each run of the
algorithm migrates away at most one VM
from the overloaded server. This does not
essentially eliminate the hot spot, but at least
reduces its temperature. If it remains a hot
spot in the next decision run, the algorithm
will repeat this process. It is possible to
design the algorithm so that it can migrate
away multiple VMs during each run. But
this can add more load on the related servers
during a period when they are already
overloaded. We decide to use this more
conservative approach and leave the system
some time to react before initiating
additional migrations.
3.3 Green Computing
When the resource utilization of
active servers is too low, some of them can
be turned off to save energy. This is handled
in our green computing algorithm. The test
here is to reduce the number of active
servers during low load without sacrificing
performance either now or in the future. We
need to avoid oscillation in the system. Our
green computing algorithm is invoked when
the average utilizations of all resources on
active servers are below the green
computing threshold. We sort the list of cold
spots in the system based on the ascending
order of their memory size. Since we need to
migrate away all its VMs before we can shut
down an underutilized server, we define the
memory size of a cold spot as the aggregate
memory size of all VMs running on it.
Recall that our model assumes all VMs
connect to a shared back-end storage.
Hence, the cost of a VM live migration is
resolute mostly by its memory footprint.
This Section in the supplementary file
explains why the memory is a good measure
in depth. We try to eliminate the cold spot
with the lowest cost first. For a cold spot p,
we check if we can migrate all its VMs

We sort the list of hot spots in the system in


descending temperature (i.e., we handle the
hottest one first). Our goal is to remove all
hot spots if possible. Otherwise, keep their
temperature as low as possible. For each
server p, we first decide which of its VMs
should be migrated away. We sort its list of
VMs based on the resulting temperature of
the server if that VM is migrated away. We
aim to travel away the VM that can reduce
the servers temperature the most. In case of
ties, we select the VM whose removal can
reduce the skewness of the server the most.
For each VM in the list, we see if we can
find a destination server to accommodate it.
The server must not become a hot spot after
accepting this VM. Among all such servers,
we select one whose skewness can be
reduced the most by accepting this VM.
Note that this reduction can be negative
which means we select the server whose
skewness increase the least. If a destination
server is found, we record the migration of
the VM to that server and update the
predicted load of related servers. Otherwise,
we move onto the next VM in the list and try
to find a destination server for it. As long as

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

309

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

somewhere else. For each VM on p, we try


to find a target server to accommodate it.
The resource utilizations of the server after
accepting the VM must be below the warm
threshold. While we can save energy by
consolidating
underutilized
servers,
overdoing it may create hot spots in the
future. The warm threshold is designed to
prevent that. If multiple servers satisfy the
above criterion, we prefer one that is not a
present cold spot. This is because increasing
load on a cold spot reduces the likelihood
that it can be eliminated. However, we will
accept a cold spot as the destination server
if necessary. All things being equal, we
select adestination server whose skewness
can be reduced the most by accepting this
VM. If we can find destination servers for
all VMs on a cold spot, we record the
sequence of migrations and update the
predict load of related servers. Otherwise,
we do not migrate any of its VMs. The list
of cold spots is also efficient because some
of them may no longer be cold due to the
proposed VM migrations in the above
process. The above consolidation adds extra
load onto the related servers. This is not as
serious a problem as in the hot spot
improvement case because green computing
is initiated only when the load in the system
is low. Nevertheless, we want to bound the
extra load due to server consolidation. We
restrict the number of cold spots that can be
eliminated in each run of the algorithm to be
no more than a certain percentage of active
servers in the system. This is called the
consolidation limit. Note that we eliminate
cold spots in the system only when the
average load of all active servers (APMs) is
below the green computing threshold.
Otherwise, we leave those cold spots there
as potential destination equipment for future
offloading. This is consistent with our
philosophy that green computing should be
conducted conservatively.
3.4 Consolidated Movements

ISBN: 378 - 26 - 138420 - 5

The movements generated in each step


above are not executed until all steps have
finished. The list of actions are then
consolidated so that each VM is moved at
most once to its final destination. For
example, hot spot mitigation may dictate a
VM to move from PM A to PM B, while
green computing dictates it to move from
PM B to PM C. In the actual execution, the
VM is moved from A to C directly.

4.System Model
There are several cloud computing
categories with this work focused on a
public cloud. A public cloud is based on the
standard cloud compute model, with service
provided by a service provider. A large
public cloud will include many nodes and
the nodes in different geographical
locations. Cloud partition is used to manage
this large cloud. A cloud partition is a
subarea of the public cloud with divisions
based on the geographic locations. The
architecture is shown in Fig.1.The load
opposite strategy is based on the cloud
partitioning concept. After creating the cloud
partitions, the load balancing then starts:
when a job arrives at

Fig. 1 Typical cloud partitioning.


the system, with the main controller
deciding which cloud partition should
receive the job. The partition load balancer
then decides how to assign the jobs to the
nodes. When the load status of a cloud
partition is normal, this partition can be
talented locally. If the cloud partition load
status is not normal, this job should be

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

310

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

transferred to another partition. The whole


process is shown in Fig.2.
4.1 Main controller and balancers
The load balance solution is done by
the main controller and the balancers. The
main controller rst assigns jobs to the
suitable
cloud
partition
and
then
communicates with the balancers in each
partition to refresh this status information.
Since the main manager deals with
information for each partition, smaller data
sets will lead to the higher processing rates.
The balancers in each partition gather the
status in order from every node and then
choose the right strategy to distribute the
jobs. The relationship between the balancers
and the main controller is shown in Fig.3.
4.2 Assigning jobs to the cloud
partition
The cloud partition status can be divided
into three types:
(1) Idle: When the percentage of idle
nodes exceeds
status.

ISBN: 378 - 26 - 138420 - 5

Fig. 3 Relationships between the main


controllers, the balancers, and the nodes.

(2) Normal: When the percentage of the


normal nodes exceeds , change to normal
load status.
(3) Overload: When the percentage of the
overloaded nodes exceeds , change to

change to idle

overloaded status. The parameters , , and


are set by the cloud partition balancers.
The main controller has to converse with the
balancers frequently to refresh the status
information. The main controller then
dispatches the jobs using the following
strategy: When job i arrives at the system,
the main controller queries the cloud divider
where job is located. If this locations status
is idle or normal, the job is handled locally.
If not, another cloud partition is found that is
not overloaded. The algorithm is shown in
Algorithm 1.
4.3 Assigning jobs to the nodes in the
cloud partition

Fig. 2 Job assignment strategy.

The cloud partition balancer gathers load in


order from every node to evaluate the cloud
partition status. This evaluation of each
nodes load status is very important. The
rst task is to dene the load degree of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

311

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

different

situations

ISBN: 378 - 26 - 138420 - 5

based

on

the

Step 4 Three nodes load status levels are


then dened as:
_ Idle When
there is no job being processed by this node
so the status is charged to Idle.
_ Normal For
the node is normal and it can process other
jobs.
_ Overloaded When
each nodes. The node load degree is related
to various static parameter and dynamic
parameters. The static parameters include
the number of CPUs, the CPU dispensation
speeds, the memory size, etc. Dynamic
parameters are the memory utilization ratio,
the CPU utilization ratio, the network
bandwidth, etc. The load degree is computed
from these parameters as below:
Step
1
Dene
a
load
paramet

the node is not available and can not receive


jobs until high it returns to the normal. The
load degree results are input into the Load
Status Tables created by the cloud partition
balancers. Each balancer has a Load Status
Table and refreshes it each xed period T .
The table is then used by the balancers to
calculate the partition status. Each partition
status has a different load balancing
solution. When a job arrives at a cloud
partition, the balancer assigns the job to the
nodes based on its current load strategy. This
strategy is changed by the balancers as the
cloud partition status changes.

with

each
parameter
being either static or dynamic. m
represents the total number of the
parameters.

5 Cloud Partition
Balancing Strategy

Step 2 Compute the load degree as:

5.1 Motivation
Good load balance will improve the
presentation of the entire cloud. However,
there is no common method that can adapt to
all possible different situations. Various
methods have been developed in improving
existing solutions to resolve new problems.
Each exacting method has advantage in a
particular area but not in all situations.
Therefore, the current model integrates
several methods and switches between the
load balance method based on the system
status. A relatively simple method can be

are weights that may differ


for different kinds of jobs. N represents the
current node.
Step 3 Dene estimate benchmarks.
Calculate the average cloud partition degree
from the node load degree statistics as:
Th
e bench mark

Load

is then set for

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

312

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

used for the divider idle state with a more


complex method for the normal state. The
load balancers then switch methods as the
status changes. Here, the idle status uses an
improved Round Robin algorithm while the
normal status uses a game theory based load
balancing strategy.

ISBN: 378 - 26 - 138420 - 5

period T . When the balance table is


refreshed, at this moment, if a job arrives at
the cloud partition, it will bring the
inconsistent problem. The system status will
have changed but the information will still
be old. This may lead to an erroneous load
strategy choice and an erroneous nodes
order. To resolve this problem, two Load
Status Tables should be created as: Load
Status Table 1 and Load Status Table 2. A
ag is also assigned to each table to indicate
Read or Write. When the ag = Read, then
the Round Robin based on the load degree
evaluation algorithm is using this table.
When the ag = Write, the table is being
refreshed, new information is written into
this table. Thus, at each moment, one table
gives the correct node locations in the queue
for the improved Round Robin algorithm,
while the other is being prepared with the
updated information. Once the data is
refreshed, the table ag is changed to
Read and the other tables ag is changed
to Write. The two tables then alternate to
solve the inconsistency. The process is
shown in Fig.4.

5.2 Load balance strategy for the idle


status
When the cloud partition is idle, many
computing resources are available and
relatively few jobs are arriving. In this
situation, this cloud partition has the ability
to process jobs as quickly as possible so a
simple load opposite method can be used.
There are many simple load balance
algorithm methods such as the Random the
Weight Round Robin, and the Dynamic
Round Robin. The Round Robin algorithm
is used here for its simplicity. The Round
Robin algorithm is one of the simplest load
balancing algorithms, which passes each
new request to the next server in the queue.
The algorithm does not record the status of
each connection so it has no status
information. In the regular Round Robin
algorithm, every node has an equal
opportunity to be chosen. However, in a
public cloud, the conguration and the
performance of each node will be not the
same; thus, this method may overload some
nodes. Thus, an improved Round Robin
algorithm is used , which called Round
Robin based on the load degree evaluation.
The algorithm is still fairly simple. Before
the Round Robin step, the nodes in the load
balancing table are ordered based on the
load degree from the lowest to the highest.
The system builds a circular queue and
walks through the queue again and again.
Jobs will then be assign to nodes with low
load degrees. The node order will be
changed when the balancer refreshes the
Load Status Table. However, there may be
read and write inconsistency at the refresh

5.3 Load balancing strategy for the


normal status
When the cloud partition is normal, jobs are
arriving much faster than in the idle state
and the situation is far more complex, so a
different strategy is used for the load
balancing. Each user wants his jobs
completed in the shortest time, so the public
cloud needs a method that can complete the
jobs of all users with reasonable response
time. Penmatsa and Chronopoulos proposed
a static load balancing strategy based on
game theory for distributed systems. And
this work provides us with a new review of
the load balance problem in the cloud
environment. As an implementation of
distributed system, the load balancing in the
cloud computing environment can be

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

313

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

viewed as a game. Game theory has noncooperative games and cooperative games.
In cooperative games, the decision makers
eventually come to an agreement which is
called a binding agreement. Each decision
maker decides by comparing notes with each
others. In non-cooperative games, each
decision maker makes decisions only for his
own benet. The system then reachs the
Nash equilibrium, where each decision
maker makes the optimized decision. The
Nash equilibrium is when each player in the
game has chosen a strategy and no player
can benet by changing his or her strategy
while the other players strategies remain
unchanged.

ISBN: 378 - 26 - 138420 - 5

environment are also distributed system,


these algorithms can also be used in grid
computing
and
cloud
computing
environments. Previous studies have shown
that the load balancing strategy for a cloud
partition in the normal load status can be
viewed as a non supportive game, as
described here. The players in the game are
the nodes and the jobs. Suppose there are n
nodes in the current cloud partition with N
jobs arriving, then dene the following
parameters:

In this model, the most important step is


nding the appropriate value of
. The
current model uses the method of Grosu et
al. called the best reply to calculate
of
each node, with a greedy algorithm then
used to calculate
for all nodes. This
procedure gives the Nash equilibrium to
minimize the response time of each job. The
strategy then changes as the nodes statuses
change.
Fig. 4 The solution of inconsistently problem.

There have been many studies in using game


theory for the load balancing. Grosu et
al.proposed a load balancing strategy based
on game theory for the distributed systems
as a non-cooperative game using the
distributed structure. They compared this
algorithm with other traditional methods to
show that their algorithm was less
complexity with better performance. Aote
and Kharat gave a energetic load balancing
model based on game theory. This model is
related on the dynamic load status of the
system with the users being the decision
makers in a non-cooperative game. Since the
grid computing and cloud computing

6.Future Work
Since this work is just a conceptual
framework, more work is needed to
implement the framework and resolve new
problems. Some important points are:
(1) Cloud division rules: Cloud division is
not a simple problem. Thus, the framework
will need a complete cloud division
methodology. For example, nodes in a
cluster may be far from other nodes or there
will be some clusters in the same geographic
area that are still far apart. The division rule
should simply be based on the geographic
location (province or state).

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

314

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

(2) How to set the refresh period: In the data


statistics analysis, the main controller and
the cloud partition balancers need to refresh
the information at a xed period. If the
period is too short, the high incidence will
inuence the system performance. If the
period is too long, the information will be
too old to make good decision. Thus, tests
and statistical tools are needed to set a
reasonable refresh periods.
(3) A better load status evaluation: A good

ISBN: 378 - 26 - 138420 - 5

[7] B. Adler, Load balancing in the cloud:


Tools,
tips
and
techniques,
http://www.rightscale.
com/info
center/whitepapers/Load-Balancing-in-the
Cloud.pdf, 2012
[8] Z. Chaczko, V. Mahadevan, S.
Aslanzadeh, and C. Mcdermid, Availability
and load balancing in cloud computing,
presented at the 2011 International
Conference on Computer and Software
Modeling, Singapore, 2011.
[9] K. Nishant, P. Sharma, V. Krishna, C.
Gupta, K. P. Singh, N. Nitin, and R. Rastogi,
Load balancing of nodes in cloud using ant
colony optimization, in Proc. 14th
International Conference on Computer
Modelling and Simulation (UKSim),
Cambridgeshire, United Kingdom,
Mar. 2012, pp. 28-30.
[10] M. Randles, D. Lamb, and A. TalebBendiab, A comparative study into
distributed load balancing algorithms for
cloud computing, in Proc. IEEE 24th
International Conference on Advanced
Information Networking and Applications,
Perth, Australia, 2010, pp. 551-556.
[11]
A.
Rouse,
Public
cloud,
http://searchcloudcomputing.techtarget.com/
denition/public-cloud, 2012.
[12] D. MacVittie, Intro to load balancing
for
developers
The
algorithms,
https://devcentral.f5.com/blogs/us/introtoloa
d-balancing-for-developers-ndash-thealgorithms,2012.
[13] S. Penmatsa and A. T. Chronopoulos,
Game-theoretic static load balancing for
distributed systems, Journalof Parallel and
Distributed Computing, vol. 71, no. 4, pp.
537-555, Apr. 2011.
[14] D. Grosu, A. T. Chronopoulos, and M.
Y. Leung, Load
balancing in distributed systems: An
approach using cooperative games, in Proc.
16th IEEE Intl. Parallel and Distributed
Processing Symp., Florida, USA, Apr. 2002,
pp. 52-61.

algorithm is needed to set


and
and the evaluation mechanism needs to be
more comprehensive.
(4) Find other load balance strategy: Other
load balance strategies may provide better
results, so tests are needed to compare
different strategies. Many tests are needed to
guarantee system availability and efciency.
References
[1] R. Hunter, The why of cloud,
http://www.gartner.com/DisplayDocument?d
oc cd=226469&ref= g noreg, 2012.
[2] M. D. Dikaiakos, D. Katsaros, P. Mehra,
G. Pallis, and A. Vakali, Cloud computing:
Distributed internet computing for IT and
scientic research, Internet Computing,
vol.13, no.5, pp.10-13, Sept.-Oct. 2009.
[3] P. Mell and T. Grance, The NIST
denition
of
cloud
computing,
http://csrc.nist.gov/publications/nistpubs/800
145/SP800-145.pdf, 2012.
[4] Microsoft Academic Research, Cloud
computing,http://libra.msra.cn/Keyword/605
1/cloudcomputing?querycloud%20computin
g, 2012.
[5] Google Trends, Cloud computing,
http://www.google.com/trends/explore#q=cl
oud%20computing, 2012.
[6] N. G. Shivaratri, P. Krueger, and M.
Singhal, Load distributing for locally
distributed systems, Computer, vol. 25, no.
12, pp. 33-44, Dec. 1992.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

315

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[15] S. Aote and M. U. Kharat, A gametheoretic model for dynamic load balancing
in distributed systems, in Proc. The
International Conference on Advances in
Computing, Communication and Control
(ICAC3 09), New York, USA, 2009, pp.
235-238.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

316

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Similarity Search in Information Networks using


Meta-Path Based between Objects
Abstract Real world physical and abstract data

in many applications. For example, in spatial

objects are interconnected, forming enormous,

database, people are interested in finding the k

interconnected networks. By structuring these

nearest neighbors for a given spatial object.

data objects and interactions between these

Object similarity is also one of the most

objects into multiple types, such networks

primitive concepts for object clustering and

become

many other data mining functions.

semi-structured

heterogeneous

information networks. Therefore, the quality


In a similar context, it is critical to provide

analysis of large heterogeneous information

effective

networks poses new challenges. In current

search for the most similar pictures for a given

distance, connectivity and co-citation. By using

relationships

between

we

measure

objects

rather

in

such as flicker, a user may be interested in

Wikipedia by reflecting all three concepts:

approach

functions

a given entity. In a network of tagged images

introduced for measuring the relationship on

current

search

information networks, to find similar entities for

system, a generalized flow based method is

the

similarity

picture. In an e-commerce system, a user would

only

be interest in search for the most similar

than

products for a given product. Different attribute-

similarities. To address these problems we

based similarity search, links play an important

introduce a novel solution meta-path based

role for

similarity searching approach for dealing with

similarity search in

information

networks, especially when the full information

heterogeneous information networks using a

about attributes for objects is difficult to obtain.

meta-path-based method. Under this framework,


similarity search and other mining tasks of the

There are a few studies leveraging link

network structure.

information in networks for similarity search,


but most of these revisions are focused on

Index terms similarity search, information

homogeneous or bipartite networks such as P-

network, and meta-path based, clustering.

PageRank and SimRank. These similarity


measures disregard the subtlety of different
types among objects and links. Adoption of such
I. INTRODUCTION

measures

to

heterogeneous

networks

his

Similarity search, which aims at locating the

significant drawbacks: even if we just want to

most relevant information for a query in large

compare objects of the same type, going through

collections of datasets, has been widely studied

link paths of different types leads to rather

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

317

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

different semantics meanings, and it makes little

connections represent different relationships

sense to mix them up and measure the similarity

between authors, each having some different

without distinguishing their semantics.

semantic meaning.

To systematically distinguish the semantics

Now the questions are, given an arbitrary

among paths connecting two objects, we

heterogeneous information network, is there any

introduce

similarity

way systematically identify all the possible

framework for objects of the same type in a

connection type between two objects types? In

heterogeneous network. A meta-path is a

order to do so, we propose two important

sequence of relations between object types,

concepts in the following.

meta-path

based

which defines a new composite relation between


a) Network Schema And Meta-Path

its starting type and ending type. The meta-path


framework provides a powerful mechanism for a

First,

user to select appropriate similarity semantics,

information network, it is necessary to provide

by choosing a proper meta-path, or learn it from

its

a set of training examples of similar objects.

understanding the network. Therefore, we

for

level

complex

description

heterogeneous

for

better

describe the Meta structure of a network.

relate it to two well-known existing link-based


functions

Meta

propose the concept of network scheme to

The meta-path based similarity framework, and

similarity

given

homogeneous

The concept of network scheme is similar to that

information networks. We define a novel

of the Entity Relationship model in database

similarity measure, PathSim that is able to find

systems, but only captures the entity type and

peer objects that are not only strongly connected

their binary relations, without considering the

with each other but also share similar visibility

attributes for each Entity type. Network schema

in the network. Moreover, we propose an

serves as a template for a network, and tells how

efficient algorithm to support online top-k

many types of objects there are in the network

queries for such similarity search.

and where the possible links exist.

II. A META-PATH BASED SIMILARITY

b) Bibliographic Scheme and Meta-Path

MEASURE
For the bibliographic network scheme, where an
The similarity between two objects in a link-

explicitly shows the direction of a relation.

based similarity function is determined by how


III. META-PATH BASED SIMILARITY

the objects are connected in a network, which

FRAMEWORK

can be described using paths. In a heterogeneous


information network, due to the heterogeneity of

Given

the types of links, the way to connect two

similarity measures can be defined for a pair of

objects can be much more diverse. The schema

objects, and according to the path instances

a user-specified meta-path,

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

318

several

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

between them following the met-path. There are

their movie styles and productivity and finding

several

similar product.

straightforward

measures

in

the

following.
This motivated us to propose a new, meta-path
Path count: the number of path instances

based similarity measure, call PathSim that

between objects.

captures the subtle of peer similarity. The insight


behind it is that two similar peer objects should

Random Walk: s(x, y) is the probability of the

not only be strongly connected, but also share

random walk that starts from x and ends with y

comparable observations. As the relation of peer

following meta-path P, which is the sum of the

should be symmetric, we confine PathSim to

probabilities of all the path instances.

symmetric meta-paths. The calculation of

Pair wise random walk: for a meta-path P that

PathSim between any two objects of the same

can be decomposed into two shorter meta-paths

type given a certain meta-path involves matrix

with the same length is then the pair wise

multiplication.

random walk probability starting from objects x

In this paper, we only consider the meta-path in

and y and reaching the same middle object.

the round trip from, to guarantee its symmetry

In general, we can define a meta-path based

and therefore the symmetry of the PathSim

similarity framework for two objects x and y.

measure.

Note that P-PageRank and SimRank, two wellknown

network

similarity

functions,

Properties of PathSim

are

weighted combinations of random walk measure

1. Symmetric.

or pair wise random walk measure, respectively,

2. Self-maximum

over meta-paths with different lengths in

3. Balance of visibility

homogeneous networks. In order to use PAlthough using meta-path based similarity we

PageRank and SimRank in heterogeneous

can define similarity between two objects given

information networks.

any round trip meta-paths.


a) A Novel Similarity Measure
As primary eigenvectors can be used as
There have been several similarity measures are

authority ranking of objects, the similarity

presented and they are partially to either highly

between two objects under an infinite meta-path

visible objects or highly concentrated objects but

can be viewed as a measure defined on their

cannot capture the semantics of peer similarity.

rankings. Two objects with more similar

However, in many scenarios, finding similar

rankings scores will have higher similarity. In

objects in networks is to find similar peers, such

the next section we discuss online query

as finding similar authors based on their fields

processing for ingle meta-path.

and reputation, finding similar actors based on

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

319

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

IV.

ISBN: 378 - 26 - 138420 - 5

check every possible object. This will be much

QUERY PROCESSING FOR SINGLE META-

more

PATH

efficient

than

Pairwise

computation

between the query and all the objects of that


Compared with P-PageRank and SimRank, the

type. We call baseline concatenation algorithm

calculation is much more efficient, as it is a local

as PathSim-baseline.

graph measure. But still involves expensive


matrix multiplication operations for top k

The PathSim-baseline algorithm is still time

search functions, as we need to calculate the

consuming if the candidate set is large. The time

similarity between a query and every object of

complexity of computing PathSim for each

the same type in the network. One possible

candidate, where is O(d) on average and O(m) in

solution is to materialize all the meta-paths.

the worst case. We now propose a co-clustering


based top-k concatenation algorithm, by which

In order to support fast online query processing


for

large-scale

networks,

we

propose

non-promising target objects are dynamically

filtered out to reduce the search space.

methodology that partially materializes short


length meta-paths and then concatenates them
online

to

derive

longer

b) Co-Clustering-Based Pruning

meta-path-based
In the baseline algorithm, the computational

similarity. First, a baseline method is proposed,

costs involve two factors. First, the more

which computes the similarity between query

candidates to check, the more time the algorithm

object x and all the candidate object y of the

will take; second, for each candidate, the dot

same type. Next, a co-clustering based pruning

product of query vector and candidate vector

method is proposed, which prunes candidate

will at most involve m operations, where m is

objects that are not promising according to their

the vector length. Based on the intuition, we

similarity upper bounds. Both algorithms return

propose, we propose a co-clustering-based path

exact top-k results the given query.

concatenation method, which first generates coclusters of two types of objects for partial

a) Baseline

relation matrix, then stores necessary statics for


Suppose we know that the relation matrix for

each of the blocks corresponding to different co-

meta-path and the diagonal vector in order to get

cluster pairs, and then uses the block statistics to

top-k objects with the highest similarity for the

prune the search space. For better picture, we

query, we need to compute the probability of

call cluster of type as target clusters, since the

objects. The straightforward baseline is: (1) first

objects are the targets for the query and call

apply vector matrix multiplication (2) calculate

clusters of type as feature clusters. Since the

probability of objects (3) sort the probability of

objects serve as features to calculate the

objects and return top-k list in the final step.

similarity between the query and the target

When a large matrix, the vector matrix

objects. By partitioning into different target

computation will be too time consuming to

clusters, if a whole target cluster is not similar to

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

320

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

the query, then all the objects in the target

extracted from DBLP and Flicker in the

cluster are likely not in the final top-k lists and

experiments.

can be pruned. By partitioning in different


The PathSim algorithm significantly improves

feature clusters, cheaper calculations on the

the query processing speed comparing with the

dimension-reduced query vector and candidate

baseline algorithm, without affecting the search

vectors can be used to derive the similarity


upper

bounds.

The

PathSim-Pruning

quality.

can

significantly improve the query processing speed

For additional case studies, we construct a

comparing with the baseline algorithm, without

Flicker network from a subset of the Flicker data

affecting the search quality.

which contains four types of objects such as


images, users, tags, and groups. We have to

c) Multiple Meta-Paths Combination

show that our algorithms improve similarity


In the previous section, we presented algorithms

search between object based on the potentiality

for similarity search using single meta-path.

and correlation between objects.

Now, we present a solution to combine multiple


VI. CONCLUSION

meta-paths. The reason why we need to combine


several meta-paths is that each meta-path

In this paper we introduced novel similarity

provides a unique angle to view the similarity

search using meta-path based similarity search

between objects, and the ground truth may be a

using baseline algorithm and co-clustering based

cause of different factors. Some useful guidance

pruning algorithms to improve the similarity

of the weight assignment includes: longer meta-

search based on the strengths and relationships

path utilize more remote relationship and thus

between objects.

should be assigned with a smaller weight, such


REFERENCES

as in P-PageRank and SimRank and meta-paths


with more important relationships should be

[1]

Jiawei Han, Lise Getoor, Wei Wang,

assigned with a higher weight. For automatically

Johannes Gehrke, Robert Grossman "Mining

determining the weights, users cloud provides

Heterogeneous

training examples of similar objects to learn the

Principles and Methodologies"

weights of different meta-paths using learning

Information

Networks

[2] Y. Koren, S.C. North, and C. Volinsky,

algorithm.

Measuring and Extracting Proximity in


Networks, Proc. 12th ACM SIGKDD Intl

V. EXPECTED RESULTS

Conf. Knowledge Discovery and Data


To show the effectiveness of the PathSim

Mining, pp. 245-255, 2006.

measure and the efficiency of the proposed


[3] M. Ito, K. Nakayama, T. Hara, and S. Nishio,

algorithms we use the bibliographic networks

Association

Thesaurus

Construction

Methods Based on Link Co-Occurrence

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

321

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Analysis for Wikipedia, Proc. 17th ACM


Conf.

Information

and

Knowledge

Management (CIKM), pp. 817-826, 2008.


[4] K. Nakayama, T. Hara, and S. Nishio,
Wikipedia Mining for an Association Web
Thesaurus Construction, Proc. Eighth Intl
Conf. Web Information Systems Eng.
(WISE), pp. 322-334, 2007.
[5] M. Yazdani and A. Popescu-Belis, A
Random Walk Framework to Compute
Textual Semantic Similarity: A Unified
Model for Three Benchmark Tasks, Proc.
IEEE

Fourth

Intl

Conf.

Semantic

Computing (ICSC), pp. 424-429, 2010.


[6] R.L. Cilibrasi and P.M.B. Vitanyi, The
Google Similarity Distance, IEEE Trans.
Knowledge and Data Eng., vol. 19, no. 3,
pp. 370-383, Mar. 2007.
[7] G. Kasneci, F.M. Suchanek, G. Ifrim, M.
Ramanath,

and

G.

Weikum,

Naga:

Searching and Ranking Knowledge, Proc.


IEEE 24th Intl Conf. Data Eng. (ICDE), pp.
953-962, 2008.
[8] R.K. Ahuja, T.L. Magnanti, and J.B. Orlin,
Network Flows: Theory, Algorithms, and
Applications. Prentice Hall, 1993.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

322

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Pinpointing performance Deviations of Subsystems in Distributed


Systems with cloud infrastructures
1

Hareesh Kamisetty M.Tech,

(Audisankara College Of Engineering And Technology,Gudur)

Abstract- Cloud Manufacturing Systems is

Four types of CMfg service platform

service oriented system (SOS) and they are used

i.e., public private, hybrid and community CMfg

to compose the various kinds of cloud related

service platforms. Compared with existing

applications. In which some of the applications

manufacturing models CMfg has the following

are failed to complete the execution of user

features or properties:

request with in the deadline. And the Cloud

i)

Service and requirement-oriented.

Manufacturing Systems are still facing the


performance

problems

and

the

Most of the manufacturing models are

existing

resource-or order-oriented, while CMfg is a

techniques considering the tracing request data

service-and requirement-oriented manufacturing

as performance data to find and diagnosis the

model. The core idea of CMfg is manufacturing

performance of the service. In this paper, we

as a service.

propose the Cloud Debugger. It is a promising

ii) Dynamic with uncertainty.

tool to diagnosis the performance problems in

The resources and services in CMfg are

Cloud Manufacturing Systems.

various and dynamic, and the solutions for


Index terms- Cloud Manufacturing System
(CMfg),

Performance

Diagnosis,

addressing manufacturing tasks are dynamic.

Service-

iii) Knowledge-based.

Oriented.

The whole life cycle of CMfg system


I.

INTRODUCTION

The Cloud

Manufacturing

needs knowledge support.


iv) Initiative.

Systems

In a Cloud manufacturing System, both

service-oriented systems and compose the

manufacturing requirements are active and they

various services. In the real world combining the


increased

advanced

technologies

such

can automatically find and match with each

as

other with the support of semantic reasoning

virtualization, advanced computing technologies

based on knowledge.

and service-oriented technologies with existing


models and new wide range of information

v) Physically

technologies, a new computing and service


manufacturing

model

called

distributed

and

logically

centralized

cloud

The physical developed resource and

manufacturing is proposed.

facility in CMfg locate in different places and

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

323

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

are

restricted

by

different

persons

or

is supreme model for utilize in production.

organizations.

There is no overhead for enabling the debugger


on this paper.

a) Existing System
The

ISBN: 378 - 26 - 138420 - 5

art

of

debugging

for

a) Cloud Trace

cloud

applications is not much more than writing out

Performance is an important feature of

diagnostic messages and spelunking the logs for

service which directly interrelated with end

them.

users satisfaction and maintenance. No on


intends to build a slow service, but it can be

When the right data is not being write to

extremely difficult to identify the core reasons of

the logs, developers have to change the code and

slowness when it happens.

redeploy the application to production. The


traditional debuggers are not well suited for

The cloud Trace helps you visualize and

cloud-based services for two reasons. First, it is

understand the time spent by application for

hard to know which method to attach to. Second,

request processing. This enables the cloud

stopping a method in production makes it hard

operator to hastily find and repair performance

to repeat an issue and gives end-users a bad

bottlenecks.

experience.

The cloud operator can easily produce a

b) Proposed System

report that shows the performance change in the


service from one release to another.

To address above mentioned problems


this paper proposes the a novel Debugger is

b) Cloud Monitoring

called as Cloud Debugger which significantly

Cloud

reduce the effort of cloud operator to identify the

Monitoring

provides

the

dashboards and alerting capabilities that help the

which method to attach and it provides a easy

cloud operator to identify and repair the

way to repeat the service based on users

performance problem quickly.

experiences.
With minimum configuration and no
II. PROPOSED WORK
The

Cloud

Debugger

separate infrastructure to

maintain,

Cloud

completely

Monitoring provides deep visibility into Cloud

changes this. It allows developers to start where

platform services when compared to the

they know best-in the code. By simply setting a

traditional approaches.

watch point on a line of code, the next time a

Finally, we can create end point checks

request on any of the cloud servers hits that line

to monitor availability and response times for

code, and we get a snapshot of all the local

end-user facing services. Endpoint checks are

variables, parameter, instance variables and a

performed by problems in Texas, Virginia

full stack trace. There is zero setup time and no

enabling monitoring of latency.

complex configuration to enable. The debugger

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

324

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

SSH to VM instantly

[2] F, Tao; L Zhang; VC Venkatesh; YL Luo; Y


Cheng (2011). "Cloud manufacturing: a

Sometimes it is unavoidable to connect

computing and service-oriented manufactu -

directly to a VM to debug or repair the

ring model". Proceedings of the Institution of

production issue. It can be a bit of pain,

Mechanical Engineers, Part B, Journal of

especially when the operator is on the road, so

Engineering

now we can do that from just about anywhere.

Manufacture.

doi:10.1177/

0954405411405575

With our new browser based SSH client can


speedily and securely connect to any of VMs

[3] Michael Rung-Tsong Lyu

from the console. It no needs to install any SDK

Toward

or any tools. This could be taken as future work.

Scalable

III.

ISBN: 378 - 26 - 138420 - 5

Fine-Grained,
Performance

and Hua Cai


Unsupervised,
Diagnosis

for

Production Cloud Computing Systems Proc.

CONCLUSION

IEEE Parallel and Distribute Systems pp


The traditional approaches diagnosis the

1245-1255, 2013.

performance problems based on the users


[4] H. Mi, H. Wang, G. Yin, H. Cai, Q. Zhou,

requesting trace data but it suffers from the


tradeoffs

between

tracing

granularity

and

and

T.

Sun,

Performance

Problems

Diagnosis in Cloud Computing Systems by

debugging effort. As a result, more effort will be

Mining Request Trace Logs, Proc. IEEE

increased in troubleshooting the performance

Network Operations and Management Symp.

problems. So that, this paper proposes the Cloud

(NOMS), pp. 893-899, 2012.

Debugger as tool for performance diagnosis with


effortless. And this can be significantly reducing

[5] P. Barham, A. Donnelly, R. Isaacs, and R.

the communication and computation overheads.

Mortier,

Using

Magpie

for

Request

Extraction and Workload Modelling, Proc.

FUTURE ENHANCEMENT

USENIX Sixth Conf. Symp. Operating


The proposed system of this paper

Systems Design and Implementation (OSDI),

considers the cloud monitoring and cloud trace

pp. 259-272, 2004.

as a techniques to find and repair the


AUTHORS

performance problems. But the performance of


the service is also depends on the system

First Author

dynamics which is leave it as future research

Second Author

work.
REFERENCES
[1] http://googlecloudplatform.blogspot.in / 20 14 /06 /enabling-developers-to-tame-product
-ion-systems-in-the-cloud.html.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

325

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Secure data storage against attacks in cloud


Environment using Defence in Depth
1

Naresh.G

(Audisankara College Of Engineering And Technology)

Abstract-

Cloud

computing

is

emerging

organizations. The outsourced storage makes

technology which offers various services with

shared data and resources much more accessible

low cost and flexible manner. One of the most

as users can access from anywhere.

important service models is Data Storage as a


On the other hand, security remains the

Service (DaaS) in which user can remotely store

important issue that concerns privacy of users. A

their data and enjoy the on demand access using

major challenge for any comprehensive access

high quality application. Cloud computing faces

control solution for outsourced data. And have

many devastating problems to ensure the proper

the ability to handle the user requests for

physical, logical and personnel security controls.

resources according to the specified security

While moving large volumes of data and

policies. Several solutions have been proposed

software, the management of the data and

in the past, but most of them dont consider

services may not be fully trustworthy. In this

protecting privacy of the policies and user access

paper, we mainly focus on the security features

patterns.

of data storage in the presence of threats and


attacks and solutions. The paper also proposes

In this paper we address the main aspects related

an effective and flexible distributed scheme with

to security of cloud storage. It presents an

two silent features opposing to its predecessors.

attempt to propose an effective and flexible

Index

terms-

Cloud

Computing,

security policy and procedure explicit to

storage

enhance the Data storage security in the cloud.

correctness, Data Storage as a Service.


I.

II. THREATS AND ATTACKS FROM STORAGE

INTRODUCTION

PERSPECTIVES
Cloud computing is the delivery of the
While the benefits of storage networks have

computing as a service rather than a product,

been widely acknowledged, consolidation of

whereby widely shared resources, software and

enterprise data on networked storage poses

information are provided to IT industry over a

significant security risks. Hackers adept at

network. Cloud can be classified as public,

exploiting network-layer vulnerabilities can now

private or hybridetc. meanwhile, the emerging

explore deeper strata of corporate information

trend of outsourcing data storages at third parties


attention from both research and industry

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

326

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Following is brief listings of some major drivers

III. SYSTEM DESIGN

to implementing security for networked storage

a) System Model

ISBN: 378 - 26 - 138420 - 5

from perspectives of challenging threats and


Cloud networking can be illustrated by three

attacks:

different network entities:


o

Perimeter

defence

strategies

focus

on
User: who have data to be stored in the cloud

protection from external threats. With the

and rely on the cloud for data computation,

number of security attacks on the rise,

consist of both individual consumers and

relying on perimeter defence alone is not

organizations?

sufficient to protect enterprise data, and a

single security breach can cripple a business

Cloud Service Provider

[7].

significant resources and expertise in building

The number of internal attacks is on the rise

and managing distributed cloud storage servers,

thereby threatening NAS/SAN deployments

owns and operates live Cloud Computing

that are part of the trusted corporate

systems.

(CSP): who

has

networks [8]. Reports such as the CSI/FBIs


Third Party Auditor (TPA): who has expertise

annual Computer Crime & Security Survey

and capabilities that users may not have, is

help quantify the significant threat caused by

trusted to assess and expose risk of cloud storage

data theft
o

services on behalf of the users upon request.

The problem of incorrectness of data storage


in the cloud

b) Adversary Model

The data stored in the cloud may be updated


There are two different sources for Security

by the users, including insertion, deletion,

threats faced by cloud data storage.

modification, appending, reordering, etc.


o

Individual users data is redundantly stored

1. CSP can be self-interested, un-trusted and

in multiple physical locations to further

possibly malicious.

reduce the data integrity threats.


o

It may move data that is rarely accessed to a

Moreover, risks due to compromised storage

lower tier of storage for monetary reasons,

range from tangible loss such as business

but

discontinuity in the form of information

It may hide a data loss incident due to

downtime, to intangibles such as the loss of

management errors, Byzantine failures and

stature as a secure business partner. With the

so on.

number of reported security attacks on the rise, a


firm

understanding

of

networked

2. Economically motivated adversary, who has

storage

the capability to compromise a number of cloud

solutions is a precursor to determining and

data storage servers in different time intervals

mitigating security risks.

and subsequently is able to modify or delete

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

327

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Layer 1 Devices on the Storage Network

users 'data while remaining undetected by CSPs


for a certain period.

The following risk-mitigation measures are


There are two types of adversary

recommended:

Weak Adversary: The adversary is interested in

Authentication schemes provisioned by the

corrupting the users data files stored on

Operating System should be evaluated.

individual servers. Once a server is comprised,

Schemes utilizing public-private key based

an adversary can pollute the original data files

authentication such as SSH or Kerberos,

by modifying or introducing its own fraudulent

which

data to prevent the original data from being

communications on the network.

retrieved by the user.

also

encrypt

authentication

Authentication using Access control Lists


(ACL) to setup role-based access and

Strong Adversary: This is the worst case

appropriate

scenario, in which we assume that the adversary

permissions

will

enhance

security,

can compromise all the storage servers so that he

can intentionally modify the data files as long as

Strong password schemes like minimum


length and periodic change of passwords

they are internally consistent.

should be enforced. The default user name


and passwords that are configured on the

IV. PROPOSED SOLUTIONS

device should be changed.


Control Access Data Storage that includes the
necessary

policies,

processes

and

Constant

control

monitoring

of

published

OS

activities for the delivery of each of the Data

vulnerabilities using database, SANS Security

service offerings. The collective control Data

Alert Consensus newsletter and the NAS

Storage encompasses the users, processes, and

vendors support site, is a

technology

necessary

to

maintain

an

environment that supports the effectiveness of

necessity to prepare for possible attacks

Logging and auditing controls should be

specific controls and the control frameworks.

implemented to prevent unauthorized use,

The Security, correctness and availability of the

track usage and for incident response

data files being stored on the distributed cloud


servers must be guaranteed by the following:
o

Layer -2 Network Connectivity

Providing Security policy and Procedure for

NAS appliances face similar vulnerabilities as IP

Data Storage

based network devices. Common techniques


used to protect IP networks are also applicable

The Defence in Depth (referred to as did in this

to Storage Network:

paper) is an excellent framework advocating a


layered approach to defending against attacks,

thereby mitigating risks.

Extending

network

perimeter

defence

strategies like using a Firewall and IDS

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

328

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

device to filter traffic reaching the NAS

localization in our challenge- response

appliance will increase protection

protocol

Use VLANs for segregating traffic to the

The response values from servers for each

NAS appliances

challenge

Separate and isolate management interface

correctness of the distributed storage, but

from

the Storage

also contain information to locate.

Network, thus enforcing out of band

b) Reliability of the analysis strategy

data

interfaces

on

not

only

determine

the

management which is more secure


o

Monitor

traffic

patterns

on

the

The reliability of secure data storage strategy

data

depends on security procedure and the backup

interfaces of the NAS devices for unusual

data coefficients. When one or more nodes

activity

cannot be accessed, the secure strategy can


Layer 3 Management Access

ensure that the data will be restored as long as


one of the k nodes can be accessed. However,

Management access is a significant source of

traditional data storage methods require all the

attack. To address the vulnerabilities, the

data in the k nodes to be retrieved. Thus, the

following guidelines provide help


o

more blocks the data are split into, the poorer the
reliability of traditional data storage

Disable the use of telnet and HTTP and


enforce management access through SSH

V. CONCLUSION

and HTTPS for encrypted communication


o

Create separate user accounts based on the

This paper suggests a methodical application of

management tasks assigned to the users

defence in depth security techniques that can

Implement

authentication

help allay security risks in networked storage.

mechanisms like two-factor authentication

More importantly, a defence in depth based

using tokens, biometrics, etc

networked storage security policy provides a

Strong password schemes like minimum

comprehensive framework to thwart future

length passwords and periodic change of

attacks as the current technologies are more

passwords should be enforced

clearly understood.

Implement

strong

authorization

using

Access

REFERENCES

Control Lists to setup role based access and


appropriate permissions

[1] What is Cloud Computing? Retrieved April

a) Correctness verification
o

6, 2011, available at:

Error localization is a key prerequisite for

http://www.microsoft.com/business/engb/sol

eliminating errors in storage systems.

utions/Pages/Cloud.aspx

We

can

correctness

do

that

by integrating the

verification

and

error

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

329

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

[2]

EMC,

Information-Centric

ISBN: 378 - 26 - 138420 - 5

Security.

http://www.idc.pt/resources/PPTs/2007/IT&
Internet_Security/12.EMC.pdf.
[3] End-User Privacy in HumanComputer
Interaction.
http://www.cs.cmu.edu/~jasonh/publications
/fnt-enduser-privacy-in-human-computerinteractionfinal. pdf.
[4] ESG White Paper, the Information-Centric
Security Architecture. http://japan.emc.com/
collateral/analystreports/emc-white-paper

v4-4-21-2006.pdf.
[5] Subashini S, Kavitha V., A survey on
security issues in service delivery models of
cloud computing, Journal of Network and
Computer Applications (2011) vol. 34 Issue
1, January 2011 pp. 1-11.
AUTHORS
First Author
Second Author

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

330

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

IMPROVED LOAD BALANCING MODEL BASED ON


PARITIONING IN CLOUD COMPUTING
G.RajivRatnakar II M.Tech AUDI SANKARACOLLEGE OF ENGINEERING
[email protected]
CH.MadhuBabu M.Tech
Associate Professor
[email protected]
Abstract: Load balancing plays a vital role in improving performance of cloud servers. Implementing better load
balancing schemes improves performance of cloud servers there by more number of users can be satisfied. As
number of cloud users increasing day by day load on cloud servers need to be maintained. In this article load
balancing schemes are analyzed and a better load balancing scheme is identified in dynamic environment

and to avoid the overloading at a certain node we


implement some algorithms.

Introduction:
Cloud computing is large group of remote
servers interconnected together that provide a facility
of sharing. The cloud computing is high utility
technology in world and having the ability to change
IT software industry. Due to its simple architecture
most of companies are adapting cloud computing.The
number of cloud service providers are increasing day
by day as the number of cloud users are increasing in
day to day life. The increase of web traffic and load
make load balancing as a big research topic.

Round Robin algorithm:


Round robin implements an equal priority to all jobs
mechanism. Based on time slice each and every
process is allotted with a certain time in a round robin
manner. It is very simple to implement.
Equally spread current execution:
ESCG work based on job priorities. It distributes the
load randomly by checking the load and then shifts
the load to virtual machine which is lightly loaded so
that to maximize throughput. It uses spread spectrum
technique in which load is spread over virtual
machines.

The term load balancing refers to


distribution of larger processing load to smaller
processing nodes for improving overall performance
of system. An ideal load balancing algorithm should
avoid overloading on a certain node in a partitioned
cloud. Security, reliability response time throughput
are some factors while choosing a good load
balancing algorithm. The main aim of good load
balancing algorithm is to improve throughput and
reduce response time of the system.

Throttled load balancing algorithm:


In throttled algorithm job manager maintains list
of virtual machines and jobs assigned to them.
Throttled algorithm finds appropriate virtual machine
for assigning a particular job. If all virtual machines
are in a heavy state it maintains jobs in job queue till
virtual machine is free to process a job.

In this article both static and dynamic


schemes are discussed and analyzed there by a good
load balancing strategy is selected. When the User
submits the job the job arrive at the main controller
the main controller allocates the jobs to the node.
Load balancing schemes:
To distribute load of multiple networks to achieve
maximum throughput minimize the response time

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

331

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

The analysis performed on static load


balancing schemes round robin, equally spread
current execution, throttled individually and the
Overall response time, data center processing
time, total revenue in dollars are listed in the
following analysis.

Related work:
Load balancing can be implemented in both
static and dynamic schemes.
Cloud analyst is a tool provided by cloud bus

organization to analyze the performance of the


cloud. In cloud analyst there are certain
terminologies as follows.

If we observe the above analysis although better


algorithms are used there is no change in
response time of the three algorithms. Not only
the response time but also the other factors are
remained unchanged. so the use of static
algorithms are not preferable.

User base: The user base represents the single


user but here user base represents a group of
users.
Data center: Datacenter manages the data
management activities virtual machines creation
destruction and other routing schemes.

Round
Robin

ESCG

Throttled

Overall
response
time

300.12

300.12

300.12

Data center
processing
time(m.sec)

0.35

0.35

0.35

0.50

0.50

0.50

16

16

16

VM
cost($)
Total cost

Table: analysis of static load

Figure 1: cloud analyst tool

Proposed work:
By using the cloud analyst tool both static and
dynamic schemes are analyzed and there by a
better load balancing model and scheme is
determined
Static schemes:

The static scheme does not analyze the


systems state. It implements a single scenario for
every situation. It is simple to implement. For
example if static system uses round robin
algorithm, for idle normal and busy states the
system uses the same algorithm. It does not
provide better results.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

332

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Implementing dynamic schemes :


Dynamic schemes analyze the systems
current status. Although dynamic schemes are
difficult to implement, dynam ic schemes
identifies the current position of the server
whether it is in idle represented with certain
thresholdor it is in normal or else it is in
overloaded state.
If the system status is in idle or
normal main controller [2] will allocate job to
virtual machine to process job or if it is busy it
will wait till virtual machine is idle.Dynamic
load balancing model may impl ements more
than on load balancing schemes.
As shown in figure belo w the load
balancing is implemented with m ore than on
scheme, it implemented round ro bin throttled
equally spread current execution. By
implementing more than one algorithm the load
on server gets reduced there by a great user
satisfaction achieved.

Fig2: model with Dynam ic schemes

The analysis of load bal ancing model in


dynamic environment using RR, ESCG, and
Throttled are as shown in below tables.

Algorithm
Start

Overall
response
time
Data center
processing
time
(m.sec)
VM cost($)

User submit a job to main controller


Identify the status of partition.
If state=idle!!normal
Processjob

Round
Robin

ESCG

Throttled

223.34

209.75

209.74

13.84

10.11

10.10

5.3

4.45

4.03

18.3

18.3

18.3

Else

Total
Search for another partition.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

333

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

References:
1. Gaochao XU Jungie pang: Load balancing
model based on partitioning public cloud.
2. Dr.Hemanth s mahalle : Load balancing on
cloud centers.
3. Bhathiya: Cloud analyst cloud sim based
visual modeller for analyzing cloud computing
environments and applications

Fig: Response time graph

Although it costs equal and high when compared


to static algorithms there is a great variation in
terms of response time the response time is very
less in dynamic schemes.

4. Ramprasad pandey, P.Gotham Prasad rao:


Load balancing in cloud computing system.
5. WWW.cloudbus.org/cloudsim

Regarding the dynamic schemes if we observe


the cost and response time, there is a similar cost
in all the three algorithms but there is a variation
in terms of response time.
Based on the response time the throttle and the
equally spread current execution are best
algorithms in dynamic load balancing
environments.

Conclusion:
Finally dynamic schemes are effective in
balancing load when compared to static
schemes. In dynamic schemes throttled and
equally spread current execution are better to
balance the load based on dynamic load
environments.

Acknowledgements:
I would thank my head of department
Prof.Mr.M.Rajendra&myguideMr.CH.MadhuBabu
sir for guiding me to this project. I thank my parents,
friends lecturers who gave a contribution for this
project.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

334

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Effective User Navigability Through Website Structure


Reorganizing using Mathematical Programming Model
B. VAMSI, CH. SAI MURALI
1PG

Scholar, Dept of CSE, Audisankara College of Engineering & Technology, Gudur, AP, India.
Prof, Dept of CSE, Audisankara College of Engineering & Technology, Gudur, AP, India.

2Asst

AbstractWebsite design is easy task but, to navigate user efficiently is big challenge, one of the reason is user

behavior is keep changing and web developer or designer not think according to users behavior. Designing well
structured websites to facilitate effective user navigation patterns has long been a hallenge in web usage mining with
various applications like navigation prediction and improvement of website management. This paper addresses how
to improve a website without introducing substantial changes. Specifically, we propose a mathematical
programming model to improve the user navigation on a website while minimizing alterations to its current
structure. Results from extensive tests conducted on a publicly available real data set indicate that our model not
only significantly improves the user navigation with very few changes, but also can be
effectively solved. We have also tested the model on large synthetic data sets to demonstrate that it scales up very
well. In addition, we define two evaluation metrics and use them to assess the performance of the improved website
using the real data set. Evaluation results confirm that the user navigation on the improved structure is indeed greatly
enhanced. More interestingly, we find that heavily disoriented users are more likely to benefit from the improved
structure than the less disorientedusers.

KEYWORDS-Website Design, User Navigatio, Web


Mining, Mathematical Programming

I.

INTRODUCTION
Previous studies on website has focused on a variety
of issues, such as understanding web structures ,
finding relevant pages of a given page , mining
informative structure of a news website, and
extracting template from WebPages . Our work, on
the other hand, is closely related to the literature that
examines how to improve website navigability
through the use of user navigation data. Various
works have made an effort to address this question
and they can be generally classified into two
categories to facilitate a particular user by
dynamically reconstituting pages based on his profile
and traversal paths, often referred as personalization,
and to modify the site structure to ease the navigation
for all users, often referred as transformation.

Facilitating effective user navigation through website


structure improvement is mainly used to improve
website navigability through the use of user
navigation data. Despite the heavy and increasing
investments in website design, it is still revealed,
however, that finding desired information in a
website is not easy and designing effective websites
is not a trivial task
A primary cause of poor website design is that the
web developers understanding of how a website
should be structured can be considerably different
from those of the users . Such differences result in
cases where users cannot easily locate the desired
information in a website. This problem is difficult to
avoid because when creating a website, web
developers may not have a clear understanding of
users preferences and can only organize pages based
on their own judgments. However, the measure of
website effectiveness should be the satisfaction of the
users rather than that of the developers. Thus,
WebPages should be organized in a way that
generally matches the users model of how pages
should be organized .

In this paper it is concerned primarily with


transformation approaches. The literature considering
transformations approaches mainly focuses on
developing methods to completely reorganize the link
structure of a website. Although there are advocates
for website reorganization approaches, their
drawbacks are obvious. First, since a complete
reorganization could radically change the location of
familiar items, the new website may disorient users .

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

335

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

Second, the reorganized website structure is highly


unpredictable, and the cost of disorienting users after
the changes remains unanalyzed. This is because a
websites structure is typically designed by experts
and bears business or organizational logic, but this
logic may no longer exist in the new structure when
the website is completely reorganized. Besides, no
prior studies have assessed the usability of a
completely reorganized website, leading to doubts on
the applicability of the reorganization approaches.
Finally, since website reorganization approaches
could dramatically change the current structure, they
cannot be frequently performed to improve the
navigability.

ISBN: 378 - 26 - 138420 - 5

the testing data using simulations to approximate the


real usage. I define two metrics and use them to
assess whether user navigation is indeed enhanced on
the improved structure. Particularly, the first metric
measures whether the average user navigation is
facilitated in the improved website, and the second
metric measures how many users can benefit from
the improved structure. Evaluation results confirm
that user navigation on the improved website is
greatly enhanced.
II.

LITERATURE REVIEW:

Web personalization is the process of tailoring


WebPages to the needs of specific users using the
information of the users navigational behavior and
profile data . Perkowitz and Etzioni describe an
approach that automatically synthesizes index pages
which contain links to pages pertaining to particular
topics based on theco-occurrence frequency of pages
in user traversals, to facilitate user navigation. The
methods proposed byMobasher et al. and Yan et al.
create clusters of users profiles from weblogs and
then dynamically generate links for users who are
classified into different categories based on their
access patterns. Nakagawa and Mobasher develop a
hybrid personalization system that can dynamically
switch between recommendation models based on
degree of connectivity and the users position in the
site. For reviews on web personalization approaches,
see [21] and [27]. Web transformation, on the other
hand, involves changing the structure of a website to
facilitate the navigation for a large set of users [28]
instead of personalizing pages for individual users.
Fu et al. [29] describe an approach to reorganize
webpages so as to provide users with their desired
information in fewer clicks. However, this approach
considers only local structures in a website rather
than the site as a whole, so the new structure may not
be necessarily optimal. Gupta et al. [19] propose a
heuristic method based on simulated annealing to
relink webpages to improve navigability. This
method makes use of the aggregate user preference
data and can be used to improve the link structure in
websites for both wired and wireless devices.
However, this approach does not yield optimal
solutions and takes relatively a long time (10 to 15
hours) to run even for a small website. Lin [20]
develops integer programming models to reorganize
a website based on the cohesion between pages to
reduce information overload and search depth for
users. In addition, a two-stage heuristic involving two
integer-programming models is developed to reduce
the computation time.

Recognizing the drawbacks of website reorganization


Approaches.
Specifically,
mathematical
programming (MP) model is developed that
facilitates user navigation on a website with minimal
changes to its current structure. This model is
particularly appropriate for informational websites
whose contents are static and relatively stable over
time. Examples of organizations that have
informational websites are universities, tourist
attractions, hospitals, federal agencies, and sports
organizations.
The number of outward links in a page, i.e., the out
degree, is an important factor in modeling web
structure. Prior studies typically model it as hard
constraints so that pages in the new structure cannot
have more links than a specified out-degree
threshold, because having too many links in a page
can cause information overload to users and is
considered undesirable. This modeling approach,
however, enforces severe restrictions on the new
structure, as it prohibits pages from having more
links than a specified threshold, even if adding these
links may greatly facilitate user navigation. This
model formulates the out-degree as a cost term in the
objective function to penalize pages that have more
links than the threshold, so a pages out-degree may
exceed the threshold if the cost of adding such links
can be justified. Extensive experiments are performed
on a data set collected from a real website. The
results indicate that this model can significantly
improve the site structure with only few changes.
Besides, the optimal solutions of the MP model are
effectively obtained, suggesting that this model is
practical to real-world websites. model with
synthetic.
To assess the user navigation on the improved
website, the entire real data set is partitioned into
training and testing sets. The training data is used to
generate improved structures which are evaluated on

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

336

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

III.

ISBN: 378 - 26 - 138420 - 5

attract and retain customers. This model allows


Webmasters to specify a goal for user navigation that
the improved structure should meet. This goal is
associated with individual target pages and is defined
as the maximum number of paths allowed to reach
the target page in a mini session. We term this goal
the path threshold for short in this paper. In other
words, in order to achieve the user navigation goal,
the website structure must be altered in a way such
that the number of paths needed to locate the targets
in the improved structure is not larger
than the path threshold.

METRIC
FOR
EVALUATING
NAVIGATION EFFECTIVENESS

Main objective is to improve the navigation


effectiveness of a website with minimal changes.
Therefore, the first question is, given a website, how
to evaluate its navigation effectiveness. Marsico and
Levialdi point out that information becomes useful
only when it is presented in a way consistent with the
target users expectation. Palmer indicates that an
easy navigated website should allow users to access
desired data without getting lost or having to
backtrack. We follow these ideas and evaluate a
websites navigation effectiveness based on how
consistently the information is organized with respect
to the users expectations.

IV.

PROBLEM

FORMULATION

The problem of improving the user navigation on a


Website while minimizing the changes to its current
Structure can then be formulated as the mathematical
Programming model.
The objective function minimizes the cost needed to
improve the website structure, where the cost consists
of two components: 1) the number of new links to be
established (the first summation), and 2) the penalties
on pages containing excessive links, i.e., more links
than the out-degree threshold in the improved structure
(the second summation).

This metric is related to the notion of information


scent developed in the context of information
foraging theory. Information foraging theory models
the cost structure of human information gathering
using the analogy of animals foraging for food and is
a widely accepted theory for addressing the
information seeking process on the web. Information
scent refers to proximal cues (e.g., the snippets of
text and graphics of links) that allow users to estimate
the location of the distal target information and
determine an appropriate path.

V.

COMPUTATIONAL
EXPERIMENTS
AND PERFORMANCE EVALUATIONS

Extensive experiments were conducted, both on a


data set collected from a real website and on
synthetic data sets. First tested the model with
varying parameters values on all data sets. Then, we
partitioned the real data into training and testing data.
The training data is used to generate improved site
structures which were evaluated on the testing data
using two metrics that are discussed in detail later.
Moreover, we compared the results of our model with
that of a heuristic.

Here backtracks are used to identify the paths that a


user has traversed, where a backtrack is defined as a
users revisit to a previously browsed page. The
intuition is that users will backtrack if they do not
find the page where they expect it . Thus, a path is
defined as a sequence of pages visited by a user
without backtracking, a concept that is similar to the
maximal forward reference defined in Chen et al.
Essentially, each backtracking point is the end of a
path.

VI.

Evaluation of the Improved Website

In addition to the extensive computational


experiments on both real and synthetic data sets, we
also perform evaluations on the improved structure to
assess whether its navigation effectiveness is indeed
enhanced by approximating its real usage.
Specifically, we partition the real data set into a
training set (first three months) and a testing set (last
month). We generate the improved structure using
the training data, and then evaluate it on the testing
data using two metrics: the average number of paths
per mini session and the percentage of mini sessions
enhanced to a specified threshold. The first metric
measures whether the improved structure can

Problem Description
Difficulty in navigation is reported as the problem
that triggers most consumers to abandon a website
and switch to a competitor. Generally, having
traversed several paths to locate a target indicates that
this user is likely to have experienced navigation
difficulty. Therefore, Webmasters can ensure
effective user navigation by improving the site
structure to help users reach targets faster. Thesis
easy navigated websites can create a positive attitude
toward the firm, and stimulate online purchases,
whereas websites with low usability are unlikely to

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

337

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

[3] Internetretailer, Web Tech Spending Static-But


High-for the Busiest E-Commerce Sites,
http://www.internetretailer.com/ dailyNews.asp?id =
23440, 2007.

facilitate users to reach their targets faster than the


current one on average, and the second metric
measures how likely users suffering navigation
difficulty can benefit from the improvements made to
the site structure.

4] D. Dhyani, W.K. Ng, and S.S. Bhowmick, A


Survey of Web Metrics, ACM Computing
Surveys, vol. 34, no. 4, pp. 469-503, 2002.

The evaluation procedure using the first metric


consists of three steps and is described as follows:
1. Apply the MP model on the training data to obtain
the set of new links and links to be improved.
2. Acquire from the testing data the mini sessions that
can be improved, i.e., having two or more paths, their
length, i.e., number of paths, and the set of candidate
links that can be used to improve them.
3. For each mini session acquired in step 2, check
whether any candidate link matches one of the links
obtained in step 1, that is, the results from the
training data. If yes, with the assumption that users
will traverse the new link or the enhanced link in the
improved structure, remove all pages (excluding the
target page) visited after the source node of the
session for the improved website, and get its updated
length information.
VII.

ISBN: 378 - 26 - 138420 - 5

[5] X. Fang and C. Holsapple, An Empirical Study


of Web Site Navigation Structures Impacts on Web
Site Usability, Decision Support Systems, vol.
43, no. 2, pp. 476-491, 2007.
6] J. Lazar, Web Usability: A User-Centered
Design Approach. Addison Wesley, 2006.
[7] D.F. Galletta, R. Henry, S. McCoy, and P. Polak,
When the Wait Isnt So Bad: The Interacting Effects
of Website Delay, Familiarity, and Breadth,
Information Systems Research, vol. 17, no. 1, pp.
20- 37, 2006.

Mini Session and Target Identification

I employed the page-stay timeout heuristic to identify


users targets and to demarcate mini sessions. The
intuition is that users spend more time on the target
pages. Page-stay time is a common implicit
measurement found to be a good indicator of
page/document relevance to the user in a number of
studies . In the context of web usage mining, the
page-stay timeout heuristic as well as other timeoriented heuristics are widely used for session
identification and are shown to be quite robust with
respect to variations of the threshold values .
The identification of target pages and mini sessions
can be affected by the choice of page-stay timeout
threshold. Because it is generally very difficult to
unerringly identify mini sessions from anonymous
user access data, we ran our experiments for different
threshold values.
REFERENCES
[1] Pingdom, Internet 2009 in Numbers,
http://royal.pingdom.
Com/2010/01/22/internet-2009-in-numbers/, 2010.
[2] J. Grau, US Retail e-Commerce: Slower But Still
Steady Growth,
http://www.emarketer.com/Report.aspx?code=emark
eter_2000492, 2008.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

338

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

AN EFFECTIVE APPROACH TO ELIMINATE TCP INCAST


COLLAPSE IN DATACENTER ENVIRONMENTS
S Anil Kumar1, G Rajesh2
1

M.Tech 2nd year, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India.

Asst Professor, Dept. of CSE, Audisankara College of Engineering & Technology, Gudur, A.P, India.

Abstract Transport

Control Protocol (TCP)

I.

many to one communication congestion happens

Transport Control Protocol (TCP) is widely used

in high-bandwidth and low-latency networks when

on the Internet and normally works fine.

two or more synchronized servers send data to the


same receiver in parallel.

However, recent works have shown that TCP does

For many key data-

not work well for many-to-one traffic patterns on

center applications such as MapReduce and


Search,

this

many-to-one traffic pattern

INTRODUCTION

high-bandwidth, low-latency networks.

is

many-to-one

Congestion occurs when many synchronized

communication congestion may severely degrade

servers under the same Gigabit Ethernet switch

their performances, e.g., by enhancing response

simultaneously send data to one receiver in

time. In this paper, we explore the many-to-one

parallel. Only after all connections have finished

communication by focusing on the relationships

the data transmission can the next round be issued.

between TCP throughput, round-trip time (RTT),

Thus, these connections are also called barrier-

and receive window. Unlike previous approaches,

synchronized. The final performance is determined

which moderate the impact of TCP incast

by the slowest TCP connection, which may suffer

congestion by using a fine-grained timeout value,

from timeout due to packet loss. The performance

this plan is to design an Incast congestion Control

collapse of these many-to-one TCP connections is

for TCP (ICTCP) scheme on the receiver side. In

called

particular, this method changes the TCP receive

networks are well structured and layered to

window proactively before packet loss occurs.

achieve high bandwidth and low latency, and the

common.

Hence

TCP

TCP

incast

congestion.

Data-center

buffer size of top-of-rack (ToR) Ethernet switches


Index terms
congestion

many-to-one communication,

control,

round

trip

time,

is usually small this is shown in the below figure.

TCP

throughput.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

339

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

time for packet loss recovery with quicker


retransmissions, or controlling switch buffer
occupation to avoid overflow by using ECN and
modified TCP on both the sender and receiver
sides. The incast congestion in data-center
networks is shown in the below figure.
This paper focuses on avoiding packet loss before
incast congestion, which is more appealing than
Fig. 1 Data-center network of a ToR switch

recovery after loss and recovery schemes can be

connected to multiple rack-mounted servers.

complementary to congestion avoidance. Our idea


is to perform incast congestion avoidance at the

A recent measurement study showed that a barriersynchronized

many-to-one

traffic

pattern

receiver side by preventing incast congestion. The

is

receiver side can adjust the receive window size of

common in data-center networks, mainly caused

each TCP connection, so the aggregate burstiness

by MapReduce and similar applications in data

of all the synchronized senders are kept under

centers this can be shown in below figure.

control. We call our design Incast congestion


Control for TCP (ICTCP).
We first perform congestion avoidance at the
system level. We then use the per-flow state to
finely tune the receive window of each connection
on the receiver side. The technical novelties of this
work are as follows:
1) To perform congestion control on the receiver
side, we use the available bandwidth on the
network interface as a quota to coordinate the
receive

window

increase

of

all

incoming

Fig. 2 Incast Congestion in Data-center application

connections.

The root cause of TCP incast collapse is that the

2) Our per-flow congestion control is performed

highly bursty traffic of multiple TCP connections

independently of the slotted time of the round-trip

overflows the Ethernet switch buffer in a short

time (RTT) of each connection, which is also the

period of time, causing intense packet loss and

control latency in its feedback loop.

thus TCP retransmission and timeouts. Prior


solutions focused on either reducing the response

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

340

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

3) Our receive window adjustment is based on the


ratio of the difference between the measured and
expected throughput over the expected. This
allows us to estimate the throughput requirements
from the sender side and adapt the receiver
window accordingly.
We also find that live RTT is necessary for
throughput estimation as we have observed that
TCP RTT in a high-bandwidth low-latency
network increases with throughput, even if link

Fig. 4 Total goodput of multiple barrier-

capacity is not reached. We have developed and

synchronized TCP connections versus the number

implemented ICTCP as a Windows Network

of senders, where the data traffic volume per

Driver Interface Specification (NDIS) filter driver.


i)

sender is a fixed amount.

TCP Incast Congestion

We first establish multiple TCP connections

Incast congestion happens when multiple sending

between all senders and the receiver, respectively.

servers under the same ToR switch send data to

Then, the receiver sends out a (very small) request

one receiver server simultaneously, as shown in

packet to ask each sender to transmit data,

Fig. 3.

respectively. The TCP connections are issued


round by round, and one round ends when all
connections on that round have finished their data
transfer to the receiver. We observe similar
goodput trends for three different traffic amounts
per server, but with slightly different transition
points. TCP throughput is severely degraded by
incast congestion since one or more TCP

Fig. 3 Scenario of incast congestion in data-enter

connections can experience timeouts caused by

Networks.

packet drops.
each

TCP variants sometimes improve performance, but

connection is relatively small. In Fig. 4, we show

cannot prevent incast congestion collapse since

the good input achieved on multiple connections

most of the timeouts are caused by full window

versus the number of sending servers.

losses due to Ethernet switch buffer overflow. The

The amount

of

data

transmitted by

TCP incast scenario is common for data-center


applications.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

341

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ii) Reasons for TCP Incast Congestion

buffer

overflow

significantly

ISBN: 378 - 26 - 138420 - 5

reduces

TCP

timeouts and saves unnecessary retransmissions.

Incast congestion happens when the switch buffer


overflows because the network pipe is not large

We focus on the typical incast scenario where

enough to contain all TCP packets injected into the

dozens of servers are connected by a Gigabit

network.

The ToR switch is usually low-end

Ethernet switch. In this scenario, the congestion

(compared to high-layer ones), and thus the queue

point happens right before the receiver. A recent

size is not very large. To constrain the number of

measurement study showed that this scenario

packets on the flight, TCP has two windows: a

exists in data-center networks, and the traffic

congestion window on the sender side and a

between servers under the same ToR switch is

receive window on the receiver side. This paper

actually one of the most significant traffic patterns

chooses the TCP receive window adjustment as its

in data centers, as locality has been considered in

solution space. If the TCP receive window sizes

job distribution.

are properly controlled, the total receive window

We observe that the TCP receive window can be

size of all connections should be no greater than

used to throttle the TCP throughput, as it can be

the base BDP plus the queue size.

leveraged to handle incast congestion even though

II. SYSTEM DESIGN

they receive window was originally designed for


flow control. The benefit of an incast congestion

Our goal is to improve TCP performance for incast

control scheme at the receiver side is that the

congestion without introducing a new transport-

receiver knows how much throughput it has

layer protocol. Our transport-layer solution keeps

achieved and how much available bandwidth

backward compatibility on the protocol and

remains. The difficulty at the receiver side is that

programming interface and makes our scheme

an overly throttled window may constrain TCP

general enough to handle the incast congestion in

performance, while an oversized window may not

future high-bandwidth and low-latency networks.

prevent incast congestion.

Therefore, the TCP connections could be incast or


not, and the coexistence of incast and non-incast

As the base RTT is hundreds of microseconds in

connections is achieved.

data centers, our algorithm is restricted to adjust


the receive window only for TCP flows with RTT

Previous work focused on how to mitigate the

less than 2 ms. this constraint is designed to focus

impact of timeouts, which are caused by a large

on low-latency flows. Based upon following

amount of packet loss on incast congestion. Given

observations, our receive-window based incast

such high bandwidth and low latency, we focus on

congestion control is intended to set a proper

how to perform congestion avoidance to prevent

receive window for all TCP connections sharing

switch buffer overflow. Avoiding unnecessary

the same last hop. Considering that there are many

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

342

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

TCP connections sharing the bottlenecked last hop

UDP or TCP, etc. Then, define the available

before incast congestion, we adjust the TCP

bandwidth to increase all of incoming connections

receive window to make those connections share

on that bandwidth BWA interface as

the bandwidth equally. This is because in a data

= max (0,

center, parallel TCP connections may belong to

the same job, where the last one finished

Where [0, 1] is a parameter to absorb potential

determines the final performance.

oversubscribed

bandwidth

adjustment. A larger

III. ICTCP ALGORITHM

during

window

[0, 1] indicates the need

to more conservatively constrain the receive


ICTCP

provides

receive-window-based

window and higher requirements for the switch

congestion control algorithm for TCP at the end-

buffer to avoid overflow; a lower indicates the

system. The receive windows of all low-RTT TCP

need to more aggressively constrain the receive

connections are jointly adjusted to control

window, but throughput could be unnecessarily

throughput on incast congestion. ICTCP algorithm

throttled. A fixed setting of BWA in ICTCP, an

closely follows the design points made. It is

available bandwidth as the quota for all incoming

described how to set the receiver window of a

connections to increase the receive window for

TCP connection.
i)

higher throughput. Each flow should estimate the

Bandwidth Estimation

potential throughput increase before its receiving

Using available bandwidth to increase all of

window is increased. Only when there is enough

incoming connections on the receiver server.

quota (BWA) can the receive window be increased,

Developing ICTCP as an NDIS driver on

and the corresponding quota is consumed to

Windows OS. Our NDIS driver intercepts TCP

prevent bandwidth oversubscription. To estimate

packets and modifies the receive window size if

the available bandwidth on the interface and

needed. It is assumed there is one network

provide a quota for a later receive window

interface on a receiver server, and define symbols

increase, we divide the time into slots. Each slot

corresponding to that interface. This algorithm can

consists of two sub slots of the same length. For

be applied to a scenario where the receiver has

each network interface, we measure all the traffic

multiple interfaces, and the connections on each

received in the first sub slot and use it to calculate

interface

algorithm

the available bandwidth as a quota for window

independently. Assume the link capacity of the

increase on the second sub slot. The receive

interface on the receiver server is L. Define the

window of any TCP connection is never increased

bandwidth of the total incoming traffic observed

at the first sub slot, but may be decreased when

on that interface as BWT, which includes all types

congestion is detected or the receive window is

of packets, i.e., broadcast, multicast, unicast of

identified as being over satisfied.

should

perform

this

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

343

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ii) Flow Stack Information

throughput and its expected throughput. The


measured throughput represents the achieved

A flow table maintains the key data structure in

throughput on a TCP connection, also implies the

the Receiver server. A flow is identified by a 5tuple:

ISBN: 378 - 26 - 138420 - 5

source/destination

IP

current requirement of the application over that

address,

TCP connection.

source/destination port, and protocol. The flow


table stores flow information for all the active

= max (

+ (1 )

flows. The packet header is parsed and the


corresponding information is updated in the flow

am represents Incoming measured throughput

table. Network driver interface specification filter

aisrepresents Sample of current throughput (on

functionalities are performed like collecting

connection i) .The expected throughput represents

network statistics,

and

our expectation of the throughput on that TCP

filtering the unauthorized ones. In ICTCP, each

connection if the throughput is only constrained by

connection adjusts its receive window only when

receive window.

monitoring activities

an ACK is sending out on that connection. No

aie = max(ami, rwndi/RTTi)

additional pure TCP ACK packets are generated


aie represents Expected throughput of i ,rwndi

solely for receive window adjustment, so that no

represents Receive window of i. The ratio of

traffic is wasted. For a TCP connection, after an

throughput

ACK is sent out, the data packet corresponding to

difference

represented as

that ACK arrives one RTT later. As a control

Tdib

(aie

of

connection

aim)/

aie.

is

system, the latency on the feedback loop is one

Our idea on receive window adjustment is to

RTT for each TCP connection, respectively.

increase window when the difference ratio of

Meanwhile, to estimate the throughput of a TCP

measured and expected throughput is small, while

connection for a receive window adjustment; the

decrease window when the difference ratio is

shortest timescale is an RTT for that connection.

large.

Therefore, the control interval for a TCP

iv) Choosing

connection is 2*RTT in ICTCP, and needed one

Fairness

among

Multiple

Connections

RTT latency for the adjusted window to take effect


and one additional RTT to measure the achieved

When the receiver detects that the available

throughput with the newly adjusted receive

bandwidth has become smaller than the threshold,

window.

ICTCP starts to decrease the receiver window of


the selected connections to prevent congestion.

iii) Receive Window Adjustment

Considering that multiple active TCP connections

For any ICTCP connection, the receive window is

typically work on the same job at the same time in

adjusted

a data center, there is a method that can achieve

based

on

its

incoming

measured

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

344

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

fair sharing for all connections without sacrificing

communication, in Proc. ACMSIGCOMM,

throughput.

2009, pp. 303314.


IV. CONCLUSION

[3] S. Kandula, S. Sengupta, A. Greenberg, P.


Patel, and R. Chaiken, The nature of data

This paper presented an effective, practical and

center traffic: Measurements & analysis, in

safe solution to eliminate TCP incast collapse in

Proc. IMC, 2009, pp. 202208.

datacenter environments. Our approach utilizes the


link bandwidth as fully as possible but without

[4] J. Dean and S. Ghemawat, MapReduce:

packet losses by limiting the round trip time value.

Simplified data processing on large clusters,

Based on the concept of bandwidth delay product,

in Proc. OSDI, 2004, p. 10.

our

technique

conservatively

estimates

the
[5] M. Alizadeh, A. Greenberg, D.Maltz, J.

reasonable number of concurrent senders. In this


work

we

used

Network

driver

Padhye, P. Patel, B.Prabhakar, S. Sengupta,

interface

and M. Sridharan, Data center TCP

specification filter functionalities like collecting


network

statistics;

monitoring,

(DCTCP), in Proc. SIGCOMM, 2010, pp. 63

filtering

74.

unauthorized ones are done. We can avoid


retransmission, safely send the data to the receiver

[6] D. Nagle, D. Serenyi, and A. Matthews, The

and also avoid congestion and traffic. Our system

PanasasActiveScale

is capable to match user preferences while

Delivering scalable high bandwidth storage,

achieving full utilization of the receivers access in

in Proc.SC, 2004, p. 53.

storage

cluster:

many different scenarios.


AUTHORS
REFERENCES

S Anil Kumar has received his


B.Tech

[1] A. Phanishayee, E. Krevat, V. Vasudevan, D.

degree

in

Computer

Andersen, G. Ganger, G.Gibson, and S.

Science & Engineering from

Seshan, Measurement and analysis of TCP

Chadalavada

throughput collapse in cluster-based storage

Engineering College, Tirupathi

systems, in Proc. USENIX FAST, 2008,

affiliated to JNTU, Anantapur in 2009 and

Article no. 12.

pursuing M.Tech degree in Computer Science &


Engineering

[2] V. Vasudevan, A. Phanishayee, H. Shah, E.

at

Audisankara

Ramanamma

College

of

Engineering & Technology, Gudur affiliated to

Krevat, D. Andersen, G. Ganger, G. Gibson,

JNTU, Anantapur in (2012-2014).

and B.Mueller, Safe and effective finegrained TCP retransmissions for datacenter

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

345

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

G Rajesh M.Tech, (Ph.D)


currently he is working as
Assistant

Professor

Audisankara

College

in
of

Engineering And Technology,


Gudur (M), Nellore, Andhra Pradesh, India.
He has seven years of experience in teaching
and two years of experience in Software
Industry. Previously he trained and worked
with

DSRC

(Data

Software

Research

Company) Chennai on Oracle Application


Functional Consultant. And he has worked
with Capgemini India Ltd Mumbai as a
Software Engineer (Oracle Apps Technical
Consultant) as a contract employee through
the Datamatics Pvt Ltd. He was doing his
Ph.D on A Cross Layer Framework for
Bandwidth Management of Wireless Mesh
Networks.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

346

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Secured and Efficient Data Scheduling of Intermediate Data Sets


in Cloud

D.TEJASWINI.,M.TECH
C.RAJENDRA.,M.TECH.,M.E.,PH.D.

AUDISANKARA COLLEGE OF ENGINEERING & TECHNOLOGY

ABSTRACT
Cloud computing is an emerging field in the development of business and
organizational environment. As it provides more computation power and storage space users
can process many applications.Due to this large number of intermediate datasets are
generated and Encryption and decryption techniques are used to preserve the intermediate
data sets in cloud. An Upper bound constraint based approach is used to identify sensitive
intermediate data sets and we apply suppression technique on sensitive data sets in order to
reduce the time and cost. The Value Generalization Hierarchy protocol is used to achieve
more security so that number of users can access the data with privacy.Along with that
Optimized Balanced Scheduling is also used for the best mapping solution to meet the system
load balance to the greatest extent or to reduce the load balancing cost
The Privacy preservation is also ensured with dynamic data size and access
frequency values. Storage space and computational requirements are optimally utilized in the
privacy preservation process. Data distribution complexity is also handled in the scheduling
process.
Keywords: Cloud computing, privacy upper bound, intermediate data sets, optimized
balanced scheduling, value generalization hierarchy protocol.

shared for multiple users but also


dynamically reallocated per demand. The
privacy issues [12] caused by retaining
intermediate data sets in cloud are
important but they were paid little
attention. For preserving privacy v[9] of
multiple data sets, we should anonymize
all data sets first and then encrypt them
before storing or sharing them in cloud.
Usually, the weightage of intermediate
data sets[11] is huge. Users will store only

1. INTRODUCTION
Cloud computing mainly relies on sharing
of resources to achieve coherence and
economies of scale similar to a utility over
a network. The basement of cloud
computing is the broader concept of
converged infrastructure and shared
services. The cloud mainly focuses on
maximizing the effectiveness of shared
resources. Cloud resources are not only

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

347

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

cloud. While our work provides a


significant first step towards Zhang et
al.[10] proposed a system named Sedic
which partitions Map Reduce computing
jobs in terms of the security labels of data
they work on and then assigns the
computation without sensitive data to a
public cloud. The sensitivity of data is
required to be labelled in advance to make
the above approaches available. Ciriani et
al.[10] has proposed an approach that
combines the encryption and data
fragmentation to achieve the privacy
protection for distributed data storage with
encrypting only part of data sets.

important datasets on cloud when


processing original data sets in dataintensive applications such as medical
diagnosis[16], in order to reduce the
overall expenses by avoiding frequent recomputation to obtain these data sets. Such
methods are quite common because data
users often re-analyse results, conduct new
analysis on intermediate data sets, and also
share some intermediate results with others
for collaboration.
Data Provenance is employed to manage
the intermediate datasets. A number of
tools for capturing provenance have been
developed in workflow systems and a
standard for provenance representation
called the Open Provenance Model (OPM)
has been designed.

3.SYSTEM ARCHITECTURE

2.RELATED WORK.
Encryption is usually integrated with other
methods to achieve cost reduction, high
data usability and privacy protection. Roy
et al. [8] investigated the data privacy
problem caused by Map Reduce and
presented a system named Airavat which
incorporates mandatory access control
with differential privacy. Puttaswamy et al.
[9] described a set of tools called
Silverline which identifies all encryptable
data and then encrypts them to protect
privacy. Encrypted data on the cloud
prevent privacy leakage to compromisedor
malicious clouds, while users can easily
access data by decrypting data locally with
keys from a trusted organization. Using
dynamic program analysis techniques
Silverline automatically identifies the
encryptable application data that can be
safely encrypted without negatively
affecting the application functionality. By
modifying the application runtime, e.g. the
PHP interpreter, we show how Silverline
can determine an optimal assignment of
encryption keys that minimizes key
management overhead and impact of key
compromise. Our applications running on
the cloud can protect their data from
security breaches or compromises in the

Fig 1: System Architecture For Secure


Transaction Using The Cloud
Our approach mainly will work by
automatically identifying the subsets of an
applications data that are not directly used
in computation, and exposing them to the
cloud only in encrypted form.
We present a technique to partition
encrypted data into parts that are accessed
by different sets of the users (groups).
Intelligent key assignment limits the
damage which is possible from a given key
compromise, and strikes a good trade off
between robustness and key management
complexity.
We present a technique that enables
clients to store and use their keys safely
while preventing cloud-based service from
stealing the keys. Our solution works

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

348

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

today on unmodified web browsers. There


are many privacy threats caused due to the
intermediate data sets so, we need to
encrypt these data sets to provide privacy
and make them secure.

ISBN: 378 - 26 - 138420 - 5

4.2 Privacy Preserved Data Scheduling


Scheme
Here multiple intermediate data set privacy
models is combined with the data
scheduling mechanism. The Privacy
preservation is ensured with the dynamic
data size and access frequency values and
along with that Storage space and
computational requirements are optimally
utilized in the privacy preservation process
and Data distribution complexity is also
handled in the scheduling process. The
Data sensitivity is considered in the
intermediate data security process.
Resource requirement levels are monitored
and controlled by the security operations.
The system is divided into five major
modules. They are data center, data
provider, intermediate data privacy,
security analysis and data scheduling. The
data center maintains the encrypted data
values for the providers. Shared data
uploading process are managed by the data
provider module. The Intermediate data
privacy module is designed to protect
intermediate results. Security analysis
module is designed to estimate the
resource and access levels. Original data
and intermediate data distribution is
planned under the data scheduling module.
Dynamic privacy management and
scheduling mechanism are integrated to
improve the data sharing with security.
Privacy preserving cost is reduced by the
joint verification mechanism.

Fig.2: A Scenario Showing Privacy


Threats Due To Intermediate Datasets

4.IMPLEMENTATION
4.1 Requirements
The problem of managing the intermediate
data which is generated during dataflow
computations, deserves deeper study as a
first-class problem. They are the following
two major requirements that any effective
intermediate storage system needs to
satisfy: availability of intermediate data,
and minimal interference on foreground
network traffic generated by the dataflow
computation.
Data Availability: A task which is in a
dataflow stage cannot be executed if the
intermediate input data is unavailable. A
system that provides higher availability for
intermediate data will suffer from fewer
delays for re-executing tasks in case of
failure. In multi-stage computations, high
availability is critical as it minimizes the
effect of cascaded re-execution.
Minimal Interference: At the same time,
the data availability cannot be pursued
over-aggressively. In particular, since
intermediate data is used immediately, and
there is high network contention for
foreground traffic of the intermediate data
transferred to the next stage. So an
intermediate data management system
needs to minimize interference.

4.3 Analysis of the Cost Problem


A cloud service provides various pricing
models to support the pay-as-you-go
model, e.g., Amazon Web Services pricing
model[4]. The Privacy-preserving cost of
the intermediate data sets can be reduced
from frequent encryption or decryption
with charged cloud services which needs
more computation power, data storage, and
other cloud services. To avoid the pricing

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

349

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

multiple data sets, it is promising to


anonymize all data sets first and then
encrypt them before storing or sharing
them in cloud. Usually, the volume of
intermediate data sets is huge. Data sets
are divided into two sets. One is sensitive
intermediate data set and another is nonsensitive intermediate data set. Sensitive
data set is denoted as SD then non
sensitive data set is denoted as NSD. The
equations, sd U NSD =D and SD NSD
= hold. The pair of (SD, NSD) as a
global privacy preserving of cloud data.
Suppression technique done only on
sensitive data sets in two ways such as
semi suppression and full suppression,
while full suppression apply on most
important sensitive intermediate data set
that is individual data set value fully
encoded then semi suppression apply on
selective sensitive data sets that is half of
the data set value will be encoded. Also
propose Value Generalization Hierarchy
(VGH) protocol to reduce cost of data

details and to focus on, combine the prices


of various services required by encryption
or decryption into one.
4.4 Proposed Framework
The technique or the new protocol which
we use for privacy protection here is the
Value Generalization Hierarchy Protocol
which has the functionality of assignment
of the common values for the unknown
and original data values for general
identification later on we add of full
suppression on the more important data
sets which enhances the complete
encryption of the entire data sets given. To
Investigate privacy aware and efficient
scheduling of intermediate data sets for
minimum cost and fast computation.
Suppression of data is done to reduce the
overall computation time and cost and
where VGH Protocol is also proposed to
achieve it. Here we secure the more
important dataset though semi suppression
only. The full suppression to achieve the
high privacy or security of original data
sets and the original data set is only
viewed by owner. Here number user can
access the data with security and to avoid
privacy leakage. The privacy protection
cost for intermediate data sets that needs to
be encoded while using an upper bound
constraint-based approach to select the
necessary subset of intermediate data sets.
The privacy concerns caused by retaining
intermediate data sets in cloud are
important Storage and computation
services in cloud are equivalent from an
economical perspective because they are
charged in proportion to their usage.
Existing
technical
approaches
for
preserving the privacy of datasets stored in
cloud mainly include encryption and
anonymization. On one hand, encrypting
all data sets, a straightforward and
effective approach, is widely adopted in
current research. However, processing on
encrypted data sets efficiently is quite a
challenging task, because most existing
applications only run on unencrypted data
sets. Thus, for preserving privacy of

4.5 Optimized Balanced Scheduling


The optimized balanced scheduling is used
for the best mapping solution to meet the
system load balance to the greatest extent
or to lower the load balancing cost. The
best scheduling solution for the current
scheduling process can be done through
genetic algorithm. First we need to
compute the cost through the ratio of the
current scheduling solution to the best
scheduling solution, and then we have to
make the best scheduling strategy
according to the cost. So that it has the
least influence on the load of the system
after scheduling and it has the lowest cost
to reach load balancing. In this way, we
can form the best scheduling strategy.

5. CONCLUSION
In this paper, focus is mainly
contributed towards identification of the
areas
where
the
most
sensitive
intermediate datasets are present in cloud.
An upper bound constraint based approach
is used where data sets needs to be

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

350

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

[5] K. Zhang, X. Zhou, Y. Chen, X. Wang,


and Y. Ruan, Sedic: Privacy-Aware Data
Intensive Computing on Hybrid Clouds,
Proc. 18th ACM Conf.
Computer and Comm. Security (CCS 11),
pp. 515-526, 2011.
[6] H. Lin and W. Tzeng, A Secure
Erasure Code-Based Cloud Storage
System with Secure Data Forwarding,
IEEE Trans. Parallel and Distributed
Systems, vol. 23, no. 6, pp. 995-1003, June
2012.
[7] G. Wang, Z. Zutao, D. Wenliang, and
T. Zhouxuan, Inference Analysis in
Privacy-Preserving Data Re-Publishing,
Proc. Eighth IEEE Intl Conf.
Data Mining (ICDM 08), pp. 1079-1084,
2008.
[8] K.P.N. Puttaswamy, C. Kruegel, and
B.Y. Zhao, Silverline: Toward Data
Confidentiality in Storage-Intensive Cloud
Applications, Proc. Second
ACM Symp. Cloud Computing (SoCC
11), 2011.
[9] I. Roy, S.T.V. Setty, A. Kilzer, V.
Shmatikov, and E. Witchel, Airavat:
Security and Privacy for Mapreduce,
Proc. Seventh USENIX Conf.
Networked
Systems
Design
and
Implementation (NSDI 10), p. 20, 2010.
[10] X. Zhang, C. Liu, J. Chen, and W.
Dou, An Upper-Bound Control Approach
for Cost-Effective Privacy Protection of
Intermediate Data Set Storage
in Cloud, Proc. Ninth IEEE Intl Conf.
Dependable, Autonomic and Secure
Computing (DASC 11), pp. 518-525,
2011.
[11] B.C.M. Fung, K. Wang, R. Chen, and
P.S. Yu, Privacy-Preserving Data
Publishing: A Survey of Recent
Developments, ACM Computing Survey,
vol. 42, no. 4, pp. 1-53, 2010.
[12] H. Lin and W. Tzeng, A Secure
Erasure Code-Based Cloud Storage
System with Secure Data Forwarding,
IEEE Trans. Parallel and Distributed
Systems, vol. 23, no. 6, pp. 995-1003, June
2012.

encoded, in order to reduce the privacy


preserving cost so we investigate privacy
aware efficient scheduling of intermediate
data sets in cloud by taking privacy
preserving as a metric together with other
metrics such as storage and computation.
Optimized balanced scheduling strategies
are expected to be developed toward
overall highly efficient privacy aware data
set scheduling and mainly in the overall
time reduction and Data delivery overhead
is reduced by the load balancing based
scheduling mechanism. Dynamic privacy
preservation model is supported by the
system and along with that a high security
provisioning is done with the help of full
suppression, semi suppression and Value
Generalization Hierarchy Protocol. This
protocol is used to assign the common
attribute for different attributes and the
Resource consumption is also controlled
by the support of the sensitive data
information graph.

6.REFERENCES
[1] L. Wang, J. Zhan, W. Shi, and Y.
Liang, In Cloud, Can Scientific
Communities Benefit from the Economies
of Scale?, IEEE Trans. Parallel and
Distributed Systems, vol. 23, no. 2, pp.
296-303, Feb.2012.
[2] Xuyun Zhang, Chang Liu, Surya
Nepal, Suraj Pandey, and Jinjun Chen, A
Privacy Leakage Upper Bound ConstraintBased Approach for Cost-Effective
Privacy Preserving of Intermediate Data
Sets in Cloud, IEEE Transactions On
Parallel And Distributed Systems, Vol. 24,
No. 6, June 2013.
[3] D. Zissis and D. Lekkas, Addressing
Cloud Computing Security Issues, Future
Generation Computer Systems, vol. 28, no.
3, pp. 583- 592, 2011.
[4] D. Yuan, Y. Yang, X. Liu, and J. Chen,
On-Demand
Minimum
Cost
Benchmarking for Intermediate Data Set
Storage in Scientific Cloud Workflow
Systems,
J.
Parallel
Distributed
Computing, vol. 71, no. 2, pp. 316-332,
2011.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

351

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

Key Reconstruction and clustering opponent nodes in Minimum Cost


Blocking Problems
D.Sireesha
ASCET
M.Tech(CSE)
[email protected]

Prof C.Rajendra
[email protected]

ABSTRACT: This paper detects malicious activities on wireless mesh networks with
wireless routing protocols. Upon the selection of route discovery malicious attacks are
ahead to compromise the nodes such as network partitioning, node isolation activities
.The cardinal covering in the routing protocols address the packets in the network.
When attacks are take place by the opponent, negotiate the attacker node and choose
another way to send the packet from source to destination .But it cant be moved from
the wireless network. Here proposing, looking into the Clustering protocols along with
id-based key update protocols, which is very predicting in wireless networks.
Cryptography techniques improvise the packet in an assured way from source to
destination and it doesnt render to secure the network/system.
Keywords: Wirelessmeshnetworks, clustering, MSR protocol, adversary nodes, updating key.
of a seamlessly connected world into reality.

1. INTORDUCTION:
Wireless mesh networks (WMNS)

Wireless mesh Networks includes mesh

issued as a bright conception to meet the

routers and mesh clients. Getting rid of the

disputes on succeeding generation wireless

wires from the wireless LANs, doubling the

networks. Providing flexibility, adaptability,

access point is known as mesh routers and it

reconfigurable the network upon the cost-

forms a backbone network. A mesh network

efficient solutions to service provider. It has

can be designed using a flooding technique

the possibility to wipe out many of these

or a routing technique.[3] When using a

disadvantages like low-cost, wireless broad

routing

band internet access for both wired and

propagated along a path, by hopping from

mobile nodes. Wireless mesh networks are

node to node until the destination is reached.

an emerging technology right now. The use of

To ensure all its paths' availability, a routing

mesh wireless networks may bring the dream

network

technique,

must

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

352

the

allow

message

for

is

continuous

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

connections

and

broken

blocked

or

reconfiguration
paths,

around

ISBN: 378 - 26 - 138420 - 5

protocols for wireless mesh networks make it

using self-

particularly

hard

for

an

adversary

to

healing algorithms. A mesh network whose

effectively launch such attacks. Attempting to

nodes are all connected to each other is

model the theoretical hardness of attacks on

a fully connected network. Mesh networks

multi-path routing protocols for mobile nodes

can be seen as one type of. Mobile adhoc

and qualify it in mathematical terms. The

networks (MANETs) [1] and mesh networks

study will impact the areas of security and

are therefore closely related. The mesh

robustness of routing protocols of wireless

routers render a rich radio connectivity in

mesh networks. [3], threshold cryptography

which importantly brings down the direct

and network coding. In previously, every

deployment cost of the network. Mesh routers

node had a unique ID randomly, routing

are usually stationary and it doesnt have

occurs through protocols such as greedy and

power constraint. Mesh clients are mobile

LP algorithms. Problems incorporated in last

nodes. Sometimes mesh routers can also act

investigations are if any opponent were

as gateways which are connected to the

hacked any of the nodes in the network, it

internet through a wired backbone.

can retrieve the packet before sending the

2. RELATED WORK:

destination. And one more consequence is


about the topological information of the

Privacy and security plays a major role in the

network, through that it may miss send the

communication network for both wired and

packet without permissions of authenticated

wireless network. Security issues related to

nodes. To reduce these complications, here

the packet transmission from the source to

introducing the clustering concepts[4] were

the destination nodes. Securing the data from

incorporated i.e. , grouping the nodes in the

the opponent without changing the main

mesh

thing. Different technologies are introduced to


secure

data

from

opponents

by

algorithms

the

to

group

the

nodes

using

MSR(multipath split routing) protocol. Again

techniques are included for protecting packet

to

from un-authorizers. It demonstrates the high

protect

data

from

adversary

nodes

introducing cryptographic key distributions

quality of multipath routing protocols above

techniques are integrated. Introducing the

traditional single path protocols in terms of

clustering algorithms to group the nodes in

resiliency under the attacks of blocking and

the mesh networks, giving different IDs to the

node isolation types of attacks , particularly in


domain.

identifying

node from the network again make clustering

network. Clustering [4] and cryptography

networks

after

misbehaving node move out the effected

using

clustering the different nodes in the mesh

wireless

networks

nodes. Grouping sensor nodes into clusters

Multi-path

has been widely investigated by researchers

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

353

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

in order to achieve the network system's

behavior, section 4.2 presents the clustering

scalability and management objectives. Every

algorithm implementation details, section 4.3

cluster would have a leader sensor node,

discuss key updating process after clustering

often referred to as the cluster-head (CH),

adversary nodes section 5 concludes the

which

paper with the future discussion.

can

be

fixed

or

variable.

CHs

aggregate the data collected by the sensor


nodes

in

its

cluster,

thus,

4. IMPLEMENTATION:

clustering

decreases the number of relayed packets. As

Wireless mesh networks integrating a key

the benefits of clustering, it conserves

technology for

communication

networks

bandwidth

and

avoids

next

showing

generation
rapid

wireless

progress

and

redundant exchange of messages among

numerous applications. In a WMN, high

sensor nodes, because of limiting the inter-

speed routers integrated with each other in a

cluster

multi-hop fashion over wireless channels and

interactions.

Therefore,

clustering

prolongs the lifetime of WSNs [2][4].

form a broadband backhaul. In the previous


analysis of the paper includes that in wireless

(A).How does a mesh network works?

mesh networks. Due to the delivery of packet

While traditional networks rely on a small

from source to destination it should be send

number of wired access points or wireless

through secure path by providing a key to the

hotspots to connect users, a wireless mesh


network

spreads

network

packet. In WMNs an adversary node involves

connection

in the network can know the details of

among dozens or even hundreds of wireless

topological information, obviously the routing

mesh nodes that "talk" to each other, sharing

details. In order to secure the packet from the

the network connection across a large area.

opponent we need to reduce the way of

Some think of the internet as being the

adversary

world's largest mesh network. When using

nodes

about

topological

information[6] and providing another key to

the internet, information travels by being

the packet. Here in this paper introducing the

bounced automatically from one router to the

clustering

next until it reaches its destination. The

and

cryptographic

key

regeneration. Analyzing the node behavior

internet has billions of potential paths across

whether it is acting in a normal way or not, it

which data can travel.

is possible by using finite state machine.

3. PAPER ORGANIZATION:

Behavior of each node is noted in the table,

The rest of this paper is organized as follows

clustering misbehaving nodes using a unique

section

implementation

id and authorized one with another ids. Upon

process, section 4.1 introducing the finite

creating different ids for misbehaving nodes

state machine to record the details of nodes

and authorized node packet details have to

discuss

the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

354

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

be modified. This is re-generating the key to

A. ROUTE DISCOVERY OF MSR:

protect

Here

Multipath split routing builds multi routes

introducing MSR protocol for generating the

using request/reply cycles. When source

maximal disjoint paths in the network, and

needs routes to the destination but no route

how the MSR is working upon the selection of

information is known, it overflows the route

routes in the network. To observer the

request message (RREQ) to the entire

behavior of the nodes incorporating finite

network. Due to the packet overflow, several

state machine model. Examine each and

replicate that spanned through different

every nodes behavior about the status of

routes

packet delivery and soon. The selfish node is

destination node selects multiple disjoint [5]

identified in the network based on data

routes and sends route reply (RREP) packets

collected by each node from its local

back to the source through the selected

message

routes.

data

from

unit

misbehaving

opponents.

(LMU).

nodes

Clustering

the

and

as

group

to

reach

the

destination.

The

authorized ones a group. Although grouping

4.2 THE FINITE STATE MACHINE MODEL:

opponents doesnt made packet in a secured

In the proposed mechanism, the messages

way. Again providing a key is an important

corresponding to a RREQ flooding and the

task

re-generating

uni-cast RREP are referred to as a message

algorithms are introduced to provide a key at

unit. It is clear that no node in the network

the packet header.

can observe all the transmission in a

to

protect

data

i.e.

message unit. The subset of a message unit

4.1 MULTI-PATH SPLIT ROUTING:

that a node can observe is referred to as the

Multipath Split Routing (MSR) protocol that

local message unit (LMU). The LMU for a

builds maximally disjoints paths. Multiple

particular node consists of the messages

routes, of which one is the shortest delay

transmitted by the node and its neighbors,

path, are discovered on demand. Established

and the messages overheard by the node.

routes are not necessarily of equal length.

The selfish node detection is done based on

Data traffic is split into multiple routes to

data collected by each node from its

avoid congestion[5] and to use network

observed

resources efficiently. We believe providing

transmission in an LMU, a node maintains a

multiple routes are beneficial in network

record of its sender, and the receiver, and the

communications,

neighbor nodes that receive the RREQ

wireless

particularly

networks

where

in

mobile

routes

LMUs.

For

each

message

broadcast sent by the node itself.

are

disconnected frequently because of mobility

The finite state machine depicts various

and poor wireless link quality.

states

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

355

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

1. Init in intial phase no RREQ is

by the source-destination pair contained in a

observed.

RREQ message. At the start of a routing

2. Unexp RREP- receiving a RREP RREQ is

session, a monitored node is at the state 1 in

not observed.

its finite state machine. As the monitor

3. Recv RREQ- acknowledgment of RREQ

node(s)

is observed.

monitored node based on the LMUs, it

4. Frwd RREQ- distribute of a RREQ is

records transitions form its initial state 1 to

observed.

one of its possible final states -- 5, 7 and 8.

observes

the

behavior

of

the

5. Timeout RREQ Timeout after receiving


RREQ.
6. Recv

RREP-

Receipt

of

RREP

is

observed.
7. LMU complete- forwarding a valid RREP
is observed.
8. Timeout RREP-Timeout after receipt of a
RREP.

Fig 1 : Finite state machine of a monitored node

In which a node may exist for each LMU [12].


The final states are shaded. Each message

When a monitor node broadcasts a RREQ, it

sent by a node causes a transition in each of

assumes that the monitored node has

its neighbors finite state machine. The finite

received it. The monitor node, therefore,

state machine in one neighbor gives only a

records a state transition 1->3 for the

local view of the activities of that node. It

monitored nodes finite state machine. If a

does not in any way, reflects the overall

monitor node observes a monitored node to

behavior of the node. The collaboration of

broadcast a RREQ, then a state transition of

each neighbor node makes it possible to get

3->4 is recorded if the RREQ message was

an accurate picture about the monitored

previously sent by the monitor node to the

nodes behavior. In the rest of the paper, a

monitored node; otherwise a transition of 1->

node being monitored by its neighbors is

4 is recorded since in this case, the RREQ

referred to as a monitored node[8], and its

was received by the monitored node from

neighbors are referred to as a monitor node.

some other neighbor. The transition to a

In the protocol, each node plays the dual role

timeout state occurs when a monitor node

of a monitor node and a monitored node. A

finds no activity of the monitored node for the

local message unit (LMU).Each monitor node

LMU before the expiry of a timer. When a

observes a series of interleaved LMUs for a

monitor node observes a monitored node to

routing session. Each LMU can be identified

forward a RREP, it records a transition to the

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

356

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

final state LMU complete (State No 7). At

Table 1: The state transitions of the neighbor

this state, the monitored node becomes a

nodes of node N

candidate for inclusion on a routing path.


4.3
When the final state is reached, the state

4.3 CLUSTERING ALGORITHM:

machine terminates and the state transitions

For clustering mesh networks need to

are stored by each node for each neighbor.

perform two phases clustering setup and

After sufficient number of events is collected,

clustering maintenance .The first phase is

a statistical analysis is performed to detect

accomplished by choosing some nodes that

the presence of any selfish nodes.

act as coordinators of the clustering process


(cluster heads). Then a cluster is formed by
associating a cluster head with some of its
neighbors that become the ordinary nodes of
the cluster. Based on the ID s and the

Fig 2: An example local message unit


(LMU) observed by node N

knowledge of the neighbors and the obtained


clustering

Above fig: depicts an example of LMU

respect

observed by the node N during the discovery

clustering

destination node D indicated by bold lines.

one.

with

Introducing

set

up

and

maintenance.

Mobility-Adaptive

Clustering

following properties.

each of its three neighbor nodes X, Y and Z.

initial

properties

(DMAC, for short) algorithm we obtain the

N and the corresponding state transitions for

the

Distributed

Table 1 shows the events observed by node

Events

to

different

clustering algorithm suitable for both the

of a route from the source node S to the

Neighbor

has

Nodes can move even in the clustering


setup DMAC is adaptive it can changes in the

State

topology of the network, remove any node

changes

from the network.

X broadcasts RREQ.

1to 4

N broadcasts RREQ.

4to 4

N sends RREP to X.

4to 6

X sends RREP to S.

6to 7

Y broadcasts RREQ.

1to 4

N broadcasts RREQ.

4to 4

Timeout
N broadcasts RREQ.

4to5
1to3

Z broadcasts RREQ.

3 to4

Z sends RREP to N

4to7

DMAC

is

fully

distributed.

node

decides its own role (i.e., cluster head or


ordinary node) solely knowing its current one
hop neighbors.
Every ordinary node always has direct
access to at least one cluster head. Thus, the
nodes in the cluster are at most two hops
apart. This guarantees fast

intra-cluster

communication

inter-cluster

and

fast

exchange of information between any pair of

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

357

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

nodes.

(2) An existing mesh router leaves the

DMAC uses a general mechanism for the

backbone network.

selection of the cluster heads. The choice is

(3) A cheater is detected in the network

now based on generic weights associated

before packets are sent by the source, if any

with the nodes.

attacker present in the network it shows an

The number of cluster heads that are

indication about the opponent.

allowed to be neighbors is a parameter of the

To prevent a mobile attacker from breaking t

algorithm

key pieces, every key piece should be


updated within a defined cycle T. Only after at

4.4

least t key pieces are obtained in the same

KEY UPDATING UPON

cycle, can the secret be reconstructed. Key

CLUSTERING CHEATER NODES:


In

the

MCBP

(minimum

cost

update involves the following four steps:

blocking

problems) whether the nodes in the network

(1) The offline CA (Administrator center) is

were attacked by the opponent are identified.

aroused in the network.

Before packet is sending a node has been

(2).The CA constructs a new group key SK 

checking whether the neighboring nodes

and selects new polynomial f (x) .The new

were attacked by the cheater. If any of the

key pieces (di, SKi) are calculated and

nodes in the network were attacked, then

delivered to n selected mesh routers. Then,

node must comprise and choose another

the CA disconnects itself from the network

path to send the packet.

and remains offline.

Even though compromising the node doesnt

3) A mesh router requests (t-1) key pieces

changes the cheaters behavior, grasping the

from other mesh routers.

topological information, Packet details. So

(4).After t key pieces are collected; the mesh

here we are introducing the key update

router reconstructs the new group key SK ,

algorithms after clustering the misbehaving

d which cheater detection and identification

node[10] from the authorized ones in the

can be carried out as described.

wireless mesh networks.


Here in the key update algorithms we are

5. CONCLUSION:

reconstructing the group key, including key

This paper concentrates on avoiding the

pieces algorithm.

opponent in the network, identifying the

i.

misbehaving nodes within the mesh networks

Key Update:

Following are several conditions in which the

clustering the misbehaving nodes and the

group key needs to be updated.

authenticated persons separately. By finite

(1) A new mesh router connects to the

state machine observing the states of the

backbone network.

adversary nodes in the network, due to these

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

358

www.iaetsd.in

INTERNATIONAL CONFERENCE ON CURRENT INNOVATIONS IN ENGINEERING AND TECHNOLOGY

ISBN: 378 - 26 - 138420 - 5

observation defines how it is behaving and

[6].A length-flexible threshold cryptosystem

comparing both the authorized and opponent.

with applications, in Proceedings of the 8th

While delivering packet from the source to the

Australasian

destination we need to secure the packet by

security and privacy, ser. ACISP03. Berlin,

using some cryptographic techniques by

Heidelberg: Springer-Verlag, 2003, pp. 350

providing key to the packet. Introducing the

364.

group key reconstruction algorithms gives

Wireless Networks, Communications and

wireless mesh networks. Further we can


on

these

group

Mobile

key

and

additional

Computing,

2005

International

Conference on, vol. 1, June 2005, pp. 69

reconstruction algorithms and the earlier


algorithms

Information

hoc networks and threshold cryptography, in

clustering the adversary node from the


simulations

on

[7]. L. Ertaul and N. Chavan, Security of ad

betterment in the delivery of packet before

make

conference

74

clustering

techniques may be introduces.

[8]. M. A. Moustafa, M. A. Youssef, and M. N.

6. REFERENCES:

El-Derini, MSR: A multipath secure reliable

[1].S. Mueller, R. Tsang, and D. Ghosal,

routing protocol for WSNs, in Computer

Multipath routing in mo-bile ad hoc networks:

Systems and Applications (AICCSA), 2011

Issues and challenges, in In Performance

9th IEEE/ACS Interna-tional Conference on,

Tools

December 2011, pp. 54 59.

and

Applications

to

Networked

Systems, volume 2965 of LNCS. Springer-

[9]. D. Ganesan, R. Govindan, S. Shenker,

Verlag, 2004, pp. 209234.


[2].Pavan

kumar

T,

Ramesh

Babu

and

B,

networks,

based routing protocol for cognitive radio

wireless

energy-

SIGMOBILE

Mob.

Comput.

2001.

[3].Ann lee and Paul A.S Ward A study of


in

Highly-resilient,

Commun. Rev., vol. 5, pp. 1125, October

wireless mesh networks.


algorithms

Estrin,

efficient multipath routing in wireless sensor

Rajasekhar rao K, Dinesh Gopalni cluster

routing

D.

mesh

[10]. J. R. Douceur, The Sybil attack, in

networks.

Peer-to-Peer Systems, First International

[4]. Nikos Dimokas, Dimitrios Katsaris, Yannis

Workshop, IPTPS 2002, Cambridge, MA,

Manolopoulos.Node clustering in wireless

USA, March 7-8, 2002, Revised Papers.

sensor networks by considering structural

Springer, 2002, pp. 251260.

characteristics of the network graph.


[5].Jaydip Sen. Innovation labs Security and
privacy issues in wireless mesh networks.

INTERNATIONAL ASSOCIATION OF ENGINEERING & TECHNOLOGY FOR SKILL DEVELOPMENT

359

www.iaetsd.in

You might also like