63
Exploiting Dynamic Resource Allocation for Efficient Parallel Data Processing in the Cloud A PROJECT REPORT Submitted by KARTHIKA.RN 21708205038 LAKSHMIPRIYA ANBU 21708205042 NAGA BHARGAVI.CH 21708205050 in partial fulfillment for the award of the degree of BACHELOR OF TECHNOLOGY In INFORMATION TECHNOLOGY

Cloud Proj

Embed Size (px)

Citation preview

Page 1: Cloud Proj

Exploiting Dynamic Resource Allocation for Efficient

Parallel Data Processing in the Cloud

A PROJECT REPORT

Submitted by

KARTHIKA.RN 21708205038

LAKSHMIPRIYA ANBU 21708205042

NAGA BHARGAVI.CH 21708205050

in partial fulfillment for the award of the degree

of

BACHELOR OF TECHNOLOGYIn

INFORMATION TECHNOLOGY

SREE SASTHA INSTITUTE OF ENGINEERING

AND TECHNOLOGY

CHEMBARAMBAKKAM ,CHENNAI-600123

ANNA UNIVERSITY: CHENNAI-600 025 APRIL-2012

Page 2: Cloud Proj

ANNA UNIVERSITY: CHENNAI 600 025BONAFIDE CERTIFICATE

Certified that this project report “Exploiting Dynamic Resource

Allocation for Efficient Parallel Data Processing in the

Cloud” is the bonafide work of

“RN.KARTHIKA” ,“LAKSHMI PRIYA

ANBU” ,“CH.NAGA BHARGAVI” who carried out the project

work under my supervision.

SIGNATURE SIGNATURE

S.MUHAMAD IBRAHIM A.RENGARAJAN

BADHUSHA

SUPERVISOR HEAD OF THE DEPARTMENT

Department of Information Department of Information

Technology Technology

Sree Sastha Institute Of Sree Sastha Institute Of

Engineering and technology Engineering and technology

Chembarambakkam, Chembarambakkam,

Chennai-600 123 Chennai-600 123

Page 3: Cloud Proj

ACKNOWLEDGEMENT

Authentic thanks to Our Ethical Founder & Chairman

Thiru Dr. “A.M.K. JAMBULINGA MUDALLIAR”, for his genuine venture

in educating us in his esteemed institution.

I would like to express my heartfelt thanks to

,Principal of our college for his kind endorsement and inspiration.

We are very obliged to thank Prof Mr. “A.RENGARAJAN”, Assistant

Professor and HOD Department Of Information And Technology ,for his

valuable counseling and evocation during the course. We are profound to him

for having reassuring us.

We are copiously beholden to our Project Co-ordinator

Mr“S.MUHAMAD IBRAHIM BADHUSHA”,LECTURER,Department Of

Information And Technology for his sustained mentorship and benefaction to

take up this project and for his zeal shown towards this project

We are gratified to express our devout owing to our Project Guide

Mr.“S.MUHAMAD IBRAHIM BADHUSHA”,Lecturer in Department Of

Information And Technology for his valuable guidance and extensive efforts

offered in making the project success.

We affirm our thanks to all our department staff members for the

unceasing benefaction and inducement for the efficacious accomplishment of

our project.

We oblige our thanks to our library staff and management for their

extensive support by providing the information and resources that helped us to

complete the project successfully

Page 4: Cloud Proj

TABLE OF CONTENTS

CHAPTER NO. TITLE PAGE NO.

List of Tables

List of Figures

List of Abbreviations

1 Introduction

1.1.1 Objective

2 System Analysis

2.1 Existing System

2.1.1 Drawbacks

2.2 Proposed System

2.2.1 Advantages

2.3 Feasibility Study

2.3.1 Economical Feasibility

2.3.2 Operational Feasibility

2.3.3 Technical Feasibility

3 System Specification

3.1 Hardware Requirements

3.2 Software Requirements

Page 5: Cloud Proj

4 Software Description

4.1 Front End

4.2 Features

5 Project Description

5.1 Problem Definition

5.2 Overview of the Project

5.3 Module Description

5.3.1 Modules

5.4 Data Flow Diagram

5.5 E-R Diagram

5.6 Database Design

5.6.1 Table 1

5.6.2 Table 2

5.7 Input Design

5.8 Output Design

6 System Testing

6.1 Unit Testing

6.2 Acceptance Testing

6.3 Test Cases

7 System Implementation

Page 6: Cloud Proj

8 Conclusion & Future Enhancements

8.1 Conclusion

8.2 Future Enhancements

9 Appendix

9.1 Source Code

9.2 Screen Shots

10 References

Page 7: Cloud Proj

LIST OF FIGURES

FIGURE NO FIGURE NAME PAGENO

1.0 FEATURES OF .NET

1.1 THE .NET FRAMEWORK

1.2 LANGUAGES SUPPORTED BY .NET

1.3 OBJECTIVES OF. NET FRAMEWORK

1.4 FEATURES OF SQL-SERVER

Page 8: Cloud Proj

1.0 FEATURES OF .NET:

Microsoft .NET is a set of Microsoft software technologies for

rapidly building and integrating XML Web services, Microsoft Windows-based

applications, and Web solutions. The .NET Framework is a language-neutral

platform for writing programs that can easily and securely interoperate. There’s

no language barrier with .NET: there are numerous languages available to the

developer including Managed C++, C#, Visual Basic and Java Script. The .NET

framework provides the foundation for components to interact seamlessly,

whether locally or remotely on different platforms. It standardizes common data

types and communications protocols so that components created in different

languages can easily interoperate. “.NET” is also the collective name given to

various software components built upon the .NET platform. These will be both

products (Visual Studio.NET and Windows.NET Server, for instance) and

services (like Passport, .NET My Services, and so on).

1.1.THE .NET FRAME WORK :

The .NET Framework has two main parts:

1. The Common Language Runtime (CLR).

2. A hierarchical set of class libraries.

The CLR is described as the “execution engine” of .NET. It provides the

environment within which programs run. The most important features are

Conversion from a low-level assembler-style language, called

Intermediate Language (IL), into code native to the platform being

executed on.

Memory management, notably including garbage collection.

Checking and enforcing security restrictions on the running code.

Page 9: Cloud Proj

Loading and executing programs, with version control and other

such features.

The following features of the .NET framework are also worth

description:

Managed Code

The code that targets .NET, and which contains certain extra Information

- “metadata” - to describe itself. Whilst both managed and unmanaged code can

run in the runtime, only managed code contains the information that allows the

CLR to guarantee, for instance, safe execution and interoperability.

Managed Data

With Managed Code comes Managed Data. CLR provides

memory allocation and Deal location facilities, and garbage collection.

Some .NET languages use Managed Data by default, such as C#, Visual

Basic.NET and JScript.NET, whereas others, namely C++, do not. Targeting

CLR can, depending on the language you’re using, impose certain constraints

on the features available. As with managed and unmanaged code, one can have

both managed and unmanaged data in .NET applications - data that doesn’t get

garbage collected but instead is looked after by unmanaged code.

Common Type System

The CLR uses something called the Common Type System (CTS) to

strictly enforce type-safety. This ensures that all classes are compatible with

each other, by describing types in a common way. CTS define how types work

within the runtime, which enables types in one language to interoperate with

types in another language, including cross-language exception handling. As

well as ensuring that types are only used in appropriate ways, the runtime also

Page 10: Cloud Proj

ensures that code doesn’t attempt to access memory that hasn’t been allocated to

it.

Common Language Specification :

The CLR provides built-in support for language interoperability. To

ensure that you can develop managed code that can be fully used by developers

using any programming language, a set of language features and rules for using

them called the Common Language Specification (CLS) has been defined.

Components that follow these rules and expose only CLS features are

considered CLS-compliant.

The class library

.NET provides a single-rooted hierarchy of classes, containing over 7000

types. The root of the namespace is called System; this contains basic types like

Byte, Double, Boolean, and String, as well as Object. All objects derive from

System. Object. As well as objects, there are value types. Value types can be

allocated on the stack, which can provide useful flexibility. There are also

efficient means of converting value types to object types if and when necessary.

The set of classes is pretty comprehensive, providing collections, file,

screen, and network I/O, threading, and so on, as well as XML and database

connectivity.

The class library is subdivided into a number of sets (or

namespaces), each providing distinct areas of functionality, with dependencies

between the namespaces kept to a minimum.

Page 11: Cloud Proj

1.2 LANGUAGES SUPPORTED BY .NET

The multi-language capability of the .NET Framework and Visual

Studio .NET enables developers to use their existing programming skills to

build all types of applications and XML Web services. The .NET framework

supports new versions of Microsoft’s old favorites Visual Basic and C++ (as

VB.NET and Managed C++), but there are also a number of new additions to

the family.

Visual Basic .NET has been updated to include many new and

improved language features that make it a powerful object-oriented

programming language. These features include inheritance, interfaces, and

overloading, among others. Visual Basic also now supports structured exception

handling, custom attributes and also supports multi-threading.

Visual Basic .NET is also CLS compliant, which means that any

CLS-compliant language can use the classes, objects, and components you

create in Visual Basic .NET.

Managed Extensions for C++ and attributed programming are just

some of the enhancements made to the C++ language. Managed Extensions

simplify the task of migrating existing C++ applications to the new .NET

Framework.

C# is Microsoft’s new language. It’s a C-style language that is

essentially “C++ for Rapid Application Development”. Unlike other languages,

its specification is just the grammar of the language. It has no standard library

of its own, and instead has been designed with the intention of using the .NET

libraries as its own.

Microsoft Visual J# .NET provides the easiest transition for Java-

language developers into the world of XML Web Services and dramatically

improves the interoperability of Java-language programs with existing software

written in a variety of other programming languages.

Page 12: Cloud Proj

Active State has created Visual Perl and Visual Python, which

enable .NET-aware applications to be built in either Perl or Python. Both

products can be integrated into the Visual Studio .NET environment. Visual

Perl includes support for Active State’s Perl Dev Kit.

Other languages for which .NET compilers are available include

FORTRAN

COBOL

Eiffel

Fig1 .Net Framework

ASP.NET

XML WEB SERVICES

Windows Forms

Base

Class Libraries

Common

Language Runtime

Operating System

C#.NET is also compliant with CLS (Common Language Specification) and

supports structured exception handling. CLS is set of rules and constructs that

are supported by the CLR (Common Language Runtime). CLR is the runtime

environment provided by the .NET Framework; it manages the execution of the

Page 13: Cloud Proj

code and also makes the development process easier by providing services.

C#.NET is a CLS-compliant language. Any objects, classes, or components that

created in C#.NET can be used in any other CLS-compliant language. In

addition, we can use objects, classes, and components created in other CLS-

compliant languages in C#.NET .The use of CLS ensures complete

interoperability among applications, regardless of the languages used to create

the application.

Constructors and destructors:

Constructors are used to initialize objects, whereas destructors are used

to destroy them. In other words, destructors are used to release the resources

allocated to the object. In C#.NET the sub finalize procedure is available. The

sub finalize procedure is used to complete the tasks that must be performed

when an object is destroyed. The sub finalize procedure is called automatically

when an object is destroyed. In addition, the sub finalize procedure can be

called only from the class it belongs to or from derived classes.

Garbage collection

Garbage Collection is another new feature in C#.NET. The .NET Framework

monitors allocated resources, such as objects and variables. In addition,

the .NET Framework automatically releases memory for reuse by destroying

objects that are no longer in use.

In C#.NET, the garbage collector checks for the objects that are not currently in

use by applications. When the garbage collector comes across an object that is

marked for garbage collection, it releases the memory occupied by the object.

Overloading

Overloading is another feature in C#. Overloading enables us to define multiple

procedures with the same name, where each procedure has a different set of

arguments. Besides using overloading for procedures, we can use it for

constructors and properties in a class.

Page 14: Cloud Proj

Multi threading:

C#.NET also supports multithreading. An application that supports

multithreading can handle multiple tasks simultaneously, we can use

multithreading to decrease the time taken by an application to respond to user

interaction.

Structured exception handling

C#.NET supports structured handling, which enables us to detect

and remove errors at runtime. In C#.NET, we need to use Try…Catch…Finally

statements to create exception handlers. Using Try…Catch…Finally statements,

we can create robust and effective exception handlers to improve the

performance of our application.

1.3 OBJECTIVES OF. NET FRAMEWORK

1. To provide a consistent object-oriented programming environment whether

object codes is stored and executed locally on Internet-distributed, or executed

remotely.

2. To provide a code-execution environment to minimizes software deployment

and guarantees safe execution of code.

3. Eliminates the performance problems.

There are different types of application, such as Windows-based applications

and Web-based applications.

Page 15: Cloud Proj

1.4 FEATURES OF SQL-SERVER

The OLAP Services feature available in SQL Server version 7.0 is

now called SQL Server 2000 Analysis Services. The term OLAP Services has

been replaced with the term Analysis Services. Analysis Services also includes

a new data mining component. The Repository component available in SQL

Server version 7.0 is now called Microsoft SQL Server 2000 Meta Data

Services. References to the component now use the term Meta Data Services.

The term repository is used only in reference to the repository engine within

Meta Data Services

SQL-SERVER database consist of six type of objects,

They are,

1. TABLE

2. QUERY

3. FORM

4. REPORT

5. MACRO

Table:

A database is a collection of data about a specific topic.

Views of table:

We can work with a table in two types,

1. Design View

2. Datasheet View

Design view

To build or modify the structure of a table we work in the

table design view. We can specify what kind of data will be hold.

Page 16: Cloud Proj

Datasheet view

To add, edit or analyses the data itself we work in tables

datasheet view mode.

Query:

A query is a question that has to be asked the data. Access gathers data

that answers the question from one or more table. The data that make up the

answer is either dynaset (if you edit it) or a snapshot (it cannot be edited).Each

time we run query, we get latest information in the dynaset. Access either

displays the dynaset or snapshot for us to view or perform an action on it, such

as deleting or updating.

Page 17: Cloud Proj

LIST OF ABBREVIATIONS

BIR Big Integer Reader

BIS Big Integer Sorter

DAG Directed Acyclic Graph.

EI Execution Instance

EV Execution vertices

HDFS Hadoop Distributed File System

HTC High-Throughput Computing .

IAAS Infrastructure-as-a-Service

JM Job Manager

JMX Java Management Extensions

MTC Many Task Computing

PACT Parallel Architectures and Compilation Techniques

TSM TeraSortMap

TM Task Manager

TSR TeraSortReduce

Page 18: Cloud Proj

VMs Virtual Machines

CHAPTER 1

INTRODUCTION

The Cloud computing paradigm is sometimes viewed as the

commercial successor of the academic Grid computing paradigm.Cloud

computing as offered by companies such as Amazon is easier to use than the

typical Grid and offers more user rights and tighter security.This has led to

much greater commercial adoption of Cloud computing than of Grid

computing.However,unlike aGrid,a Cloud doesn’t not assist its users with the

selection and use of different remote Cloud sites.Due to commercial interests,a

Cloud provider usually has a proprietary system where a user creates an image

that only runs on the provider’s site.This Represents a significant step back

compared to grid computing.While many cross-site technologies applied in grid

computing such as virtual organization management and single sign-on can

quite easily be adopted by Cloud computing,the acyual configuration and

management of user software in the cloud is significantly different compared to

the traditional Grid.

In The Cloud,Users expect to be able to install complex software

With super-user privileges on-demand.If the cloud paradigm is to be extended

to facilitate cross-provider utilization of resources,several challenges need to be

solved.One of the requirements for Cross-Cloud computing discussed in this

paper is the capability to create and efficiently distribute interoperable virtual

machines.First,We resent a Cross-Cloud virtual machine creation solution that

is based on a layered approach to allow a single user image to be deployed to

multiple cloud sites,including desktop Cloud sites.Next,we analyse methods for

Page 19: Cloud Proj

the bulk transefer of these layered images to the cloud sites.Finally,We present a

study of the performance of our system with and without security.

1.1 OBJECTIVE:

For efficient parallel data processing in cloud environments

and presented Nephele framework, the first data processing framework to

exploit the dynamic resource provisioning offered by today’s IaaS clouds.

ABSTRACT:

     In recent years ad-hoc parallel data processing has emerged to be one of the

killer applications for Infrastructure-as-a-Service (IaaS) clouds. Major Cloud

computing companies have started to integrate frameworks for parallel data

processing in their product portfolio, making it easy for customers to access

these services and to deploy their programs. However, the processing

frameworks which are currently used have been designed for static,

homogeneous cluster setups and disregard the particular nature of a cloud.

Consequently, the allocated compute resources may be inadequate for big parts

of the submitted job and unnecessarily increase processing time and cost. In this

paper we discuss the opportunities and challenges for efficient parallel data

processing in clouds and present our research project. It is the first data

processing framework to explicitly exploit the dynamic resource allocation

offered by today’s IaaS clouds for both, task scheduling and execution.

Particular tasks of a processing job can be assigned to different types of virtual

machines which are automatically instantiated and terminated during the job

execution.

Page 20: Cloud Proj

CHAPTER 2

SYSTEM ANALYSIS

2.1 Existing System

A growing number of companies have to process huge amounts of

data in a cost-efficient manner. Classic representatives for these companies are

operators of Internet search engines. The vast amount of data they have to deal

with every day has made traditional database solutions prohibitively

expensive .Instead, these companies have popularized an architectural paradigm

based on a large number of commodity servers. Problems like processing

crawled documents or regenerating a web index are split into several

independent subtasks, distributed among the available nodes, and computed in

parallel.

2.1.1 Disadvantages of Existing System:

The cloud’s virtualized nature helps to enable promising new use

cases for efficient parallel data processing. ,However, it also imposes

new challenges compared to classic cluster setups.

The major challenge we see is the cloud’s opaqueness with prospect

to exploiting data locality.

Page 21: Cloud Proj

2.2 Proposed System: 

In recent years a variety of systems to facilitate MTC has been developed.

Although these systems typically share common goals (e.g. to hide issues of

parallelism or fault tolerance), they aim at different fields of application. Map

Reduce is designed to run data analysis jobs on a large amount of data, which is

expected to be stored across a large set of share-nothing commodity servers.

Once a user has fit his program into the required map and reduce pattern, the

execution framework takes care of splitting the job into subtasks, distributing

and executing them. A single Map Reduce job always consists of a distinct map

and reduce program.

2.2.1 Advantages of Proposed System:

The first data processing framework to exploit the dynamic resource

provisioning offered by today’s IaaS clouds .The performance evaluation gives

a first impression on how the ability to assign specific virtual machine types to

specific tasks of a processing job, as well as the possibility to automatically

allocate/deallocate virtual machines in the course of a job execution, can help to

improve the overall resource utilization and, consequently, reduce the

processing cost.

Page 22: Cloud Proj

2.3 Feasibility Study:

Feasibility studies aim to objectively and rationally uncover the

strengths and weaknesses of the existing business or proposed venture,

opportunities and threats as presented by the environment, the resources

required to carry through, and ultimately the prospects forsuccess.In its simplest

term, the two criteria to judge feasibility are cost required and value to be

attained. As such, a well-designed feasibility study should provide a historical

background of the business or project, description of the product or service,

accounting statements, details of the operations and management, marketing

research and policies, financial data, legal requirements and tax

obligations.Generally, feasibility studies precede technical development and

project implementation

2.3.1 Economical Feasibility:

Economic analysis is the most frequently used method for evaluating the

effectiveness of a new system. More commonly known as cost/benefit analysis,

the procedure is to determine the benefits and savings that are expected from a

candidate system and compare them with costs. If benefits outweigh costs, then

the decision is made to design and implement the system. An entrepreneur must

accurately weigh the cost versus benefits before taking an action.

Cost-based study: It is important to identify cost and benefit factors, which can

be categorized as follows: 1. Development costs; and 2. Operating costs. This is

Page 23: Cloud Proj

an analysis of the costs to be incurred in the system and the benefits derivable

out of the system.

Time-based study: This is an analysis of the time required to achieve a return on

investments. The future value of a project is also a factor.

2.3.2 Operational Feasibility:

Operational feasibility is a measure of how well a proposed system solves

the problems, and takes advantage of the opportunities identified during scope

definition and how it satisfies the requirements identified in the requirements

analysis phase of system development

2.3.3 Technical Feasibility:

The assessment is based on an outline design of system requirements in

terms of Input, Processes, Output, Fields, Programs, and Procedures. This can

be quantified in terms of volumes of data, trends, frequency of updating, etc. in

order to estimate whether the new system will perform adequately or not.

Technological feasibility is carried out to determine whether the company has

the capability, in terms of software, hardware, personnel and expertise, to

handle the completion of the project. When writing a feasibility report the

following should be taken to consideration:

A brief description of the business to assess more possible factor/s which

could affect the study

The part of the business being examined

The human and economic factor

The possible solutions to the problems

Page 24: Cloud Proj

At this level, the concern is whether the proposal is

both technically and legally feasible (assuming moderate cost).

CHAPTER 3

SYSTEM SPECIFICATION

HARDWARE REQUIREMENTS: 

System    :   Pentium IV 2.4 GHz.

Hard Disk  :   40 GB.

Floppy Drive :   1.44 Mb.

Monitor   :   15 VGA Colour.

Mouse    :   Logitech.

Ram    :   512 MB.

SOFTWARE REQUIREMENTS: 

Operating system   : Windows XP Professional.

Coding Language  : ASP .Net,C#

Database : Sql Server 2005.

Page 25: Cloud Proj

CHAPTER 4

SOFTWARE DESCRIPTION

4.1 Front End:

ASP.NET is a Web application framework developed and marketed by

Microsoft to allow programmers to build dynamic Web sites, Web applications

and Web services. It was first released in January 2002 with version 1.0 of

the .NET Framework, and is the successor to Microsoft's Active Server Pages

(ASP) technology. ASP.NET is built on the Common Language Runtime

(CLR), allowing programmers to write ASP.NET code using any

supported .NET language. The ASP.NET SOAP extension framework allows

ASP.NET components to process SOAP messages.

4.2 Features Of .Net :

Microsoft .NET is a set of Microsoft software technologies for rapidly

building and integrating XML Web services, Microsoft Windows-based

applications, and Web solutions. The .NET Framework is a language-neutral

platform for writing programs that can easily and securely interoperate. There’s

no language barrier with .NET: there are numerous languages available to the

developer including Managed C++, C#, Visual Basic and Java Script. The .NET

framework provides the foundation for components to interact seamlessly,

whether locally or remotely on different platforms. It standardizes common data

types and communications protocols so that components created in different

Page 26: Cloud Proj

languages can easily interoperate.“.NET” is also the collective name given to

various software components built upon the .NET platform. These will be both

products (Visual Studio.NET and Windows.NET Server, for instance) and

services (like Passport, .NET My Services, and so on).

CHAPTER 5

PROJECT DESCRIPTION

5.1 Problem Definition:

5.2 Overview of the Project:

Abstract

Existing System

Proposed System

Hardware Requirements

Software Requirements

The Nephele Execution Engine

Nephele Architecture

Demonstration of an Example Task

5.3 Module Description :

5.3.1 . NETWORK MODULE:

       Server - Client computing or networking is a distributed application

architecture that partitions tasks or workloads between service providers

(servers) and service requesters, called clients. Often clients and servers operate

over a computer network on separate hardware. A server machine is a high-

Page 27: Cloud Proj

performance host that is running one or more server programs which share its

resources with clients. A client also shares any of its resources; Clients therefore

initiate communication sessions with servers which await (listen to) incoming

requests.

5.3.2. LBS SERVICES:

       In particular, users are reluctant to use LBSs, since revealing their

position may link to their identity. Even though a user may create a fake ID to

access the service, her location alone may disclose her actual identity. Linking a

position to an individual is possible by various means, such as publicly

available information city maps. When a user u wishes to pose a query, she

sends her location to a trusted server, the anonymizer through a secure

connection (SSL). The latter obfuscates her location, replacing it with an

anonymizing spatial region (ASR) that encloses u. The ASR is then forwarded

to the LS. Ignoring where exactly u is, the LS retrieves (and reports to the AZ) a

candidate set (CS) that is guaranteed to contain the query results for any

possible user location inside the ASR. The AZ receives the CS and reports to u

the subset of candidates that corresponds to her original query. 

5.3.3. SYSTEM MODEL: 

       The ASR construction at the  anonymization process abides by the

user’s privacy requirements. Particularly, specified an anonymity degree K by

u, the ASR satisfies two properties: (i) it contains u and at least another K * 1

users, and (ii) even if the LS knew the exact locations of all users in the system.

We propose an edge ordering anonymization approach for users in road

networks, which guarantees K-anonymity under the strict reciprocity

requirement (described later).

Page 28: Cloud Proj

We identify the crucial concept of border nodes, an important indicator of

the CS size and of the query processing cost at the LS.

We consider various edge orderings, and qualitatively assess their query

performance based on border nodes.

We design efficient query processing mechanisms that exploit existing

network database infrastructure, and guarantee CS inclusiveness and

minimality. Furthermore, they apply to various network storage schemes.

We devise batch execution techniques for anonymous queries that

significantly reduce the overhead of the LS by computation sharing.

5.3. 4. SCHEDULED TASK:

       Recently, considerable research interest has focused on preventing

identity inference in location-based services. Proposing spatial cloaking

techniques. In the following, we describe existing techniques for ASR

computation (at the AZ) and query processing (at the LS). At the end, we cover

alternative location privacy approaches and discuss why they are inappropriate

to our problem setting. This offers privacy protection in the sense that the actual

user position u cannot be distinguished from others in the ASR, even when

malicious LS is equipped/advanced enough to possess all user locations. This

spatial K-anonymity model is most widely used in location privacy

research/applications, even though alternative models are emerging. 

5.3.5. QUERY PROCESSING: 

       Processing is based on implementation of the theorem uses

(network-based) search operations as off the shelf building blocks. Thus, the

NAP query evaluation methodology is readily deployable on existing systems,

and can be easily adapted to different network storage schemes. In this case, the

queries are evaluated in a batch. we propose the network-based anonymization

Page 29: Cloud Proj

and processing (NAP) framework, the first system for K- anonymous query

processing in road networks. NAP relies on a global user ordering and

bucketization that satisfies reciprocity and guarantees K-anonymity. We

identify the ordering characteristics that affect subsequent processing, and

qualitatively compare alternatives. Then, we propose query evaluation

techniques that exploit these characteristics. In addition to user privacy, NAP

achieves low computational and communication costs, and quick responses

overall. It is readily deployable, requiring only basic network operations. 

5.4 Data Flow Diagram:

Join Request       Nodes

Address &routing  Un addressed dataTable

Network Tree

Address

Nodes

Cache

Routing cache Network Path

                               New Node

ReOrganizer

Address Resolver

Routing system

Page 30: Cloud Proj

5.5 E-R Diagram:

5.6 Database Design:

5.6.1 Table 15.6.2 Table 2

User Mode of transport

Personal information

Location

Gives

Pin      Area

Distance

    Flight

    Train

License id

Bus Personal travel

Name Phone no

Email id

Page 31: Cloud Proj

5.7 Input Design5.8 Output Design

CHAPTER 6

SYSTEM TESTING

6.1 Unit Testing:

Unit testing is conducted to verify the functional performance of

each modular component of the software.Unit testing focuses on the smallest

unit of the software design (i.e), the module. The white-box testing techniques

were heavily employed for unit testing.

6.2 Acceptance Testing:

Acceptance testing is a term used in agile software development

methodologies, particularly Extreme Programming, referring to the functional

testing of a user story by the software development team during the

implementation phase.

The customer specifies scenarios to test when a user story has been correctly

implemented. A story can have one or many acceptance tests, whatever it takes

to ensure the functionality works. Acceptance tests are black box system tests.

Each acceptance test represents some expected result from the system.

Customers are responsible for verifying the correctness of the acceptance tests

and reviewing test scores to decide which failed tests are of highest priority.

Acceptance tests are also used as regression tests prior to a production release.

A user story is not considered complete until it has passed its acceptance tests.

This means that new acceptance tests must be created for each iteration or the

development team will report zero progress

Page 32: Cloud Proj

6.3 Test Cases:

A test case in software engineering is a set of conditions or

variables under which a tester will determine whether an application

or software system  is working correctly or not. The mechanism for determining

whether a software program or system has passed or failed such a test is known

as a test oracle. In some settings, an oracle could be a requirement or use case,

while in others it could be a heuristic. It may take many test cases to determine

that a software program or system is considered sufficiently scrutinized to be

released. Test cases are often referred to as test scripts, particularly when

written. Written test cases are usually collected into test suites.

Page 33: Cloud Proj

CHAPTER 7

SYSTEM IMPLEMENTATION

Implementation is the stage of the project when the theortical

design is turned out into a working system.Thus it can be considered to be the

most critical stage in achieving a successful new system and in giving the

user,confidence thtat the new system will work and be effective.

The Implementation stage involves carefull planning,investigation

of the existing system and it’s constraints on implementation,designing of

methods to achieve change over and evaluation of change over methods.

A hierarchial structuring of relations may result in more classes and a

more complicated structure to implement.Therefore it is advisable to transform

the hierarchial relation structure to a simpler structure such as a classical flat

one.It is ratherstraight forward to transform the developed hierarchial model

into a bipartite,flat model,consisting of classes on the one hand and flat relations

on the other.

Flat relations are preferred at the design level for reasons of simplicity

and implementation ease.there is no identity or functionality associated with a

flat relation.A flat relation corresponds with the relation concept of entity-

relationship modeling and many object oriented methods

Page 34: Cloud Proj

CHAPTER 8

CONCLUSION AND FUTURE ENHANCEMENT

8.1 CONCLUSIONS:

For efficient parallel data processing incloud environments and

presented Nephele, the first data processing framework to exploit the dynamic

resource provisioning offered by today’s IaaS clouds. Wehave described

Nephele’s basic architecture and presented a performance comparison to the

well-established data processing framework Hadoop. The performance

evaluation gives a first impression on how the ability to assign specific virtual

machine types to specific tasks of a processing job, as well as the possibility to

automatically allocate/deallocate virtual machines in the course of a job

execution, can help to improve the overall resource utilization

and,consequently, reduce the processing cost.

8.2 FUTURE ENHANCEMENT:

With a framework like Nephele at hand, there are a variety

of open research issues, which we plan to address for future work. In particular,

we are interested in improving Nephele’s ability to adapt to resource overload

or underutilization during the job execution automatically. Our current profiling

approach builds a valuable basis for this, however, at the moment the system

still requires a reasonable amount of user annotations.

Page 35: Cloud Proj

CHAPTER 9

APPENDIX

9.1 Source Code:

Page 36: Cloud Proj

9.2 Screen Shots:

Page 37: Cloud Proj
Page 38: Cloud Proj
Page 39: Cloud Proj
Page 40: Cloud Proj
Page 41: Cloud Proj
Page 42: Cloud Proj
Page 43: Cloud Proj
Page 44: Cloud Proj
Page 45: Cloud Proj
Page 46: Cloud Proj
Page 47: Cloud Proj