Professional Documents
Culture Documents
***********Course Name*************
Done By
**************
******College Name******
External
ACKNOWLEDGEMENT
At every outset I express my gratitude to almighty lord for showering his grace and blessings
upon me to complete this project.
Although our name appears on the cover of this book, many people had contributed in some form
or the other form to this project Development. We could not done this project without the
assistance or support of each of the following we thank you all.
I wish to place on my record my deep sense of gratitude to my project guide, Mr. ******, *****
Software Solutions, for his constant motivation and valuable help through the project work.
Express my gratitude to Mr. ******, Director of ******* Institute of Management & Computer
Sciences for his valuable suggestions and advices throughout the ***** course. I also extend my
thanks to other Faculties for their Cooperation during my Course.
Finally I would like to thank my friends for their cooperation to complete this project.
ABSTRACT
Conventional spatial queries, such as range search and nearest neighbour retrieval, involve only
conditions on objects’ geometric properties. Today, many modern applications call for novel
forms of queries that aim to find objects satisfying both a spatial predicate, and a predicate on
their associated texts. For example, instead of considering all the restaurants, a nearest neighbour
query would instead ask for the restaurant that is the closest among those whose menus contain
“steak, spaghetti, brandy” all at the same time. Currently, the best solution to such queries is
based on the IR2 -tree, which, as shown in this paper, has a few deficiencies that seriously
impact its efficiency. Motivated by this, we develop a new access method called the spatial
inverted index that extends the conventional inverted index to cope with multidimensional data,
and comes with algorithms that can answer nearest neighbour queries with keywords in real
time. As verified by experiments, the proposed techniques outperform the IR2 -tree in query
response time significantly, often by a factor of orders of magnitude.
INDEX
CHAPTERS Page No:
1 INTRODUCTION
2 LITERATURE SURVEY
2.1 INTRODUCTION
2.2 DEFINITIONS
2.3 CHARACTERISTICS
2.4 PROBLEMS ASSOCIATED
2.5 TECHNIQUES AND ALGORITHMS APPLICABLE
2.6 OTHER SUGGESTED SOLUTIONS
2.7 APPLICATIONS
4. SYSTEM DESIGN
5. TESTING
6. OUTPUT SCREENS( NOT MORE THAN 10 PAGES( AT MOST TWO SCREENS PER
PAGE))
APPENDIX
REFERENCES
LIST OF TABLES
CurLsr Table
CurNu Table
LIST OF FIGURES
Generally, data mining (sometimes called data or knowledge discovery) is the process of
analyzing data from different perspectives and summarizing it into useful information -
information that can be used to increase revenue, cuts costs, or both. Data mining software is one
of a number of analytical tools for analyzing data. It allows users to analyze data from many
different dimensions or angles, categorize it, and summarize the relationships identified.
Technically, data mining is the process of finding correlations or patterns among dozens of fields
in large relational databases.
While large-scale information technology has been evolving separate transaction and analytical
systems, data mining provides the link between the two. Data mining software analyzes
relationships and patterns in stored transaction data based on open-ended user queries. Several
types of analytical software are available: statistical, machine learning, and neural
networks.Generally, any of four types of relationships are sought:
Classes: Stored data is used to locate data in predetermined groups. For example, a
restaurant chain could mine customer purchase data to determine when customers visit
and what they typically order. This information could be used to increase traffic by
having daily specials.
Sequential patterns: Data is mined to anticipate behavior patterns and trends. For
example, an outdoor equipment retailer could predict the likelihood of a backpack being
purchased based on a consumer's purchase of sleeping bags and hiking shoes.
1) Extract, transform, and load transaction data onto the data warehouse system.
2) Store and manage the data in a multidimensional database system.
3) Provide data access to business analysts and information technology professionals.
4) Analyze the data by application software.
5) Present the data in a useful format, such as a graph or table.
Artificial neural networks: Non-linear predictive models that learn through training and
resemble biological neural networks in structure.
Genetic algorithms: Optimization techniques that use process such as genetic
combination, mutation, and natural selection in a design based on the concepts of natural
evolution.
Decision trees: Tree-shaped structures that represent sets of decisions. These decisions
generate rules for the classification of a dataset. Specific decision tree methods include
Classification and Regression Trees (CART) and Chi Square Automatic Interaction
Detection (CHAID). CART and CHAID are decision tree techniques used for
classification of a dataset. They provide a set of rules that you can apply to a new
(unclassified) dataset to predict which records will have a given outcome. CART
segments a dataset by creating 2-way splits while CHAID segments using chi square
Tests to create multi-way splits. CART typically requires less data preparation than
CHAID.
Nearest neighbor method: A technique that classifies each record in a dataset based on a
combination of the classes of the k record(s) most similar to it in a historical dataset
(where k=1). Sometimes called the k-nearest neighbor technique.
Rule induction: The extraction of useful if-then rules from data based on statistical
significance.
1) It’s one of the most effective services that are available today. With the help of data
mining, one can discover precious information about the customers and their behavior for
a specific set of products and evaluate and analyze, store, mine and load data related to
them
2) An analytical CRM model and strategic business related decisions can be made with the
help of data mining as it helps in providing a complete synopsis of customers
3) An endless number of organizations have installed data mining projects and it has helped
them see their own companies make an unprecedented improvement in their marketing
strategies (Campaigns)
4) Data mining is generally used by organizations with a solid customer focus. For its
flexible nature as far as applicability is concerned is being used vehemently in
applications to foresee crucial data including industry analysis and consumer buying
behaviors
5) Fast paced and prompt access to data along with economic processing techniques have
made data mining one of the most suitable services that a company seek
1. Marketing / Retail:
Data mining helps marketing companies build models based on historical data to predict who
will respond to the new marketing campaigns such as direct mail, online marketing
campaign…etc. Through the results, marketers will have appropriate approach to sell profitable
products to targeted customers.
Data mining brings a lot of benefits to retail companies in the same way as marketing.
Through market basket analysis, a store can have an appropriate production arrangement in a
way that customers can buy frequent buying products together with pleasant. In addition, it also
helps the retail companies offer certain discounts for particular products that will attract more
customers.
2. Finance / Banking
Data mining gives financial institutions information about loan information and credit
reporting. By building a model from historical customer’s data, the bank and financial institution
can determine good and bad loans. In addition, data mining helps banks detect fraudulent credit
card transactions to protect credit card’s owner.
3. Manufacturing
By applying data mining in operational engineering data, manufacturers can detect faulty
equipments and determine optimal control parameters. For example semi-conductor
manufacturers has a challenge that even the conditions of manufacturing environments at
different wafer production plants are similar, the quality of wafer are lot the same and some
for unknown reasons even has defects. Data mining has been applying to determine the
ranges of
control parameters that lead to the production of golden wafer. Then those optimal control
parameters are used to manufacture wafers with desired quality.
4. Governments
Data mining helps government agency by digging and analyzing records of financial
transaction to build patterns that can detect money laundering or criminal activities.
5. Law enforcement:
Data mining can aid law enforcers in identifying criminal suspects as well as apprehending
these criminals by examining trends in location, crime type, habit, and other patterns of
behaviors.
6. Researchers:
Data mining can assist researchers by speeding up their data analyzing process; thus,
allowing those more time to work on other projects.
The real nearest neighbour lies quite far away from the query point, while all the closer
neighbours are missing at least one of the query keywords.
Spatial queries with keywords have not been extensively explored. In the past years, the
community has sparked enthusiasm in studying keyword search in relational databases.
It is until recently that attention was diverted to multidimensional data. The best method
to date for nearest neighbour search with keywords is due to Felipe et al.. They nicely
integrate two well-known concepts: R-tree, a popular spatial index, and signature file, an
effective method for keyword-based document retrieval. By doing so they develop a
structure called the IR2 -tree, which has the strengths of both R-trees and signature files.
Like R-trees, the IR2 - tree preserves objects’ spatial proximity, which is the key to
solving spatial queries efficiently. On the other hand, like signature files, the IR2 -tree is
able to filter a considerable portion of the objects that do not contain all the query
keywords, thus significantly reducing the number of objects to be examined.
1.3 MOTIVATION
Integration tests are designed to test integrated software components to determine if they
actually run as one program. Testing is event driven and is more concerned with the
basic outcome of screens or fields. Integration tests demonstrate that although the
components were individually satisfaction, as shown by successfully unit testing, the
combination of components is correct and consistent. Integration testing is specifically
aimed at exposing the problems that arise from the combination of components.
The real nearest neighbour lies quite far away from the query point, while all the closer
neighbours are missing at least one of the query keywords.
In this paper, we design a variant of inverted index that is optimized for multidimensional
points, and is thus named the spatial inverted index (SI-index). This access method
successfully incorporates point coordinates into a conventional inverted index with small
extra space, owing to a delicate compact storage scheme.
Meanwhile, an SI-index preserves the spatial locality of data points, and comes with an
R-tree built on every inverted list at little space overhead. As a result, it offers two
competing ways for query processing.
We can (sequentially) merge multiple lists very much like merging traditional inverted
lists by ids. Alternatively, we can also leverage the R-trees to browse the points of all
relevant lists in ascending order of their distances to the query point. As demonstrated by
experiments, the SI-index significantly outperforms the IR2 -tree in query efficiency,
often by a factor of orders of magnitude.
1.5 SCOPE
Integration tests are designed to test integrated software components to determine if they
actually run as one program. Testing is event driven and is more concerned with the
basic outcome of screens or fields. Integration tests demonstrate that although the
components were individually satisfaction, as shown by successfully unit testing, the
combination of components is correct and consistent. Integration testing is specifically
aimed at exposing the problems that arise from the combination of components.
Distance browsing is easy with R-trees. In fact, the best-first algorithm is exactly
designed to output data points in ascending order of their distances
MODULES:
System Model
Map View
Distance Search
Neighbor Search
MODULES DESCRIPTION
System Model
In this module a User have to register first, then only he/she has to access the data base.
In this module, any of the above mentioned person have to login, they should login by
giving their email id and password.
In this module Admin registers the location along with its famous place. Also he
measures the distance of the corresponding place from the corresponding source place by
using spatial distance of Google map
It means that the user can give the key in which place that the city/location is famous for
.This results in the list of menu items displayed.
Map View:
The User can see the view of their locality by Google Map (such as map view, satellite
view).
As our goal is to combine keyword search with the existing location-finding services on
facilities such as hospitals, restaurants, hotels, etc., we will focus on dimensionality 2, but
our technique can be extended to arbitrary dimensionalities with no technical obstacle.
Note that the list of each word maintains a sorted order of point ids, which provides
considerable convenience in query processing by allowing an efficient merge step. For
example, assume that we want to find the points that have words c and d. This is
essentially to compute the intersection of the two words’ inverted lists.
Distance Search:
The User can measure the distance and calculate time that takes them to reach the
destination by giving speed. Chart will be prepared by using these values. These are done
by the use of Google Maps.
Traditional nearest neighbor search returns the data point closest to a query point.
We consider that the data set does not fit in memory, and needs to be indexed by efficient
access methods in order to minimize the number of I/Os in answering a query
Neighbor Search
In this module we implement our neighbor Search. The other problem with this
search algorithm is that the indexing information has to be replicated in the broadcast
cycle to enable twice scanning.
The first scan is for deciding the search range, and the second scan is for retrieving k
objects based on the search range.
Therefore, we propose the Nearest Neighbor query approach to improve the
preceding on-air query algorithm.
The system attempts to verify the validity of k objects by processing results obtained
from several peers.
In this module we implement our neighbor Search. The other problem with this
search algorithm is that the indexing information has to be replicated in the broadcast
cycle to enable twice scanning.
1.4 Objective:
Conventional spatial queries, such as range search and nearest neighbour retrieval,Involve only
conditions on objects’ geometric properties. Today, many modern Applications call for novel
forms of queries that aim to find objects satisfying both a spatial predicate, and a predicate on
their associated texts. For example, instead of considering all the restaurants, a nearest
neighbour query would instead ask for the restaurant that is the closest among those whose
menus contain “steak, spaghetti, brandy” all at the same time. Currently, the best solution to
such queries is based on the IR2 -tree, which, as shown in this paper, has a few deficiencies
that Seriously impact its efficiency. Motivated by this, we develop a new access method Called
the spatial inverted index that extends the conventional inverted index to Cope with
multidimensional data, and comes with algorithms that can answer Nearest neighbour queries
with keywords in real time.
LITERATURE SURVEY
We propose two algorithms that compute LkPT queries. Empirical studies with real-world spatial
data demonstrate that LkPT queries are more effective in retrieving web objects than a previous
approach that does not consider the effects of nearby objects; and they show that the proposed
algorithms are scalable and outperform a baseline approach significantly.
Simple
Architecture neutral
Object oriented
Portable
Distributed
High performance
Interpreted
Multithreaded
Robust
Dynamic
Secure
With most programming languages, you either compile or interpret a program so that you
can run it on your computer. The Java programming language is unusual in that a program is
both compiled and interpreted. With the compiler, first you translate a program into an
intermediate language called Java byte codes —the platform-independent codes interpreted by
the interpreter on the Java platform. The interpreter parses and runs each Java byte code
instruction on the computer. Compilation happens just once; interpretation occurs each time the
program is executed. The following figure illustrates how this works.
You can think of Java byte codes as the machine code instructions for the Java Virtual
Machine (Java VM). Every Java interpreter, whether it’s a development tool or a Web browser
that can run applets, is an implementation of the Java VM. Java byte codes help make “write
once, run anywhere” possible. You can compile your program into byte codes on any platform
that has a Java compiler. The byte codes can then be run on any implementation of the Java VM.
That means that as long as a computer has a Java VM, the same program written in the Java
programming language can run on Windows 2000, a Solaris workstation, or on an iMac.
Fig 3.4 Java Complier
From a programming perspective, the beauty of ODBC is that the application can be
written to use the same set of function calls to interface with any data source, regardless of the
database vendor. The source code of the application doesn’t change whether it talks to Oracle or
SQL Server. We only mention these two as an example. There are ODBC drivers available for
several dozen popular database systems. Even Excel spreadsheets and plain text files can be
turned into data sources. The operating system uses the Registry information written by ODBC
Administrator to determine which low-level ODBC drivers are needed to talk to the data source
(such as the interface to Oracle or SQL Server). The loading of the ODBC drivers is transparent
to the ODBC application program. In a client/server environment, the ODBC API even handles
many of the network issues for the application programmer.
The advantages of this scheme are so numerous that you are probably thinking there must
be some catch. The only disadvantage of ODBC is that it isn’t as efficient as talking directly to
the native database interface. ODBC has had many detractors make the charge that it is too slow.
Microsoft has always claimed that the critical factor in performance is the quality of the driver
software that is used. In our humble opinion, this is true. The availability of good ODBC drivers
has improved a great deal recently. And anyway, the criticism about performance is somewhat
analogous to those who said that compilers would never match the speed of pure assembly
language. Maybe not, but the compiler (or ODBC) gives you the opportunity to write cleaner
programs, which means you finish sooner. Meanwhile, computers get faster every year.
JDBC
In an effort to set an independent database standard API for Java; Sun Microsystems
developed Java Database Connectivity, or JDBC. JDBC offers a generic SQL database access
mechanism that provides a consistent interface to a variety of RDBMSs. This consistent interface
is achieved through the use of “plug-in” database connectivity modules, or drivers. If a database
vendor wishes to have JDBC support, he or she must provide the driver for each platform that the
database and Java run on.
To gain a wider acceptance of JDBC, Sun based JDBC’s framework on ODBC. As you
discovered earlier in this chapter, ODBC has widespread support on a variety of platforms.
Basing JDBC on ODBC will allow vendors to bring JDBC drivers to market much faster than
developing a completely new connectivity solution.
JDBC was announced in March of 1996. It was released for a 90 day public review that
ended June 8, 1996. Because of user input, the final JDBC v1.0 specification was released soon
after.
The remainder of this section will cover enough information about JDBC for you to know what it
is about and how to use it effectively. This is by no means a complete overview of JDBC. That
would fill an entire book.
JDBC Goals
Few software packages are designed without goals in mind. JDBC is one that, because of
its many goals, drove the development of the API. These goals, in conjunction with early
reviewer feedback, have finalized the JDBC class library into a solid framework for building
database applications in Java.
The goals that were set for JDBC are important. They will give you some insight as to why
certain classes and functionalities behave the way they do. The eight design goals for JDBC are
as follows:
Simple Architecture-neutral
Object-oriented Portable
Distributed High-performance
Interpreted multithreaded
Robust Dynamic
Secure
Java is also unusual in that each Java program is both compiled and interpreted.
With a compile you translate a Java program into an intermediate language called Java
byte codes the platform-independent code instruction is passed and run on the
computer.
Compilation happens just once; interpretation occurs each time the program is
executed. The figure illustrates how this works.
JavaProgram Interpreter
Compilers My Program
You can think of Java byte codes as the machine code instructions for the Java
Virtual Machine (Java VM). Every Java interpreter, whether it’s a Java development
tool or a Web browser that can run Java applets, is an implementation of the Java VM.
The Java VM can also be implemented in hardware.
Java byte codes help make “write once, run anywhere” possible. You can compile
your Java program into byte codes on my platform that has a Java compiler. The byte
codes can then be run any implementation of the Java VM. For example, the same
Java program can run Windows NT, Solaris, and Macintosh.
Networking
TCP/IP stack
The TCP/IP stack is shorter than the OSI one:
TCP
TCP supplies logic to give a reliable connection-oriented protocol above IP. It
provides a virtual circuit that two processes can use to communicate.
Internet addresses
In order to use a service, you must be able to find it. The Internet uses an address
scheme for machines so that they can be located. The address is a 32 bit integer which
gives the IP address. This encodes a network ID and more addressing. The network ID
falls into various classes according to the size of the network address.
Network address
Class A uses 8 bits for the network address with 24 bits left over for other
addressing. Class B uses 16 bit network addressing. Class C uses 24 bit network
addressing and class D uses all 32.
Subnet address
Internally, the UNIX network is divided into sub networks. Building 11 is currently
on one sub network and uses 10-bit addressing, allowing 1024 different hosts.
Host address
8 bits are finally used for host addresses within our subnet. This places a limit of
256 machines that can be on the subnet.
Total address
Port addresses
A service exists on a host, and is identified by its port. This is a 16 bit number. To
send a message to a server, you send it to the port for that service of the host that it is
running on. This is not location transparency! Certain of these ports are "well known".
Sockets
A socket is a data structure maintained by the system to handle network
connections. A socket is created using the call socket. It returns an integer that is like a
file descriptor. In fact, under Windows, this handle can be used with Read File and
Write File functions.
#include <sys/types.h>
#include <sys/socket.h>
int socket(int family, int type, int protocol);
Here "family" will be AF_INET for IP communications, protocol will be zero, and
type will depend on whether TCP or UDP is used. Two processes wishing to
communicate over a network create a socket each. These are similar to two ends of a
pipe - but the actual pipe does not yet exist.
JFree Chart
JFreeChart is a free 100% Java chart library that makes it easy for developers to
display professional quality charts in their applications. JFreeChart's extensive feature set
includes:
A consistent and well-documented API, supporting a wide range of chart types;
A flexible design that is easy to extend, and targets both server-side and client-
side applications;
Support for many output types, including Swing components, image files
(including PNG and JPEG), and vector graphics file formats (including PDF, EPS and
SVG);
JFreeChart is "open source" or, more specifically, free software. It is distributed
under the terms of the GNU Lesser General Public Licence (LGPL), which permits use in
proprietary applications.
1. Map Visualizations
Charts showing values that relate to geographical areas. Some examples include:
(a) population density in each state of the United States, (b) income per capita for each
country in Europe, (c) life expectancy in each country of the world. The tasks in this
project include:
Sourcing freely redistributable vector outlines for the countries of the world,
states/provinces in particular countries (USA in particular, but also other areas);
Creating an appropriate dataset interface (plus default implementation), a
rendered, and integrating this with the existing XYPlot class in JFreeChart;
Testing, documenting, testing some more, documenting some more.
2. Time Series Chart Interactivity
Implement a new (to JFreeChart) feature for interactive time series charts --- to display a
separate control that shows a small version of ALL the time series data, with a sliding "view"
rectangle that allows you to select the subset of the time series data to display in the main
chart.
3. Dashboards
There is currently a lot of interest in dashboard displays. Create a flexible dashboard
mechanism that supports a subset of JFreeChart chart types (dials, pies, thermometers, bars,
and lines/time series) that can be delivered easily via both Java Web Start and an applet.
4. Property Editors
The property editor mechanism in JFreeChart only handles a small subset of the
properties that can be set for charts. Extend (or reimplement) this mechanism to provide
greater end-user control over the appearance of the charts.
SYSTEM REQUIREMENTS:
HARDWARE REQUIREMENTS:
SOFTWARE REQUIREMENTS:
Input design is a part of overall system design. The main objective during the input
design is as given below:
Data recording
Data transcription
Data conversion
Data verification
Data control
Data transmission
Data validation
Data correction
INPUT TYPES:
It is necessary to determine the various types of inputs. Inputs can be categorized as follows:
INPUT MEDIA:
At this stage choice has to be made about the input media. To conclude about the input
media consideration has to be given to;
Type of input
Flexibility of format
Speed
Accuracy
Verification methods
Rejection rates
Ease of correction
Storage and handling requirements
Security
Easy to use
Portability
Keeping in view the above description of the input types and input media, it can be said
that most of the inputs are of the form of internal and interactive. As
Input data is to be the directly keyed in by the user, the keyboard can be considered to be the
most suitable input device.
ERROR AVOIDANCE
At this stage care is to be taken to ensure that input data remains accurate form the stage
at which it is recorded up to the stage in which the data is accepted by the system. This can be
achieved only by means of careful control each time the data is handled.
ERROR DETECTION
Even though every effort is make to avoid the occurrence of errors, still a small
proportion of errors is always likely to occur, these types of errors can be discovered by using
validations to check the input data.
DATA VALIDATION
Procedures are designed to detect errors in data at a lower level of detail. Data
validations have been included in the system in almost every area where there is a possibility for
the user to commit errors. The system will not accept invalid data. Whenever an invalid data is
keyed in, the system immediately prompts the user and the user has to again key in the data and
the system will accept the data only if the data is correct. Validations have been included where
necessary.
The system is designed to be a user friendly one. In other words the system has been
designed to communicate effectively with the user. The system has been designed with popup
menus.
FEASIBILITY STUDY:
Preliminary investigation examine project feasibility, the likelihood the system will be
useful to the organization. A Project Feasibility Study is an exercise that involves documenting
each of the potential solutions to a particular business problem or opportunity. Feasibility Studies
can be undertaken by any type of business, project or team and they are a critical part of the
Project Life Cycle.
The purpose of a Feasibility Study is to identify the likelihood of one or more solutions
meeting the stated business requirements. In other words, if you are unsure whether your solution
will deliver the outcome you want, then a Project Feasibility Study will help gain that clarity.
During the Feasibility Study, a variety of 'assessment' methods are undertaken. The outcome of
the Feasibility Study is a confirmed solution for implementation.
The main objective of the feasibility study is to test the Technical, Operational and
Economical feasibility for adding new modules and debugging old running system. All system is
feasible if they are unlimited resources and infinite time. There are aspects in the feasibility study
portion of the preliminary investigation:
Technical Feasibility
Economical Feasibility
Operation Feasibility
Technical Feasibility:
In the feasibility study first step is that the organization or company has to decide that
what technologies are suitable to develop by considering existing system.
Here in this application used the technologies like Java and Oracle. These are free
software that would be downloaded from web.
Java IDE –it is tool or technology.
For any system if the expected benefits equal or exceed the expected costs, the system can be
judged to be economically feasible. In economic feasibility, cost benefit analysis is done in
which expected costs and benefits are evaluated. Economic analysis is used for evaluating the
effectiveness of the proposed system.
In economic feasibility, the most important is Cost-benefit analysis. As the name suggests, it is
an analysis of the costs to be incurred in the system and benefits derivable out of the system.
Click on the link below which will get you to the page that explains what cost benefit analysis is
and how you can perform a cost benefit analysis.
Operational Feasibility
Not only must an application make economic and technical sense, it must also make operational
sense.
What
What tools are documenta
needed to support tion will
operations? users be
What skills will given?
operators need to be What
trained in? training
What processes need will users
to be created and/or be given?
updated? How will
What documentation change
does operations requests
need? be
managed?
Very often you will need to improve the existing operations, maintenance, and support
infrastructure to support the operation of the new application that you intend to develop. To
determine what the impact will be you will need to understand both the current operations and
support infrastructure of your organization and the operations and support characteristics of your
new application.
To operate this application END-TO-END VMS. The user no need to require any
technical knowledge that we are used to develop this project is JSP, Java. That the application
providing rich user interface by user can do the operation in flexible manner.
Non-functional requirements describe user-visible aspects of the system that are not
directly related to functionality of the system.
User Interface
Performance Constraints
In case of User Error, the System should display a meaningful error message to the user,
such that the user can correct his Error.
The high level components in proposed system should handle exceptions that
occur while connecting to various database servers, IO Exceptions etc.
Quality Issues
Quality issues mainly refers to reliable, available and robust system developing the
proposed system the developer must be able to guarantee the reliability transactions so that they
will be processed completely and accurately.
The ability of system to detect failures and recovery from those failures refers to the
availability of system.
3.4 COMMUNICATION INTERFACES
What is Java EE?
Java EE (Enterprise Edition) is a widely used platform containing a set of coordinated
technologies that significantly reduce the cost and complexity of developing, deploying, and
managing multi-tier, server-centric applications. Java EE builds upon the Java SE platform and
provides a set of APIs (application programming interfaces) for developing and running portable,
robust, scalable, reliable and secure server-side applications.
Some of the fundamental components of Java EE include:
Enterprise JavaBeans (EJB): a managed, server-side component architecture used to
encapsulate the business logic of an application. EJB technology enables rapid and
simplified development of distributed, transactional, secure and portable applications
based on Java technology.
Java Persistence API (JPA): a framework that allows developers to manage data using
object-relational mapping (ORM) in applications built on the Java Platform.
MIME Type or Content Type: If you see above sample HTTP response header, it contains tag
“Content-Type”. It’s also called MIME type and server sends it to client to let them know the
kind of data it’s sending. It helps client in rendering the data for user. Some of the mostly used
mime types are text/html, text/xml, application/xml etc.
Understanding URL
URL is acronym of Universal Resource Locator and it’s used to locate the server and resource.
Every resource on the web has it’s own unique address. Let’s see parts of URL with an example.
http://localhost:8080/FirstServletProject/jsps/hello.jsp
http:// – This is the first part of URL and provides the communication protocol to be used in
server-client communication.
localhost – The unique address of the server, most of the times it’s the hostname of the server
that maps to unique IP address. Sometimes multiple hostnames point to same IP addresses and
web server virtual host takes care of sending request to the particular server instance.
8080 – This is the port on which server is listening, it’s optional and if we don’t provide it in
URL then request goes to the default port of the protocol. Port numbers 0 to 1023 are reserved
ports for well known services, for example 80 for HTTP, 443 for HTTPS, 21 for FTP etc.
Web Container
Tomcat is a web container, when a request is made from Client to web server, it passes the
request to web container and it’s web container job to find the correct resource to handle the
request (servlet or JSP) and then use the response from the resource to generate the response and
provide it to web server. Then web server sends the response back to the client.
When web container gets the request and if it’s for servlet then container creates two Objects
HTTPServletRequest and HTTPServletResponse. Then it finds the correct servlet based on the
URL and creates a thread for the request. Then it invokes the servlet service() method and based
on the HTTP method service() method invokes doGet() or doPost() methods. Servlet methods
generate the dynamic page and write it to response. Once servlet thread is complete, container
converts the response to HTTP response and send it back to client.
Some of the important work done by web container are:
Communication Support – Container provides easy way of communication between
web server and the servlets and JSPs. Because of container, we don’t need to build a
server socket to listen for any request from web server, parse the request and generate
response. All these important and complex tasks are done by container and all we need to
focus is on our business logic for our applications.
Lifecycle and Resource Management – Container takes care of managing the life cycle
of servlet. Container takes care of loading the servlets into memory, initializing servlets,
invoking servlet methods and destroying them. Container also provides utility like JNDI
for resource pooling and management.
Multithreading Support – Container creates new thread for every request to the servlet
and when it’s processed the thread dies. So servlets are not initialized for each request
and saves time and memory.
JSP Support – JSPs doesn’t look like normal java classes and web container provides
support for JSP. Every JSP in the application is compiled by container and converted to
Servlet and then container manages them like other servlets.
Miscellaneous Task – Web container manages the resource pool, does memory
optimizations, run garbage collector, provides security configurations, support for
multiple applications, hot deployment and several other tasks behind the scene that makes
our life easier.
MySQL:
MySQL, the most popular Open Source SQL database management system, is developed,
distributed, and supported by Oracle Corporation.
The MySQL Web site (http://www.mysql.com/) provides the latest information about MySQL
software.
MySQL Server was originally developed to handle large databases much faster than
existing solutions and has been successfully used in highly demanding production
environments for several years. Although under constant development, MySQL Server
today offers a rich and useful set of functions. Its connectivity, speed, and security make
MySQL Server highly suited for accessing databases on the Internet.
The official way to pronounce “MySQL” is “My EssQue Ell” (not “my sequel”), but we do not
mind if you pronounce it as “my sequel” or in some other localized way.
4.SYSTEM DESIGN
1. The DFD is also called as bubble chart. It is a simple graphical formalism that can be
used to represent a system in terms of input data to the system, various processing carried
out on this data, and the output data is generated by this system.
2. The data flow diagram (DFD) is one of the most important modeling tools. It is used to
model the system components. These components are the system process, the data used
by the process, an external entity that interacts with the system and the information flows
in the system.
3. DFD shows how the information moves through the system and how it is modified by a
series of transformations. It is a graphical technique that depicts information flow and the
transformations that are applied as data moves from input to output.
4. DFD is also known as bubble chart. A DFD may be used to represent a system at any
level of abstraction. DFD may be partitioned into levels that represent increasing
information flow and functional detail.
DFD Diagrams:
Messaging
UI SCREENS
user
Reports
User
System
No
Validates
Data
Validates
Data
Tbl_Location Tbl_Location
Tbl_Location
Tbl_Location
Verifies Details
View City
2.0.1 View Hotels
2.0.3 Data Storage
Validates
Data
Validates
Data
Verifies Details
Add City
2.0.1 Add Hotels
2.0.3 Data Storage
Validates
Data
GOALS:
The Primary goals in the design of the UML are as follows:
1. Provide users a ready-to-use, expressive visual modeling Language so that they can
develop and exchange meaningful models.
2. Provide extendibility and specialization mechanisms to extend the core concepts.
3. Be independent of particular programming languages and development process.
4. Provide a formal basis for understanding the modeling language.
5. Encourage the growth of OO tools market.
6. Support higher level development concepts such as collaborations, frameworks, patterns
and components.
7. Integrate best practices.
send();
view(); request();
response();
A use case diagram in the Unified Modeling Language (UML) is a type of behavioral
diagram defined by and created from a Use-case analysis. Its purpose is to present a graphical
overview of the functionality provided by a system in terms of actors, their goals (represented as
use cases), and any dependencies between those use cases. The main purpose of a use case
diagram is to show what system functions are performed for which actor. Roles of the actors in
the system can be depicted.
Register
Login
view place
User
Logout
User Server
Create Account
Login
View inbo
1 : Enter UserId,Pwd()
2 :InsertData ()
3 :Connetion()
4 :Actionmethod()
5 : Response()
6 : Show Result()
1 : Login()
2 :AcceptRequests()
3 :Connetion()
4 :GetData()
5 : Response()
6 : Show Result()
User Login
4 :Connectmethod()
DAL:clsSql DB
Helper
3 : Connected()
BAL:clsLogi
n
2 : Check Data()
5 : Response()
frmLogin
1 : Enter UserId,Pwd()
Admin
User
Registe r
Login
NO
View places
A state chart diagram shows the behavior of classes in response to external stimuli. This diagram
models the dynamic flow of control from state to state within a system.
States: States represent situations during the life of an object. You can easily illustrate a state in
Smart Draw by using a rectangle with rounded corners.
Transition: A solid arrow represents the path between different states of an object. Label the
transition with the event that triggered it and the action that results from it.
Initial State: A filled circle followed by an arrow represents the object's initial state.
Final State: An arrow pointing to a filled circle nested inside another circle represents the
object's final state.
Synchronization and Splitting of Control:A short heavy bar with two transitions entering it
represents a synchronization of control. A short heavy bar with two transitions leaving it
represents a splitting of control that creates multiple states.
State Chat Diagram:-
Submit
View Map
Submit Data
Home
OLEDB
Map
The software engineering process can be viewed as a spiral. Initially system engineering
defines the role of software and leads to software requirements analysis where the information
domain , functions , behavior , performance , constraints and validation criteria for software are
established. moving inward along the spiral , we come to design and finally to coding . To
develop computer software we spiral in along streamlines that decreases the level of abstraction
on each item.
A Strategy for software testing may also be viewed in the context of the spiral. Unit testing
begins at the vertex of the spiral and concentrates on each unit of the software as implemented in
source code. Testing will progress by moving outward along the spiral to integration testing ,
where the focus on the design and the concentration of the software architecture. Talking another
turn on outward on the spiral we encounter validation testing where requirements established as
part of software requirements analysis are validated against the software that has been
constructed . Finally we arrive at system testing , where the software and other system elements
are tested as a whole .
UNIT TESTING
UNUNI
MODULE
Component
SUB-SYSTEM
SYSTEM TESTING
Integration Testing
ACCEPTANCE
User Testing:
Testing is the process of finding difference between the expected behavior specified by system
models and the observed behavior of the implemented system.
Testing Activities
Different levels of testing are used in the testing process , each level of testing aims to test
different aspects of the system. the basic levels are:
Unit testing
Integration testing
System testing
Acceptance testing
Unit Testing
Unit testing focuses on the building blocks of the software system, that is, objects and sub
system . There are three motivations behind focusing on components. First, unit testing reduces
the complexity of the overall tests activities, allowing us to focus on smaller units of the system.
Second , unit testing makes it easier to pinpoint and correct faults given that few components are
involved in this test . Third , Unit testing allows parallelism in the testing activities , that is each
component can be tested independently of one another . Hence the goal is to test the internal
logic of the module.
Integration Testing
In the integration testing, many test modules are combined into sub systems , which are
then tested . The goal here is to see if the modules can be integrated properly, the emphasis being
on testing module interaction.
After structural testing and functional testing we get error free modules. These modules
are to be integrated to get the required results of the system. After checking a module, another
module is tested and is integrated with the previous module. After the integration, the test cases
are generated and the results are tested.
System Testing
In system testing the entire software is tested . The reference document for this process is
the requirement document and the goal is to see whether the software meets its requirements.
The system was tested for various test cases with various inputs.
Acceptance Testing
Acceptance testing is sometimes performed with realistic data of the client to demonstrate
that the software is working satisfactory. Testing here focus on the external behavior of the
system , the internal logic of the program is not emphasized . In acceptance testing the system is
tested for various inputs.
Types of Testing
1. Black box or functional testing
2. White box testing or structural testing
As shown in the following figure of Black box testing , we are not thinking of the internal
workings , just we think about
What is the output to our system?
What is the output for given input to our system?
Input Output
?
The Black box is an imaginary box that hides its internal workings
box testing is concerned with testing the implementation of the program. the intent of structural
is not to exercise all the inputs or outputs but to exercise the different programming and data
structure used in the program. Thus structural testing aims to achieve test cases that will force the
desire coverage of different structures . Two types of path testing are statement testing coverage
and branch testing coverage.
Input Output
INTERNAL
WORKING
Test Data:
Here all test cases that are used for the system testing are specified. The goal is to test the
different functional requirements specified in Software Requirements Specifications (SRS)
document.
Unit Testing:
Each individual module has been tested against the requirement with some test data.
Test Report:
The module is working properly provided the user has to enter information. All data entry
forms have tested with specified test cases and all data entry forms are working properly.
Error Report:
If the user does not enter data in specified order then the user will be prompted with error
messages. Error handling was done to handle the expected and unexpected errors.
Test cases :
A Test case is a set of input data and expected results that exercises a component with the
purpose of causing failure and detecting faults .test case is an explicit set of instructions designed
to detect a particular class of defect in a software system , by bringing about a failure . A Test
case can give rise to many tests.
5.1TEST CASES:
Test cases can be divided in to two types. First one is Positive test cases and second one
is negative test cases. In positive test cases are conducted by the developer intention is to get the
output. In negative test cases are conducted by the developer intention is to don’t get the output.
1 Create the new user New user created Update personal True
registration process successfully info in to oracle
database
We have seen plenty of applications calling for a search engine that is able to efficiently
support novel forms of spatial queries that are integrated with keyword search. The existing
solutions to such queries either incur prohibitive space consumption or are unable to give real
time answers. In this paper, we have remedied the situation by developing an access method
called the spatial inverted index (SI-index). Not only that the SI-index is fairly space economical,
but also it has the ability to perform keyword-augmented nearest neighbor search in time that is
at the order of dozens of milli-seconds. Furthermore, as the SI-index is based on the conventional
technology of inverted index, it is readily incorporable in a commercial search engine that
applies massive parallelism, implying its immediate industrial merits.
Our Project Fast Nearest Neighbor Search with Keywords is extremely effective for searching
nearest restaurant from user with expected menus. It does this by IR2 tree algorithm-
Compression, Merging and Distance Browsing, and GPS System. In this we can add features like
After selecting Hotel it will display menu card of that Hotel Implement this application for PC’s
and Desktops.We have so many applications that can be used as search engine which is able to
efficiently support novel forms of spatial queries that are integrated with keyword search. In this
project we have developed an access method called the Spatial Inverted Index (SIIndex). This
method is fairly space economical and it has ability to perform keyword augmented
nearestneighbor search in real time. This method is based on conventional technology of Inverted
Index. It is readily incorporable in a commercial search engine.
REFERENCES
[1] S. Agrawal, S. Chaudhuri, and G. Das, “Dbxplorer: A System for Keyword-Based Search
over Relational Databases,” Proc. Int’l Conf. Data Eng. (ICDE), pp. 5-16, 2002.
[2] N. Beckmann, H. Kriegel, R. Schneider, and B. Seeger, “The R - tree: An Efficient and
Robust Access Method for Points and Rectangles,” Proc. ACM SIGMOD Int’l Conf.
Management of Data, pp. 322-331, 1990.
[4] X. Cao, L. Chen, G. Cong, C.S. Jensen, Q. Qu, A. Skovsgaard, D. Wu, and M.L. Yiu,
“Spatial Keyword Querying,” Proc. 31st Int’l Conf. Conceptual Modeling (ER), pp. 16-29, 2012.
[5] X. Cao, G. Cong, and C.S. Jensen, “Retrieving Top-k Prestige- Based Relevant Spatial Web
Objects,” Proc. VLDB Endowment, vol. 3, no. 1, pp. 373-384, 2010.
[6] X. Cao, G. Cong, C.S. Jensen, and B.C. Ooi, “Collective Spatial Keyword Querying,” Proc.
ACM SIGMOD Int’l Conf. Management of Data, pp. 373-384, 2011.
[7] B. Chazelle, J. Kilian, R. Rubinfeld, and A. Tal, “The Bloomier Filter: An Efficient Data
Structure for Static Support Lookup Tables,” Proc. Ann. ACM-SIAM Symp. Discrete
Algorithms (SODA), pp. 30- 39, 2004.
[8] Y.-Y. Chen, T. Suel, and A. Markowetz, “Efficient Query Processing in Geographic Web
Search Engines,” Proc. ACM SIGMOD Int’l Conf. Management of Data, pp. 277-288, 2006.
[9] E. Chu, A. Baid, X. Chai, A. Doan, and J. Naughton, “Combining Keyword Search and
Forms for Ad Hoc Querying of Databases,” Proc. ACM SIGMOD Int’l Conf. Management of
Data, 2009.
[10] G. Cong, C.S. Jensen, and D. Wu, “Efficient Retrieval of the Top-k Most Relevant Spatial
Web Objects,” PVLDB, vol. 2, no. 1, pp. 337- 348, 2009.
[11] C. Faloutsos and S. Christodoulakis, “Signature Files: An Access Method for Documents
and Its Analytical Performance Evaluation,” ACM Trans. Information Systems, vol. 2, no. 4, pp.
267-288, 1984.
[12] I.D. Felipe, V. Hristidis, and N. Rishe, “Keyword Search on Spatial Databases,” Proc. Int’l
Conf. Data Eng. (ICDE), pp. 656-665, 2008.
[13] R. Hariharan, B. Hore, C. Li, and S. Mehrotra, “Processing Spatial- Keyword (SK) Queries
in Geographic Information Retrieval (GIR) Systems,” Proc. Scientific and Statistical Database
Management (SSDBM), 2007.
[14] G.R. Hjaltason and H. Samet, “Distance Browsing in Spatial Databases,” ACM Trans.
Database Systems, vol. 24, no. 2, pp. 265-318, 1999.
[16] I. Kamel and C. Faloutsos, “Hilbert R-Tree: An Improved R-Tree Using Fractals,” Proc.
Very Large Data Bases (VLDB), pp. 500-509, 1994.
[17] J. Lu, Y. Lu, and G. Cong, “Reverse Spatial and Textual k Nearest Neighbor Search,” Proc.
ACM SIGMOD Int’l Conf. Management of Data, pp. 349-360, 2011.
[19] J.S. Vitter, “Algorithms and Data Structures for External Memory,” Foundation and Trends
in Theoretical Computer Science, vol. 2, no. 4, pp. 305-474, 2006.
[20] D. Zhang, Y.M. Chee, A. Mondal, A.K.H. Tung, and M. Kitsuregawa, “Keyword Search in
Spatial Databases: Towards Searching by Document,” Proc. Int’l Conf. Data Eng. (ICDE), pp.
688-699, 2009.
[21] Y. Zhou, X. Xie, C. Wang, Y. Gong, and W.-Y.Ma, “Hybrid Index Structures for Location-
Based Web Search,” Proc. Conf. Information and Knowledge Management (CIKM), pp. 155-
162, 2005.