Professional Documents
Culture Documents
Q1. Describe the different levels of knowledge that are necessary for
natural languagae processing.
Solution: There are 5 main types of knowledge representation in Artificial
Intelligence.
1. Meta Knowledge - its knowledge about knowledge and how to gain them
2. Heuristic Knowledge - Representing knowledge of some expert in a field
or
subject.
3. Procedural Knowledge - Gives information/ knowledge about how to
achieve
something
4. Declarative Knowledge - Its about statements that describe a
particular
object and its attributes, including some behavior in relation with it
5. Structural Knowledge - Describes what relationship exists between
concepts/ objects.
Q2. Define Augmented Transition Network.
Solution: An augmented transition network (ATN) is a type of graph
theoretic structure used in the operational definition of formal languages,
used especially in parsing relatively complex natural languages, and having
wide application in artificial intelligence. An ATN can, theoretically, analyze
the structure of any sentence, however complicated.
ATNs build on the idea of using finite state machines (Markov model) to parse
sentences. W. A. Woods in "Transition Network Grammars for Natural
Language Analysis" claims that by adding a recursive mechanism to a finite
state model, parsing can be achieved much more efficiently. Instead of
building an automaton for a particular sentence, a collection of transition
graphs are built. A grammatically correct sentence is parsed by reaching a
final state in any state graph. Transitions between these graphs are simply
subroutine calls from one state to any initial state on any graph in the
network. A sentence is determined to be grammatically correct if a final state
is reached by the last word in the sentence.
This model meets many of the goals set forth by the nature of language in
that it captures the regularities of the language. That is, if there is a process
that operates in a number of environments, the grammar should encapsulate
the process in a single structure. Such encapsulation not only simplifies the
grammar, but has the added bonus of efficiency of operation. Another
advantage of such a model is the ability to postpone decisions. Many
grammars use guessing when an ambiguity comes up. This means that not
enough is yet known about the sentence. By the use of recursion, ATNs solve
this inefficiency by postponing decisions until more is known about a
sentence.
Q6. What are the characteristics of expert system? Write the various
application area of expert system.
Solution: Characteristics of an Expert System
Match: In this first phase, the left-hand sides of all productions are
matched against the contents of working memory. As a result a conflict
set is obtained, which consists of instantiations of all satisfied
productions. An instantiation of a production is an ordered list of
working memory elements that satisfies the left-hand side of the
production.
Act: In this third phase, the actions of the production selected in the
conflict-resolution phase are executed. These actions may change
the contents of working memory. At the end of this phase,
execution returns to the first phase.
has a peak at 18 units. Heuristic Dendral would use this input mass
and the knowledge of atomic mass numbers and valence rules, to
determine the possible combinations of atomic constituents whose
mass would add up to 18.As the weight increases and the molecules
become more complex, the number of possible compounds increases
drastically. Thus, a program that is able to reduce this number of
candidate solutions through the process of hypothesis formation is
essential.
MYCIN
MYCIN was an early expert system that used artificial intelligence to
identify bacteria causing severe infections, such as bacteremia and
meningitis, and to recommend antibiotics, with the dosage adjusted for
patient's body weight .The Mycin system was also used for the
diagnosis of blood clotting diseases. MYCIN operated using a fairly
simple inference engine, and a knowledge base of ~600 rules. It would
query the physician running the program via a long series of simple
yes/no or textual questions. At the end, it provided a list of possible
culprit bacteria ranked from high to low based on the probability of
each diagnosis, its confidence in each diagnosis' probability, the
reasoning behind each diagnosis (that is, MYCIN would also list the
questions and rules which led it to rank a diagnosis a particular way),
and its recommended course of drug treatment.
PROSPECTOR
Consultation system to assist geologists working in mineral
exploration
Attempts to represent the knowledge and reasoning processes of
experts in the geological domain
System has been kept domain independent
It matches data from a site against models describing regional and
local characteristics favorable for specific ore deposits
The input data are assumed to be incomplete and uncertain
PROSPECTOR performs a consultation to determine such things as
Which model best fits the data?
Where the most favorable drilling sites are located?
What additional data would be most helpful in reaching firmer
conclusions?
What is the basis for these conclusions and recommendations?
R1
ASSIGNMENT FOUR
Q1. How does an artificial neural network model the brain? Describe
two major classes of learning paradigms: supervised and
unsupervised learning. What are the features that distinguish these
paradigms from each other?
Solution: In particular, the most basic element of the human brain is a
specific type of cell which, unlike the rest of the body, doesn't appear to
regenerate. Because this type of cell is the only part of the body that isn't
slowly replaced, it is assumed that these cells are what provides us with our
abilities to remember, think, and apply previous experiences to our every
action. These cells, all 100 billion of them, are known as neurons. Each of
these neurons can connect with up to 200,000 other neurons, although 1,000
to 10,000 is typical.
The power of the human mind comes from the sheer numbers of these basic
components and the multiple connections between them. It also comes from
genetic programming and learning.
The individual neurons are complicated. They have a myriad of parts, subsystems, and control mechanisms. They convey information via a host of
electrochemical pathways. There are over one hundred different classes of
neurons, depending on the classification method used. Together these
neurons and their connections form a process which is not binary, not stable,
and not synchronous. In short, it is nothing like the currently available
electronic computers, or even artificial neural networks.
These artificial neural networks try to replicate only the most basic elements
of this complicated, versatile, and powerful organism. They do it in a primitive
way
Supervised learning which incorporates an external teacher, so that each
output unit is told what its desired response to input signals ought to be.
During the learning process global information may be required. Paradigms of
supervised learning include error-correction learning, reinforcement learning
and
stochastic
learning.
An important issue concerning supervised learning is the problem of error
convergence, ie the minimization of error between the desired and computed
unit values. The aim is to determine a set of weights which minimizes the
error. One well-known method, which is common to many learning
paradigms, is the least mean square (LMS) convergence.
Unsupervised learning uses no external teacher and is based upon only local
information. It is also referred to as self-organization, in the sense that it selforganizes data presented to the network and detects their emergent
collective properties. Paradigms of unsupervised learning are Hebbian
learning and competitive learning.
It is obvious that the input units use the identity function. Sometimes a
constant is multiplied by the net input to form a linear function.
(2.4)
This kind of function is often used in single layer networks.
(2.5)
This function is especially advantageous for use in neural networks
trained by back-propagation; because it is easy to differentiate, and thus
can dramatically reduce the computation burden for training. It applies to
applications whose desired output values are between 0 and 1.
(2.6)
This function has similar properties with the sigmoid function. It works
well for applications that yield output values in the range of [-1,1].
Q4. Discuss how neural networks help in solving AI problems. Discuss about different
types of functions of an ANN.
or
discriminant analysis,
In many cases, simple neural network configurations yield the same solution
as many traditional statistical applications. For example, a single-layer,
feedforward neural network with linear activation for its output perceptron is
equivalent to a general linear regression fit. Neural networks can provide
more accurate and robust solutions for problems where traditional methods
do not completely apply.
A neural network is defined not only by its architecture and flow, or
interconnections, but also by computations used to transmit information from
one node or input to another node. These computations are determined by
Types of functions:
Step Function
A step function is a function like that used by the original Perceptron. The
output is a certain value, A1, if the input sum is above a certain threshold and
A0 if the input sum is below a certain threshold. The values used by the
Perceptron were A1 = 1 and A0 = 0.
These kinds of step activation functions are useful for binary classification
schemes
Linear combination
A linear combination is where the weighted sum input of the neuron plus a
linearly dependant bias becomes the system output. Specifically:
In these cases, the sign of the output is considered to be equivalent to the 1
or 0 of the step function systems, which enables the two methods, be to
equivalent if
Continuous Log-Sigmoid Function
A log-sigmoid function, also known as a logistic function, is given by the
relationship:
Softmax Function
The softmax activation function is useful predominantly in the output layer of
a clustering system. Softmax functions convert a raw value into a posterior
probability. This provides a measure of certainty. The softmax activation
function is given as: