You are on page 1of 25

Computer network

Q1 (a) A host with Ip Addresses 128.23.67.3 sends a message to a host


with IP Address 193.45.23.7. Does the message travel through any
router? Assume no sub netting.
Ans)
Yes Class B to C. This message must travel through a router because the netids of the two
addresses are different (128.23 versus 193.45.23).Since the address 128.23.67.3 lies in the
range of class B, and the address 193.45.23.7 lies in the range of class A.

Q1 (b) Why ATM cells have a small and fixed size?


Ans)
1. Allows better statistical multiplexing of information on a given medium than does
the use of larger packets of variable length.
2. With longer variable-length packets, high-speed switching would be more
complex.
3. The limited amount of information in the header allows ATM cells to be processed
at a very high rate from 150 Mbps to several Gbps.
4. Provides speed, flexibility and compatibility with many different applications and
physical media.
5. The reduced size of the internal buffers guarantees a minimal delay and delay jitter.
This is why ATM can handle both constant rate traffic (audio, video) and variable
rate traffic (data) easily.
6. T o limit the queuing delays in internal buffers.

Q1 (c) A channel has a bit rate of 4kbps and a propagation delay of 20


msec. For what range of frame sizes does stop-and-wait give an
efficiency of at least 50 percent?
Ans) You need to use two formulas - firstly a= propagation delay/transmission time. We'll
denote this by tprop and ttrans.
tprop is given in the question, it is 20 x 10-3
ttrans is Distance/Data Rate
So ttrans = L / 4 x 103
ttrans = 80/L
(Because the powers of ten cancel each other out so you just multiply 20 x 4.
we use powers of ten because we need the same format as the propogation delay
a=80/L

2. Use the formula 1/1+2a >= .5 = 1/1+ 160/L>=.5


First 1>= .5(1+160/L)

second 1>= .5 + 80/L


third 80/L >= 1 x .5
fourth .5 >= 80/L
L >= 80/.5
L >= 160
Efficiency will be 50% when the time to transmit the frame equals the round trip
propagation delay. At a transmission rate of 4bits/ms, 160 bits takes 40 ms. For
frame sizes above 160 bits, stop-and-wait is reasonably efficient.

Q1(d) Compare LAN standards 802.3,802.4,802.5.


Ans)
802.3 Ethernet
802.3 is the standard which Ethernet operates by. It is the standard for CSMA/CD
(Carrier Sense Multiple Access with Collision Detection). This standard encompasses
both the MAC and Physical Layer standards.
CSMA/CD is what Ethernet uses to control access to the network medium (network
cable). If there is no data, any node may attempt to transmit, if the nodes detect a
collision, both stop transmitting and wait a random amount of time before
retransmitting the data.
The original 802.3 standard is 10 Mbps (Megabits per second). 802.3u defined the 100
Mbps (Fast Ethernet) standard, 802.3z/802.3ab defined 1000 Mbps Gigabit Ethernet,
and 802.3ae define 10 Gigabit Ethernet.
Commonly, Ethernet networks transmit data in packets, or small bits of information. A
packet can be a minimum size of 72 bytes or a maximum of 1518 bytes.
The most common topology for Ethernet is the star topology.
802.5 Token Ring
Token ring is designed to use the ring topology and utilizes a token to control the
transmission of data on the network.
The token is a special frame which is designed to travel from node to node around the
ring. When it does not have any data attached to it, a node on the network can modify
the frame, attach its data and transmit. Each node on the network checks the token as it
passes to see if the data is intended for that node, if it is; it accepts the data and
transmits a new token. If it is not intended for that node, it retransmits the token on to
the next node.
The token ring network is designed in such a way that each node on the network is
guaranteed access to the token at some point. This equalizes the data transfer on the
network. This is different from an Ethernet network where each workstation has equal
access to grab the available bandwidth, with the possible of a node using more
bandwidth than other nodes.Originally, token ring operated at a speed of about 4 Mbps
and 16 Mbps. 802.5t allows for 100 Mbps speeds and 802.5v provides for 1 Gbps over
fibber.

Token ring can be run over a star topology as well as the ring topology.There are three
major cable types for token ring: Unshielded twisted pair (UTP), Shielded twisted pair
(STP), and fibber.
The IEEE 802.4 standard describes a token bus
Token bus is a network implementing the token ring protocol over a "virtual ring" on
a coaxial cable A token is passed around the network nodes and only the node possessing
the token may transmit. If a node doesn't have anything to send, the token is passed on to
the next node on the virtual ring. Each node must know the address of its neighbor in the
ring, so a special protocol is needed to notify the other nodes of connections to, and
disconnections from, the ring. Token bus combines the physical configuration of Ethernet
and the collision free feature of Token ring.
This is an application of the concepts used in token ring networks. The main difference is
that the endpoints of the bus do not meet to form a physical ring.

Token Bus

Q1(e) Explain ARP and RARP


Ans)
Address Resolution Protocol: A TCP/IP protocol used to obtain a node's physical address.
A client station broadcasts an ARP request onto the network with the IP address of the target
node it wishes to communicate with, and the node with that address responds by sending
back its physical address so that packets can be transmitted. ARP returns the layer 2 address
for a layer 3 address.

ARP'ing
The IP protocol broadcasts the IP address of the destination station onto the network, and the
node with that address responds.ARP -Meaning of ARP "Address Resolution Protocol", is
used to map ip Network addresses to the hardware (Media Access Control sub layer)
addresses used by the data link protocol. The ARP protocol operates between the
network layer and the data link layer in the Open System Interconnection (osi) model.
RARP- RARP (Reverse Address Resolution Protocol) is a protocol by which a physical
machine in a local area network can request to learn its IP address from a gateway server's
Address Resolution Protocol (ARP) table or cache. A network administrator creates a table in
a local area network's gateway router that maps the physical machine (or Media Access
Control - MAC address) addresses to corresponding Internet Protocol addresses. When a new
machine is set up, its RARP client program requests from the RARP server on the router to be
sent its IP address. Assuming that an entry has been set up in the router table, the RARP
server will return the IP address to the machine which can store it for future use. RARP is
available for Ethernet, Fiber Distributed-Data Interface, and token ring LANs.

Q2(a) Explain the following with their advantages and disadvantages


(1)

Star Topology
Advantages:
1. Good performance
2. Reliable (if one connection fails, it doesn't affect others)
3. Easy to replace or remove hosts or other devices
4. Easy to install and wire.
5. No disruptions to the network when connecting or removing devices
6. Easy to detect faults and to remove parts.
Disadvantages of Star Network Topology
1. In star network topology, data communication depends on HUB. If central
hub fails, then whole network fails.
2. Since each computer will be connected with HUB by means of a separate
wire, star network topology needs more cable to connect computers.
3. It is more expensive due to more wires.

4. More expensive than linear bus topologies because of the cost of the hubs
5. Expensive to install
6. Extra hardware required
(2) Bus Topology
Advantages of Bus Network Topology
1. It is very simple topology.
2. It is easy to use.
3. It needs small amount of wire for connecting computers.
4. It is less expensive due to small wire needed.
5. If one computer fails, it does not disturb the other computers in network.
Other computers will continue to share information and other resources with
other connected computers.
Disadvantages of Bus Network Topology
1.
2.

Only small number of computers can be connected in a bus network.


Network speed slows down as the number of computer increases in bus
topology.
3. Finding a fault is difficult in bus topology.
4. Entire network shuts of there is a break in main cable.
(3)

Ring Topology

Advantages of Ring Network Topology


1. It is relatively less expensive than a star topology network.
2. In a Ring topology, every computer has an equal access to the network.
3. Performs better than a bus topology under heavy network load
Disadvantages of Ring Network Topology
1. Failure of one computer in the ring can affect the whole network.
2. It is difficult to find faults in a ring network topology.
3. Adding or removing computers will also affect the whole network since every
computer is connected with previous and next computer.
4. Sending a message from one computer to another takes time according to the
number of nodes between the two computers. Communication delay is directly
proportional to number of nodes in the network.
(4)

Mesh Topology

Advantages of Mesh Network Topology


1. Since, there are many links to transfer data, Mesh topology gets rid of the traffic problem.
Data may be transferred through different links.
2. If one link becomes unusable, it does not disturb the whole system. Other links can be
used for communication.

3. Since each node has physical connection with other nodes, therefore, one node can
transfer data to many nodes at the same time.
Disadvantages of Mesh Network Topology
It is very expensive due to implementation of multiple links for each node.
2. It is difficult to install and reconfigure.
3. Adding or removing a computer is difficult
Q 2(b) Why cell switching is used in ATM? Explain ATM reference model.
Ans) ATM transfers information in fixed-size units called cells. Each cell consists of
53 octets, or bytes as shown in Fig. 4.6.4. The first 5 bytes contain cell-header
information, and the remaining 48 contain the payload (user information). Small,
fixed-length cells are well suited to transfer voice and video traffic because such
traffic is intolerant to delays that result from having to wait for a large data packet to
download, among other things.

An ATM cell header can be one of two formats: UNI or NNI. The UNI header is
used for communication between ATM endpoints and ATM switches in private ATM
networks. The NNI header is used for communication between ATM switches.
Unlike the UNI, the NNI header does not include the Generic Flow Control (GFC)
field. Additionally, the NNI header has a Virtual Path Identifier (VPI) field that
occupies the first 12 bits, allowing for larger trunks between public ATM switches.
ATM Cell Header Fields
The following descriptions summarize the ATM cell header fields:
Generic Flow Control (GFC)Provides local functions, such as identifying
multiple stations that share a single ATM interface. This field is typically not used
and is set to its default value of 0 (binary 0000).
Virtual Path Identifier (VPI)In conjunction with the VCI, identifies the next
destination of a cell as it passes through a series of ATM switches on the way to
its destination.

Virtual Channel Identifier (VCI)In conjunction with the VPI, identifies the
next destination of a cell as it passes through a series of ATM switches on the way
to its destination.
Payload Type (PT)Indicates in the first bit whether the cell contains user data
or control data. If the cell contains user data, the bit is set to 0. If it contains
control data, it is set to 1. The second bit indicates congestion (0 = no congestion,
1 = congestion), and the third bit indicates whether the cell is the last in a series of
cells that represent a single AAL5 frame (1 = last cell for the frame).
Cell Loss Priority (CLP)Indicates whether the cell should be discarded if it
encounters extreme congestion as it moves through the network. If the CLP bit
equals 1, the cell should be discarded in preference to cells with the CLP bit equal
to 0.
Header Error Control (HEC)Calculates checksum only on the first 4 bytes of
the header. HEC can correct a single bit error in these bytes, thereby preserving
the cell rather than discard

The ATM reference model is composed of the following ATM layers:

Physical layer - Analogous to the physical layer of the OSI reference model, the
ATM physical layer manages the medium-dependent transmission.
ATM layer - Combined with the ATM adaptation layer, the ATM layer is roughly
analogous to the data link layer of the OSI reference model. The ATM layer is
responsible for the simultaneous sharing of virtual circuits over a physical link (cell
multiplexing) and passing cells through the ATM network (cell relay). To do this, it uses
the VPI and VCI information in the header of each ATM cell.
ATM adaptation layer (AAL) - Combined with the ATM layer, the AAL is roughly
analogous to the data link layer of the OSI model. The AAL is responsible for isolating
higher-layer protocols from the details of the ATM processes. The adaptation layer
prepares user data for conversion into cells and segments the data into 48-byte cell
payloads.
Finally, the higher layers residing above the AAL accept user data, arrange it into
packets, and hand it to the AAL.

The ATM Physical Layer


The ATM physical layer has four functions: Cells are converted into a bitstream, the
transmission and receipt of bits on the physical medium are controlled, ATM cell boundaries
are tracked, and cells are packaged into the appropriate types of frames for the physical
medium. For example, cells are packaged differently for SONET than for DS-3/E-3 media
types.
The ATM physical layer is divided into two parts: the physical medium-dependent (PMD)
sublayer and the transmission convergence (TC) sublayer.
The PMD sublayer provides two key functions. First, it synchronizes transmission and
reception by sending and receiving a continuous flow of bits with associated timing
information. Second, it specifies the physical media for the physical medium used,
including connector types and cable.
The TC sub layer has four functions: cell delineation, header error control (HEC) sequence
generation and verification, cell-rate decoupling, and transmission frame adaptation. The
cell delineation function maintains ATM cell boundaries, allowing devices to locate cells
within a stream of bits. HEC sequence generation and verification generates and checks the
header error control code to ensure valid data. Cell-rate decoupling maintains
synchronization and inserts or suppresses idle (unassigned) ATM cells to adapt the rate of
valid ATM cells to the payload capacity of the transmission system. Transmission frame
adaptation packages ATM cells into frames acceptable to the particular physical layer
implementation.
ATM Adaptation Layers: AAL1
AAL1, a connection-oriented service, is suitable for handling constant bit rate sources
(CBR), such as voice and videoconferencing. ATM transports CBR traffic using circuitemulation services. Circuit-emulation service also accommodates the attachment of
equipment currently using leased lines to an ATM backbone network. AAL1 requires timing

synchronization between the source and the destination. For this reason, AAL1 depends on
a medium, such as SONET, that supports clocking.
ATM Adaptation Layers: AAL2
Another traffic type has timing requirements like CBR but tends to be bursty in nature. This
is called variable bit rate (VBR) traffic. This typically includes services characterized as
packetized voice or video that do not have a constant data transmission speed but that do
have requirements similar to constant bit rate services. AAL2 is suitable for VBR traffic.
The AAL2 process uses 44 bytes of the cell payload for user data and reserves 4 bytes of the
payload to support the AAL2 processes.
VBR traffic is characterized as either real-time (VBR-RT) or as non-real-time (VBR-NRT).
AAL2 supports both types of VBR traffic.
ATM Adaptation Layers: AAL3/4
AAL3/4 supports both connection-oriented and connectionless data. It was designed for
network service providers and is closely aligned with Switched Multimegabit Data Service
(SMDS). AAL3/4 is used to transmit SMDS packets over an ATM network.
An AAL 3/4 SAR PDU header consists of Type, Sequence Number, and Multiplexing
Identifier fields. Type fields identify whether a cell is the beginning, continuation, or end of
a message. Sequence number fields identify the order in which cells should be reassembled.
The Multiplexing Identifier field determines which cells from different traffic sources are
interleaved on the same virtual circuit connection (VCC) so that the correct cells are
reassembled at the destination.
ATM Adaptation Layers: AAL5
AAL5 is the primary AAL for data and supports both connection-oriented and
connectionless data. It is used to transfer most non-SMDS data, such as classical IP over
ATM and LAN Emulation (LANE). AAL5 also is known as the simple and efficient
adaptation layer (SEAL) because the SAR sublayer simply accepts the CS-PDU and
segments it into 48-octet SAR-PDUs without reserving any bytes in each cell.
Q 2(c) What are advantages of using Fiber optic cable compared to copper cable.
Thinner - Optical fibers can be drawn to smaller diameters than copper wire.
Higher carrying capacity - Because optical fibers are thinner than copper wires, more
fibers can be bundled into a given-diameter cable than copper wires. This allows more
phone lines to go over the same cable or more channels to come through the cable into your
cable TV box.
Less signal degradation - The loss of signal in optical fiber is less than in copper wire
Light signals Unlike electrical signals in copper wires, light signals from one fiber do not
interfere with those of other fibers in the same cable. This means clearer phone
conversations or TV reception
Low power - Because signals in optical fibers degrade less, lower-power transmitters can
be used instead of the high-voltage electrical transmitters needed for copper wires. Again,
this saves your provider and you money.

Digital signals - Optical fibers are ideally suited for carrying digital information, which is
especially useful in computer networks.
Non-flammable - Because no electricity is passed through optical fibers, there is no fire
hazard
.Lightweight - An optical cable weighs less than a comparable copper wire cable. Fiberoptic cables take up less space in the ground.
Flexible - Because fiber optics are so flexible and can transmit and receive light, they are
used in many flexible digital cameras for the following purposes:
Speed: Fiber optic networks operate at speeds up to 10 gigabits per second or higher, as
opposed to 1.54 megabits per second for copper. A fiber optic system is now capable of
transmitting the equivalent of an entire encyclopedia (24 volumes) of information in one
second. Fiber can carry information so fast that you could transmit three television episodes
in one second.
Bandwidth: Taken in bulk, it would take 33 tons of copper to transmit the same amount of
information handled by 1/4 pound of optical fiber.
Resistance: Fiber optic cables have a greater resistance to electromagnetic noise such as
radios, motors or other nearby cables. Because optical fibers carry beams of light, they are
free of electrical noise and interference.
Q 3(a) In class A, the first address is 20.0.0.0.What is the 220000th address.
For address 220,000 in x.y.z.t notation is 0.3.91.96.In order to solve the problem we add this to 20.0.0.0,
and the 220,001th address is 20.3.91.96.
So, the 220,000th address is 20.3.91.95.

Q 3(b) Give an advantage and disadvantage of NT12 in an ISDN network.

NT12

Q 3(c) Compare and ISO-OSI with TCP/IP reference model.


SIMILARITIES

The main similarities between the two models include the following:
They share similar architecture. - Both of the models share a similar architecture.
This can be illustrated by the fact that both of them are constructed with layers.
They share a common application layer.- Both of the models share a common
"application layer". However in practice this layer includes different services depending
upon each model.
Both models have comparable transport and network layers.- This can be illustrated
by the fact that whatever functions are performed between the presentation and network
layer of the OSI model similar functions are performed at the Transport layer of the
TCP/IP model.
Knowledge of both models is required by networking professionals.- According to
article obtained from the internet networking professionals "need to know both models".
(Source:
Both models assume that packets are switched.- Basically this means that individual
packets may take differing paths in order to reach the same destination.
DIFFERENCES
The main differences between the two models are as follows:
TCP/IP Protocols are considered to be standards around which the internet has
developed. The OSI model however is a "generic, protocol- independent
standard." (www.netfact.com/crs)
TCP/IP combines the presentation and session layer issues into its application layer.
TCP/IP combines the OSI data link and physical layers into the network access layer.
TCP/IP appears to be a more simpler model and this is mainly due to the fact that it has
fewer layers.
TCP/IP is considered to be a more credible model- This is mainly due to the fact because
TCP/IP protocols are the standards around which the internet was developed therefore it
mainly gains creditability due to this reason. Where as in contrast networks are not
usually built around the OSI model as it is merely used as a guidance tool.
The OSI model consists of 7 architectural layers whereas the TCP/IP only has 4 layers.
Q 4(a) Explain the working of 3-bit sliding window protocol with suitable example.

The simplex stop and wait ARQ protocol using sequence numbers and sequence
number acknowledgments will ensure an error free communications channel for
higher levels. However, waiting for an acknowledgment for each frame in turn is
very wasteful and a more efficient alternative is to use a sliding window protocol
which enables a number of frames to be transmitted and separately
acknowledged.
In a sliding window protocol each outbound frame contains a sequence number in
the range 0 to some maximum (MaxSeq). If n bits are allocated in the header to
store a sequence number the number range would be from 0 to 2" - 1, eg if a 3 bit
number is used the sequence numbers would range from 0 to 7. The sender and
receiver maintain a window:

Sending window

is a list of consecutive frame sequence numbers that can be sent by the sender
or that have been sent and acknowledgments are waited for. When
an ack arrives and all previous frames have already been acknowledged the
window can be advanced and a new message obtained from the host to be
transmitted with the next highest available sequence number. If a ack arrives
for a frame that is not within the window it is discarded, eg an extra ack for
a frame that has already been acknowledged.

is a list of sequence numbers for frames that can be accepted by the receiver.
When a valid frame arrives and all previous frames have already arrived the
Receiving window
window is advanced. If a frame arrives that is not within the window it is
discarded.

Although the IMPs (interface message processor) now have more freedom in the
order in which frames are sent or received the higher layers of the destination host
must get the messages in the same order that the source host supplied them. In
addition, the physical layer must still appear to be a piece of wire to the data link
layer.
Since frames currently within the senders window could be lost the sender has to
have sufficient buffer space to store n frames. When all the buffers are full
(waiting for acks) the IMP must stop the host from passing more messages until a
buffer is free. Must make sure that no more than 2n - 1 frames are un-acked at any point,not 2n frames (even though there are 2n distinct frame
numbers).
Consider 3 bit frame number, where frames numbered 0..7 (8 distinct
numbers)

1. 8 un-ack-ed frames bad:


transmit 8 frames numbered 0..7
If received ok, will get ack=7
If error, will get ack=7 too! (Ack up to last ok frame)
Next transmission 8 frames 0..7
ack=7 comes back
is that ack-ing the new batch, or just re-ack-ing the old?
2. 7 un-ack-ed frames ok:
If transmit 7 frames 0..6
If error, ack=7
If received, ack=6
Then transmit 7 frames 7,0..5
If error, ack=6
If received, ack=5
That is, after transmitting 7 packets, there are 8 possible results: Anywhere from 0 to 7 packets
could have been received successfully. This is 8 possibilities, and the transmitter needs enough
information in the acknowledgment to distinguish them all.If the transmitter sent 8 packets
without waiting for acknowledgment, it could find itself in a quandary similar to the stop-and-wait
case: does the acknowledgment mean that all 8 packets were received successfully, or none of
them?

Q 4(b) Consider an error-free 64-kbps satellite channel used to send 512-byte data
frames in one direction with very short acknowledgements coming back the other way.
What is the maximum throughput for window sizes of 1, 7, 15 and 127?
A 512 byte (4096 bits) data frame has a duration of 4096/64000 seconds that is 64 msec.
Assume that the satellite is at 36000 km distance. This leads to roundtrip propagation time
of 240 msec. We should also add 64 msec to this to account for transmission time. Hence
with a window size of 1, 4096 bits can be sent every 240+64=304 msec. This equates to the
throughput of 4096 bits/304 msec or 13.5 kbps. For a window size greater than 5, the full 64
kbps is used. 3- Two adjacent nodes (A and B) use a sliding window protocol and a 3-bit
sequence

OR
Sliding window send specified amount of frames, wait for one ack for that set of frames
512 bytes x 8 bits/B = 4096 bits per frame
4096/64000 bps = 64 msec to send one frame
Round trip delay = 540 msec
Window size 1: send 4096 bits per 540msec
4096 bits / 540 msec = 7.585 x 103 bps throughput
Window size 7: 7585 x 7 = 53096 bps
Window size 9 and greater: 7585 x 9 = 68265 bps but the maximum capacity is 64 kbps so
for window sizes greater than 9 the maximum throughput is 64 kbps.
Q 5(a) Why framing is necessary? Explain framing techniques.

Framing determines start and end of packets (all we see is a string of bits) and their order by
inserting packet headers The main problem is to decide where successive packets start and
end. Therefore, the DLC encapsulates the packet into a frame by adding its own header and
trailer.
Moreover, if there is a period of idle lls between successive ,it becomes also necessary to
separate idle lls from frames. Even when idle lls are replaced by dead periods, the
problem is not simplied, because often there are no dead periods.
Types of Framing:
1 Character based framing: Character based codes, such as ASCII (7 bits and 1 parity bit),
provide binary representation for keyboard characters and terminal control characters. The
idea in character based framing is that such codes can also provide representation for
various communication characters:
One disadvantage of character based framing is that the frame must contain aninteger
number of characters; otherwise, the characters cannot be read correctly.
2 Fixed length packets/framesIn networks with xed length packets, the length eld is
implicit (not needed) because all packets (and hence frames) have the same size. An
example of such networks is ATM, where all packets are 53 bytes long. This framing
technique requires synchronization upon initialization. Another drawback is that message
length often not a multiple of packet size.
3 Bit oriented framing
In bit oriented framing, a special binary ag indicates the end of the frame, and it is
avoided within the frame itself by using a technique called bit stung. Therefore, the
concept is thesame, but the main dierence is that a ag.

Q 5(b) Explain Sliding window protocol. Also explain error control at DLL.
A sliding window protocol is a feature of packet-based data transmission protocols. Sliding
window protocols are used where reliable in-order delivery of packets is required, such as in
the Data Link Layer (OSI model) as well as in the Transmission Control Protocol (TCP).
Conceptually, each portion of the transmission (packets in most data link layers, but bytes in
TCP) is assigned a unique consecutive sequence number, and the receiver uses the numbers
to place received packets in the correct order, discarding duplicate packets and identifying
missing ones. The problem with this is that there is no limit of the size of the sequence
numbers that can be required.
By placing limits on the number of packets that can be transmitted or received at any given
time, a sliding window protocol allows an unlimited number of packets to be communicated
using fixed-size sequence numbers. The term "window" on transmitter side represents the
logical boundary of the total number of packets yet to be acknowledged by the receiver. The
receiver informs the transmitter in each acknowledgment packet the current maximum
receiver buffer size (window boundary). The TCP header uses a 16 bit field to report the

receive window size to the sender. Therefore, the largest window that can be used is
2^16=65k bytes. In slow-start mode, the transmitter starts with low packet count and
increases the number of packets in each transmission after receiving acknowledgment
packets from receiver. For every ack packet received, the window slides by one packet
(logically) to transmit one new packet. When the window threshold is reached, the
transmitter sends one packet for one ack packet received. If the window limit is 10 packets
then in slow start mode the transmitter may start transmitting one packet followed by two
packets (before transmitting two packets, one packet ack has to be received), followed by
three packets and so on until 10 packets. But after reaching 10 packets, further
transmissions are restricted to one packet transmitted for one ack packet received. In a
simulation this appears as if the window is moving by one packet distance for every ack
packet received. On the receiver side also the window moves one packet for every packet
received. The sliding window method ensures that traffic congestion on the network is
avoided. The application layer will still be offering data for transmission to TCP without
worrying about the network traffic congestion issues as the TCP on sender and receiver side
implement sliding windows of packet buffer. The window size may vary dynamically
depending on network traffic.
For the highest possible throughput, it is important that the transmitter is not forced to stop
sending by the sliding window protocol earlier than one round-trip delay time (RTT). The
limit on the amount of data that it can send before stopping to wait for
an acknowledgment should be larger than the bandwidth-delay product of the
communications link. If it is not, the protocol will limit the effective bandwidth of the link.
1. Sender window might grow as it receives more frames to send and still has ones

un-ack'ed. Starts with nothing to send, then NA gives it frames to send.


Later, window may shrink as frames are ack-ed and NA has no more.
2. Receiver window constant size Receiver window size 1 means will only accept
them in order. Size n means will receive out of order (e.g. receive later ones after
earlier frame is lost) and then must buffer them before sending to NB (must send to
NB in order).

Examples of Error Detecting methods:

Parity bit: Simple example of error detection technique is parity bit. The parity bit
is chosen that the number of 1 bits in the code word is either even( for even parity)
or odd (for odd parity). For example when 10110101 is transmitted then for even
parity an 1 will be appended to the data and for odd parity a 0 will be appended. This
scheme can detect only single bits. So if two or more bits are changed then that can
not be detected.
Longitudinal Redundancy Checksum: Longitudinal Redundancy Checksum is an
error detecting scheme which overcomes the problem of two erroneous bits. In this
concept of parity bit is used but with slightly more intelligence. With each byte we
send one parity bit then send one additional byte which have the parity
corresponding to the each bit position of the sent bytes. So the parity bit is set in
both horizontal and vertical direction. If one bit get flipped we can tell which row

and column have error then we find the intersection of the two and determine the
erroneous bit. If 2 bits are in error and they are in the different column and row then
they can be detected. If the error are in the same column then the row will
differentiate and vice versa. Parity can detect the only odd number of errors. If they
are even and distributed in a fashion that in all direction then LRC may not be able
to find the error.

Cyclic

Redundancy

Checksum

(CRC):

We have an n-bit message. The sender adds a k-bit Frame Check Sequence (FCS) to
this message before sending. The resulting (n+k) bit message is divisible by some
(k+1) bit number. The receiver divides the message ((n+k)-bit) by the same (k+1)-bit
number and if there is no remainder, assumes that there was no error. How do we
choose
this
number?
For example, if k=12 then 1000000000000 (13-bit number) can be chosen, but this
is a pretty crappy choice. Because it will result in a zero remainder for all (n+k) bit
messages with the last 12 bits zero. Thus, any bits flipping beyond the last 12 go
undetected. If k=12, and we take 1110001000110 as the 13-bit number (incidentally,
in decimal representation this turns out to be 7238). This will be unable to detect
errors only if the corrupt message and original message have a difference of a
multiple of 7238. The probablilty of this is low, much lower than the probability that
anything beyond the last 12-bits flips. In practice, this number is chosen after
analyzing common network transmission errors and then selecting a number which
is likely to detect these common errors.
Q 6(a) Explain network devices repeaters, switches, hubs, and bridges.
Repeaters and hubs
A repeater is an electronic device that receives a signal, cleans it of unnecessary noise,
regenerates it, and retransmits it at a higher power level, or so that the signal can cover
longer distances without degradation. In most twisted pair Ethernet configurations, repeaters
are required for cable that runs longer than 100 meters. A repeater with multiple ports is
known as a hub. Repeaters work on the Physical Layer of the OSI model. Repeaters require
a small amount of time to regenerate the signal. This can cause a propagation delay which
can affect network communication when there are several repeaters in a row. Many network
architectures limit the number of repeaters that can be used in a row (e.g. Ethernet's 5-4-3
rule).
Bridges
A network bridge connects multiple network segments at the data link layer (layer 2) of
the OSI model. Bridges broadcast to all ports except the port on which the broadcast was
received. However, bridges do not promiscuously copy traffic to all ports, as hubs do, but
learn which MAC addresses are reachable through specific ports. Once the bridge associates
a port and an address, it will send traffic for that address to that port only.
Bridges learn the association of ports and addresses by examining the source address of
frames that it sees on various ports. Once a frame arrives through a port, its source address
is stored and the bridge assumes that MAC address is associated with that port. The first
time that a previously unknown destination address is seen, the bridge will forward the
frame to all ports other than the one on which the frame arrived.
Bridges come in three basic types:

Local bridges: Directly connect LANs


Remote bridges: Can be used to create a wide area network (WAN) link between
LANs. Remote bridges, where the connecting link is slower than the end networks,
largely have been replaced with routers.
Wireless bridges: Can be used to join LANs or connect remote stations to LANs.

Switches
A network switch is a device that forwards and filters OSI layer 2 datagrams (chunks of
data communication) between ports (connected cables) based on the MAC addresses in the
packets.[15] A switch is distinct from a hub in that it only forwards the frames to the ports
involved in the communication rather than all ports connected. A switch breaks the collision
domain but represents itself as a broadcast domain. Switches make forwarding decisions of
frames on the basis of MAC addresses. A switch normally has numerous ports, facilitating a
star topology for devices, and cascading additional switches.[16] Some switches are capable
of routing based on Layer 3 addressing or additional logical levels; these are called multilayer switches. The term switch is used loosely in marketing to encompass devices
including routers and bridges, as well as devices that may distribute traffic on load or by
application content (e.g., a Web URL identifier).
Repeater
Forwards every frame it receives
it is a generator,not an amplifier(i.e it removes noise & regenerates signal )

Bi-directional in nature

Useful in increasing ethernet size/length

Maximum of 5 Repeaters in an Ethernet

Hub

basically a multiport repeater

can be used to divide a single LAN into multiple levels of Hierarchy

Bridge

connect similar/dissimilar LANS


Designed to store and forward frame

Protocol independent

Transparent to End Stations

Operates in Layer-1 & Layer-2

uses a table for filtering/routing

does not change the Mac address in the frame

Q 6(b) Explain MAC sublayer of token ring.


802.5 Token Ring protocols specify that the MAC sub-layer must supply a 48-bit (6 byte)
address. The MAC address is most frequently represented as 12 hexadecimal digits. The MAC

address uniquely identifies a specificnetwork device and MAC addresses must be unique on a
given LAN (a network of computing devices in a single subnet of IP addresses). The first 12bit portion of the MAC address identifies the vendor of the network device, the last 12-bit portion
identifies the unique id of the device itself. When looking at a hexadecimal representation of the
MAC address, the first six hexadecimal digits identify the vendor and the last
six hexadecimal digits identify the specific network interface card.

Token Ring and IEEE 802.5 support two basic frame types: tokens and data/command
frames. Tokens are 3 bytes in length and consist of a start delimiter, an access control
byte, and an end delimiter. Data/command frames vary in size, depending on the size of
the Information field. Data frames carry information for upper-layer protocols, while
command frames contain control information and have no data for upper-layer protocols.

Token Frame Fields


Start Delimiter

Access Control

Ending delimiter

Token Frame contains three fields, each of which is 1 byte in length:

Start delimiter (1 byte): Alerts each station of the arrival of a token (or
data/command frame). This field includes signals that distinguish the byte from
the rest of the frame by violating the encoding scheme used elsewhere in the
frame.
Access-control (1 byte): Contains the Priority field (the most significant 3 bits)
and the Reservation field (the least significant 3 bits), as well as a token bit (used
to differentiate a token from a data/command frame) and a monitor bit (used by
the active monitor to determine whether a frame is circling the ring endlessly).
End delimiter (1 byte): Signals the end of the token or data/command frame.
This field also contains bits to indicate a damaged frame and identify the frame
that is the last in a logical sequence.

Data/Command Frame Fields


Start
Delimiter

Access
Control

Frame
Control

Destination
address

Source
address

Data

Frame check
sequence

End
Delimiter

Frame
Status

Data/command frames have the same three fields as Token Frames, plus several others.
The Data/command frame fields are described below:

Frame-control byte (1 byte)Indicates whether the frame contains data or control


information. In control frames, this byte specifies the type of control information.
Destination and source addresses (2-6 bytes)Consists of two 6-byte address
fields that identify the destination and source station addresses.
Data (up to 4500 bytes)Indicates that the length of field is limited by the ring
token holding time, which defines the maximum time a station can hold the token.
Frame-check sequence (FCS- 4 byte)Is filed by the source station with a
calculated value dependent on the frame contents. The destination station recalculates
the value to determine whether the frame was damaged in transit. If so, the frame is
discarded.
Frame Status (1 byte)This is the terminating field of a command/data frame. The
Frame Status field includes the address-recognized indicator and frame-copied
indicator.

Q 6(c) Explain FDDI


Fiber distributed data interface (FDDI), which is an optical data communication standard used
for long distance networks provides communication with fiber optic lines up to 200 kilometers at
a speed of 100 megabit per second (Mbps).
FDDI uses a dual-ring architecture with traffic on each ring flowing in opposite directions called
counter-rotating. The dual-rings consist of a primary and a secondary ring. The primary ring
oers up to 100 Mbps capacity. During normal operation, the primary ring is used for data
transmission, and the secondary ring remains idle and available for backup. The primary purpose

of the dual rings is to provide superior reliability and robustness. FDDI was later extended to
FDDI-2 for long distance voice and multimedia communication.
FDDI can have two types of Network Interface Cards, A and B, that connect to it.
1) Class A Network Interface Cards connect to both rings while class B Network Interface Cards
connect to only one ring.
2) Only class A cards can be used to heal broken rings. Thus the number of class A cards define
the fault tolerant characteristics of the network.
3) When an error occurs the nearest computer routes frames from the inner ring to the outer ring.
Q 7 (a) Imagine two LAN bridges, both connecting a pair of 802.4 networks . The first
bridge is faced with 1000 512-byte frames per second that must be forwarded. The second
is faced with 200 4096-byte frames per second. Which bridge do you think will need the
faster CPU? Discuss.
The bridge with 1000 512-byte frames will require a faster CPU since it has to do 1000
forwarding lookup and learning operations compared to 200 for the other bridge. It is the number
if frame headers not the size of the frame which governs how busy the bridge microprocessor is.
The 100 frames/sec bridge woul need a faster CPU. Although the other one has a higher
throughput,
the
1000 frames/sec bridge
has
more interrupts, more process
switches, more frames passed and more of everything that needs the CPU.
Q 7(b) Explain the disadvantages of CDMA?
Main disadvantage of CDMA is the degradation if the number of user will be increases an RF
channel because in cdma 10 users can acess at a time.
One major problem in CDMA technology is channel pollution, where signals from too many cell
sites are present in the subscribers phone but none of them is dominant. When this situation
arises the quality of the audio degrades. Another disadvantage in this technology when compared
to GSM is the lack of international roaming capabilities. The ability to upgrade or change to
another handset is not easy with this technology because the network service information for the
phone is put in the actual phone unlike GSM which uses SIM card for this. One another
disadvantage is the limited variety of the handset, because at present the major mobile companies
use GSM technology. Some other drawbacks :
Due to its proprietary nature, all of CDMA's flaws are not known to the engineering
community.
CDMA is relatively new, and the network is not as mature as GSM.
CDMA cannot offer international roaming, a large GSM advantage.
Q 7(c) Explain ALOHA

Q 8(a) What is the maximum burst length on an 155.2Mbps ATM ABR connection whose
PCR value is 200,000 & whose L value is 25 micro sec?
Q 8(b) Give three examples of protocol parameters that might be negotiated when a
connection is set up.
The negotiation can be on setting the following :
1)
2)
3)
4)

Window size
Maximum packet size (MTU or MSS)
Timer values.
The protocol parameters that may be negotiated during the setup of a connection
could be
A) Maximum packet size
B) Maximum transmission/reception speed
C) Quality of service standards.

Q 9 (a) Give two examples for which connection oriented service is appropriate. Now give
two examples for which connectionless service is best.
Connection-oriented communication is
a
data
communication
mode
in telecommunications whereby the devices at the end points use a protocol to establish an endto-end logical or physical connection before any data may be sent. Connection-oriented

protocol services are often but not always reliable network services, that provide
acknowledgment after successful delivery, and automatic repeat request functions in case of
missing data or detected bit-errors.
Circuit
mode communication,
for
example
the public
switched
telephone
network, ISDN, SONET/SDH and optical mesh networks, are examples of connection-oriented
communication. Circuit mode communication provides guarantees that data will arrive with
constant bandwidth and at constant delay.
Packet mode communication may also be connection-oriented, which is called virtual
circuit mode communication. Due to the packet switching, the communication may suffer from
variable bit rate and delay, due to varying traffic load and packet queue lengths. . Examples of
connection-oriented packet mode communication, i.e. virtual circuit mode communication:

The Transmission Control Protocol (TCP) is a connection-oriented reliable


protocol that is based on a datagram protocol (the IP protocol).
X.25 is a connection-oriented reliable network protocol.

Frame relay is a connection-oriented unreliable data link layer protocol.

GPRS

Asynchronous Transfer Mode

Multiprotocol Label Switching

The alternative to connection-oriented transmission is connectionless communication, also


known as datagram communication, in which data is sent in form of packets from one end point
to another without prior arrangement or signaling. Each data packet must contain complete
address information, since packets are routed individually and independently of each other,
possibly transmitted along different network paths.
Connectionless protocols are usually described as stateless because the end points have no
protocol-defined

List of connectionless protocols

Hypertext Transfer Protocol

IP

UDP

ICMP

IPX

TIPC

NetBEUI

DNS

SNMP

Q 9(b) Give an argument why the leaky bucket algorithm should allow just one packet per
tick, independent of how large the packet is.

. When variable sized packet are being used, it is often better to allow a fixed
number of bytes per tick rather than just one packet the leaky bucket consists of a
finite queue. When a packet aarives if there ia a room on the queue it is appended
to the queue, otherwise it is discarded. At every clock tick, one packet is
transmitted. The host is aalowed to put one packet per clock tick onto the network.
Again this can be enforced by the interface card or by OS this algorithm is mostly
used for traffic shaping. If leaky bucket allow more than one packet then
congestion will occurs.
If we send more than one packet per tick, independent of packet size, then
eventually you might see a potential increase of processing going on at the
nearest router by the leaky bucket. The more packets you are putting on the line
per tick for the router to handle, the more time it has to give to handle multiple
packets per clock tick. The router might get behind its work and become
congested.
Therefore, we would only want to put out one packet per clock tick.

You might also like