
Data Structure
Networking
RDBMS
Operating System
Java
MS Excel
iOS
HTML
CSS
Android
Python
C Programming
C++
C#
MongoDB
MySQL
Javascript
PHP
- Selected Reading
- UPSC IAS Exams Notes
- Developer's Best Practices
- Questions and Answers
- Effective Resume Writing
- HR Interview Questions
- Computer Glossary
- Who is Who
Found 1217 Articles for MCA

5K+ Views
Fletcher checksum is an error – detection technique that uses two checksums to determine single-bit errors in a message transmitted over network channels. It is a block code technique that was devised by John G. Fletcher in 1970s at Lawrence Livermore Labs, USA.The checksums are created based on the data values in the data blocks to be transmitted and appended to the data. When the receiver gets this data, the checksums are re-calculated and compared with the existing checksums. A non-match indicates an error.The error-detection capabilities of this method is nearly same as that of Cyclic Redundancy Check (CRC) but ... Read More

8K+ Views
Interleaving is a tool that is used to enhance existing error correcting codes so that they can be used to perform burst error corrections as well.Most error correcting codes (ECCs) are designed to correct random errors, i.e. error caused by additive noise that is independent of each other. Burst error are the errors that occur in a sequence or as groups. They are caused due to defects in storage media or disruption in communication signals due to external factors like lightning etc. Interleaving modifies the ECC or does some processing on the data after they are encoded by ECCs.Interleaving ProcedureDuring ... Read More

52K+ Views
A parity bit is a check bit, which is added to a block of data for error detection purposes. It is used to validate the integrity of the data. The value of the parity bit is assigned either 0 or 1 that makes the number of 1s in the message block either even or odd depending upon the type of parity. Parity check is suitable for single bit error detection only.The two types of parity checking areEven Parity − Here the total number of bits in the message is made even.Odd Parity − Here the total number of bits in ... Read More

8K+ Views
Errors and Error DetectionWhen bits are transmitted over the computer network, they are subject to get corrupted due to interference and network problems. The corrupted bits leads to spurious data being received by the receiver and are called errors.Error detection techniques are responsible for checking whether any error has occurred or not in the frame that has been transmitted via network. It does not take into account the number of error bits and the type of error.For error detection, the sender needs to send some additional bits along with the data bits. The receiver performs necessary checks based upon the ... Read More

48K+ Views
Errors and Error DetectionWhen bits are transmitted over the computer network, they are subject to get corrupted due to interference and network problems. The corrupted bits leads to spurious data being received by the receiver and are called errors.Error detection techniques are responsible for checking whether any error has occurred or not in the frame that has been transmitted via network. It does not take into account the number of error bits and the type of error.For error detection, the sender needs to send some additional bits along with the data bits. The receiver performs necessary checks based upon the ... Read More

26K+ Views
Error-detecting codes are a sequence of numbers generated by specific procedures for detecting errors in data that has been transmitted over computer networks.When bits are transmitted over the computer network, they are subject to get corrupted due to interference and network problems. The corrupted bits leads to spurious data being received by the receiver and are called errors.Error – detecting codes ensures messages to be encoded before they are sent over noisy channels. The encoding is done in a manner so that the decoder at the receiving end can detect whether there are errors in the incoming signal with high ... Read More

11K+ Views
Low - density parity check (LDPC) code is a linear error-correcting block code, suitable for error correction in large block sizes transmitted via very noisy channels.LDPC was developed by Robert G. Gallager, in his doctoral dissertation at the Massachusetts Institute of Technology in 1960. So, these codes are also known as Gallager codes.Encoding by Low-Density Parity Check CodesA low - density parity check (LFPC) code is specified by a parity-check matrix containing mostly 0s and a low density of 1s. The rows of the matrix represent the equations and the columns represent the bits in the codeword, i.e. code symbols.A ... Read More

10K+ Views
A challenging task in error correction is decoding the codewords that have been received via noisy channels. Before data is transmitted, the sender adds redundant bits or parity bits to the message forming codewords. The codewords are then transmitted via computer networks. The receiver checks the incoming codewords and performs the decoding or error correction process to retrieve the original data.If there are no errors, i.e. the codewords find an exact matching, then it is easy to decode the data by eliminating the parity bits. However, if a match is not found, then more complex decoding mechanisms are adopted.The two ... Read More

4K+ Views
Convolutional CodesConvolutional codes are error correcting codes where the data streams of indefinite lengths are encoded before transmission over noisy channels. The message streams are encoded by the sliding application of Boolean functions that generate a sequence of output bits.Convolutional codes were first introduced in 1955, by Elias. After that, there were many interim researches by many mathematicians. In 1973, Viterbi developed an algorithm for maximum likelihood decoding scheme, called Viterbi scheme that lead to modern convolutional codes.Parameters in Convolutional CodesFor generating a convolutional code, the information is passed sequentially through a linear finite-state shift register. The shift register comprises ... Read More

16K+ Views
Hamming code is a block code that is capable of detecting up to two simultaneous bit errors and correcting single-bit errors. It was developed by R.W. Hamming for error correction.In this coding method, the source encodes the message by inserting redundant bits within the message. These redundant bits are extra bits that are generated and inserted at specific positions in the message itself to enable error detection and correction. When the destination receives this message, it performs recalculations to detect errors and find the bit position that has error.Hamming Code for Single Error CorrectionThe procedure for single error correction by ... Read More