Skip to main content

Posts

Showing posts from April, 2008

Random Access Memory

Random Access Memory (RAM) is a type of computer data storage. Its mainly used as main memory of a computer. RAM allows to access the data in any order, i.e random. The word random thus refers to the fact that any piece of data can be returned in a constant time, regardless of its physical location and whether or not it is related to the previous piece of data. You can access any memory cell directly if you know the row and column that intersect at that cell.     Most of the RAM chips are volatile types of memory, where the information is lost after the power is switched off. There are some non-volatile types such as, ROM, NOR-Flash. SRAM: Static Random Access Memory SRAM is static, which doesn't need to be periodically refreshed, as SRAM uses bistable latching circuitry to store each bit. SRAM is volatile memory. Each bit in an SRAM is stored on four transistors that form two cross-coupled inverters. This storage cell has two stable states which are used to denote...

Direct Memory Access

Direct memory access (DMA) is a feature of modern computers that allows certain hardware subsystems within the computer to access system memory for reading and/or writing independently of the central processing unit. Computers that have DMA channels can transfer data to and from devices with much less CPU overhead than computers without a DMA channel. Principle of DMA DMA is an essential feature of all modern computers, as it allows devices to transfer data without subjecting the CPU to a heavy overhead. Otherwise, the CPU would have to copy each piece of data from the source to the destination. This is typically slower than copying normal blocks of memory since access to I/O devices over a peripheral bus is generally slower than normal system RAM. During this time the CPU would be unavailable for any other tasks involving CPU bus access, although it could continue doing any work which did not require bus access. A DMA transfer essentially copies a block of memory from one device to an...

Sitemap

Digital Design Interview Questions - 1 Digital Design Interview Questions - 2 Digital Design Interview Questions - 3 Digital Design Interview Questions - 4 Digital Design Interview Questions - 5 Microprocessor Interview Questions - 1 Microprocessor Interview Questions - 2 Microprocessor Interview Questions - 3 Microprocessor Interview Questions - 4 Verilog Interview Questions - 1 Verilog Interview Questions - 2 Verilog Interview Questions - 3 VLSI Interview Questions - 1 VLSI Interview Questions - 2 VLSI Interview Questions - 3 VLSI Interview Questions - 4 VLSI Interview Questions - 5 Introduction to Digital Logic Design Introduction Mealy and Moore FSMs One-hot Encoding Introduction to Verilog HDL Basics: Lexical Tokens Basics: Data Types Modules Ports List Of Operators Gate-Level Modeling Dataflow Modeling Behavioral Modeling Tasks and Functions The VLSI Design Flow Complex Programmable Logic Device Direct Memory Access Field-Programmable Gate Array FPGA vs ASIC Parallel vs Seri...

Setup and Hold TIme

Every flip-flop has restrictive time regions around the active clock edge in which input should not change. We call them restrictive because any change in the input in this regions the output may be the expected one (*see below). It may be derived from either the old input, the new input, or even in between the two. Here we define, two very important terms in the digital clocking. Setup and Hold time. The setup time is the interval before the clock where the data must be held stable. The hold time is the interval after the clock where the data must be held stable. Hold time can be negative, which means the data can change slightly before the clock edge and still be properly captured. Most of the current day flip-flops has zero or negative hold time. In the above figure, the shaded region is the restricted region. The shaded region is divided into two parts by the dashed line. The left hand side part of shaded region is the setup time period and the right hand side part is the h...

Parallel vs Serial Data Transmission

Parallel and serial data transmission are most widely used data transfer techniques. Parallel transfer have been the preferred way for transfer data. But with serial data transmission we can achieve high speed and with some other advantages. In parallel transmission n bits are transfered simultaneously, hence we have to process each bit separately and line up them in an order at the receiver. Hence we have to convert parallel to serial form. This is known as overhead in parallel transmission. Signal skewing is the another problem with parallel data transmission. In the parallel communication, n bits leave at a time, but may not be received at the receiver at the same time, some may reach late than others. To overcome this problem, receiving end has to synchronize with the transmitter and must wait until all the bits are received. The greater the skew the greater the delay, if delay is increased that effects the speed. Another problem associated with parallel transmission is crosstalk. ...