Field-Programmable Gate Array
Field-Programmable Gate Array
"FPGA" redirects here. It is not to be confused with Flip-chip pin grid array.
1Technical design
2History
o 2.1Integration
o 2.2Soft Core
o 2.3Timelines
2.3.1Gates
2.3.2Market size
2.3.3Design starts
3Comparisons
o 3.1To ASICs
3.1.1Trends
o 3.2Complex Programmable Logic Devices (CPLD)
o 3.3Security considerations
4Applications
o 4.1Common applications
5Architecture
o 5.1Logic blocks
o 5.2Hard blocks
o 5.3Clocking
o 5.43D architectures
6Design and programming
7Basic process technology types
8Major manufacturers
9See also
10Notes
11References
12Further reading
13External links
Technical design[edit]
Contemporary field-programmable gate arrays (FPGAs) have large resources of logic
gates and RAM blocks to implement complex digital computations.[2] As FPGA designs employ
very fast I/O rates and bidirectional data buses, it becomes a challenge to verify correct timing
of valid data within setup time and hold time.
Floor planning enables resource allocation within FPGAs to meet these time constraints.
FPGAs can be used to implement any logical function that an ASIC can perform. The ability to
update the functionality after shipping, partial re-configuration of a portion of the design[3] and
the low non-recurring engineering costs relative to an ASIC design (notwithstanding the
generally higher unit cost), offer advantages for many applications.[1]
Some FPGAs have analog features in addition to digital functions. The most common analog
feature is a programmable slew rate on each output pin, allowing the engineer to set low rates
on lightly loaded pins that would otherwise ring or couple unacceptably, and to set higher rates
on heavily loaded pins on high-speed channels that would otherwise run too slowly.[4][5] Also
common are quartz-crystal oscillators, on-chip resistance-capacitance oscillators, and phase-
locked loops with embedded voltage-controlled oscillators used for clock generation and
management and for high-speed serializer-deserializer (SERDES) transmit clocks and receiver
clock recovery. Fairly common are differential comparators on input pins designed to be
connected to differential signaling channels. A few "mixed signal FPGAs" have integrated
peripheral analog-to-digital converters (ADCs) and digital-to-analog converters (DACs) with
analog signal conditioning blocks allowing them to operate as a system-on-a-
chip (SoC).[6] Such devices blur the line between an FPGA, which carries digital ones and
zeros on its internal programmable interconnect fabric, and field-programmable analog
array (FPAA), which carries analog values on its internal programmable interconnect fabric.
History[edit]
The FPGA industry sprouted from programmable read-only memory (PROM)
and programmable logic devices (PLDs). PROMs and PLDs both had the option of being
programmed in batches in a factory or in the field (field-programmable). However,
programmable logic was hard-wired between logic gates.[7]
Altera was founded in 1983 and delivered the industry's first reprogrammable logic device in
1984 – the EP300 – which featured a quartz window in the package that allowed users to shine
an ultra-violet lamp on the die to erase the EPROM cells that held the device configuration.[8] In
December 2015, Intel acquired Altera.
Xilinx co-founders Ross Freeman and Bernard Vonderschmitt invented the first commercially
viable field-programmable gate array in 1985 – the XC2064.[9] The XC2064 had programmable
gates and programmable interconnects between gates, the beginnings of a new technology
and market.[10] The XC2064 had 64 configurable logic blocks (CLBs), with two three-
input lookup tables (LUTs).[11] More than 20 years later, Freeman was entered into the National
Inventors Hall of Fame for his invention.[12][13]
In 1987, the Naval Surface Warfare Center funded an experiment proposed by Steve
Casselman to develop a computer that would implement 600,000 reprogrammable gates.
Casselman was successful and a patent related to the system was issued in 1992.[7]
Altera and Xilinx continued unchallenged and quickly grew from 1985 to the mid-1990s, when
competitors sprouted up, eroding significant market share. By 1993, Actel (now Microsemi)
was serving about 18 percent of the market.[10] By 2013, Altera (31 percent), Actel (10 percent)
and Xilinx (36 percent) together represented approximately 77 percent of the FPGA market.[14]
The 1990s were a period of rapid growth for FPGAs, both in circuit sophistication and the
volume of production. In the early 1990s, FPGAs were primarily used
in telecommunicationsand networking. By the end of the decade, FPGAs found their way into
consumer, automotive, and industrial applications.[15]
Companies like Microsoft have started to use FPGAs to accelerate high-performance,
computationally intensive systems (like the data centers that operate their Bing search engine),
due to the performance per watt advantage FPGAs deliver.[16] Microsoft began using FPGAs
to accelerate Bing in 2014, and in 2018 began deploying FPGAs across other data center
workloads for their Azure cloud computing platform.[17]
Integration[edit]
This section needs additional citations for verification. Please
help improve this article by adding citations to reliable sources.
Unsourced material may be challenged and removed.
Find sources: "Field-programmable gate
array" – news · newspapers · books · scholar · JSTOR (June 2017) (Learn how and
when to remove this template message)
In 2012 the coarse-grained architectural approach was taken a step further by combining
the logic blocks and interconnects of traditional FPGAs with embedded microprocessorsand
related peripherals to form a complete "system on a programmable chip". This work mirrors the
architecture created by Ron Perlof and Hana Potash of Burroughs Advanced Systems Group
in 1982 which combined a reconfigurable CPU architecture on a single chip called the
SB24.[citation needed]
Examples of such hybrid technologies can be found in the Xilinx Zynq-7000 All Programmable
SoC,[18] which includes a 1.0 GHz dual-core ARM Cortex-A9 MPCore
processor embedded within the FPGA's logic fabric[19] or in the Altera Arria V FPGA, which
includes an 800 MHz dual-core ARM Cortex-A9 MPCore. The Atmel FPSLIC is another such
device, which uses an AVR processor in combination with Atmel's programmable logic
architecture. The Microsemi SmartFusion devices incorporate an ARM Cortex-M3 hard
processor core (with up to 512 kB of flash and 64 kB of RAM) and analog peripherals such as
a multi-channel analog-to-digital converters and digital-to-analog converters to their flash
memory-based FPGA fabric.
Soft Core[edit]
An alternate approach to using hard-macro processors is to make use of soft processor IP
cores that are implemented within the FPGA logic. Nios II, MicroBlaze and Mico32 are
examples of popular softcore processors. Many modern FPGAs are programmed at "run time",
which has led to the idea of reconfigurable computing or reconfigurable systems – CPUs that
reconfigure themselves to suit the task at hand. Additionally, new, non-FPGA architectures are
beginning to emerge. Software-configurable microprocessors such as the Stretch S5000 adopt
a hybrid approach by providing an array of processor cores and FPGA-like programmable
cores on the same chip.
Timelines[edit]
Gates[edit]
2005: 80,000[23]
2008: 90,000[24]
Comparisons[edit]
To ASICs[edit]
Historically, FPGAs have been slower, less energy efficient and generally achieved less
functionality than their fixed ASIC counterparts. An older study[when?] showed that designs
implemented on FPGAs need on average 40 times as much area, draw 12 times as much
dynamic power, and run at one third the speed of corresponding ASIC implementations.[citation
needed]
More recently, FPGAs such as the Xilinx Virtex-7 or the Altera Stratix 5 have come to rival
corresponding ASIC and ASSP ("Application-specific standard part", such as a standalone
USB interface chip[25]) solutions by providing significantly reduced power usage, increased
speed, lower materials cost, minimal implementation real-estate, and increased possibilities for
re-configuration 'on-the-fly'. Where previously[when?] a design may have included 6 to 10 ASICs,
the same design can now be achieved using only one FPGA.[26]
Advantages of FPGAs include the ability to re-program when already deployed (i.e. "in the
field") to fix bugs, and often include shorter time to market and lower non-recurring
engineering costs. Vendors can also take a middle road via FPGA prototyping: developing their
prototype hardware on FPGAs, but manufacture their final version as an ASIC so that it can no
longer be modified after the design has been committed.
Trends[edit]
This article needs to be updated. In particular: It relies on a report
from February 2009, nearly 10 years ago; in which many changes have
occurred. Please update this section to reflect recent events or newly
available information. (December 2018)
Xilinx claimed that several market and technology dynamics are changing the ASIC/FPGA
paradigm as of February 2009:[27]
Applications[edit]
See also: Hardware acceleration
An FPGA can be used to solve any problem which is computable. This is trivially proven by the
fact that FPGAs can be used to implement a soft microprocessor, such as the
Xilinx MicroBlaze or Altera Nios II. Their advantage lies in that they are significantly faster for
some applications because of their parallel nature and optimality in terms of the number of
gates used for certain processes.[34]
FPGAs originally began as competitors to CPLDs to implement glue logic for printed circuit
boards. As their size, capabilities, and speed increased, FPGAs took over additional functions
to the point where some are now marketed as full systems on chips (SoCs). Particularly with
the introduction of dedicated multipliers into FPGA architectures in the late 1990s, applications
which had traditionally been the sole reserve of digital signal processor hardware (DSPs)
began to incorporate FPGAs instead.[35][36]
Another trend in the use of FPGAs is hardware acceleration, where one can use the FPGA to
accelerate certain parts of an algorithm and share part of the computation between the FPGA
and a generic processor.[2] The search engine Bing is noted for adopting FPGA acceleration for
its search algorithm in 2014.[37] As of 2018, FPGAs are seeing increased use as AI
accelerators including Microsoft's so-termed "Project Catapult"[17] and for accelerating artificial
neural networks for machine learning applications.
Traditionally,[when?] FPGAs have been reserved for specific vertical applications where the
volume of production is small. For these low-volume applications, the premium that companies
pay in hardware cost per unit for a programmable chip is more affordable than the
development resources spent on creating an ASIC. As of 2017, new cost and performance
dynamics have broadened the range of viable applications.
Common applications[edit]
This is a dynamic list and may never be able to satisfy particular standards for completeness.
You can help by expanding it with reliably sourced entries.
Architecture[edit]
Logic blocks[edit]
Main article: Logic block
Simplified example illustration of a logic cell (LUT – Lookup table, FA – Full adder, DFF – D-type flip-flop)
The most common FPGA architecture consists of an array of logic blocks,[note 1] I/O pads, and
routing channels.[1]Generally, all the routing channels have the same width (number of wires).
Multiple I/O pads may fit into the height of one row or the width of one column in the array.
An application circuit must be mapped into an FPGA with adequate resources. While the
number of CLBs/LABs and I/Os required is easily determined from the design, the number of
routing tracks needed may vary considerably even among designs with the same amount of
logic.[note 2]
For example, a crossbar switch requires much more routing than a systolic array with the same
gate count. Since unused routing tracks increase the cost (and decrease the performance) of
the part without providing any benefit, FPGA manufacturers try to provide just enough tracks so
that most designs that will fit in terms of lookup tables (LUTs) and I/Os can be routed.[note 2] This
is determined by estimates such as those derived from Rent's rule or by experiments with
existing designs. As of 2018, network-on-chip architectures for routing and interconnection are
being developed.
In general, a logic block consists of a few logical cells (called ALM, LE, slice etc.). A typical cell
consists of a 4-input LUT[timeframe?], a full adder (FA) and a D-type flip-flop, as shown above. The
LUTs are in this figure split into two 3-input LUTs. In normal mode those are combined into a 4-
input LUT through the left multiplexer (mux). In arithmetic mode, their outputs are fed to the
adder. The selection of mode is programmed into the middle MUX. The output can be
either synchronous or asynchronous, depending on the programming of the mux to the right, in
the figure example. In practice, entire or parts of the adder are stored as functions into the
LUTs in order to save space.[40][41][42]
Hard blocks[edit]
Modern FPGA families expand upon the above capabilities to include higher level functionality
fixed in silicon. Having these common functions embedded in the circuit reduces the area
required and gives those functions increased speed compared to building them from logical
primitives. Examples of these include multipliers, generic DSP blocks, embedded processors,
high speed I/O logic and embedded memories.
Higher-end FPGAs can contain high speed multi-gigabit transceivers and hard IP cores such
as processor cores, Ethernet medium access control units, PCI/PCI Express controllers, and
external memory controllers. These cores exist alongside the programmable fabric, but they
are built out of transistors instead of LUTs so they have ASIC-level performance and power
consumption without consuming a significant amount of fabric resources, leaving more of the
fabric free for the application-specific logic. The multi-gigabit transceivers also contain high
performance analog input and output circuitry along with high-speed serializers and
deserializers, components which cannot be built out of LUTs. Higher-level PHY[definition needed] layer
functionality such as line coding may or may not be implemented alongside the serializers and
deserializers in hard logic, depending on the FPGA.
Clocking[edit]
Most of the circuitry built inside of an FPGA is synchronous circuitry that requires a clock
signal. FPGAs contain dedicated global and regional routing networks for clock and reset so
they can be delivered with minimal skew. Also, FPGAs generally contain analog phase-locked
loop and/or delay-locked loop components to synthesize new clock frequencies as well as
attenuate jitter. Complex designs can use multiple clocks with different frequency and phase
relationships, each forming separate clock domains. These clock signals can be generated
locally by an oscillator or they can be recovered from a high speed serial data stream. Care
must be taken when building clock domain crossing circuitry to avoid metastability. FPGAs
generally contain block RAMs that are capable of working as dual port RAMs with different
clocks, aiding in the construction of building FIFOs and dual port buffers that connect differing
clock domains.
3D architectures[edit]
To shrink the size and power consumption of FPGAs, vendors such as Tabula and Xilinx have
introduced 3D or stacked architectures.[43][44] Following the introduction of its 28 nm 7-series
FPGAs, Xilinx said that several of the highest-density parts in those FPGA product lines will be
constructed using multiple dies in one package, employing technology developed for 3D
construction and stacked-die assemblies.
Xilinx's approach stacks several (three or four) active FPGA dies side-by-side on a
silicon interposer – a single piece of silicon that carries passive interconnect.[44][45] The multi-die
construction also allows different parts of the FPGA to be created with different process
technologies, as the process requirements are different between the FPGA fabric itself and the
very high speed 28 Gbit/s serial transceivers. An FPGA built in this way is called
a heterogeneous FPGA.[46]
Altera's heterogeneous approach involves using a single monolithic FPGA die and connecting
other die/technologies to the FPGA using Intel's embedded multi-die interconnect bridge
(EMIB) technology.[47]
Major manufacturers[edit]
In 2016, long-time industry rivals Xilinx and Altera (now an Intel subsidiary) were the FPGA
market leaders.[49] At that time, they controlled nearly 90 percent of the market.
Both Xilinx and Altera[note 3] provide proprietary electronic design automation software
for Windows and Linux (ISE/Vivado and Quartus) which enables engineers to design,
analyze, simulate, and synthesize (compile) their designs.[50][51]
Other manufacturers include:
Microchip:
o Microsemi (previously Actel), producing antifuse, flash-based, mixed-signal FPGAs;
acquired by Microchip in 2018
o Atmel, a second source of some Altera-compatible devices; also FPSLIC[clarification
needed]
mentioned above;[52] acquired by Microchip in 2016
Lattice Semiconductor, which manufactures low-power SRAM-based FPGAs featuring
integrated configuration flash, instant-on and live reconfiguration
o SiliconBlue Technologies, which provides extremely low power SRAM-based FPGAs
with optional integrated nonvolatile configuration memory; acquired by Lattice in 2011
QuickLogic,[53] which manufactures Ultra Low Power Sensor Hubs, extremely low powered,
low density SRAM-based FPGAs, with display bridges MIPI & RGB inputs, MIPI, RGB and
LVDS outputs
Achronix, manufacturing SRAM based FPGAS with 1.5 GHz fabric speed[54]
In March 2010, Tabula announced their FPGA technology that uses time-multiplexed logic and
interconnect that claims potential cost savings for high-density applications.[55] On March 24,
2015, Tabula officially shut down.[56]
On June 1, 2015, Intel announced it would acquire Altera for approximately $16.7 billion and
completed the acquisition on December 30, 2015.[57]