71 |
Electronic circuits designed to improve the time resoluion in nuclear lifetime studiesCraig, Edwin L. 03 June 2011 (has links)
In this research a unique linear pulse amplifier was developed that significantly reduces the spread in pulse amplitudes of those pulses selected from a Ge(Li) detector by a single channel analyzer. This circuit utilizes an operational amplifier with its closed-loop gain automatically controlled by a P-channel junction field.-effect transistor. The amplification is adjusted for each pulse such that the output pulses are constant in amplitude. The performance of the system was analyzed with a multichannel analyzer and it was shown that an improvement in pulse amplitude variation of as much as 29.1 percent was achieved.Ball State UniversityMuncie, IN 47306
|
72 |
Reliable high-throughput FPGA interconnect using source-synchronous surfing and wave pipeliningTeehan, Paul Leonard 05 1900 (has links)
FPGA clock frequencies are slow enough that only a fraction of the interconnect’s
bandwidth is used. By exploiting this bandwidth, the transfer of large amounts of
data can be greatly accelerated. Alternatively, it may also be possible to save area
on fixed-bandwidth links by using on-chip serial signaling. For datapath-intensive
designs which operate on words instead of bits, this can reduce wiring congestion
as well. This thesis proposes relatively simple circuit-level modifications to FPGA
interconnect to enable high-bandwidth communication. High-level area estimates
indicate a potential interconnect area savings of 10 to 60% when serial links are used.
Two interconnect pipelining techniques, wave pipelining and surfing, are adapted
to FPGAs and compared against each other and against regular FPGA interconnect
in terms of throughput, reliability, area, power, and latency. Source-synchronous
signaling is used to achieve high data rates with simple receiver design. Statistical
models for high-frequency power supply noise are developed and used to estimate the
probability of error of wave pipelined and surfing links as a function of link length
and operating speed. Surfing is generally found to be more reliable and less sensitive
to noise than wave pipelining. Simulation results in a 65nm process demonstrate a
throughput of 3Gbps per wire across a 50-stage, 25mm link.
|
73 |
Built-in proactive tuning for circuit aging and process variation resilienceShah, Nimay Shamik 15 May 2009 (has links)
VLSI circuits in nanometer VLSI technology experience significant variations -
intrinsic process variations and variations brought about by transistor degradation or
aging. These are generally embodied by yield loss or performance degradation over
operation time. Although the degradation can be compensated by the worst-case scenario
based over-design approach, it induces remarkable power overhead which is undesirable
in tightly power-constrained designs. Dynamic voltage scaling (DVS) is a more powerefficient
approach. However, its coarse granularity implies difficulty in handling finegrained
variations. These factors have contributed to the growing interest in poweraware
robust circuit design.
In this thesis, we propose a Built-In Proactive Tuning (BIPT) system, a lowpower
typical case design methodology based on dynamic prediction and prevention of
possible circuit timing errors. BIPT makes use of the canary circuit to predict the
variation induced performance degradation. The approach presented allows each circuit
block to autonomously tune its performance according to its own degree of variation.
The tuning is conducted offline, either at power on or periodically. A test pattern generator is included to reduce the uncertainty of the aging prediction due to different
input vectors.
The BIPT system is validated through SPICE simulations on benchmark circuits
with consideration of process variations and NBTI, a static stress based PMOS aging
effect. The experimental results indicate that to achieve the same variation resilience,
proposed BIPT system leads to 33% power savings in case of process variations as
compared to the over-design approach. In the case of aging resilience, the approach
proposed in this thesis leads to 40% less power than the approach of over-design while
30% less power as compared to DVS with NBTI effect modeling.
|
74 |
Algorithms for VLSI Circuit Optimization and GPU-Based ParallelizationLiu, Yifang 2010 May 1900 (has links)
This research addresses some critical challenges in various problems of VLSI design
automation, including sophisticated solution search on DAG topology, simultaneous
multi-stage design optimization, optimization on multi-scenario and multi-core
designs, and GPU-based parallel computing for runtime acceleration.
Discrete optimization for VLSI design automation problems is often quite complex,
due to the inconsistency and interference between solutions on reconvergent
paths in directed acyclic graph (DAG). This research proposes a systematic solution
search guided by a global view of the solution space. The key idea of the proposal
is joint relaxation and restriction (JRR), which is similar in spirit to mathematical
relaxation techniques, such as Lagrangian relaxation. Here, the relaxation and
restriction together provides a global view, and iteratively improves the solution.
Traditionally, circuit optimization is carried out in a sequence of separate optimization
stages. The problem with sequential optimization is that the best solution
in one stage may be worse for another. To overcome this difficulty, we take the approach
of performing multiple optimization techniques simultaneously. By searching
in the combined solution space of multiple optimization techniques, a broader view
of the problem leads to the overall better optimization result. This research takes
this approach on two problems, namely, simultaneous technology mapping and cell placement, and simultaneous gate sizing and threshold voltage assignment.
Modern processors have multiple working modes, which trade off between power
consumption and performance, or to maintain certain performance level in a powerefficient
way. As a result, the design of a circuit needs to accommodate different
scenarios, such as different supply voltage settings. This research deals with this
multi-scenario optimization problem with Lagrangian relaxation technique. Multiple
scenarios are taken care of simultaneously through the balance by Lagrangian multipliers.
Similarly, multiple objective and constraints are simultaneously dealt with by
Lagrangian relaxation. This research proposed a new method to calculate the subgradients
of the Lagrangian function, and solve the Lagrangian dual problem more
effectively.
Multi-core architecture also poses new problems and challenges to design automation.
For example, multiple cores on the same chip may have identical design
in some part, while differ from each other in the rest. In the case of buffer insertion,
the identical part have to be carefully optimized for all the cores with different
environmental parameters. This problem has much higher complexity compared to
buffer insertion on single cores. This research proposes an algorithm that optimizes
the buffering solution for multiple cores simultaneously, based on critical component
analysis.
Under the intensifying time-to-market pressure, circuit optimization not only
needs to find high quality solutions, but also has to come up with the result fast.
Recent advance in general purpose graphics processing unit (GPGPU) technology
provides massive parallel computing power. This research turns the complex computation
task of circuit optimization into many subtasks processed by parallel threads.
The proposed task partitioning and scheduling methods take advantage of the GPU
computing power, achieve significant speedup without sacrifice on the solution quality.
|
75 |
Design Techniques for High Speed Low Voltage and Low Power Non-Calibrated Pipeline Analog to Digital ConvertersAssaad, Rida Shawky 2009 December 1900 (has links)
The profound digitization of modern microelectronic modules made Analog-to-
Digital converters (ADC) key components in many systems. With resolutions up to
14bits and sampling rates in the 100s of MHz, the pipeline ADC is a prime candidate for
a wide range of applications such as instrumentation, communications and consumer
electronics. However, while past work focused on enhancing the performance of the
pipeline ADC from an architectural standpoint, little has been done to individually
address its fundamental building blocks. This work aims to achieve the latter by
proposing design techniques to improve the performance of these blocks with minimal
power consumption in low voltage environments, such that collectively high
performance is achieved in the pipeline ADC.
Towards this goal, a Recycling Folded Cascode (RFC) amplifier is proposed as
an enhancement to the general performance of the conventional folded cascode. Tested
in Taiwan Semiconductor Manufacturing Company (TSMC) 0.18?m Complementary
Metal Oxide Semiconductor (CMOS) technology, the RFC provides twice the
bandwidth, 8-10dB additional gain, more than twice the slew rate and improved noise performance over the conventional folded cascode-all at no additional power or silicon
area. The direct auto-zeroing offset cancellation scheme is optimized for low voltage
environments using a dual level common mode feedback (CMFB) circuit, and amplifier
differential offsets up to 50mV are effectively cancelled. Together with the RFC, the
dual level CMFB was used to implement a sample and hold amplifier driving a singleended
load of 1.4pF and using only 2.6mA; at 200MS/s better than 9bit linearity is
achieved. Finally a power conscious technique is proposed to reduce the kickback noise
of dynamic comparators without resorting to the use of pre-amplifiers. When all
techniques are collectively used to implement a 1Vpp 10bit 160MS/s pipeline ADC in
Semiconductor Manufacturing International Corporation (SMIC) 0.18[mu]m CMOS, 9.2
effective number of bits (ENOB) is achieved with a near Nyquist-rate full scale signal.
The ADC uses an area of 1.1mm2 and consumes 42mW in its analog core. Compared to
recent state-of-the-art implementations in the 100-200MS/s range, the presented pipeline
ADC uses the least power per conversion rated at 0.45pJ/conversion-step.
|
76 |
Analog integrated circuit design techniques for high-speed signal processing in communications systemsHernandez Garduno, David 15 May 2009 (has links)
This work presents design techniques for the implementation of high-speed analog
integrated circuits for wireless and wireline communications systems.
Limitations commonly found in high-speed switched-capacitor (SC) circuits used
for intermediate frequency (IF) filters in wireless receivers are explored. A model
to analyze the aliasing effects due to periodical non-uniform individual sampling,
a technique used in high-Q high-speed SC filters, is presented along with practical
expressions that estimate the power of the generated alias components. The results
are verified through circuit simulation of a 10.7MHz bandpass SC filter in TSMC
0.35mu-m CMOS technology. Implications on the use of this technique on the design of
IF filters are discussed.
To improve the speed at which SC networks can operate, a continuous-time
common-mode feedback (CMFB) with reduced loading capacitance is proposed. This
increases the achievable gain-bandwidth product (GBW) of fully-differential ampli-
fiers. The performance of the CMFB is demonstrated in the implementation of a
second-order 10.7MHz bandpass SC filter and compared with that of an identical
filter using the conventional switched-capacitor CMFB (SC-CMFB). The filter using
the continuous-time CMFB reduces the error due to finite GBW and slew rate to less
than 1% for clock frequencies up to 72MHz while providing a dynamic range of 59dB and a PSRR- > 22dB. The design of high-speed transversal equalizers for wireline transceivers requires the implementation of broadband delay lines. A delay line based on a third-order
linear-phase filter is presented for the implementation of a fractionally-spaced 1Gb/s
transversal equalizer. Two topologies for a broadband summing node which enable
the placement of the parasitic poles at the output of the transversal equalizer beyond
650MHz are presented. Using these cells, a 5-tap 1Gb/s equalizer was implemented
in TSMC 0.35mu-m CMOS technology. The results show a programmable frequency
response able to compensate up to 25dB loss at 500MHz. The eye-pattern diagrams
at 1Gb/s demonstrate the equalization of 15 meters and 23 meters of CAT5e twistedpair
cable, with a vertical eye-opening improvement from 0% (before the equalizer)
to 58% (after the equalizer) in the second case. The equalizer consumes 96mW and
an area of 630mu-m x 490mu-m.
|
77 |
Statistical design, analysis, and diagnosis of digital systems and embedded RF circuitsMatoglu, Erdem, January 2003 (has links) (PDF)
Thesis (Ph. D.)--School of Electrical and Computer Engineering, Georgia Institute of Technology, 2004. Directed by Madhaven Swaminathan. / Vita. Includes bibliographical references (leaves 154-163).
|
78 |
Physical design automation of structured high-performance integrated circuitsWard, Samuel Isaac 06 February 2014 (has links)
During the last forty years, advancements have pushed state-of-the-art placers to impressive performance placing modern multimillion gate designs in under an hour.
Wide industry adoption of the analytical framework indicates the quality of these approaches. However, modern designs present significant challenges to address the multi objective requirements for multi GHz designs. As devices continue to scale, wires become more resistive and power constraints significantly dampen performance gains, continued improvement in placement quality is necessary. Additionally, placement has become more challenging with the integration of multi-objective constraints such as routability, timing and reliability. These constraints intensify the challenge of producing quality placement solutions and must be handled carefully. Exasperating the issue, shrinking schedules and budgets are requiring increased automation by blurring the boundary between manual and automated placement. An example of this new hybrid design style is the integration of structured placement constraints within traditional ASIC style circuit structures.
Structure aware placement is a significant challenge to modern high performance physical design flows. The goal of this dissertation is to develop enhancements to state-of-the-art placement flows overcoming inadequacies for structured circuits.
A key observation is that specific structures exist where modern analytical placement frameworks significantly underperform. Accurately measuring suboptimality of a particular placement solution however is very challenging. As such, this work begins by designing a series of structured placement benchmarks. Generating placement for the benchmarks manually offers the opportunity to accurately quantify placer performance. Then, the latest generation of academic placers is compared to evaluate how the placers performed for these design styles. Results of this work lead to discoveries in three key aspects of modern physical design flows.
Datapath placement is the first aspect to be examined.
This work narrows the focus to specifically target datapath style circuits that contain high fanout nets. As the datapath benchmarks showed, these high fanout nets misdirect analytical placement flows. To effectively handle these circuit styles, this work proposes a new unified placement flow that simultaneously places random-logic and datapath cells. The flow is built on top of a leading academic force-directed placer and significantly improves the quality of datapath placement while leveraging the speed and flexibility of existing algorithms.
Effectively placing these circuits is not enough because in modern high performance designs, datapath circuits are often embedded within a larger ASIC style circuit and thus are unknown. As such, the next aspect of structured placement applies novel data learning techniques to train, predict, and evaluate potential structured circuits. Extracted circuits are mapped to groups that are aligned and simultaneously placed with random logic.
The third aspect that can be enhanced with improved structured placement impacts local clock tree synthesis. Performance and power requirements for multi-GHz microprocessors necessitate the use of a grid-based clock network methodology, wherein a global clock grid is overlaid on the entire die area followed by local buffered clock trees. This clock mesh methodology is driven by three key reasons: First, full trees do not offer enough performance for modern microprocessors. Second, clock trees offer significant power savings over full clock meshes. Third, local clock trees reduce the local clock wiring demands compared to full meshes at lower level metal layers. To meet these demands, a shift in latch placement methodology is proposed by using structured placement templates.
Placement configurations are identified a priori with significantly lower capacitance and the solutions are developed into placement templates.
Results through careful experimentation demonstrate the effectiveness of these approaches and the impact potential for modern high-speed designs. / text
|
79 |
Cost-effective test at system-levelKim, Hyun-moo, 1970- 09 June 2011 (has links)
Not available / text
|
80 |
Reliable high-throughput FPGA interconnect using source-synchronous surfing and wave pipeliningTeehan, Paul Leonard 05 1900 (has links)
FPGA clock frequencies are slow enough that only a fraction of the interconnect’s
bandwidth is used. By exploiting this bandwidth, the transfer of large amounts of
data can be greatly accelerated. Alternatively, it may also be possible to save area
on fixed-bandwidth links by using on-chip serial signaling. For datapath-intensive
designs which operate on words instead of bits, this can reduce wiring congestion
as well. This thesis proposes relatively simple circuit-level modifications to FPGA
interconnect to enable high-bandwidth communication. High-level area estimates
indicate a potential interconnect area savings of 10 to 60% when serial links are used.
Two interconnect pipelining techniques, wave pipelining and surfing, are adapted
to FPGAs and compared against each other and against regular FPGA interconnect
in terms of throughput, reliability, area, power, and latency. Source-synchronous
signaling is used to achieve high data rates with simple receiver design. Statistical
models for high-frequency power supply noise are developed and used to estimate the
probability of error of wave pipelined and surfing links as a function of link length
and operating speed. Surfing is generally found to be more reliable and less sensitive
to noise than wave pipelining. Simulation results in a 65nm process demonstrate a
throughput of 3Gbps per wire across a 50-stage, 25mm link.
|
Page generated in 0.0424 seconds