CV

Basics

Name Yibo Liu

Skills

Programming Languages
Python
TCL
Shell
Matlab
Verilog
JMP
EDA Tools
Synopsys PrimeTime
Synopsys VCS
Synopsys Design Compiler
Synopsys HSpice
Vivado HLS
Machine Learning
PyTorch
Data Mining
Information Retrieval
Recommender Systems
Data Science
Information Retrieval
Pandas
Embedded Systems
Linux Kernel
Scheduling
GPU and Paralel Programming
CUDA
GPGPU Architecture

Work

  • 2022.06 - 2022.09
    R&D Intern
    Synopsys Inc
    PrimeTime, PrimeShield, Workload-dependent Aged STA, Aging(BTI/HCI) Mission Profile
    • Assisted the development of a machine learning-accelerated workload-dependent aging-aware STA approach. The new approach counted in the DVFS usage by supporting the scalability of the BTI/HCI aging mission profiles, provided more accurate STA simulation on the actual path, and preventing overly pessimistic aging derates in the current PrimeTime tool. Meanwhile, a one-shot pipeline was developed with Python to overcome the productivity bottleneck from months to one day.

Education

  • 2019.09 - Present

    California, USA

    PhD
    University of California, Riverside
    Electrical Engineering
  • 2017.09 - 2019.06

    California, USA

    MS
    University of California, Riverside
    Computer Engineering
  • 2013.00 - 2017.06

    Wuhan, China

    BS
    Huazhong University of Science and Technology
    Electrical Engineering

Projects

  • 2021.09 - 2023.09
    Machine Learning (ML)-Accelerated On-chip Power Grid EM-Aware Voltage Failure Fixing
    Electronic Design Automation (EDA), Copper Interconnect Reliability, Electromigration(EM), EM/IR, On-chip Power Grid, Machine Learning, Generative AI Models, Optimization
    • Utilized Physics Informed Neural Network (PINN) to develop a model for solving partial differential equation based Korhonen equations, enabling efficient Electromigration(EM) stress analysis.
    • Applied generative AI models, including generative adversarial networks (GAN), variational autoencoder (VAE) and Vi-Transformer to quickly predict the power grid's EM-aware voltage.
    • Modeled EM-aware on-chip power grid fixing scenario as an optimization problem, accelerate the optimization solving process by Machine Learning model acquired sensitivity data (PyTorch AutoGradient) to skip the circuit analysis-based sensitivity calculation.
  • 2021.09 - 2023.09
    Improving Device Aging Reliability (TDDB, BTI/HCI) with Approximate Computing (AC) Divider
    EDA, Design Automation, Silicon Device Reliability (TDDB, BTI/HCI), Approximate Computing
    • Proposed a SOTA approximate stochastic computing divider design that achieves the highest accuracy (close to the theoretical upper limit) and the lowest energy cost among all divider designs.
    • Implemented the proposed divider design with Verilog. Functionality test by Synopsys VCS, Synthesis ASIC design with Synopsys Design Compiler.
  • 2019.09 - 2021.02
    Improving ML Hardware Accelerator Aging Reliability and Performance with AC Multiplier
    EDA, Silicon Device Reliability, Machine Learning Hardware Accelerator, Neural Network Quantization, FPGA
    • Proposed an approximate stochastic computing multiplier, which can compensate for the delay increase caused by BTI/HCI by reducing the demand computation cycles, and mitigating TDDB-induced errors by including error tolerance encoding.
    • Embedded the proposed approximate computing multiplier in the neural network accelerator, which enables the neural network accelerator to trade-off among throughput, accuracy, and power during the inference stage by adjusting the bit-width.
    • Started from neural network quantization by PyTorch, and implemented hardware design on Vivado FPGA with C++.

Languages

English
Fluent
Chinese
Native