BoYang Hsueh

BoYang Hsueh joined NVIDIA in 2019 as a developer technology engineer. He focuses on large language model inference optimization and works as a core developer of NVIDIA TensorRT-LLM.
Bo Yang Hsueh

Posts by BoYang Hsueh

Generative AI / LLMs

Seamlessly Deploying a Swarm of LoRA Adapters with NVIDIA NIM

The latest state-of-the-art foundation large language models (LLMs) have billions of parameters and are pretrained on trillions of tokens of input text. They... 11 MIN READ
Simulation / Modeling / Design

Full-Stack Innovation Fuels Highest MLPerf Inference 2.1 Results for NVIDIA

Today’s AI-powered applications are enabling richer experiences, fueled by both larger and more complex AI models as well as the application of many models in... 14 MIN READ
Data Science

Accelerated Inference for Large Transformer Models Using NVIDIA Triton Inference Server

This is the first part of a two-part series discussing the NVIDIA Triton Inference Server’s FasterTransformer (FT) library, one of the fastest libraries for... 10 MIN READ
Data Science

Deploying GPT-J and T5 with NVIDIA Triton Inference Server

This is the second part of a two-part series about NVIDIA tools that allow you to run large transformer models for accelerated inference. For an introduction to... 16 MIN READ