Skip to content

Blog

[Linux] Efficient Directory Navigation On Linux: Mastering The `z` Command

Introduction

z command is something I’ve wanted to write about for a long time! However, I’ve been busy with AI training (company work) and model acceleration (personal interest), so I haven’t had the time. Let’s put it this way, if someone asks me to recommend essential tools for a Linux system, I would undoubtedly place z in my top ten list.

Read More »[Linux] Efficient Directory Navigation On Linux: Mastering The `z` Command

Troubleshooting Accelerated Inference of Gemma-2 on V100 GPUs Using vLLM

Problem Description

Recently, I’ve achieved some good application results by fine-tuning Gemma-2. However, I encountered various errors when deploying it on the client’s equipment, which was quite frustrating. Currently, there isn’t a systematic troubleshooting guide online, so I’m documenting it here.

Read More »Troubleshooting Accelerated Inference of Gemma-2 on V100 GPUs Using vLLM

OpenAI Triton Note (1): Vector Addition

Introduction

Triton is an open-source GPU programming language compiler released by OpenAI in 2021. Over recent years, it has become increasingly popular among developers for writing and optimizing parallel programs on GPUs. Compared to traditional libraries such as CUDA or OpenCL, Triton offers a Python-like syntax, making it more readable and easier to learn.

Read More »OpenAI Triton Note (1): Vector Addition

[PyTorch] BERT Architecture Implementation Note

Introduction

My advisor used to tell me, “Don’t just use other people’s libraries; you have to write your own to truly understand.” Back then, I didn’t have much time to implement various technologies I was interested in since I was fully occupied with my dissertation. However, I often recall his earnest advice even now, and it prompted me to finally attempt the implementation of BERT, a classic encoder-only transformer model.

Read More »[PyTorch] BERT Architecture Implementation Note