Skip to content

January 2024

[Paper Reading] Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection

Last Updated on 2024-07-25 by Clay

Introduction

RAG-based LLM is a well-known architecture in current usage of Large Language Models (LLM). It involves "retrieval" to provide the model with prior knowledge that it lacks during training, enabling the model to answer questions in the context of specific information.

Read More »[Paper Reading] Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection

[Paper Reading] The Forward-Forward Algorithm: Some Preliminary Investigation

Last Updated on 2024-07-25 by Clay

Introduction

Paper link: https://arxiv.org/abs/2212.13345

The author of this research work is the renowned figure in the field of deep learning, Geoffrey Hinton, who was originally a researcher at Google Brain when he initially wrote this paper (he left in 2023).

Read More »[Paper Reading] The Forward-Forward Algorithm: Some Preliminary Investigation

[Python] Creating and Auto-Removing Temporary Directories with Python's `tempfile`

Last Updated on 2024-01-07 by Clay

Introduction

Today I was reading the training source code of DreamBooth, and I found the tempfile built-in module; I happened to reconstruct a script for model's layer merging, suddenly thought that the module would make the code more elegant, so I made this note.

Read More »[Python] Creating and Auto-Removing Temporary Directories with Python's `tempfile`

[Solved] Mistral Cannot Generate eos_token `<|im_end|>` After SFTTrainer Fine-tuning

Last Updated on 2024-01-02 by Clay

Problem

HuggingFace has published an article stating that the current LLM is best trained according to the ChatML format. In normal case, it will be generated according to three different roles of system, user and assistant. The format is as follows:

Read More »[Solved] Mistral Cannot Generate eos_token `<|im_end|>` After SFTTrainer Fine-tuning
Exit mobile version