paint-brush
QLoRA: Fine-Tuning Your LLMs With a Single GPUby@aibites
224 reads

QLoRA: Fine-Tuning Your LLMs With a Single GPU

by Shrinivasan Sankar6mJanuary 2nd, 2024
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

To fine-tune a LLAMA 65 billion parameter model, we need 780 GB of GPU memory. That is about 16 A40 GPUs. The solution lies with QLoRA where Q stands for Quantisation.

Company Mentioned

Mention Thumbnail
featured image - QLoRA: Fine-Tuning Your LLMs With a Single GPU
Shrinivasan Sankar HackerNoon profile picture
Shrinivasan Sankar

Shrinivasan Sankar

@aibites

I am an AI Reseach Engineer. I was formerly a researcher @Oxford VGG before founding the AI Bites YouTube channel.

0-item

STORY’S CREDIBILITY

Original Reporting

Original Reporting

This story contains new, firsthand information uncovered by the writer.

L O A D I N G
. . . comments & more!

About Author

Shrinivasan Sankar HackerNoon profile picture
Shrinivasan Sankar@aibites
I am an AI Reseach Engineer. I was formerly a researcher @Oxford VGG before founding the AI Bites YouTube channel.

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite
Also published here