Example Code and instructions to reproduce the Single GPU optimization demo on GPT2 shown in the repo main page #5259
Unanswered
LCChen
asked this question in
Community | Q&A
Replies: 1 comment
-
Some update : I have been able to run 1.2B model (gpt_xl) in examples/language/gpt/gemini directory using two GPUs. But I never be able to run gpt2-10b model using the same example code. Here is the configuration : ( I have minimize the batch to 1 to save the memory) root@a370836de98a:/workspace/ColossalAI/examples/language/gpt/gemini# bash run_gemini.sh
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Dear experts,
I have a two RTX3090 cards locally, and is excited about the data shown in the picture bellow (single GPU could train 18B model). I would like to try it locally myself on the same demo and help me understand the mechanism of the Colossal AI better. Can I get your helps to find the example code and instructions to run the 18B GPT2 model on one GPUs (3090) ?
Thank u a lot for the help.
Beta Was this translation helpful? Give feedback.
All reactions