Skip to content

lliu606/COSMOS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 

Repository files navigation

COSMOS

This is the official (preliminary) implementation of the COSMOS optimizer. For example, to use COSMOS for the pretraining of huggingface default GPT2 model, copy the COSMOS.py file to your codebase and use COSMOS optimizer in the following fashion:

from COSMOS import COSMOS_for_gpt

optimizer = COSMOS_for_gpt(lr = 1e-3, betas=(0.9, 0.95, 0.95), lr_ratio=0.25)

Here lr_ratio is the ratio of learning rate for hidden layers to that for embedding layer.

For llama pretraining the code is similar:

from COSMOS import COSMOS_for_llama

optimizer = COSMOS_for_llama(lr = 1e-3, betas=(0.9, 0.95, 0.95), lr_ratio=0.25)

We will release an improved version of the optimizer with lower memory usage and overhead.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages