Skip to content

n1o/LLMDistilery

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

llm_distil

Idea

Since there is this paper MiniLLM I do not see any problem of having an general framework for Knowledge Distillation. The only limitation is that 2 models should share the same Tokenizer, and maybe booth have to have the same CausalLM Denoising objective, but I thing it would be enough that they predict token by token.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published