Skip to content

AreelKhan/llm-lab

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

10 Commits
 
 
 
 

Repository files navigation

Welcome to my LLM lab!

This is a repo of my experiments and notes while learning about LLMs. I'm starting with a decent theoretical understanding of neural networks, and hands on experience training large models on distributed systems. I'm very comfortable with data and ML engineering.

What's done

I've completed:

I've read:

What's up next

Here are all the things I'd like to do:

Implementations:

  • Implement FlashAttention myself (in Cuda maybe?)
  • Implement FSDP myself (no idea how!?)

Experiments

  • Model efficiency experiments. Try out the following and benchmark performance changes:
    • Speculative decoding
    • Knowledge distillation
    • Quantization
    • Pruning
    • Sparsity low ran compression
    • etc
  • Play around with LLAMA models locally

Readings:

  • Depthwise Seperable Convolutions for NMT
  • One Model To Learn Them All
  • Self-Attention with Relative Position Representations
  • GANs
  • Stable Diffusion
  • KANs
  • Explore LLM evaluation
  • Explore LLM interpretability

About

Documented LLM learning journey.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published