Thanks to visit codestin.com
Credit goes to Github.com

Skip to content
View tspeterkim's full-sized avatar

Organizations

@EnzymeAD

Block or report tspeterkim

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Pinned Loading

  1. flash-attention-minimal flash-attention-minimal Public

    Flash Attention in ~100 lines of CUDA (forward pass only)

    Cuda 1k 101

  2. cuda-1brc cuda-1brc Public

    My CUDA solution to the 1BRC

    Cuda 10 3

  3. mixed-precision-from-scratch mixed-precision-from-scratch Public

    Mixed precision training from scratch with Tensors and CUDA

    Python 28 4

  4. paged-attention-minimal paged-attention-minimal Public

    a minimal cache manager for PagedAttention, on top of llama3.

    Python 127 11

  5. insta-chat insta-chat Public

    DIY Instagram Chat Automation with Google Sheets

    HTML 243 31

  6. HazyResearch/ThunderKittens HazyResearch/ThunderKittens Public

    Tile primitives for speedy kernels

    Cuda 3k 217