Dataset Viewer
Auto-converted to Parquet Duplicate
repo_name
stringlengths
2
22
repo_link
stringlengths
28
60
category
stringlengths
3
39
github_about_section
stringlengths
22
415
homepage_link
stringlengths
14
89
github_topic_closest_fit
stringlengths
3
28
contributors_all
int64
2
7.04k
contributors_2025
int64
0
2.38k
contributors_2024
int64
0
2.13k
contributors_2023
int64
0
1.92k
llvm-project
https://github.com/llvm/llvm-project
compiler
The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.
http://llvm.org
compiler
7,043
2,378
2,130
1,920
vllm
https://github.com/vllm-project/vllm
inference engine
A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
inference
2,302
1,369
579
145
pytorch
https://github.com/pytorch/pytorch
machine learning framework
Tensors and Dynamic neural networks in Python with strong GPU acceleration
https://pytorch.org
machine-learning
5,662
1,187
1,090
1,024
transformers
https://github.com/huggingface/transformers
multi-purpose library
Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
https://huggingface.co/transformers
machine-learning
3,728
860
769
758
sglang
https://github.com/sgl-project/sglang
inference engine
SGLang is a fast serving framework for large language models and vision language models.
https://docs.sglang.ai
inference
1,229
796
189
1
hhvm
https://github.com/facebook/hhvm
virtual machine
A virtual machine for executing programs written in Hack.
https://hhvm.com
virtual-machine
2,773
692
648
604
llama.cpp
https://github.com/ggml-org/llama.cpp
inference engine
LLM inference in C/C++
https://ggml.ai
inference
1,540
535
575
461
kubernetes
https://github.com/kubernetes/kubernetes
container orchestration
Production-Grade Container Scheduling and Management
https://kubernetes.io
kubernetes
5,157
541
499
565
tensorflow
https://github.com/tensorflow/tensorflow
machine learning framework
An Open Source Machine Learning Framework for Everyone
https://tensorflow.org
machine-learning
4,672
506
523
630
verl
https://github.com/volcengine/verl
reinforcement learning
verl: Volcano Engine Reinforcement Learning for LLMs
https://verl.readthedocs.io
deep-reinforcement-learning
567
454
10
0
rocm-systems
https://github.com/ROCm/rocm-systems
multi-purpose library
super repo for rocm systems projects
https://amd.com/en/products/software/rocm.html
amd
1,142
486
351
213
ray
https://github.com/ray-project/ray
multi-purpose library
Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.
https://ray.io
machine-learning
1,458
397
223
230
spark
https://github.com/apache/spark
data processing
Apache Spark - A unified analytics engine for large-scale data processing
https://spark.apache.org
data-processing
3,129
322
300
336
goose
https://github.com/block/goose
agent
an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM
https://block.github.io/goose
ai-agents
417
319
32
0
elasticsearch
https://github.com/elastic/elasticsearch
search engine
Free and Open Source, Distributed, RESTful Search Engine
https://elastic.co/products/elasticsearch
search-engine
2,337
316
284
270
jax
https://github.com/jax-ml/jax
scientific computing
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
https://docs.jax.dev
scientific-computing
1,033
316
280
202
modelcontextprotocol
https://github.com/modelcontextprotocol/modelcontextprotocol
mcp
Specification and documentation for the Model Context Protocol
https://modelcontextprotocol.io
mcp
361
301
42
0
executorch
https://github.com/pytorch/executorch
model compiler
On-device AI across mobile, embedded and edge for PyTorch
https://executorch.ai
inference
493
267
243
77
numpy
https://github.com/numpy/numpy
scientific computing
The fundamental package for scientific computing with Python.
https://numpy.org
scientific-computing
2,210
237
233
252
triton
https://github.com/triton-lang/triton
parallel computing dsl
Development repository for the Triton language and compiler
https://triton-lang.org
parallel-programming
558
233
206
159
modular
https://github.com/modular/modular
parallel computing
The Modular Platform (includes MAX & Mojo)
https://docs.modular.com
parallel-programming
418
222
205
99
scipy
https://github.com/scipy/scipy
scientific computing
SciPy library main repository
https://scipy.org
scientific-computing
2,008
213
251
245
ollama
https://github.com/ollama/ollama
inference engine
Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.
https://ollama.com
inference
596
202
314
97
trl
https://github.com/huggingface/trl
reinforcement learning
Train transformer language models with reinforcement learning.
http://hf.co/docs/trl
reinforcement-learning
472
189
154
122
flashinfer
https://github.com/flashinfer-ai/flashinfer
gpu kernels
FlashInfer: Kernel Library for LLM Serving
https://flashinfer.ai
attention
259
158
50
11
aiter
https://github.com/ROCm/aiter
gpu kernels
AI Tensor Engine for ROCm
https://rocm.blogs.amd.com/software-tools-optimization/aiter-ai-tensor-engine/README.html
null
216
145
10
0
LMCache
https://github.com/LMCache/LMCache
inference
Supercharge Your LLM with the Fastest KV Cache Layer
https://lmcache.ai
null
174
144
18
0
Mooncake
https://github.com/kvcache-ai/Mooncake
inference
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
https://kvcache-ai.github.io/Mooncake
inference
188
133
13
0
torchtitan
https://github.com/pytorch/torchtitan
training framework
A PyTorch native platform for training generative AI models
https://arxiv.org/abs/2410.06511
null
181
119
43
1
ao
https://github.com/pytorch/ao
quantization
PyTorch native quantization and sparsity for training and inference
https://pytorch.org/ao
quantization
211
114
100
5
ComfyUI
https://github.com/comfyanonymous/ComfyUI
user interface
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
https://comfy.org
stable-diffusion
309
108
119
94
unsloth
https://github.com/unslothai/unsloth
fine tuning
Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.
https://docs.unsloth.ai
fine-tuning
175
108
29
3
accelerate
https://github.com/huggingface/accelerate
training framework
A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support.
https://huggingface.co/docs/accelerate
null
410
97
124
149
terminal-bench
https://github.com/laude-institute/terminal-bench
benchmark
A benchmark for LLMs on complicated tasks in the terminal
https://tbench.ai
benchmark
96
96
0
0
DeepSpeed
https://github.com/deepspeedai/DeepSpeed
training framework
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
https://deepspeed.ai
null
458
96
134
165
milvus
https://github.com/milvus-io/milvus
vector database
Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search
https://milvus.io
vector-search
398
95
84
72
cutlass
https://github.com/NVIDIA/cutlass
parallel computing
CUDA Templates and Python DSLs for High-Performance Linear Algebra
https://docs.nvidia.com/cutlass/index.html
parallel-programming
264
94
64
66
tilelang
https://github.com/tile-ai/tilelang
parallel computing dsl
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
https://tilelang.com
parallel-programming
117
89
1
0
monarch
https://github.com/meta-pytorch/monarch
distributed computing
PyTorch Single Controller
https://meta-pytorch.org/monarch
null
101
85
0
0
Liger-Kernel
https://github.com/linkedin/Liger-Kernel
kernel examples
Efficient Triton Kernels for LLM Training
https://openreview.net/pdf?id=36SjAIT42G
triton
138
78
61
0
hipBLASLt
https://github.com/AMD-AGI/hipBLASLt
Basic Linear Algebra Subprograms (BLAS)
hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditional BLAS library
https://rocm.docs.amd.com/projects/hipBLASLt
matrix-multiplication
111
69
70
35
peft
https://github.com/huggingface/peft
fine tuning
PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
https://huggingface.co/docs/peft
null
286
69
111
115
ROCm
https://github.com/ROCm/ROCm
multi-purpose library
AMD ROCm Software - GitHub Home
https://rocm.docs.amd.com
null
167
67
61
44
mcp-agent
https://github.com/lastmile-ai/mcp-agent
mcp
Build effective agents using Model Context Protocol and simple workflow patterns
null
mcp
64
63
1
0
onnx
https://github.com/onnx/onnx
machine learning interoperability
Open standard for machine learning interoperability
https://onnx.ai
onnx
380
56
45
61
letta
https://github.com/letta-ai/letta
agent
Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.
https://docs.letta.com
ai-agents
159
57
75
47
helion
https://github.com/pytorch/helion
parallel computing dsl
A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.
https://helionlang.com
parallel-programming
66
49
0
0
openevolve
https://github.com/codelion/openevolve
evolutionary algorithm
Open-source implementation of AlphaEvolve
null
genetic-algorithm
51
46
0
0
lightning-thunder
https://github.com/Lightning-AI/lightning-thunder
model compiler
PyTorch compiler that accelerates training and inference. Get built-in optimizations for performance, memory, parallelism, and easily write your own.
null
null
79
44
47
29
truss
https://github.com/basetenlabs/truss
inference engine
The simplest way to serve AI/ML models in production
https://truss.baseten.co
inference
84
44
30
21
cuda-python
https://github.com/NVIDIA/cuda-python
middleware
CUDA Python: Performance meets Productivity
https://nvidia.github.io/cuda-python
parallel-programming
54
41
12
1
warp
https://github.com/NVIDIA/warp
spatial computing
A Python framework for accelerated simulation, data generation and spatial computing.
https://nvidia.github.io/warp
physics-simulation
89
40
29
17
metaflow
https://github.com/Netflix/metaflow
container orchestration
Build, Manage and Deploy AI/ML Systems
https://metaflow.org
null
132
37
35
28
numba
https://github.com/numba/numba
compiler
NumPy aware dynamic Python compiler using LLVM
https://numba.pydata.org
null
446
40
32
55
SWE-bench
https://github.com/SWE-bench/SWE-bench
benchmark
SWE-bench: Can Language Models Resolve Real-world Github Issues?
https://swebench.com
benchmark
66
33
37
9
Triton-distributed
https://github.com/ByteDance-Seed/Triton-distributed
distributed computing
Distributed Compiler based on Triton for Parallel Systems
https://triton-distributed.readthedocs.io
null
35
30
0
0
ThunderKittens
https://github.com/HazyResearch/ThunderKittens
parallel computing
Tile primitives for speedy kernels
https://hazyresearch.stanford.edu/blog/2024-10-29-tk2
parallel-programming
37
29
13
0
dstack
https://github.com/dstackai/dstack
container orchestration
dstack is an open-source control plane for running development, training, and inference jobs on GPUs-across hyperscalers, neoclouds, or on-prem.
https://dstack.ai
orchestration
69
28
42
14
ome
https://github.com/sgl-project/ome
container orchestration
OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)
http://docs.sglang.ai/ome
k8s
31
28
0
0
server
https://github.com/triton-inference-server/server
inference server
The Triton Inference Server provides an optimized cloud and edge inferencing solution.
https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html
inference
149
24
36
34
ccache
https://github.com/ccache/ccache
compiler
ccache - a fast compiler cache
https://ccache.dev
null
224
20
28
22
lapack
https://github.com/Reference-LAPACK/lapack
linear algebra
LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.
https://netlib.org/lapack
linear-algebra
187
23
25
42
quack
https://github.com/Dao-AILab/quack
kernel examples
A Quirky Assortment of CuTe Kernels
null
null
31
17
0
0
KernelBench
https://github.com/ScalingIntelligence/KernelBench
benchmark
KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems
https://scalingintelligence.stanford.edu/blogs/kernelbench
benchmark
21
16
3
0
reference-kernels
https://github.com/gpu-mode/reference-kernels
kernel examples
Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!
https://gpumode.com
null
23
16
0
0
synthetic-data-kit
https://github.com/meta-llama/synthetic-data-kit
synthetic data generation
Tool for generating high quality Synthetic datasets
https://pypi.org/project/synthetic-data-kit
synthetic-dataset-generation
15
15
0
0
tritonparse
https://github.com/meta-pytorch/tritonparse
performance testing
TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels
https://meta-pytorch.org/tritonparse
null
26
15
0
0
kernels
https://github.com/huggingface/kernels
gpu kernels
Load compute kernels from the Hub
null
null
23
14
2
0
Wan2.2
https://github.com/Wan-Video/Wan2.2
video generation
Wan: Open and Advanced Large-Scale Video Generative Models
https://wan.video
diffusion-models
16
14
0
0
Primus-Turbo
https://github.com/AMD-AGI/Primus-Turbo
training framework
Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision comp...
null
null
14
12
0
0
flashinfer-bench
https://github.com/flashinfer-ai/flashinfer-bench
benchmark
Building the Virtuous Cycle for AI-driven LLM Systems
https://bench.flashinfer.ai
benchmark
16
11
0
0
FTorch
https://github.com/Cambridge-ICCS/FTorch
middleware
A library for directly calling PyTorch ML models from Fortran.
https://cambridge-iccs.github.io/FTorch
machine-learning
22
12
8
9
TensorRT
https://github.com/NVIDIA/TensorRT
inference engine
NVIDIA TensorRT is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
https://developer.nvidia.com/tensorrt
null
104
10
18
19
TileIR
https://github.com/microsoft/TileIR
parallel computing dsl
TileIR (tile-ir) is a concise domain-specific IR designed to streamline the development of high-performance GPU/CPU kernels (e.g., GEMM, Dequant GEMM, FlashAttention, LinearAttention). By employing a Pythonic syntax with an underlying compiler infrastructure on top of TVM, TileIR allows developers to focus on productiv...
null
parallel-programming
10
10
1
0
kernels-community
https://github.com/huggingface/kernels-community
gpu kernels
Kernel sources for https://huggingface.co/kernels-community
https://huggingface.co/kernels-community
null
14
9
0
0
GEAK-agent
https://github.com/AMD-AGI/GEAK-agent
agent
It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.
null
ai-agents
17
9
0
0
intelliperf
https://github.com/AMDResearch/intelliperf
performance testing
Automated bottleneck detection and solution orchestration
https://arxiv.org/html/2508.20258v1
profiling
7
7
0
0
cudnn-frontend
https://github.com/NVIDIA/cudnn-frontend
parallel computing
cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it
https://developer.nvidia.com/cudnn
parallel-programming
14
6
5
1
BitBLAS
https://github.com/microsoft/BitBLAS
Basic Linear Algebra Subprograms (BLAS)
BitBLAS is a library to support mixed-precision matrix multiplications, especially for quantized LLM deployment.
null
matrix-multiplication
17
5
14
0
Self-Forcing
https://github.com/guandeh17/Self-Forcing
video generation
Official codebase for "Self Forcing: Bridging Training and Inference in Autoregressive Video Diffusion" (NeurIPS 2025 Spotlight)
https://self-forcing.github.io
diffusion-models
4
4
0
0
TritonBench
https://github.com/thunlp/TritonBench
benchmark
TritonBench: Benchmarking Large Language Model Capabilities for Generating Triton Operators
https://arxiv.org/abs/2502.14752
benchmark
3
3
0
0
hatchet
https://github.com/LLNL/hatchet
performance testing
Graph-indexed Pandas DataFrames for analyzing hierarchical performance data
https://llnl-hatchet.readthedocs.io
profiling
25
3
6
8
streamv2v
https://github.com/Jeff-LiangF/streamv2v
video generation
Official Pytorch implementation of StreamV2V.
https://jeff-liangf.github.io/projects/streamv2v
diffusion-models
7
3
6
0
mistral-inference
https://github.com/mistralai/mistral-inference
inference engine
Official inference library for Mistral models
https://mistral.ai
inference
30
2
17
14
omnitrace
https://github.com/ROCm/omnitrace
performance testing
Omnitrace: Application Profiling, Tracing, and Analysis
https://rocm.docs.amd.com/projects/omnitrace
profiling
16
2
12
2
IMO2025
https://github.com/harmonic-ai/IMO2025
formal mathematical reasoning
Harmonic's model Aristotle achieved gold medal performance, solving 5 problems. This repository contains the lean statement files and proofs for Problems 1-5.
https://harmonic.fun
lean
2
2
0
0
RaBitQ
https://github.com/gaoj0017/RaBitQ
quantization
[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search
https://github.com/VectorDB-NTU/RaBitQ-Library
nearest-neighbor-search
2
2
1
0
torchdendrite
https://github.com/sandialabs/torchdendrite
machine learning framework
Dendrites for PyTorch and SNNTorch neural networks
null
null
2
1
1
0
triton-runner
https://github.com/toyaix/triton-runner
debugger
Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.
https://triton-runner.org
null
2
1
0
0
triSYCL
https://github.com/triSYCL/triSYCL
parallel computing
Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group
https://trisycl.github.io/triSYCL/Doxygen/triSYCL/html/index.html
parallel-programming
31
0
1
3
StreamDiffusion
https://github.com/cumulo-autumn/StreamDiffusion
image generation
StreamDiffusion: A Pipeline-Level Solution for Real-Time Interactive Generation
https://arxiv.org/abs/2312.12491
diffusion-models
29
0
9
25
wandb
https://github.com/wandb/wandb
ml visualization
The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.
https://wandb.ai
null
235
46
67
62
aws-neuron-sdk
https://github.com/aws-neuron/aws-neuron-sdk
sdk
Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services
https://aws.amazon.com/ai/machine-learning/neuron
null
142
33
37
32
onnxruntime
https://github.com/microsoft/onnxruntime
machine learning interoperability
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
https://onnxruntime.ai
null
876
237
213
213
ort
https://github.com/pykeio/ort
machine learning interoperability
Fast ML inference & training for ONNX models in Rust
https://ort.pyke.io
null
70
25
20
21
Triton-distributed
https://github.com/ByteDance-Seed/Triton-distributed
distributed computing
Distributed Compiler based on Triton for Parallel Systems
https://triton-distributed.readthedocs.io
null
35
30
0
0
gemlite
https://github.com/dropbox/gemlite
gpu kernels
Fast low-bit matmul kernels in Triton
null
null
5
1
5
0
cutile-python
https://github.com/NVIDIA/cutile-python
parallel computing
cuTile is a programming model for writing parallel kernels for NVIDIA GPUs
https://docs.nvidia.com/cuda/cutile-python
null
19
10
0
0
tilus
https://github.com/NVIDIA/tilus
parallel computing
Tilus is a tile-level kernel programming language with explicit control over shared memory and registers.
https://nvidia.github.io/tilus
null
6
4
0
0
triton-windows
https://github.com/woct0rdho/triton-windows
parallel computing dsl
Fork of the Triton language and compiler for Windows support and easy installation
null
null
537
233
207
159
End of preview. Expand in Data Studio

PyTorch Conference 2025 GitHub Repos

I created a list of every GitHub repo mentioned during PyTorch Conference 2025 and Open Source AI Week.

Usage

Script used to update unique contributor count: https://github.com/Tyler-Hilbert/Update_GitHub_Dataset_PyTorchConference2025

Downloads last month
70