TylerHilbert commited on
Commit
f0be80c
·
1 Parent(s): e92fe6c

Updated contributors values.

Browse files
PyTorchConference2025_GithubRepos.json CHANGED
@@ -6,7 +6,7 @@
6
  "github_about_section": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
7
  "homepage_link": "http://llvm.org",
8
  "github_topic_closest_fit": "compiler",
9
- "contributors_all": 7086,
10
  "contributors_2025": 2378,
11
  "contributors_2024": 2130,
12
  "contributors_2023": 1920,
@@ -19,7 +19,7 @@
19
  "github_about_section": "A high-throughput and memory-efficient inference and serving engine for LLMs",
20
  "homepage_link": "https://docs.vllm.ai",
21
  "github_topic_closest_fit": "inference",
22
- "contributors_all": 2351,
23
  "contributors_2025": 1369,
24
  "contributors_2024": 579,
25
  "contributors_2023": 145,
@@ -32,7 +32,7 @@
32
  "github_about_section": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
33
  "homepage_link": "https://pytorch.org",
34
  "github_topic_closest_fit": "machine-learning",
35
- "contributors_all": 5690,
36
  "contributors_2025": 1187,
37
  "contributors_2024": 1090,
38
  "contributors_2023": 1024,
@@ -45,7 +45,7 @@
45
  "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
46
  "homepage_link": "https://huggingface.co/transformers",
47
  "github_topic_closest_fit": "machine-learning",
48
- "contributors_all": 3742,
49
  "contributors_2025": 860,
50
  "contributors_2024": 769,
51
  "contributors_2023": 758,
@@ -58,7 +58,7 @@
58
  "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
59
  "homepage_link": "https://docs.sglang.ai",
60
  "github_topic_closest_fit": "inference",
61
- "contributors_all": 1267,
62
  "contributors_2025": 796,
63
  "contributors_2024": 189,
64
  "contributors_2023": 1,
@@ -84,7 +84,7 @@
84
  "github_about_section": "LLM inference in C/C++",
85
  "homepage_link": "https://ggml.ai",
86
  "github_topic_closest_fit": "inference",
87
- "contributors_all": 1573,
88
  "contributors_2025": 535,
89
  "contributors_2024": 575,
90
  "contributors_2023": 461,
@@ -97,11 +97,11 @@
97
  "github_about_section": "Production-Grade Container Scheduling and Management",
98
  "homepage_link": "https://kubernetes.io",
99
  "github_topic_closest_fit": "kubernetes",
100
- "contributors_all": 5158,
101
- "contributors_2025": 542,
102
  "contributors_2024": 499,
103
  "contributors_2023": 565,
104
- "contributors_2026_q1": 233
105
  },
106
  {
107
  "repo_name": "tensorflow",
@@ -110,7 +110,7 @@
110
  "github_about_section": "An Open Source Machine Learning Framework for Everyone",
111
  "homepage_link": "https://tensorflow.org",
112
  "github_topic_closest_fit": "machine-learning",
113
- "contributors_all": 4679,
114
  "contributors_2025": 506,
115
  "contributors_2024": 523,
116
  "contributors_2023": 630,
@@ -123,7 +123,7 @@
123
  "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
124
  "homepage_link": "https://verl.readthedocs.io",
125
  "github_topic_closest_fit": "deep-reinforcement-learning",
126
- "contributors_all": 584,
127
  "contributors_2025": 454,
128
  "contributors_2024": 10,
129
  "contributors_2023": 0,
@@ -136,11 +136,11 @@
136
  "github_about_section": "super repo for rocm systems projects",
137
  "homepage_link": "https://amd.com/en/products/software/rocm.html",
138
  "github_topic_closest_fit": "amd",
139
- "contributors_all": 1174,
140
- "contributors_2025": 498,
141
  "contributors_2024": 351,
142
  "contributors_2023": 213,
143
- "contributors_2026_q1": 250
144
  },
145
  {
146
  "repo_name": "ray",
@@ -149,7 +149,7 @@
149
  "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
150
  "homepage_link": "https://ray.io",
151
  "github_topic_closest_fit": "machine-learning",
152
- "contributors_all": 1473,
153
  "contributors_2025": 397,
154
  "contributors_2024": 223,
155
  "contributors_2023": 230,
@@ -162,7 +162,7 @@
162
  "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
163
  "homepage_link": "https://spark.apache.org",
164
  "github_topic_closest_fit": "data-processing",
165
- "contributors_all": 3139,
166
  "contributors_2025": 322,
167
  "contributors_2024": 300,
168
  "contributors_2023": 336,
@@ -175,7 +175,7 @@
175
  "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
176
  "homepage_link": "https://block.github.io/goose",
177
  "github_topic_closest_fit": "ai-agents",
178
- "contributors_all": 439,
179
  "contributors_2025": 319,
180
  "contributors_2024": 32,
181
  "contributors_2023": 0,
@@ -188,11 +188,11 @@
188
  "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
189
  "homepage_link": "https://elastic.co/products/elasticsearch",
190
  "github_topic_closest_fit": "search-engine",
191
- "contributors_all": 2344,
192
- "contributors_2025": 316,
193
  "contributors_2024": 284,
194
  "contributors_2023": 270,
195
- "contributors_2026_q1": 200
196
  },
197
  {
198
  "repo_name": "jax",
@@ -201,11 +201,11 @@
201
  "github_about_section": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
202
  "homepage_link": "https://docs.jax.dev",
203
  "github_topic_closest_fit": "scientific-computing",
204
- "contributors_all": 1037,
205
  "contributors_2025": 317,
206
  "contributors_2024": 280,
207
  "contributors_2023": 202,
208
- "contributors_2026_q1": 130
209
  },
210
  {
211
  "repo_name": "modelcontextprotocol",
@@ -214,11 +214,11 @@
214
  "github_about_section": "Specification and documentation for the Model Context Protocol",
215
  "homepage_link": "https://modelcontextprotocol.io",
216
  "github_topic_closest_fit": "mcp",
217
- "contributors_all": 368,
218
  "contributors_2025": 301,
219
  "contributors_2024": 42,
220
  "contributors_2023": 0,
221
- "contributors_2026_q1": 67
222
  },
223
  {
224
  "repo_name": "executorch",
@@ -227,7 +227,7 @@
227
  "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
228
  "homepage_link": "https://executorch.ai",
229
  "github_topic_closest_fit": "inference",
230
- "contributors_all": 503,
231
  "contributors_2025": 267,
232
  "contributors_2024": 243,
233
  "contributors_2023": 77,
@@ -240,7 +240,7 @@
240
  "github_about_section": "The fundamental package for scientific computing with Python.",
241
  "homepage_link": "https://numpy.org",
242
  "github_topic_closest_fit": "scientific-computing",
243
- "contributors_all": 2217,
244
  "contributors_2025": 237,
245
  "contributors_2024": 233,
246
  "contributors_2023": 252,
@@ -253,7 +253,7 @@
253
  "github_about_section": "Development repository for the Triton language and compiler",
254
  "homepage_link": "https://triton-lang.org",
255
  "github_topic_closest_fit": "parallel-programming",
256
- "contributors_all": 562,
257
  "contributors_2025": 233,
258
  "contributors_2024": 206,
259
  "contributors_2023": 159,
@@ -266,7 +266,7 @@
266
  "github_about_section": "The Modular Platform (includes MAX & Mojo)",
267
  "homepage_link": "https://docs.modular.com",
268
  "github_topic_closest_fit": "parallel-programming",
269
- "contributors_all": 419,
270
  "contributors_2025": 222,
271
  "contributors_2024": 205,
272
  "contributors_2023": 99,
@@ -279,11 +279,11 @@
279
  "github_about_section": "SciPy library main repository",
280
  "homepage_link": "https://scipy.org",
281
  "github_topic_closest_fit": "scientific-computing",
282
- "contributors_all": 2011,
283
  "contributors_2025": 213,
284
  "contributors_2024": 251,
285
  "contributors_2023": 245,
286
- "contributors_2026_q1": 74
287
  },
288
  {
289
  "repo_name": "ollama",
@@ -292,7 +292,7 @@
292
  "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
293
  "homepage_link": "https://ollama.com",
294
  "github_topic_closest_fit": "inference",
295
- "contributors_all": 599,
296
  "contributors_2025": 202,
297
  "contributors_2024": 314,
298
  "contributors_2023": 97,
@@ -305,7 +305,7 @@
305
  "github_about_section": "Train transformer language models with reinforcement learning.",
306
  "homepage_link": "http://hf.co/docs/trl",
307
  "github_topic_closest_fit": "reinforcement-learning",
308
- "contributors_all": 474,
309
  "contributors_2025": 189,
310
  "contributors_2024": 154,
311
  "contributors_2023": 122,
@@ -318,7 +318,7 @@
318
  "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
319
  "homepage_link": "https://flashinfer.ai",
320
  "github_topic_closest_fit": "attention",
321
- "contributors_all": 268,
322
  "contributors_2025": 158,
323
  "contributors_2024": 50,
324
  "contributors_2023": 11,
@@ -330,7 +330,7 @@
330
  "category": "gpu kernels",
331
  "github_about_section": "AI Tensor Engine for ROCm",
332
  "homepage_link": "https://rocm.blogs.amd.com/software-tools-optimization/aiter-ai-tensor-engine/README.html",
333
- "contributors_all": 227,
334
  "contributors_2025": 145,
335
  "contributors_2024": 10,
336
  "contributors_2023": 0,
@@ -342,7 +342,7 @@
342
  "category": "inference",
343
  "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
344
  "homepage_link": "https://lmcache.ai",
345
- "contributors_all": 175,
346
  "contributors_2025": 144,
347
  "contributors_2024": 18,
348
  "contributors_2023": 0,
@@ -355,7 +355,7 @@
355
  "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
356
  "homepage_link": "https://kvcache-ai.github.io/Mooncake",
357
  "github_topic_closest_fit": "inference",
358
- "contributors_all": 195,
359
  "contributors_2025": 133,
360
  "contributors_2024": 13,
361
  "contributors_2023": 0,
@@ -367,7 +367,7 @@
367
  "category": "training framework",
368
  "github_about_section": "A PyTorch native platform for training generative AI models",
369
  "homepage_link": "https://arxiv.org/abs/2410.06511",
370
- "contributors_all": 187,
371
  "contributors_2025": 119,
372
  "contributors_2024": 43,
373
  "contributors_2023": 1,
@@ -380,7 +380,7 @@
380
  "github_about_section": "PyTorch native quantization and sparsity for training and inference",
381
  "homepage_link": "https://pytorch.org/ao",
382
  "github_topic_closest_fit": "quantization",
383
- "contributors_all": 219,
384
  "contributors_2025": 114,
385
  "contributors_2024": 100,
386
  "contributors_2023": 5,
@@ -393,7 +393,7 @@
393
  "github_about_section": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
394
  "homepage_link": "https://comfy.org",
395
  "github_topic_closest_fit": "stable-diffusion",
396
- "contributors_all": 312,
397
  "contributors_2025": 108,
398
  "contributors_2024": 119,
399
  "contributors_2023": 94,
@@ -406,7 +406,7 @@
406
  "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
407
  "homepage_link": "https://docs.unsloth.ai",
408
  "github_topic_closest_fit": "fine-tuning",
409
- "contributors_all": 186,
410
  "contributors_2025": 108,
411
  "contributors_2024": 29,
412
  "contributors_2023": 3,
@@ -443,7 +443,7 @@
443
  "category": "training framework",
444
  "github_about_section": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
445
  "homepage_link": "https://deepspeed.ai",
446
- "contributors_all": 460,
447
  "contributors_2025": 96,
448
  "contributors_2024": 134,
449
  "contributors_2023": 165,
@@ -456,7 +456,7 @@
456
  "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
457
  "homepage_link": "https://milvus.io",
458
  "github_topic_closest_fit": "vector-search",
459
- "contributors_all": 399,
460
  "contributors_2025": 95,
461
  "contributors_2024": 84,
462
  "contributors_2023": 72,
@@ -469,7 +469,7 @@
469
  "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
470
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
471
  "github_topic_closest_fit": "parallel-programming",
472
- "contributors_all": 266,
473
  "contributors_2025": 94,
474
  "contributors_2024": 64,
475
  "contributors_2023": 66,
@@ -482,7 +482,7 @@
482
  "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
483
  "homepage_link": "https://tilelang.com",
484
  "github_topic_closest_fit": "parallel-programming",
485
- "contributors_all": 121,
486
  "contributors_2025": 89,
487
  "contributors_2024": 1,
488
  "contributors_2023": 0,
@@ -494,7 +494,7 @@
494
  "category": "distributed computing",
495
  "github_about_section": "PyTorch Single Controller",
496
  "homepage_link": "https://meta-pytorch.org/monarch",
497
- "contributors_all": 103,
498
  "contributors_2025": 85,
499
  "contributors_2024": 0,
500
  "contributors_2023": 0,
@@ -507,7 +507,7 @@
507
  "github_about_section": "Efficient Triton Kernels for LLM Training",
508
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
509
  "github_topic_closest_fit": "triton",
510
- "contributors_all": 140,
511
  "contributors_2025": 78,
512
  "contributors_2024": 61,
513
  "contributors_2023": 0,
@@ -532,7 +532,7 @@
532
  "category": "fine tuning",
533
  "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
534
  "homepage_link": "https://huggingface.co/docs/peft",
535
- "contributors_all": 292,
536
  "contributors_2025": 69,
537
  "contributors_2024": 111,
538
  "contributors_2023": 115,
@@ -544,7 +544,7 @@
544
  "category": "multi-purpose library",
545
  "github_about_section": "AMD ROCm Software - GitHub Home",
546
  "homepage_link": "https://rocm.docs.amd.com",
547
- "contributors_all": 168,
548
  "contributors_2025": 67,
549
  "contributors_2024": 61,
550
  "contributors_2023": 44,
@@ -569,7 +569,7 @@
569
  "github_about_section": "Open standard for machine learning interoperability",
570
  "homepage_link": "https://onnx.ai",
571
  "github_topic_closest_fit": "onnx",
572
- "contributors_all": 382,
573
  "contributors_2025": 56,
574
  "contributors_2024": 45,
575
  "contributors_2023": 61,
@@ -595,7 +595,7 @@
595
  "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
596
  "homepage_link": "https://helionlang.com",
597
  "github_topic_closest_fit": "parallel-programming",
598
- "contributors_all": 70,
599
  "contributors_2025": 49,
600
  "contributors_2024": 0,
601
  "contributors_2023": 0,
@@ -631,7 +631,7 @@
631
  "github_about_section": "The simplest way to serve AI/ML models in production",
632
  "homepage_link": "https://truss.baseten.co",
633
  "github_topic_closest_fit": "inference",
634
- "contributors_all": 84,
635
  "contributors_2025": 44,
636
  "contributors_2024": 30,
637
  "contributors_2023": 21,
@@ -644,7 +644,7 @@
644
  "github_about_section": "CUDA Python: Performance meets Productivity",
645
  "homepage_link": "https://nvidia.github.io/cuda-python",
646
  "github_topic_closest_fit": "parallel-programming",
647
- "contributors_all": 54,
648
  "contributors_2025": 41,
649
  "contributors_2024": 12,
650
  "contributors_2023": 1,
@@ -657,7 +657,7 @@
657
  "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
658
  "homepage_link": "https://nvidia.github.io/warp",
659
  "github_topic_closest_fit": "physics-simulation",
660
- "contributors_all": 90,
661
  "contributors_2025": 40,
662
  "contributors_2024": 29,
663
  "contributors_2023": 17,
@@ -669,7 +669,7 @@
669
  "category": "container orchestration",
670
  "github_about_section": "Build, Manage and Deploy AI/ML Systems",
671
  "homepage_link": "https://metaflow.org",
672
- "contributors_all": 132,
673
  "contributors_2025": 37,
674
  "contributors_2024": 35,
675
  "contributors_2023": 28,
@@ -681,7 +681,7 @@
681
  "category": "compiler",
682
  "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
683
  "homepage_link": "https://numba.pydata.org",
684
- "contributors_all": 449,
685
  "contributors_2025": 40,
686
  "contributors_2024": 32,
687
  "contributors_2023": 55,
@@ -745,7 +745,7 @@
745
  "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
746
  "homepage_link": "http://docs.sglang.ai/ome",
747
  "github_topic_closest_fit": "k8s",
748
- "contributors_all": 31,
749
  "contributors_2025": 28,
750
  "contributors_2024": 0,
751
  "contributors_2023": 0,
@@ -783,7 +783,7 @@
783
  "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
784
  "homepage_link": "https://netlib.org/lapack",
785
  "github_topic_closest_fit": "linear-algebra",
786
- "contributors_all": 187,
787
  "contributors_2025": 23,
788
  "contributors_2024": 25,
789
  "contributors_2023": 42,
@@ -794,7 +794,7 @@
794
  "repo_link": "https://github.com/Dao-AILab/quack",
795
  "category": "kernel examples",
796
  "github_about_section": "A Quirky Assortment of CuTe Kernels",
797
- "contributors_all": 35,
798
  "contributors_2025": 17,
799
  "contributors_2024": 0,
800
  "contributors_2023": 0,
@@ -819,7 +819,7 @@
819
  "category": "kernel examples",
820
  "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
821
  "homepage_link": "https://gpumode.com",
822
- "contributors_all": 24,
823
  "contributors_2025": 16,
824
  "contributors_2024": 0,
825
  "contributors_2023": 0,
@@ -844,7 +844,7 @@
844
  "category": "performance testing",
845
  "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
846
  "homepage_link": "https://meta-pytorch.org/tritonparse",
847
- "contributors_all": 27,
848
  "contributors_2025": 15,
849
  "contributors_2024": 0,
850
  "contributors_2023": 0,
@@ -855,7 +855,7 @@
855
  "repo_link": "https://github.com/huggingface/kernels",
856
  "category": "gpu kernels",
857
  "github_about_section": "Load compute kernels from the Hub",
858
- "contributors_all": 27,
859
  "contributors_2025": 14,
860
  "contributors_2024": 2,
861
  "contributors_2023": 0,
@@ -879,7 +879,7 @@
879
  "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
880
  "category": "training framework",
881
  "github_about_section": "Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision computation (FP8), and compute–communication overlap kernels.",
882
- "contributors_all": 14,
883
  "contributors_2025": 12,
884
  "contributors_2024": 0,
885
  "contributors_2023": 0,
@@ -941,7 +941,7 @@
941
  "category": "gpu kernels",
942
  "homepage_link": "https://huggingface.co/kernels-community",
943
  "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
944
- "contributors_all": 15,
945
  "contributors_2025": 9,
946
  "contributors_2024": 0,
947
  "contributors_2023": 0,
@@ -953,7 +953,7 @@
953
  "category": "agent",
954
  "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
955
  "github_topic_closest_fit": "ai-agents",
956
- "contributors_all": 20,
957
  "contributors_2025": 9,
958
  "contributors_2024": 0,
959
  "contributors_2023": 0,
@@ -979,7 +979,7 @@
979
  "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
980
  "homepage_link": "https://developer.nvidia.com/cudnn",
981
  "github_topic_closest_fit": "parallel-programming",
982
- "contributors_all": 14,
983
  "contributors_2025": 6,
984
  "contributors_2024": 5,
985
  "contributors_2023": 1,
@@ -1056,11 +1056,11 @@
1056
  "github_about_section": "Official inference library for Mistral models",
1057
  "homepage_link": "https://mistral.ai",
1058
  "github_topic_closest_fit": "inference",
1059
- "contributors_all": 30,
1060
  "contributors_2025": 2,
1061
  "contributors_2024": 17,
1062
  "contributors_2023": 14,
1063
- "contributors_2026_q1": 1
1064
  },
1065
  {
1066
  "repo_name": "omnitrace",
@@ -1156,7 +1156,7 @@
1156
  "category": "ml visualization",
1157
  "github_about_section": "The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.",
1158
  "homepage_link": "https://wandb.ai",
1159
- "contributors_all": 238,
1160
  "contributors_2025": 46,
1161
  "contributors_2024": 67,
1162
  "contributors_2023": 62,
@@ -1168,7 +1168,7 @@
1168
  "category": "sdk",
1169
  "github_about_section": "Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services",
1170
  "homepage_link": "https://aws.amazon.com/ai/machine-learning/neuron",
1171
- "contributors_all": 145,
1172
  "contributors_2025": 33,
1173
  "contributors_2024": 37,
1174
  "contributors_2023": 32,
@@ -1180,7 +1180,7 @@
1180
  "category": "machine learning interoperability",
1181
  "github_about_section": "ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator",
1182
  "homepage_link": "https://onnxruntime.ai",
1183
- "contributors_all": 877,
1184
  "contributors_2025": 237,
1185
  "contributors_2024": 213,
1186
  "contributors_2023": 213,
@@ -1192,7 +1192,7 @@
1192
  "category": "machine learning interoperability",
1193
  "github_about_section": "Fast ML inference & training for ONNX models in Rust",
1194
  "homepage_link": "https://ort.pyke.io",
1195
- "contributors_all": 70,
1196
  "contributors_2025": 25,
1197
  "contributors_2024": 20,
1198
  "contributors_2023": 21,
@@ -1227,7 +1227,7 @@
1227
  "category": "parallel computing",
1228
  "github_about_section": "cuTile is a programming model for writing parallel kernels for NVIDIA GPUs",
1229
  "homepage_link": "https://docs.nvidia.com/cuda/cutile-python",
1230
- "contributors_all": 20,
1231
  "contributors_2025": 10,
1232
  "contributors_2024": 0,
1233
  "contributors_2023": 0,
@@ -1261,7 +1261,7 @@
1261
  "repo_link": "https://github.com/fla-org/flash-linear-attention",
1262
  "category": "gpu kernels",
1263
  "github_about_section": "Efficient implementations of state-of-the-art linear attention models",
1264
- "contributors_all": 93,
1265
  "contributors_2025": 64,
1266
  "contributors_2024": 22,
1267
  "contributors_2023": 3,
@@ -1273,7 +1273,7 @@
1273
  "category": "distributed computing",
1274
  "github_about_section": "Optimized primitives for collective multi-GPU communication",
1275
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
1276
- "contributors_all": 84,
1277
  "contributors_2026_q1": 26,
1278
  "contributors_2025": 10,
1279
  "contributors_2024": 5,
@@ -1320,7 +1320,7 @@
1320
  "category": "data compression",
1321
  "github_about_section": "A novel data compression framework",
1322
  "homepage_link": "https://openzl.org",
1323
- "contributors_all": 39,
1324
  "contributors_2026_q1": 24,
1325
  "contributors_2025": 22,
1326
  "contributors_2024": 0,
@@ -1344,7 +1344,7 @@
1344
  "category": "reinforcement learning",
1345
  "github_about_section": "AllenAI's post-training codebase",
1346
  "homepage_link": "https://allenai.github.io/open-instruct/",
1347
- "contributors_all": 57,
1348
  "contributors_2026_q1": 12,
1349
  "contributors_2025": 28,
1350
  "contributors_2024": 24,
@@ -1355,7 +1355,7 @@
1355
  "repo_link": "https://github.com/PrimeIntellect-ai/prime-rl",
1356
  "category": "reinforcement learning",
1357
  "github_about_section": "Agentic RL Training at Scale",
1358
- "contributors_all": 58,
1359
  "contributors_2026_q1": 29,
1360
  "contributors_2025": 40,
1361
  "contributors_2024": 8,
@@ -1367,7 +1367,7 @@
1367
  "category": "reinforcement learning",
1368
  "github_about_section": "SkyRL: A Modular Full-stack RL Library for LLMs",
1369
  "homepage_link": "https://docs.skyrl.ai/docs",
1370
- "contributors_all": 77,
1371
  "contributors_2026_q1": 29,
1372
  "contributors_2025": 57,
1373
  "contributors_2024": 0,
@@ -1403,7 +1403,7 @@
1403
  "category": "world model",
1404
  "github_about_section": "Cosmos-Predict2.5, the latest version of the Cosmos World Foundation Models (WFMs) family, specialized for simulating and predicting the future state of the world in the form of video.",
1405
  "homepage_link": "https://research.nvidia.com/labs/cosmos-lab/cosmos-predict2.5",
1406
- "contributors_all": 13,
1407
  "contributors_2026_q1": 7,
1408
  "contributors_2025": 9,
1409
  "contributors_2024": 0,
@@ -1415,7 +1415,7 @@
1415
  "category": "reinforcement learning",
1416
  "github_about_section": "The RL Bridge for LLM-based Agent Applications. Made Simple & Flexible.",
1417
  "homepage_link": "https://www.inclusion-ai.org/AReaL",
1418
- "contributors_all": 89,
1419
  "contributors_2026_q1": 37,
1420
  "contributors_2025": 63,
1421
  "contributors_2024": 0,
@@ -1427,7 +1427,7 @@
1427
  "category": "reinforcement learning",
1428
  "github_about_section": "RLinf: Reinforcement Learning Infrastructure for Embodied and Agentic AI",
1429
  "homepage_link": "https://rlinf.readthedocs.io",
1430
- "contributors_all": 76,
1431
  "contributors_2026_q1": 52,
1432
  "contributors_2025": 35,
1433
  "contributors_2024": 0,
@@ -1439,7 +1439,7 @@
1439
  "category": "reinforcement learning",
1440
  "github_about_section": "An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models",
1441
  "homepage_link": "https://alibaba.github.io/ROLL/",
1442
- "contributors_all": 78,
1443
  "contributors_2026_q1": 30,
1444
  "contributors_2025": 60,
1445
  "contributors_2024": 0,
@@ -1450,6 +1450,11 @@
1450
  "repo_link": "https://github.com/allenai/OLMo-core",
1451
  "category": "training framework",
1452
  "github_about_section": "PyTorch building blocks for the OLMo ecosystem",
1453
- "homepage_link": "https://olmo-core.readthedocs.io/en/latest/"
 
 
 
 
 
1454
  }
1455
  ]
 
6
  "github_about_section": "The LLVM Project is a collection of modular and reusable compiler and toolchain technologies.",
7
  "homepage_link": "http://llvm.org",
8
  "github_topic_closest_fit": "compiler",
9
+ "contributors_all": 7170,
10
  "contributors_2025": 2378,
11
  "contributors_2024": 2130,
12
  "contributors_2023": 1920,
 
19
  "github_about_section": "A high-throughput and memory-efficient inference and serving engine for LLMs",
20
  "homepage_link": "https://docs.vllm.ai",
21
  "github_topic_closest_fit": "inference",
22
+ "contributors_all": 2436,
23
  "contributors_2025": 1369,
24
  "contributors_2024": 579,
25
  "contributors_2023": 145,
 
32
  "github_about_section": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
33
  "homepage_link": "https://pytorch.org",
34
  "github_topic_closest_fit": "machine-learning",
35
+ "contributors_all": 5769,
36
  "contributors_2025": 1187,
37
  "contributors_2024": 1090,
38
  "contributors_2023": 1024,
 
45
  "github_about_section": "Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.",
46
  "homepage_link": "https://huggingface.co/transformers",
47
  "github_topic_closest_fit": "machine-learning",
48
+ "contributors_all": 3783,
49
  "contributors_2025": 860,
50
  "contributors_2024": 769,
51
  "contributors_2023": 758,
 
58
  "github_about_section": "SGLang is a fast serving framework for large language models and vision language models.",
59
  "homepage_link": "https://docs.sglang.ai",
60
  "github_topic_closest_fit": "inference",
61
+ "contributors_all": 1318,
62
  "contributors_2025": 796,
63
  "contributors_2024": 189,
64
  "contributors_2023": 1,
 
84
  "github_about_section": "LLM inference in C/C++",
85
  "homepage_link": "https://ggml.ai",
86
  "github_topic_closest_fit": "inference",
87
+ "contributors_all": 1617,
88
  "contributors_2025": 535,
89
  "contributors_2024": 575,
90
  "contributors_2023": 461,
 
97
  "github_about_section": "Production-Grade Container Scheduling and Management",
98
  "homepage_link": "https://kubernetes.io",
99
  "github_topic_closest_fit": "kubernetes",
100
+ "contributors_all": 5203,
101
+ "contributors_2025": 544,
102
  "contributors_2024": 499,
103
  "contributors_2023": 565,
104
+ "contributors_2026_q1": 250
105
  },
106
  {
107
  "repo_name": "tensorflow",
 
110
  "github_about_section": "An Open Source Machine Learning Framework for Everyone",
111
  "homepage_link": "https://tensorflow.org",
112
  "github_topic_closest_fit": "machine-learning",
113
+ "contributors_all": 4698,
114
  "contributors_2025": 506,
115
  "contributors_2024": 523,
116
  "contributors_2023": 630,
 
123
  "github_about_section": "verl: Volcano Engine Reinforcement Learning for LLMs",
124
  "homepage_link": "https://verl.readthedocs.io",
125
  "github_topic_closest_fit": "deep-reinforcement-learning",
126
+ "contributors_all": 604,
127
  "contributors_2025": 454,
128
  "contributors_2024": 10,
129
  "contributors_2023": 0,
 
136
  "github_about_section": "super repo for rocm systems projects",
137
  "homepage_link": "https://amd.com/en/products/software/rocm.html",
138
  "github_topic_closest_fit": "amd",
139
+ "contributors_all": 1200,
140
+ "contributors_2025": 500,
141
  "contributors_2024": 351,
142
  "contributors_2023": 213,
143
+ "contributors_2026_q1": 257
144
  },
145
  {
146
  "repo_name": "ray",
 
149
  "github_about_section": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
150
  "homepage_link": "https://ray.io",
151
  "github_topic_closest_fit": "machine-learning",
152
+ "contributors_all": 1489,
153
  "contributors_2025": 397,
154
  "contributors_2024": 223,
155
  "contributors_2023": 230,
 
162
  "github_about_section": "Apache Spark - A unified analytics engine for large-scale data processing",
163
  "homepage_link": "https://spark.apache.org",
164
  "github_topic_closest_fit": "data-processing",
165
+ "contributors_all": 3155,
166
  "contributors_2025": 322,
167
  "contributors_2024": 300,
168
  "contributors_2023": 336,
 
175
  "github_about_section": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
176
  "homepage_link": "https://block.github.io/goose",
177
  "github_topic_closest_fit": "ai-agents",
178
+ "contributors_all": 452,
179
  "contributors_2025": 319,
180
  "contributors_2024": 32,
181
  "contributors_2023": 0,
 
188
  "github_about_section": "Free and Open Source, Distributed, RESTful Search Engine",
189
  "homepage_link": "https://elastic.co/products/elasticsearch",
190
  "github_topic_closest_fit": "search-engine",
191
+ "contributors_all": 2354,
192
+ "contributors_2025": 317,
193
  "contributors_2024": 284,
194
  "contributors_2023": 270,
195
+ "contributors_2026_q1": 201
196
  },
197
  {
198
  "repo_name": "jax",
 
201
  "github_about_section": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
202
  "homepage_link": "https://docs.jax.dev",
203
  "github_topic_closest_fit": "scientific-computing",
204
+ "contributors_all": 1050,
205
  "contributors_2025": 317,
206
  "contributors_2024": 280,
207
  "contributors_2023": 202,
208
+ "contributors_2026_q1": 132
209
  },
210
  {
211
  "repo_name": "modelcontextprotocol",
 
214
  "github_about_section": "Specification and documentation for the Model Context Protocol",
215
  "homepage_link": "https://modelcontextprotocol.io",
216
  "github_topic_closest_fit": "mcp",
217
+ "contributors_all": 373,
218
  "contributors_2025": 301,
219
  "contributors_2024": 42,
220
  "contributors_2023": 0,
221
+ "contributors_2026_q1": 69
222
  },
223
  {
224
  "repo_name": "executorch",
 
227
  "github_about_section": "On-device AI across mobile, embedded and edge for PyTorch",
228
  "homepage_link": "https://executorch.ai",
229
  "github_topic_closest_fit": "inference",
230
+ "contributors_all": 517,
231
  "contributors_2025": 267,
232
  "contributors_2024": 243,
233
  "contributors_2023": 77,
 
240
  "github_about_section": "The fundamental package for scientific computing with Python.",
241
  "homepage_link": "https://numpy.org",
242
  "github_topic_closest_fit": "scientific-computing",
243
+ "contributors_all": 2225,
244
  "contributors_2025": 237,
245
  "contributors_2024": 233,
246
  "contributors_2023": 252,
 
253
  "github_about_section": "Development repository for the Triton language and compiler",
254
  "homepage_link": "https://triton-lang.org",
255
  "github_topic_closest_fit": "parallel-programming",
256
+ "contributors_all": 577,
257
  "contributors_2025": 233,
258
  "contributors_2024": 206,
259
  "contributors_2023": 159,
 
266
  "github_about_section": "The Modular Platform (includes MAX & Mojo)",
267
  "homepage_link": "https://docs.modular.com",
268
  "github_topic_closest_fit": "parallel-programming",
269
+ "contributors_all": 423,
270
  "contributors_2025": 222,
271
  "contributors_2024": 205,
272
  "contributors_2023": 99,
 
279
  "github_about_section": "SciPy library main repository",
280
  "homepage_link": "https://scipy.org",
281
  "github_topic_closest_fit": "scientific-computing",
282
+ "contributors_all": 2022,
283
  "contributors_2025": 213,
284
  "contributors_2024": 251,
285
  "contributors_2023": 245,
286
+ "contributors_2026_q1": 76
287
  },
288
  {
289
  "repo_name": "ollama",
 
292
  "github_about_section": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
293
  "homepage_link": "https://ollama.com",
294
  "github_topic_closest_fit": "inference",
295
+ "contributors_all": 602,
296
  "contributors_2025": 202,
297
  "contributors_2024": 314,
298
  "contributors_2023": 97,
 
305
  "github_about_section": "Train transformer language models with reinforcement learning.",
306
  "homepage_link": "http://hf.co/docs/trl",
307
  "github_topic_closest_fit": "reinforcement-learning",
308
+ "contributors_all": 484,
309
  "contributors_2025": 189,
310
  "contributors_2024": 154,
311
  "contributors_2023": 122,
 
318
  "github_about_section": "FlashInfer: Kernel Library for LLM Serving",
319
  "homepage_link": "https://flashinfer.ai",
320
  "github_topic_closest_fit": "attention",
321
+ "contributors_all": 285,
322
  "contributors_2025": 158,
323
  "contributors_2024": 50,
324
  "contributors_2023": 11,
 
330
  "category": "gpu kernels",
331
  "github_about_section": "AI Tensor Engine for ROCm",
332
  "homepage_link": "https://rocm.blogs.amd.com/software-tools-optimization/aiter-ai-tensor-engine/README.html",
333
+ "contributors_all": 242,
334
  "contributors_2025": 145,
335
  "contributors_2024": 10,
336
  "contributors_2023": 0,
 
342
  "category": "inference",
343
  "github_about_section": "Supercharge Your LLM with the Fastest KV Cache Layer",
344
  "homepage_link": "https://lmcache.ai",
345
+ "contributors_all": 184,
346
  "contributors_2025": 144,
347
  "contributors_2024": 18,
348
  "contributors_2023": 0,
 
355
  "github_about_section": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
356
  "homepage_link": "https://kvcache-ai.github.io/Mooncake",
357
  "github_topic_closest_fit": "inference",
358
+ "contributors_all": 214,
359
  "contributors_2025": 133,
360
  "contributors_2024": 13,
361
  "contributors_2023": 0,
 
367
  "category": "training framework",
368
  "github_about_section": "A PyTorch native platform for training generative AI models",
369
  "homepage_link": "https://arxiv.org/abs/2410.06511",
370
+ "contributors_all": 198,
371
  "contributors_2025": 119,
372
  "contributors_2024": 43,
373
  "contributors_2023": 1,
 
380
  "github_about_section": "PyTorch native quantization and sparsity for training and inference",
381
  "homepage_link": "https://pytorch.org/ao",
382
  "github_topic_closest_fit": "quantization",
383
+ "contributors_all": 225,
384
  "contributors_2025": 114,
385
  "contributors_2024": 100,
386
  "contributors_2023": 5,
 
393
  "github_about_section": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
394
  "homepage_link": "https://comfy.org",
395
  "github_topic_closest_fit": "stable-diffusion",
396
+ "contributors_all": 323,
397
  "contributors_2025": 108,
398
  "contributors_2024": 119,
399
  "contributors_2023": 94,
 
406
  "github_about_section": "Fine-tuning & Reinforcement Learning for LLMs. Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
407
  "homepage_link": "https://docs.unsloth.ai",
408
  "github_topic_closest_fit": "fine-tuning",
409
+ "contributors_all": 192,
410
  "contributors_2025": 108,
411
  "contributors_2024": 29,
412
  "contributors_2023": 3,
 
443
  "category": "training framework",
444
  "github_about_section": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
445
  "homepage_link": "https://deepspeed.ai",
446
+ "contributors_all": 465,
447
  "contributors_2025": 96,
448
  "contributors_2024": 134,
449
  "contributors_2023": 165,
 
456
  "github_about_section": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
457
  "homepage_link": "https://milvus.io",
458
  "github_topic_closest_fit": "vector-search",
459
+ "contributors_all": 402,
460
  "contributors_2025": 95,
461
  "contributors_2024": 84,
462
  "contributors_2023": 72,
 
469
  "github_about_section": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
470
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
471
  "github_topic_closest_fit": "parallel-programming",
472
+ "contributors_all": 268,
473
  "contributors_2025": 94,
474
  "contributors_2024": 64,
475
  "contributors_2023": 66,
 
482
  "github_about_section": "Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels",
483
  "homepage_link": "https://tilelang.com",
484
  "github_topic_closest_fit": "parallel-programming",
485
+ "contributors_all": 132,
486
  "contributors_2025": 89,
487
  "contributors_2024": 1,
488
  "contributors_2023": 0,
 
494
  "category": "distributed computing",
495
  "github_about_section": "PyTorch Single Controller",
496
  "homepage_link": "https://meta-pytorch.org/monarch",
497
+ "contributors_all": 116,
498
  "contributors_2025": 85,
499
  "contributors_2024": 0,
500
  "contributors_2023": 0,
 
507
  "github_about_section": "Efficient Triton Kernels for LLM Training",
508
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
509
  "github_topic_closest_fit": "triton",
510
+ "contributors_all": 144,
511
  "contributors_2025": 78,
512
  "contributors_2024": 61,
513
  "contributors_2023": 0,
 
532
  "category": "fine tuning",
533
  "github_about_section": "PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
534
  "homepage_link": "https://huggingface.co/docs/peft",
535
+ "contributors_all": 297,
536
  "contributors_2025": 69,
537
  "contributors_2024": 111,
538
  "contributors_2023": 115,
 
544
  "category": "multi-purpose library",
545
  "github_about_section": "AMD ROCm Software - GitHub Home",
546
  "homepage_link": "https://rocm.docs.amd.com",
547
+ "contributors_all": 170,
548
  "contributors_2025": 67,
549
  "contributors_2024": 61,
550
  "contributors_2023": 44,
 
569
  "github_about_section": "Open standard for machine learning interoperability",
570
  "homepage_link": "https://onnx.ai",
571
  "github_topic_closest_fit": "onnx",
572
+ "contributors_all": 388,
573
  "contributors_2025": 56,
574
  "contributors_2024": 45,
575
  "contributors_2023": 61,
 
595
  "github_about_section": "A Python-embedded DSL that makes it easy to write fast, scalable ML kernels with minimal boilerplate.",
596
  "homepage_link": "https://helionlang.com",
597
  "github_topic_closest_fit": "parallel-programming",
598
+ "contributors_all": 75,
599
  "contributors_2025": 49,
600
  "contributors_2024": 0,
601
  "contributors_2023": 0,
 
631
  "github_about_section": "The simplest way to serve AI/ML models in production",
632
  "homepage_link": "https://truss.baseten.co",
633
  "github_topic_closest_fit": "inference",
634
+ "contributors_all": 88,
635
  "contributors_2025": 44,
636
  "contributors_2024": 30,
637
  "contributors_2023": 21,
 
644
  "github_about_section": "CUDA Python: Performance meets Productivity",
645
  "homepage_link": "https://nvidia.github.io/cuda-python",
646
  "github_topic_closest_fit": "parallel-programming",
647
+ "contributors_all": 55,
648
  "contributors_2025": 41,
649
  "contributors_2024": 12,
650
  "contributors_2023": 1,
 
657
  "github_about_section": "A Python framework for accelerated simulation, data generation and spatial computing.",
658
  "homepage_link": "https://nvidia.github.io/warp",
659
  "github_topic_closest_fit": "physics-simulation",
660
+ "contributors_all": 93,
661
  "contributors_2025": 40,
662
  "contributors_2024": 29,
663
  "contributors_2023": 17,
 
669
  "category": "container orchestration",
670
  "github_about_section": "Build, Manage and Deploy AI/ML Systems",
671
  "homepage_link": "https://metaflow.org",
672
+ "contributors_all": 137,
673
  "contributors_2025": 37,
674
  "contributors_2024": 35,
675
  "contributors_2023": 28,
 
681
  "category": "compiler",
682
  "github_about_section": "NumPy aware dynamic Python compiler using LLVM",
683
  "homepage_link": "https://numba.pydata.org",
684
+ "contributors_all": 452,
685
  "contributors_2025": 40,
686
  "contributors_2024": 32,
687
  "contributors_2023": 55,
 
745
  "github_about_section": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
746
  "homepage_link": "http://docs.sglang.ai/ome",
747
  "github_topic_closest_fit": "k8s",
748
+ "contributors_all": 32,
749
  "contributors_2025": 28,
750
  "contributors_2024": 0,
751
  "contributors_2023": 0,
 
783
  "github_about_section": "LAPACK is a library of Fortran subroutines for solving the most commonly occurring problems in numerical linear algebra.",
784
  "homepage_link": "https://netlib.org/lapack",
785
  "github_topic_closest_fit": "linear-algebra",
786
+ "contributors_all": 188,
787
  "contributors_2025": 23,
788
  "contributors_2024": 25,
789
  "contributors_2023": 42,
 
794
  "repo_link": "https://github.com/Dao-AILab/quack",
795
  "category": "kernel examples",
796
  "github_about_section": "A Quirky Assortment of CuTe Kernels",
797
+ "contributors_all": 39,
798
  "contributors_2025": 17,
799
  "contributors_2024": 0,
800
  "contributors_2023": 0,
 
819
  "category": "kernel examples",
820
  "github_about_section": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
821
  "homepage_link": "https://gpumode.com",
822
+ "contributors_all": 25,
823
  "contributors_2025": 16,
824
  "contributors_2024": 0,
825
  "contributors_2023": 0,
 
844
  "category": "performance testing",
845
  "github_about_section": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
846
  "homepage_link": "https://meta-pytorch.org/tritonparse",
847
+ "contributors_all": 29,
848
  "contributors_2025": 15,
849
  "contributors_2024": 0,
850
  "contributors_2023": 0,
 
855
  "repo_link": "https://github.com/huggingface/kernels",
856
  "category": "gpu kernels",
857
  "github_about_section": "Load compute kernels from the Hub",
858
+ "contributors_all": 29,
859
  "contributors_2025": 14,
860
  "contributors_2024": 2,
861
  "contributors_2023": 0,
 
879
  "repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
880
  "category": "training framework",
881
  "github_about_section": "Primus-Turbo is a high-performance acceleration library dedicated to large-scale model training on AMD GPUs. Built and optimized for the AMD ROCm platform, it covers the full training stack — including core compute operators (GEMM, Attention, GroupedGEMM), communication primitives, optimizer modules, low-precision computation (FP8), and compute–communication overlap kernels.",
882
+ "contributors_all": 16,
883
  "contributors_2025": 12,
884
  "contributors_2024": 0,
885
  "contributors_2023": 0,
 
941
  "category": "gpu kernels",
942
  "homepage_link": "https://huggingface.co/kernels-community",
943
  "github_about_section": "Kernel sources for https://huggingface.co/kernels-community",
944
+ "contributors_all": 16,
945
  "contributors_2025": 9,
946
  "contributors_2024": 0,
947
  "contributors_2023": 0,
 
953
  "category": "agent",
954
  "github_about_section": "It is an LLM-based AI agent, which can write correct and efficient gpu kernels automatically.",
955
  "github_topic_closest_fit": "ai-agents",
956
+ "contributors_all": 26,
957
  "contributors_2025": 9,
958
  "contributors_2024": 0,
959
  "contributors_2023": 0,
 
979
  "github_about_section": "cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it",
980
  "homepage_link": "https://developer.nvidia.com/cudnn",
981
  "github_topic_closest_fit": "parallel-programming",
982
+ "contributors_all": 15,
983
  "contributors_2025": 6,
984
  "contributors_2024": 5,
985
  "contributors_2023": 1,
 
1056
  "github_about_section": "Official inference library for Mistral models",
1057
  "homepage_link": "https://mistral.ai",
1058
  "github_topic_closest_fit": "inference",
1059
+ "contributors_all": 31,
1060
  "contributors_2025": 2,
1061
  "contributors_2024": 17,
1062
  "contributors_2023": 14,
1063
+ "contributors_2026_q1": 2
1064
  },
1065
  {
1066
  "repo_name": "omnitrace",
 
1156
  "category": "ml visualization",
1157
  "github_about_section": "The AI developer platform. Use Weights & Biases to train and fine-tune models, and manage models from experimentation to production.",
1158
  "homepage_link": "https://wandb.ai",
1159
+ "contributors_all": 241,
1160
  "contributors_2025": 46,
1161
  "contributors_2024": 67,
1162
  "contributors_2023": 62,
 
1168
  "category": "sdk",
1169
  "github_about_section": "Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and integrated with your favorite AWS services",
1170
  "homepage_link": "https://aws.amazon.com/ai/machine-learning/neuron",
1171
+ "contributors_all": 147,
1172
  "contributors_2025": 33,
1173
  "contributors_2024": 37,
1174
  "contributors_2023": 32,
 
1180
  "category": "machine learning interoperability",
1181
  "github_about_section": "ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator",
1182
  "homepage_link": "https://onnxruntime.ai",
1183
+ "contributors_all": 887,
1184
  "contributors_2025": 237,
1185
  "contributors_2024": 213,
1186
  "contributors_2023": 213,
 
1192
  "category": "machine learning interoperability",
1193
  "github_about_section": "Fast ML inference & training for ONNX models in Rust",
1194
  "homepage_link": "https://ort.pyke.io",
1195
+ "contributors_all": 72,
1196
  "contributors_2025": 25,
1197
  "contributors_2024": 20,
1198
  "contributors_2023": 21,
 
1227
  "category": "parallel computing",
1228
  "github_about_section": "cuTile is a programming model for writing parallel kernels for NVIDIA GPUs",
1229
  "homepage_link": "https://docs.nvidia.com/cuda/cutile-python",
1230
+ "contributors_all": 21,
1231
  "contributors_2025": 10,
1232
  "contributors_2024": 0,
1233
  "contributors_2023": 0,
 
1261
  "repo_link": "https://github.com/fla-org/flash-linear-attention",
1262
  "category": "gpu kernels",
1263
  "github_about_section": "Efficient implementations of state-of-the-art linear attention models",
1264
+ "contributors_all": 103,
1265
  "contributors_2025": 64,
1266
  "contributors_2024": 22,
1267
  "contributors_2023": 3,
 
1273
  "category": "distributed computing",
1274
  "github_about_section": "Optimized primitives for collective multi-GPU communication",
1275
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
1276
+ "contributors_all": 85,
1277
  "contributors_2026_q1": 26,
1278
  "contributors_2025": 10,
1279
  "contributors_2024": 5,
 
1320
  "category": "data compression",
1321
  "github_about_section": "A novel data compression framework",
1322
  "homepage_link": "https://openzl.org",
1323
+ "contributors_all": 42,
1324
  "contributors_2026_q1": 24,
1325
  "contributors_2025": 22,
1326
  "contributors_2024": 0,
 
1344
  "category": "reinforcement learning",
1345
  "github_about_section": "AllenAI's post-training codebase",
1346
  "homepage_link": "https://allenai.github.io/open-instruct/",
1347
+ "contributors_all": 58,
1348
  "contributors_2026_q1": 12,
1349
  "contributors_2025": 28,
1350
  "contributors_2024": 24,
 
1355
  "repo_link": "https://github.com/PrimeIntellect-ai/prime-rl",
1356
  "category": "reinforcement learning",
1357
  "github_about_section": "Agentic RL Training at Scale",
1358
+ "contributors_all": 60,
1359
  "contributors_2026_q1": 29,
1360
  "contributors_2025": 40,
1361
  "contributors_2024": 8,
 
1367
  "category": "reinforcement learning",
1368
  "github_about_section": "SkyRL: A Modular Full-stack RL Library for LLMs",
1369
  "homepage_link": "https://docs.skyrl.ai/docs",
1370
+ "contributors_all": 79,
1371
  "contributors_2026_q1": 29,
1372
  "contributors_2025": 57,
1373
  "contributors_2024": 0,
 
1403
  "category": "world model",
1404
  "github_about_section": "Cosmos-Predict2.5, the latest version of the Cosmos World Foundation Models (WFMs) family, specialized for simulating and predicting the future state of the world in the form of video.",
1405
  "homepage_link": "https://research.nvidia.com/labs/cosmos-lab/cosmos-predict2.5",
1406
+ "contributors_all": 14,
1407
  "contributors_2026_q1": 7,
1408
  "contributors_2025": 9,
1409
  "contributors_2024": 0,
 
1415
  "category": "reinforcement learning",
1416
  "github_about_section": "The RL Bridge for LLM-based Agent Applications. Made Simple & Flexible.",
1417
  "homepage_link": "https://www.inclusion-ai.org/AReaL",
1418
+ "contributors_all": 96,
1419
  "contributors_2026_q1": 37,
1420
  "contributors_2025": 63,
1421
  "contributors_2024": 0,
 
1427
  "category": "reinforcement learning",
1428
  "github_about_section": "RLinf: Reinforcement Learning Infrastructure for Embodied and Agentic AI",
1429
  "homepage_link": "https://rlinf.readthedocs.io",
1430
+ "contributors_all": 82,
1431
  "contributors_2026_q1": 52,
1432
  "contributors_2025": 35,
1433
  "contributors_2024": 0,
 
1439
  "category": "reinforcement learning",
1440
  "github_about_section": "An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Models",
1441
  "homepage_link": "https://alibaba.github.io/ROLL/",
1442
+ "contributors_all": 79,
1443
  "contributors_2026_q1": 30,
1444
  "contributors_2025": 60,
1445
  "contributors_2024": 0,
 
1450
  "repo_link": "https://github.com/allenai/OLMo-core",
1451
  "category": "training framework",
1452
  "github_about_section": "PyTorch building blocks for the OLMo ecosystem",
1453
+ "homepage_link": "https://olmo-core.readthedocs.io/en/latest/",
1454
+ "contributors_all": 31,
1455
+ "contributors_2026_q1": 9,
1456
+ "contributors_2025": 23,
1457
+ "contributors_2024": 9,
1458
+ "contributors_2023": 0
1459
  }
1460
  ]