TylerHilbert commited on
Commit
58cc89e
·
verified ·
1 Parent(s): 07c7dc0

Updated repo_tags / repo_tag / closest_github_tag

Browse files
Files changed (2) hide show
  1. PyTorchConference2025_GithubRepos.json +76 -312
  2. README.md +1 -4
PyTorchConference2025_GithubRepos.json CHANGED
@@ -4,14 +4,14 @@
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
- "repo_tag": "machine-learning",
8
  "category": "machine learning framework"
9
  },
10
  {
11
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
12
  "repo_name": "llama.cpp",
13
  "repo_description": "LLM inference in C/C++",
14
- "repo_tag": "ggml",
15
  "category": "inference engine"
16
  },
17
  {
@@ -19,21 +19,21 @@
19
  "repo_name": "onnx",
20
  "repo_description": "Open standard for machine learning interoperability",
21
  "homepage_link": "https://onnx.ai/",
22
- "repo_tag": "deep-learning"
23
  },
24
  {
25
  "github_repo_link": "https://github.com/ray-project/ray",
26
  "repo_name": "ray",
27
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
28
  "homepage_link": "https://ray.io",
29
- "repo_tag": "deep-learning"
30
  },
31
  {
32
  "github_repo_link": "https://github.com/vllm-project/vllm",
33
  "repo_name": "vllm",
34
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
35
  "homepage_link": "https://docs.vllm.ai",
36
- "repo_tag": "inference",
37
  "category": "inference engine"
38
  },
39
  {
@@ -41,7 +41,7 @@
41
  "repo_name": "ollama",
42
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
43
  "homepage_link": "https://ollama.com",
44
- "repo_tag": "llms",
45
  "category": "inference engine"
46
  },
47
  {
@@ -49,7 +49,7 @@
49
  "repo_name": "sglang",
50
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
51
  "homepage_link": "https://docs.sglang.ai/",
52
- "repo_tag": "inference",
53
  "category": "inference engine"
54
  },
55
  {
@@ -57,14 +57,14 @@
57
  "repo_name": "modular",
58
  "repo_description": "The Modular Platform (includes MAX & Mojo)",
59
  "homepage_link": "https://docs.modular.com/",
60
- "repo_tag": "mojo"
61
  },
62
  {
63
  "github_repo_link": "https://github.com/pytorch/ao",
64
  "repo_name": "ao",
65
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
66
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
67
- "repo_tag": "quantization"
68
  },
69
  {
70
  "github_repo_link": "https://github.com/triton-lang/triton",
@@ -83,7 +83,7 @@
83
  "repo_name": "reference-kernels",
84
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
85
  "homepage_link": "https://gpumode.com",
86
- "repo_tag": "gpu",
87
  "category": "kernels"
88
  },
89
  {
@@ -91,7 +91,7 @@
91
  "repo_name": "executorch",
92
  "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
93
  "homepage_link": "https://executorch.ai",
94
- "repo_tag": "mobile",
95
  "category": "model compiler"
96
  },
97
  {
@@ -110,7 +110,7 @@
110
  "repo_name": "ComfyUI",
111
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
112
  "homepage_link": "https://www.comfy.org/",
113
- "repo_tag": "stable-diffusion"
114
  },
115
  {
116
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
@@ -123,14 +123,14 @@
123
  "repo_name": "letta",
124
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
125
  "homepage_link": "https://docs.letta.com/",
126
- "repo_tag": "ai-agents"
127
  },
128
  {
129
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
130
  "repo_name": "jupyterlab",
131
  "repo_description": "JupyterLab computational environment.",
132
  "homepage_link": "https://jupyterlab.readthedocs.io/",
133
- "repo_tag": "jupyter",
134
  "category": "ui"
135
  },
136
  {
@@ -143,7 +143,7 @@
143
  "repo_name": "cutlass",
144
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
145
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
146
- "repo_tag": "cuda"
147
  },
148
  {
149
  "github_repo_link": "https://github.com/pytorch/helion",
@@ -156,14 +156,14 @@
156
  "repo_name": "jax",
157
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
158
  "homepage_link": "https://docs.jax.dev",
159
- "repo_tag": "jax"
160
  },
161
  {
162
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
163
  "repo_name": "tensorflow",
164
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
165
  "homepage_link": "https://tensorflow.org",
166
- "repo_tag": "deep-learning",
167
  "category": "machine learning framework"
168
  },
169
  {
@@ -171,21 +171,21 @@
171
  "repo_name": "DeepSpeed",
172
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
173
  "homepage_link": "https://www.deepspeed.ai/",
174
- "repo_tag": "gpu"
175
  },
176
  {
177
  "github_repo_link": "https://github.com/triton-inference-server/server",
178
  "repo_name": "server",
179
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
180
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
181
- "repo_tag": "inference"
182
  },
183
  {
184
  "github_repo_link": "https://github.com/ROCm/ROCm",
185
  "repo_name": "ROCm",
186
  "repo_description": "AMD ROCm™ Software - GitHub Home",
187
  "homepage_link": "https://rocm.docs.amd.com",
188
- "repo_tag": "documentation"
189
  },
190
  {
191
  "github_repo_link": "https://github.com/llvm/llvm-project",
@@ -205,7 +205,7 @@
205
  "repo_name": "hatchet",
206
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
207
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
208
- "repo_tag": "performance",
209
  "category": "profiler"
210
  },
211
  {
@@ -213,7 +213,7 @@
213
  "repo_name": "triton-runner",
214
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
215
  "homepage_link": "https://triton-runner.org",
216
- "repo_tag": "triton"
217
  },
218
  {
219
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
@@ -227,7 +227,7 @@
227
  "repo_name": "Liger-Kernel",
228
  "repo_description": "Efficient Triton Kernels for LLM Training",
229
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
230
- "repo_tag": "triton",
231
  "category": "kernels"
232
  },
233
  {
@@ -241,14 +241,14 @@
241
  "repo_name": "tritonparse",
242
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
243
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
244
- "repo_tag": "triton"
245
  },
246
  {
247
  "github_repo_link": "https://github.com/elastic/elasticsearch",
248
  "repo_name": "elasticsearch",
249
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
250
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
251
- "repo_tag": "search-engine",
252
  "category": "search engine"
253
  },
254
  {
@@ -256,7 +256,7 @@
256
  "repo_name": "kubernetes",
257
  "repo_description": "Production-Grade Container Scheduling and Management",
258
  "homepage_link": "https://kubernetes.io",
259
- "repo_tag": "containers"
260
  },
261
  {
262
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
@@ -268,14 +268,14 @@
268
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
269
  "repo_name": "mcp-agent",
270
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
271
- "repo_tag": "ai-agents"
272
  },
273
  {
274
  "github_repo_link": "https://github.com/milvus-io/milvus",
275
  "repo_name": "milvus",
276
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
277
  "homepage_link": "https://milvus.io",
278
- "repo_tag": "vector-search",
279
  "category": "vector databse"
280
  },
281
  {
@@ -283,7 +283,7 @@
283
  "repo_name": "RaBitQ",
284
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
285
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
286
- "repo_tag": "nearest-neighbor-search"
287
  },
288
  {
289
  "github_repo_link": "https://github.com/Airtable/airtable.js",
@@ -295,7 +295,7 @@
295
  "repo_name": "mistral-inference",
296
  "repo_description": "Official inference library for Mistral models",
297
  "homepage_link": "https://mistral.ai/",
298
- "repo_tag": "llm-inference",
299
  "category": "inference engine"
300
  },
301
  {
@@ -303,31 +303,14 @@
303
  "repo_name": "dstack",
304
  "repo_description": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.",
305
  "homepage_link": "https://dstack.ai",
306
- "repo_tags": [
307
- "amd",
308
- "cloud",
309
- "containers",
310
- "docker",
311
- "fine-tuning",
312
- "gpu",
313
- "inference",
314
- "k8s",
315
- "kubernetes",
316
- "llms",
317
- "machine-learning",
318
- "nvidia",
319
- "orchestration",
320
- "python",
321
- "slurm",
322
- "training"
323
- ]
324
  },
325
  {
326
  "github_repo_link": "https://github.com/numpy/numpy",
327
  "repo_name": "numpy",
328
  "repo_description": "The fundamental package for scientific computing with Python.",
329
  "homepage_link": "https://numpy.org",
330
- "repo_tag": "python",
331
  "category": "python library"
332
  },
333
  {
@@ -335,7 +318,7 @@
335
  "repo_name": "scipy",
336
  "repo_description": "SciPy library main repository",
337
  "homepage_link": "https://scipy.org",
338
- "repo_tag": "python",
339
  "category": "python library"
340
  },
341
  {
@@ -343,13 +326,13 @@
343
  "repo_name": "numba",
344
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
345
  "homepage_link": "https://numba.pydata.org/",
346
- "repo_tag": "compiler"
347
  },
348
  {
349
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
350
  "repo_name": "torchdendrite",
351
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
352
- "repo_tag": "scr-3078",
353
  "category": "machine learning framework"
354
  },
355
  {
@@ -387,27 +370,14 @@
387
  "repo_name": "nccl",
388
  "repo_description": "Optimized primitives for collective multi-GPU communication",
389
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
390
- "repo_tag": "cuda"
391
  },
392
  {
393
  "github_repo_link": "https://github.com/sgl-project/ome",
394
  "repo_name": "ome",
395
  "repo_description": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
396
  "homepage_link": "http://docs.sglang.ai/ome/",
397
- "repo_tags": [
398
- "deepseek",
399
- "k8s",
400
- "kimi-k2",
401
- "llama",
402
- "llm",
403
- "llm-inference",
404
- "model-as-a-service",
405
- "model-serving",
406
- "multi-node-kubernetes",
407
- "oracle-cloud",
408
- "sgalng",
409
- "sglang"
410
- ]
411
  },
412
  {
413
  "github_repo_link": "https://github.com/volcengine/verl",
@@ -436,50 +406,42 @@
436
  "repo_name": "LMCache",
437
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
438
  "homepage_link": "https://lmcache.ai/",
439
- "repo_tag": "inference"
440
  },
441
  {
442
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
443
  "repo_name": "rdma-core",
444
  "repo_description": "RDMA core userspace libraries and daemons",
445
  "homepage_link": null,
446
- "repo_tags": [
447
- "infiniband",
448
- "iwarp",
449
- "kernel-rdma-drivers",
450
- "linux-kernel",
451
- "rdma",
452
- "roce",
453
- "userspace-libraries"
454
- ]
455
  },
456
  {
457
  "github_repo_link": "https://github.com/NVIDIA/TensorRT",
458
  "repo_name": "TensorRT",
459
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
460
  "homepage_link": "https://developer.nvidia.com/tensorrt",
461
- "repo_tag": "inference"
462
  },
463
  {
464
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
465
  "repo_name": "FTorch",
466
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
467
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
468
- "repo_tag": "deep-learning"
469
  },
470
  {
471
  "github_repo_link": "https://github.com/facebook/hhvm",
472
  "repo_name": "hhvm",
473
  "repo_description": "A virtual machine for executing programs written in Hack.",
474
  "homepage_link": "https://hhvm.com",
475
- "repo_tag": "hack"
476
  },
477
  {
478
  "github_repo_link": "https://github.com/vosen/ZLUDA",
479
  "repo_name": "ZLUDA",
480
  "repo_description": "CUDA on non-NVIDIA GPUs",
481
  "homepage_link": "https://vosen.github.io/ZLUDA/",
482
- "repo_tag": "cuda"
483
  },
484
  {
485
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
@@ -492,48 +454,28 @@
492
  "repo_name": "pocl",
493
  "repo_description": "pocl - Portable Computing Language",
494
  "homepage_link": "https://portablecl.org",
495
- "repo_tag": "opencl"
496
  },
497
  {
498
  "github_repo_link": "https://github.com/apache/spark",
499
  "repo_name": "spark",
500
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
501
  "homepage_link": "https://spark.apache.org/",
502
- "repo_tag": "big-data"
503
  },
504
  {
505
  "github_repo_link": "https://github.com/codelion/openevolve",
506
  "repo_name": "openevolve",
507
  "repo_description": "Open-source implementation of AlphaEvolve",
508
  "homepage_link": "",
509
- "repo_tags": [
510
- "alpha-evolve",
511
- "alphacode",
512
- "alphaevolve",
513
- "coding-agent",
514
- "deepmind",
515
- "deepmind-lab",
516
- "discovery",
517
- "distributed-evolutionary-algorithms",
518
- "evolutionary-algorithms",
519
- "evolutionary-computation",
520
- "genetic-algorithm",
521
- "genetic-algorithms",
522
- "iterative-methods",
523
- "iterative-refinement",
524
- "llm-engineering",
525
- "llm-ensemble",
526
- "llm-inference",
527
- "openevolve",
528
- "optimize"
529
- ]
530
  },
531
  {
532
  "github_repo_link": "https://github.com/ROCm/hipBLAS",
533
  "repo_name": "hipBLAS",
534
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
535
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
536
- "repo_tag": "hip"
537
  },
538
  {
539
  "github_repo_link": "https://github.com/ROCm/roctracer",
@@ -546,25 +488,14 @@
546
  "repo_name": "peft",
547
  "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
548
  "homepage_link": "https://huggingface.co/docs/peft",
549
- "repo_tags": [
550
- "adapter",
551
- "diffusion",
552
- "fine-tuning",
553
- "llm",
554
- "lora",
555
- "parameter-efficient-learning",
556
- "peft",
557
- "python",
558
- "pytorch",
559
- "transformers"
560
- ]
561
  },
562
  {
563
  "github_repo_link": "https://github.com/ROCm/hip",
564
  "repo_name": "hip",
565
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
566
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
567
- "repo_tag": "hip"
568
  },
569
  {
570
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
@@ -581,16 +512,7 @@
581
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
582
  "repo_name": "intelliperf",
583
  "repo_description": "Automated bottleneck detection and solution orchestration",
584
- "repo_tags": [
585
- "amd",
586
- "genai",
587
- "gpu",
588
- "hip",
589
- "instinct",
590
- "llm",
591
- "performance",
592
- "rocm"
593
- ]
594
  },
595
  {
596
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
@@ -635,7 +557,7 @@
635
  "repo_name": "opencv",
636
  "repo_description": "Open Source Computer Vision Library",
637
  "homepage_link": "https://opencv.org",
638
- "repo_tag": "image-processing"
639
  },
640
  {
641
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
@@ -647,26 +569,7 @@
647
  "repo_name": "burn",
648
  "repo_description": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
649
  "homepage_link": "https://burn.dev",
650
- "repo_tags": [
651
- "autodiff",
652
- "cross-platform",
653
- "cuda",
654
- "deep-learning",
655
- "kernel-fusion",
656
- "machine-learning",
657
- "metal",
658
- "ndarray",
659
- "neural-network",
660
- "onnx",
661
- "pytorch",
662
- "rocm",
663
- "rust",
664
- "scientific-computing",
665
- "tensor",
666
- "vulkan",
667
- "wasm",
668
- "webgpu"
669
- ]
670
  },
671
  {
672
  "github_repo_link": "https://github.com/huggingface/kernels-community",
@@ -686,37 +589,21 @@
686
  "repo_name": "ondemand",
687
  "repo_description": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
688
  "homepage_link": "https://openondemand.org/",
689
- "repo_tags": [
690
- "gateway",
691
- "hacktoberfest",
692
- "hpc",
693
- "hpc-applications"
694
- ]
695
  },
696
  {
697
  "github_repo_link": "https://github.com/flashinfer-ai/flashinfer",
698
  "repo_name": "flashinfer",
699
  "repo_description": "FlashInfer: Kernel Library for LLM Serving",
700
  "homepage_link": "https://flashinfer.ai",
701
- "repo_tags": [
702
- "attention",
703
- "cuda",
704
- "distributed-inference",
705
- "gpu",
706
- "jit",
707
- "large-large-models",
708
- "llm-inference",
709
- "moe",
710
- "nvidia",
711
- "pytorch"
712
- ]
713
  },
714
  {
715
  "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
716
  "repo_name": "KernelBench",
717
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
718
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
719
- "repo_tag": "benchmark",
720
  "category": "benchmark"
721
  },
722
  {
@@ -735,28 +622,7 @@
735
  "repo_name": "metaflow",
736
  "repo_description": "Build, Manage and Deploy AI/ML Systems",
737
  "homepage_link": "https://metaflow.org",
738
- "repo_tags": [
739
- "agents",
740
- "ai",
741
- "aws",
742
- "azure",
743
- "cost-optimization",
744
- "datascience",
745
- "distributed-training",
746
- "gcp",
747
- "generative-ai",
748
- "high-performance-computing",
749
- "kubernetes",
750
- "llm",
751
- "llmops",
752
- "machine-learning",
753
- "ml",
754
- "ml-infrastructure",
755
- "ml-platform",
756
- "mlops",
757
- "model-management",
758
- "python"
759
- ]
760
  },
761
  {
762
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
@@ -767,22 +633,14 @@
767
  "repo_name": "lean4",
768
  "repo_description": "Lean 4 programming language and theorem prover",
769
  "homepage_link": "https://lean-lang.org",
770
- "repo_tag": "lean"
771
  },
772
  {
773
  "github_repo_link": "https://github.com/NVIDIA/warp",
774
  "repo_name": "warp",
775
  "repo_description": "A Python framework for accelerated simulation, data generation and spatial computing.",
776
  "homepage_link": "https://nvidia.github.io/warp/",
777
- "repo_tags": [
778
- "cuda",
779
- "differentiable-programming",
780
- "gpu",
781
- "gpu-acceleration",
782
- "nvidia",
783
- "nvidia-warp",
784
- "python"
785
- ]
786
  },
787
  {
788
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
@@ -795,20 +653,7 @@
795
  "repo_name": "truss",
796
  "repo_description": "The simplest way to serve AI/ML models in production",
797
  "homepage_link": "https://truss.baseten.co",
798
- "repo_tags": [
799
- "artificial-intelligence",
800
- "easy-to-use",
801
- "falcon",
802
- "inference-api",
803
- "inference-server",
804
- "machine-learning",
805
- "model-serving",
806
- "open-source",
807
- "packaging",
808
- "stable-diffusion",
809
- "whisper",
810
- "wizardlm"
811
- ]
812
  },
813
  {
814
  "github_repo_link": "https://github.com/laude-institute/terminal-bench",
@@ -822,7 +667,7 @@
822
  "repo_name": "goose",
823
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
824
  "homepage_link": "https://block.github.io/goose/",
825
- "repo_tag": "mcp",
826
  "category": "agent"
827
  },
828
  {
@@ -830,26 +675,14 @@
830
  "repo_name": "Mooncake",
831
  "repo_description": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
832
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
833
- "repo_tags": [
834
- "disaggregation",
835
- "inference",
836
- "kvcache",
837
- "llm",
838
- "rdma",
839
- "sglang",
840
- "vllm"
841
- ]
842
  },
843
  {
844
  "github_repo_link": "https://github.com/SWE-bench/SWE-bench",
845
  "repo_name": "SWE-bench",
846
  "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
847
  "homepage_link": "https://www.swebench.com",
848
- "repo_tags": [
849
- "benchmark",
850
- "language-model",
851
- "software-engineering"
852
- ],
853
  "category": "benchmark"
854
  },
855
  {
@@ -867,24 +700,14 @@
867
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
868
  "repo_name": "triSYCL",
869
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
870
- "repo_tags": [
871
- "cpp",
872
- "cpp20",
873
- "fpga",
874
- "gpu-computing",
875
- "heterogeneous-parallel-programming",
876
- "opencl",
877
- "spir",
878
- "sycl",
879
- "trisycl"
880
- ]
881
  },
882
  {
883
  "github_repo_link": "https://github.com/pybind/pybind11",
884
  "repo_name": "pybind11",
885
  "repo_description": "Seamless operability between C++11 and Python",
886
  "homepage_link": "https://pybind11.readthedocs.io/",
887
- "repo_tag": "bindings"
888
  },
889
  {
890
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
@@ -897,40 +720,21 @@
897
  "repo_name": "rocSOLVER",
898
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
899
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
900
- "repo_tag": "rocm"
901
  },
902
  {
903
  "github_repo_link": "https://github.com/ROCm/Tensile",
904
  "repo_name": "Tensile",
905
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
906
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
907
- "repo_tags": [
908
- "amd",
909
- "assembly",
910
- "auto-tuning",
911
- "blas",
912
- "dnn",
913
- "gemm",
914
- "gpu",
915
- "gpu-acceleration",
916
- "gpu-computing",
917
- "hip",
918
- "machine-learning",
919
- "matrix-multiplication",
920
- "neural-networks",
921
- "opencl",
922
- "python",
923
- "radeon",
924
- "tensor-contraction",
925
- "tensors"
926
- ]
927
  },
928
  {
929
  "github_repo_link": "https://github.com/ROCm/rocPRIM",
930
  "repo_name": "rocPRIM",
931
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
932
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
933
- "repo_tag": "hip"
934
  },
935
  {
936
  "github_repo_link": "https://github.com/ROCm/hipCUB",
@@ -943,7 +747,7 @@
943
  "repo_name": "rocFFT",
944
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
945
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
946
- "repo_tag": "hip"
947
  },
948
  {
949
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
@@ -956,48 +760,26 @@
956
  "repo_name": "rocRAND",
957
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
958
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
959
- "repo_tag": "hip"
960
  },
961
  {
962
  "github_repo_link": "https://github.com/ROCm/MIOpen",
963
  "repo_name": "MIOpen",
964
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
965
- "homepage_link": "https://github.com/ROCm/rocm-libraries",
966
  },
967
  {
968
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
969
  "repo_name": "lapack",
970
  "repo_description": "LAPACK development repository",
971
- "repo_tags": [
972
- "blas",
973
- "eigenvalues",
974
- "eigenvectors",
975
- "lapack",
976
- "lapacke",
977
- "linear-algebra",
978
- "linear-equations",
979
- "matrix-factorization",
980
- "singular-values",
981
- "svd"
982
- ]
983
  },
984
  {
985
  "github_repo_link": "https://github.com/ccache/ccache",
986
  "repo_name": "ccache",
987
  "repo_description": "ccache – a fast compiler cache",
988
  "homepage_link": "https://ccache.dev",
989
- "repo_tags": [
990
- "c",
991
- "c-plus-plus",
992
- "cache",
993
- "ccache",
994
- "clang",
995
- "compiler",
996
- "cplusplus",
997
- "cpp",
998
- "gcc",
999
- "msvc"
1000
- ],
1001
  "category": "compiler"
1002
  },
1003
  {
@@ -1005,25 +787,7 @@
1005
  "repo_name": "omnitrace",
1006
  "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
1007
  "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
1008
- "repo_tags": [
1009
- "binary-instrumentation",
1010
- "code-coverage",
1011
- "cpu-profiler",
1012
- "dynamic-instrumentation",
1013
- "gpu-profiler",
1014
- "hardware-counters",
1015
- "instrumentation-profiler",
1016
- "linux",
1017
- "performance-analysis",
1018
- "performance-metrics",
1019
- "performance-monitoring",
1020
- "profiler",
1021
- "profiling",
1022
- "python",
1023
- "python-profiler",
1024
- "sampling-profiler",
1025
- "tracing"
1026
- ]
1027
  },
1028
  {
1029
  "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
@@ -1035,14 +799,14 @@
1035
  "repo_name": "synthetic-data-kit",
1036
  "repo_description": "Tool for generating high quality Synthetic datasets",
1037
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
1038
- "repo_tag": "generation"
1039
  },
1040
  {
1041
  "github_repo_link": "https://github.com/unslothai/unsloth",
1042
  "repo_name": "unsloth",
1043
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
1044
  "homepage_link": "https://docs.unsloth.ai/",
1045
- "repo_tag": "unsloth"
1046
  },
1047
  {
1048
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
@@ -1059,7 +823,7 @@
1059
  "repo_name": "Wan2.2",
1060
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
1061
  "homepage_link": "https://wan.video",
1062
- "repo_tag": "video-generation"
1063
  },
1064
  {
1065
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
 
4
  "repo_name": "pytorch",
5
  "repo_description": "Tensors and Dynamic neural networks in Python with strong GPU acceleration",
6
  "homepage_link": "https://pytorch.org",
7
+ "closest_github_tag": "machine-learning",
8
  "category": "machine learning framework"
9
  },
10
  {
11
  "github_repo_link": "https://github.com/ggml-org/llama.cpp",
12
  "repo_name": "llama.cpp",
13
  "repo_description": "LLM inference in C/C++",
14
+ "closest_github_tag": "ggml",
15
  "category": "inference engine"
16
  },
17
  {
 
19
  "repo_name": "onnx",
20
  "repo_description": "Open standard for machine learning interoperability",
21
  "homepage_link": "https://onnx.ai/",
22
+ "closest_github_tag": "deep-learning"
23
  },
24
  {
25
  "github_repo_link": "https://github.com/ray-project/ray",
26
  "repo_name": "ray",
27
  "repo_description": "Ray is an AI compute engine. Ray consists of a core distributed runtime and a set of AI Libraries for accelerating ML workloads.",
28
  "homepage_link": "https://ray.io",
29
+ "closest_github_tag": "deep-learning"
30
  },
31
  {
32
  "github_repo_link": "https://github.com/vllm-project/vllm",
33
  "repo_name": "vllm",
34
  "repo_description": "A high-throughput and memory-efficient inference and serving engine for LLMs",
35
  "homepage_link": "https://docs.vllm.ai",
36
+ "closest_github_tag": "inference",
37
  "category": "inference engine"
38
  },
39
  {
 
41
  "repo_name": "ollama",
42
  "repo_description": "Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.",
43
  "homepage_link": "https://ollama.com",
44
+ "closest_github_tag": "llms",
45
  "category": "inference engine"
46
  },
47
  {
 
49
  "repo_name": "sglang",
50
  "repo_description": "SGLang is a fast serving framework for large language models and vision language models.",
51
  "homepage_link": "https://docs.sglang.ai/",
52
+ "closest_github_tag": "inference",
53
  "category": "inference engine"
54
  },
55
  {
 
57
  "repo_name": "modular",
58
  "repo_description": "The Modular Platform (includes MAX & Mojo)",
59
  "homepage_link": "https://docs.modular.com/",
60
+ "closest_github_tag": "mojo"
61
  },
62
  {
63
  "github_repo_link": "https://github.com/pytorch/ao",
64
  "repo_name": "ao",
65
  "repo_description": "PyTorch native quantization and sparsity for training and inference",
66
  "homepage_link": "https://pytorch.org/ao/stable/index.html",
67
+ "closest_github_tag": "quantization"
68
  },
69
  {
70
  "github_repo_link": "https://github.com/triton-lang/triton",
 
83
  "repo_name": "reference-kernels",
84
  "repo_description": "Official Problem Sets / Reference Kernels for the GPU MODE Leaderboard!",
85
  "homepage_link": "https://gpumode.com",
86
+ "closest_github_tag": "gpu",
87
  "category": "kernels"
88
  },
89
  {
 
91
  "repo_name": "executorch",
92
  "repo_description": "On-device AI across mobile, embedded and edge for PyTorch",
93
  "homepage_link": "https://executorch.ai",
94
+ "closest_github_tag": "mobile",
95
  "category": "model compiler"
96
  },
97
  {
 
110
  "repo_name": "ComfyUI",
111
  "repo_description": "The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.",
112
  "homepage_link": "https://www.comfy.org/",
113
+ "closest_github_tag": "stable-diffusion"
114
  },
115
  {
116
  "github_repo_link": "https://github.com/Jeff-LiangF/streamv2v",
 
123
  "repo_name": "letta",
124
  "repo_description": "Letta is the platform for building stateful agents: open AI with advanced memory that can learn and self-improve over time.",
125
  "homepage_link": "https://docs.letta.com/",
126
+ "closest_github_tag": "ai-agents"
127
  },
128
  {
129
  "github_repo_link": "https://github.com/jupyterlab/jupyterlab",
130
  "repo_name": "jupyterlab",
131
  "repo_description": "JupyterLab computational environment.",
132
  "homepage_link": "https://jupyterlab.readthedocs.io/",
133
+ "closest_github_tag": "jupyter",
134
  "category": "ui"
135
  },
136
  {
 
143
  "repo_name": "cutlass",
144
  "repo_description": "CUDA Templates and Python DSLs for High-Performance Linear Algebra",
145
  "homepage_link": "https://docs.nvidia.com/cutlass/index.html",
146
+ "closest_github_tag": "cuda"
147
  },
148
  {
149
  "github_repo_link": "https://github.com/pytorch/helion",
 
156
  "repo_name": "jax",
157
  "repo_description": "Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more",
158
  "homepage_link": "https://docs.jax.dev",
159
+ "closest_github_tag": "jax"
160
  },
161
  {
162
  "github_repo_link": "https://github.com/tensorflow/tensorflow",
163
  "repo_name": "tensorflow",
164
  "repo_description": "An Open Source Machine Learning Framework for Everyone",
165
  "homepage_link": "https://tensorflow.org",
166
+ "closest_github_tag": "deep-learning",
167
  "category": "machine learning framework"
168
  },
169
  {
 
171
  "repo_name": "DeepSpeed",
172
  "repo_description": "DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.",
173
  "homepage_link": "https://www.deepspeed.ai/",
174
+ "closest_github_tag": "gpu"
175
  },
176
  {
177
  "github_repo_link": "https://github.com/triton-inference-server/server",
178
  "repo_name": "server",
179
  "repo_description": "The Triton Inference Server provides an optimized cloud and edge inferencing solution. ",
180
  "homepage_link": "https://docs.nvidia.com/deeplearning/triton-inference-server/user-guide/docs/index.html",
181
+ "closest_github_tag": "inference"
182
  },
183
  {
184
  "github_repo_link": "https://github.com/ROCm/ROCm",
185
  "repo_name": "ROCm",
186
  "repo_description": "AMD ROCm™ Software - GitHub Home",
187
  "homepage_link": "https://rocm.docs.amd.com",
188
+ "closest_github_tag": "documentation"
189
  },
190
  {
191
  "github_repo_link": "https://github.com/llvm/llvm-project",
 
205
  "repo_name": "hatchet",
206
  "repo_description": "Graph-indexed Pandas DataFrames for analyzing hierarchical performance data",
207
  "homepage_link": "https://llnl-hatchet.readthedocs.io",
208
+ "closest_github_tag": "performance",
209
  "category": "profiler"
210
  },
211
  {
 
213
  "repo_name": "triton-runner",
214
  "repo_description": "Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.",
215
  "homepage_link": "https://triton-runner.org",
216
+ "closest_github_tag": "triton"
217
  },
218
  {
219
  "github_repo_link": "https://github.com/ByteDance-Seed/Triton-distributed",
 
227
  "repo_name": "Liger-Kernel",
228
  "repo_description": "Efficient Triton Kernels for LLM Training",
229
  "homepage_link": "https://openreview.net/pdf?id=36SjAIT42G",
230
+ "closest_github_tag": "triton",
231
  "category": "kernels"
232
  },
233
  {
 
241
  "repo_name": "tritonparse",
242
  "repo_description": "TritonParse: A Compiler Tracer, Visualizer, and Reproducer for Triton Kernels",
243
  "homepage_link": "https://meta-pytorch.org/tritonparse/",
244
+ "closest_github_tag": "triton"
245
  },
246
  {
247
  "github_repo_link": "https://github.com/elastic/elasticsearch",
248
  "repo_name": "elasticsearch",
249
  "repo_description": "Free and Open Source, Distributed, RESTful Search Engine",
250
  "homepage_link": "https://www.elastic.co/products/elasticsearch",
251
+ "closest_github_tag": "search-engine",
252
  "category": "search engine"
253
  },
254
  {
 
256
  "repo_name": "kubernetes",
257
  "repo_description": "Production-Grade Container Scheduling and Management",
258
  "homepage_link": "https://kubernetes.io",
259
+ "closest_github_tag": "containers"
260
  },
261
  {
262
  "github_repo_link": "https://github.com/modelcontextprotocol/modelcontextprotocol",
 
268
  "github_repo_link": "https://github.com/lastmile-ai/mcp-agent",
269
  "repo_name": "mcp-agent",
270
  "repo_description": "Build effective agents using Model Context Protocol and simple workflow patterns",
271
+ "closest_github_tag": "ai-agents"
272
  },
273
  {
274
  "github_repo_link": "https://github.com/milvus-io/milvus",
275
  "repo_name": "milvus",
276
  "repo_description": "Milvus is a high-performance, cloud-native vector database built for scalable vector ANN search",
277
  "homepage_link": "https://milvus.io",
278
+ "closest_github_tag": "vector-search",
279
  "category": "vector databse"
280
  },
281
  {
 
283
  "repo_name": "RaBitQ",
284
  "repo_description": "[SIGMOD 2024] RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search",
285
  "homepage_link": "https://github.com/VectorDB-NTU/RaBitQ-Library",
286
+ "closest_github_tag": "nearest-neighbor-search"
287
  },
288
  {
289
  "github_repo_link": "https://github.com/Airtable/airtable.js",
 
295
  "repo_name": "mistral-inference",
296
  "repo_description": "Official inference library for Mistral models",
297
  "homepage_link": "https://mistral.ai/",
298
+ "closest_github_tag": "llm-inference",
299
  "category": "inference engine"
300
  },
301
  {
 
303
  "repo_name": "dstack",
304
  "repo_description": "dstack is an open-source control plane for running development, training, and inference jobs on GPUs—across hyperscalers, neoclouds, or on-prem.",
305
  "homepage_link": "https://dstack.ai",
306
+ "closest_github_tag": "orchestration"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
307
  },
308
  {
309
  "github_repo_link": "https://github.com/numpy/numpy",
310
  "repo_name": "numpy",
311
  "repo_description": "The fundamental package for scientific computing with Python.",
312
  "homepage_link": "https://numpy.org",
313
+ "closest_github_tag": "python",
314
  "category": "python library"
315
  },
316
  {
 
318
  "repo_name": "scipy",
319
  "repo_description": "SciPy library main repository",
320
  "homepage_link": "https://scipy.org",
321
+ "closest_github_tag": "python",
322
  "category": "python library"
323
  },
324
  {
 
326
  "repo_name": "numba",
327
  "repo_description": "NumPy aware dynamic Python compiler using LLVM",
328
  "homepage_link": "https://numba.pydata.org/",
329
+ "closest_github_tag": "compiler"
330
  },
331
  {
332
  "github_repo_link": "https://github.com/sandialabs/torchdendrite",
333
  "repo_name": "torchdendrite",
334
  "repo_description": "Dendrites for PyTorch and SNNTorch neural networks ",
335
+ "closest_github_tag": "scr-3078",
336
  "category": "machine learning framework"
337
  },
338
  {
 
370
  "repo_name": "nccl",
371
  "repo_description": "Optimized primitives for collective multi-GPU communication",
372
  "homepage_link": "https://docs.nvidia.com/deeplearning/nccl/user-guide/docs/index.html",
373
+ "closest_github_tag": "cuda"
374
  },
375
  {
376
  "github_repo_link": "https://github.com/sgl-project/ome",
377
  "repo_name": "ome",
378
  "repo_description": "OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)",
379
  "homepage_link": "http://docs.sglang.ai/ome/",
380
+ "closest_github_tag": "k8s"
 
 
 
 
 
 
 
 
 
 
 
 
 
381
  },
382
  {
383
  "github_repo_link": "https://github.com/volcengine/verl",
 
406
  "repo_name": "LMCache",
407
  "repo_description": "Supercharge Your LLM with the Fastest KV Cache Layer",
408
  "homepage_link": "https://lmcache.ai/",
409
+ "closest_github_tag": "inference"
410
  },
411
  {
412
  "github_repo_link": "https://github.com/linux-rdma/rdma-core",
413
  "repo_name": "rdma-core",
414
  "repo_description": "RDMA core userspace libraries and daemons",
415
  "homepage_link": null,
416
+ "closest_github_tag": "linux-kernel"
 
 
 
 
 
 
 
 
417
  },
418
  {
419
  "github_repo_link": "https://github.com/NVIDIA/TensorRT",
420
  "repo_name": "TensorRT",
421
  "repo_description": "NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.",
422
  "homepage_link": "https://developer.nvidia.com/tensorrt",
423
+ "closest_github_tag": "inference"
424
  },
425
  {
426
  "github_repo_link": "https://github.com/Cambridge-ICCS/FTorch",
427
  "repo_name": "FTorch",
428
  "repo_description": "A library for directly calling PyTorch ML models from Fortran.",
429
  "homepage_link": "https://cambridge-iccs.github.io/FTorch/",
430
+ "closest_github_tag": "deep-learning"
431
  },
432
  {
433
  "github_repo_link": "https://github.com/facebook/hhvm",
434
  "repo_name": "hhvm",
435
  "repo_description": "A virtual machine for executing programs written in Hack.",
436
  "homepage_link": "https://hhvm.com",
437
+ "closest_github_tag": "hack"
438
  },
439
  {
440
  "github_repo_link": "https://github.com/vosen/ZLUDA",
441
  "repo_name": "ZLUDA",
442
  "repo_description": "CUDA on non-NVIDIA GPUs",
443
  "homepage_link": "https://vosen.github.io/ZLUDA/",
444
+ "closest_github_tag": "cuda"
445
  },
446
  {
447
  "github_repo_link": "https://github.com/vtsynergy/CU2CL",
 
454
  "repo_name": "pocl",
455
  "repo_description": "pocl - Portable Computing Language",
456
  "homepage_link": "https://portablecl.org",
457
+ "closest_github_tag": "opencl"
458
  },
459
  {
460
  "github_repo_link": "https://github.com/apache/spark",
461
  "repo_name": "spark",
462
  "repo_description": "Apache Spark - A unified analytics engine for large-scale data processing",
463
  "homepage_link": "https://spark.apache.org/",
464
+ "closest_github_tag": "big-data"
465
  },
466
  {
467
  "github_repo_link": "https://github.com/codelion/openevolve",
468
  "repo_name": "openevolve",
469
  "repo_description": "Open-source implementation of AlphaEvolve",
470
  "homepage_link": "",
471
+ "closest_github_tag": "genetic-algorithm"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
472
  },
473
  {
474
  "github_repo_link": "https://github.com/ROCm/hipBLAS",
475
  "repo_name": "hipBLAS",
476
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
477
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
478
+ "closest_github_tag": "hip"
479
  },
480
  {
481
  "github_repo_link": "https://github.com/ROCm/roctracer",
 
488
  "repo_name": "peft",
489
  "repo_description": "🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.",
490
  "homepage_link": "https://huggingface.co/docs/peft",
491
+ "closest_github_tag": "lora"
 
 
 
 
 
 
 
 
 
 
 
492
  },
493
  {
494
  "github_repo_link": "https://github.com/ROCm/hip",
495
  "repo_name": "hip",
496
  "repo_description": "HIP: C++ Heterogeneous-Compute Interface for Portability",
497
  "homepage_link": "https://rocmdocs.amd.com/projects/HIP/",
498
+ "closest_github_tag": "hip"
499
  },
500
  {
501
  "github_repo_link": "https://github.com/ROCm/composable_kernel",
 
512
  "github_repo_link": "https://github.com/AMDResearch/intelliperf",
513
  "repo_name": "intelliperf",
514
  "repo_description": "Automated bottleneck detection and solution orchestration",
515
+ "closest_github_tag": "performance"
 
 
 
 
 
 
 
 
 
516
  },
517
  {
518
  "github_repo_link": "https://github.com/AMD-AGI/GEAK-agent",
 
557
  "repo_name": "opencv",
558
  "repo_description": "Open Source Computer Vision Library",
559
  "homepage_link": "https://opencv.org",
560
+ "closest_github_tag": "image-processing"
561
  },
562
  {
563
  "github_repo_link": "https://github.com/Lightning-AI/lightning-thunder",
 
569
  "repo_name": "burn",
570
  "repo_description": "Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.",
571
  "homepage_link": "https://burn.dev",
572
+ "closest_github_tag": "machine-learning"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
573
  },
574
  {
575
  "github_repo_link": "https://github.com/huggingface/kernels-community",
 
589
  "repo_name": "ondemand",
590
  "repo_description": "Supercomputing. Seamlessly. Open, Interactive HPC Via the Web",
591
  "homepage_link": "https://openondemand.org/",
592
+ "closest_github_tag": "hpc"
 
 
 
 
 
593
  },
594
  {
595
  "github_repo_link": "https://github.com/flashinfer-ai/flashinfer",
596
  "repo_name": "flashinfer",
597
  "repo_description": "FlashInfer: Kernel Library for LLM Serving",
598
  "homepage_link": "https://flashinfer.ai",
599
+ "closest_github_tag": "attention"
 
 
 
 
 
 
 
 
 
 
 
600
  },
601
  {
602
  "github_repo_link": "https://github.com/ScalingIntelligence/KernelBench",
603
  "repo_name": "KernelBench",
604
  "repo_description": "KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems",
605
  "homepage_link": "https://scalingintelligence.stanford.edu/blogs/kernelbench/",
606
+ "closest_github_tag": "benchmark",
607
  "category": "benchmark"
608
  },
609
  {
 
622
  "repo_name": "metaflow",
623
  "repo_description": "Build, Manage and Deploy AI/ML Systems",
624
  "homepage_link": "https://metaflow.org",
625
+ "closest_github_tag": "machine-learning"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
626
  },
627
  {
628
  "github_repo_link": "https://github.com/harmonic-ai/IMO2025",
 
633
  "repo_name": "lean4",
634
  "repo_description": "Lean 4 programming language and theorem prover",
635
  "homepage_link": "https://lean-lang.org",
636
+ "closest_github_tag": "lean"
637
  },
638
  {
639
  "github_repo_link": "https://github.com/NVIDIA/warp",
640
  "repo_name": "warp",
641
  "repo_description": "A Python framework for accelerated simulation, data generation and spatial computing.",
642
  "homepage_link": "https://nvidia.github.io/warp/",
643
+ "closest_github_tag": "gpu"
 
 
 
 
 
 
 
 
644
  },
645
  {
646
  "github_repo_link": "https://github.com/NVIDIA/cuda-python",
 
653
  "repo_name": "truss",
654
  "repo_description": "The simplest way to serve AI/ML models in production",
655
  "homepage_link": "https://truss.baseten.co",
656
+ "closest_github_tag": "machine-learning"
 
 
 
 
 
 
 
 
 
 
 
 
 
657
  },
658
  {
659
  "github_repo_link": "https://github.com/laude-institute/terminal-bench",
 
667
  "repo_name": "goose",
668
  "repo_description": "an open source, extensible AI agent that goes beyond code suggestions - install, execute, edit, and test with any LLM",
669
  "homepage_link": "https://block.github.io/goose/",
670
+ "closest_github_tag": "mcp",
671
  "category": "agent"
672
  },
673
  {
 
675
  "repo_name": "Mooncake",
676
  "repo_description": "Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.",
677
  "homepage_link": "https://kvcache-ai.github.io/Mooncake/",
678
+ "closest_github_tag": "inference"
 
 
 
 
 
 
 
 
679
  },
680
  {
681
  "github_repo_link": "https://github.com/SWE-bench/SWE-bench",
682
  "repo_name": "SWE-bench",
683
  "repo_description": "SWE-bench: Can Language Models Resolve Real-world Github Issues?",
684
  "homepage_link": "https://www.swebench.com",
685
+ "closest_github_tag": "benchmark",
 
 
 
 
686
  "category": "benchmark"
687
  },
688
  {
 
700
  "github_repo_link": "https://github.com/triSYCL/triSYCL",
701
  "repo_name": "triSYCL",
702
  "repo_description": " Generic system-wide modern C++ for heterogeneous platforms with SYCL from Khronos Group",
703
+ "closest_github_tag": "opencl"
 
 
 
 
 
 
 
 
 
 
704
  },
705
  {
706
  "github_repo_link": "https://github.com/pybind/pybind11",
707
  "repo_name": "pybind11",
708
  "repo_description": "Seamless operability between C++11 and Python",
709
  "homepage_link": "https://pybind11.readthedocs.io/",
710
+ "closest_github_tag": "bindings"
711
  },
712
  {
713
  "github_repo_link": "https://github.com/andreinechaev/nvcc4jupyter",
 
720
  "repo_name": "rocSOLVER",
721
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
722
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
723
+ "closest_github_tag": "rocm"
724
  },
725
  {
726
  "github_repo_link": "https://github.com/ROCm/Tensile",
727
  "repo_name": "Tensile",
728
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
729
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
730
+ "closest_github_tag": "gpu"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
731
  },
732
  {
733
  "github_repo_link": "https://github.com/ROCm/rocPRIM",
734
  "repo_name": "rocPRIM",
735
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
736
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
737
+ "closest_github_tag": "hip"
738
  },
739
  {
740
  "github_repo_link": "https://github.com/ROCm/hipCUB",
 
747
  "repo_name": "rocFFT",
748
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
749
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
750
+ "closest_github_tag": "hip"
751
  },
752
  {
753
  "github_repo_link": "https://github.com/ROCm/rocSPARSE",
 
760
  "repo_name": "rocRAND",
761
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo ",
762
  "homepage_link": "https://github.com/ROCm/rocm-libraries",
763
+ "closest_github_tag": "hip"
764
  },
765
  {
766
  "github_repo_link": "https://github.com/ROCm/MIOpen",
767
  "repo_name": "MIOpen",
768
  "repo_description": "[DEPRECATED] Moved to ROCm/rocm-libraries repo",
769
+ "homepage_link": "https://github.com/ROCm/rocm-libraries"
770
  },
771
  {
772
  "github_repo_link": "https://github.com/Reference-LAPACK/lapack",
773
  "repo_name": "lapack",
774
  "repo_description": "LAPACK development repository",
775
+ "closest_github_tag": "linear-algebra"
 
 
 
 
 
 
 
 
 
 
 
776
  },
777
  {
778
  "github_repo_link": "https://github.com/ccache/ccache",
779
  "repo_name": "ccache",
780
  "repo_description": "ccache – a fast compiler cache",
781
  "homepage_link": "https://ccache.dev",
782
+ "closest_github_tag": "compiler",
 
 
 
 
 
 
 
 
 
 
 
783
  "category": "compiler"
784
  },
785
  {
 
787
  "repo_name": "omnitrace",
788
  "repo_description": "Omnitrace: Application Profiling, Tracing, and Analysis",
789
  "homepage_link": "https://rocm.docs.amd.com/projects/omnitrace/en/docs-6.2.4/",
790
+ "closest_github_tag": "performance-analysis"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
791
  },
792
  {
793
  "github_repo_link": "https://github.com/KhronosGroup/OpenCL-SDK",
 
799
  "repo_name": "synthetic-data-kit",
800
  "repo_description": "Tool for generating high quality Synthetic datasets",
801
  "homepage_link": "https://pypi.org/project/synthetic-data-kit/",
802
+ "closest_github_tag": "generation"
803
  },
804
  {
805
  "github_repo_link": "https://github.com/unslothai/unsloth",
806
  "repo_name": "unsloth",
807
  "repo_description": "Fine-tuning & Reinforcement Learning for LLMs. 🦥 Train OpenAI gpt-oss, DeepSeek-R1, Qwen3, Gemma 3, TTS 2x faster with 70% less VRAM.",
808
  "homepage_link": "https://docs.unsloth.ai/",
809
+ "closest_github_tag": "unsloth"
810
  },
811
  {
812
  "github_repo_link": "https://github.com/KhronosGroup/Vulkan-Docs",
 
823
  "repo_name": "Wan2.2",
824
  "repo_description": "Wan: Open and Advanced Large-Scale Video Generative Models",
825
  "homepage_link": "https://wan.video",
826
+ "closest_github_tag": "video-generation"
827
  },
828
  {
829
  "github_repo_link": "https://github.com/AMD-AGI/Primus-Turbo",
README.md CHANGED
@@ -4,7 +4,4 @@ language:
4
  pretty_name: PyTorch Conference 2025 GitHub Repos
5
  ---
6
  # PyTorch Conference 2025 GitHub Repos
7
- I created a list of every GitHub repo mentioned during PyTorch Conference 2025 and Open Source AI Week.
8
-
9
- # AI Generated Content!
10
- The fields `repo_description`, `homepage_link` and `repo_tags` come from AI generated code.
 
4
  pretty_name: PyTorch Conference 2025 GitHub Repos
5
  ---
6
  # PyTorch Conference 2025 GitHub Repos
7
+ I created a list of every GitHub repo mentioned during PyTorch Conference 2025 and Open Source AI Week.