-
Intel
- Shanghai
-
20:06
(UTC +08:00)
Pinned Loading
-
huggingface/transformers
huggingface/transformers Public🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.
-
llm-d/llm-d
llm-d/llm-d PublicAchieve state of the art inference performance with modern accelerators on Kubernetes
-
intel/neural-compressor
intel/neural-compressor PublicSOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, and ONNX Runtime
-
ai-dynamo/dynamo
ai-dynamo/dynamo PublicA Datacenter Scale Distributed Inference Serving Framework
-
ai-dynamo/aiconfigurator
ai-dynamo/aiconfigurator PublicOffline optimization of your disaggregated Dynamo graph
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.




