gptkbp:instanceOf
|
machine learning serving software
|
gptkbp:deployment
|
gptkb:cloud_service
on-premises
edge
|
gptkbp:developedBy
|
gptkb:NVIDIA
|
gptkbp:enables
|
gptkb:model
gptkb:HTTP/REST_API
gptkb:gRPC_API
model monitoring
model versioning
model explainability
C API
dynamic batching
model ensemble
multi-model serving
metrics export
|
gptkbp:firstReleased
|
2019
|
https://www.w3.org/2000/01/rdf-schema#label
|
Triton Inference Server
|
gptkbp:integratesWith
|
gptkb:Amazon_SageMaker
gptkb:Prometheus
gptkb:NVIDIA_Clara
gptkb:Google_Vertex_AI
gptkb:Kubeflow
gptkb:NVIDIA_DeepStream
gptkb:NVIDIA_NGC
gptkb:Grafana
gptkb:Kubernetes
gptkb:Azure_ML
gptkb:MLflow
|
gptkbp:latestReleaseVersion
|
2024.1
|
gptkbp:license
|
gptkb:Apache_License_2.0
|
gptkbp:officialWebsite
|
https://developer.nvidia.com/nvidia-triton-inference-server
|
gptkbp:platform
|
gptkb:Python
gptkb:TensorFlow
gptkb:DALI
gptkb:FIL
gptkb:NVIDIA_GPU
gptkb:ONNX_Runtime
gptkb:OpenVINO
gptkb:TensorRT
gptkb:ARM_CPU
gptkb:x86_CPU
gptkb:JAX
gptkb:PyTorch
gptkb:Hugging_Face_Transformers
|
gptkbp:repository
|
https://github.com/triton-inference-server/server
|
gptkbp:usedFor
|
AI inference
batch inference
real-time inference
deep learning model deployment
multi-modal inference
scalable inference
|
gptkbp:bfsParent
|
gptkb:NVIDIA_TAO_Toolkit
gptkb:Nvidia_DeepStream
gptkb:Transformers_library
gptkb:英伟达
|
gptkbp:bfsLayer
|
6
|