NVIDIA T4

                    FLEXIBLE DESIGN, BREAKTHROUGH PERFORMANCE

                    中文字幕亂倫免賛視頻

                    Next-Level Acceleration Has Arrived

                    We’re racing toward the future where every customer interaction, every product, and every service offering will be touched and improved by AI. Realizing that the future requires a computing platform that can accelerate the full diversity of modern AI, enabling businesses to create new customer experiences, reimagine how they meet—and exceed—customer demands, and cost-effectively scale their AI-based products and services.

                    The NVIDIA? T4 GPU accelerates diverse cloud workloads, including high-performance computing, deep learning training and inference, machine learning, data analytics, and graphics. Based on the new NVIDIA Turing? architecture and packaged in an energy-efficient 70-watt, small PCIe form factor, T4 is optimized for mainstream computing environments and features multi-precision Turing Tensor Cores and new RT Cores. Combined with accelerated containerized software stacks from NGC, T4 delivers revolutionary performance at scale.

                    Virtual Compute-Intensive Server Workloads

                    Virtual Compute-Intensive Server Workloads

                    NVIDIA vComputeServer provides the ability to virtualize GPUs and accelerate compute-intensive server workloads, including AI, Deep Learning, and Data Science. With this, automotive manufacturers can use the latest in simulation and compute technologies to create the most fuel efficient and stylish designs and researchers can analyze the function of genes to develop medical treatments more quickly.

                    A new partnership is advancing the hybrid cloud to power modern enterprise workloads.

                    Breakthrough Performance

                    T4 introduces the revolutionary Turing Tensor Core technology with multi-precision computing to handle diverse workloads. Powering breakthrough performance from FP32 to FP16 to INT8, as well as INT4 precisions, T4 delivers up to 40X higher performance than CPUs.

                    Breakthrough Inference Performance

                    SEE HOW YOU CAN ACCELERATE YOUR AI MODELS WITH MIXED PRECISION ON TENSOR CORES

                    STATE-OF-THE-ART INFERENCE IN REAL-TIME

                    Responsiveness is key to user engagement for services such as conversational AI, recommender systems, and visual search. As models increase in accuracy and complexity, delivering the right answer right now requires exponentially larger compute capability. T4 delivers up to 40X times better throughput, so more requests can be served in real time.

                    T4 INFERENCE PERFORMANCE

                    Resnet50

                    DeepSpeech2

                    GNMT

                    Video Transcoding Performance

                    As the volume of online videos continues to grow exponentially, demand for solutions to efficiently search and gain insights from video continues to grow as well. T4 delivers breakthrough performance for AI video applications, with dedicated hardware transcoding engines that bring twice the decoding performance of prior-generation GPUs. T4 can decode up to 38 full-HD video streams, making it easy to integrate scalable deep learning into video pipelines to deliver innovative, smart video services.

                    NVIDIA T4 Specifications

                     

                    Performance

                    Turing Tensor Cores
                    320

                    NVIDIA CUDA? cores
                    2,560

                    Single Precision Performance (FP32)
                    8.1 TFLOPS

                    Mixed Precision (FP16/FP32)
                    65 FP16 TFLOPS

                    INT8 Precision
                    130 INT8 TOPS

                    INT4 Precision
                    260 INT4 TOPS

                     
                     

                    Interconnect

                    Gen3
                    x16 PCIe

                     
                     

                    Memory

                    Capacity
                    16 GB GDDR6

                    Bandwidth
                    320+ GB/s

                     
                     

                    Power

                    70 watts

                     

                    NVIDIA AI Inference Platform

                    Explore the World's Most Advanced Inference Platform.