The SC3 accelerator card is equipped with a BM1682 chip, which provides 3TFLOPS single-precision floating-point computing capability, and the actual computing power utilization rate is significantly higher than similar competitors. SC3 provides two product forms of active / passive heat dissipation, which can be deployed on the server or industrial computer production environment on demand in the cloud and on the edge. It is suitable for business scenarios that have strict requirements on computing accuracy, such as industrial, medical, and dangerous goods management etc.
Each BM1682 chip has 64 NPU processing units, and each NPU has 32 EU arithmetic units. A single BM1682 chip can provide up to 3TFLOPs of single-precision peak computing power. At the same time, the chip has up to 16MB of on-chip SRAM, which can greatly reduce data handling during model calculation, improve performance and reduce power consumption.
High-performance deep learning accelerator card with completely independent intellectual property rights
On-chip hardware decoding engine supports HD video stream decoding from 1080P@240fps to 4K@60fps
Rich tool chain, support Caffe / TensorFlow / Pytorch / Mxnet and other deep learning frameworks
Passed CE / FCC and other international standard certification
Support PCIE 3.0 interface, compatible with mainstream x86 servers, easy to apply and expand
Passive cooling, fanless design
Sophon SC3 deep learning accelerator card can be used in various artificial intelligence, machine vision, high-performance computing environments, supporting facial feature detection, extraction, tracking, recognition, comparison, machine vision, and video structured analysis and processing Video structured applications such as image search and track tracking.
BMNNSDK (SOPHON Neural Network SDK) one-stop toolkit provides a series of software tools such as the underlying driver environment, compiler, inference deployment tool and so on. Easy to use and convenient, covering the model optimization, efficient runtime support and other capabilities required for the neural network inference stage, providing easy-to-use and efficient full-stack solutions for deep learning application development and deployment. BMNNSDK minimizes the development cycle and cost of algorithms and software. Users can quickly deploy deep learning algorithms on various AI hardware products of Fortune Group to facilitate intelligent applications.
NPU Core number
PCI Express 3.0 x8
Video Decoder format
H.264/H.265 / HEVC / MPEG1 / 2 / 4 / DivX / XviD / H.263 / VC-1 / Sorenson / VP8 / AVS
Video decoder performance
1080p @ 240fps or 4K @60fps
Ubuntu16.04 / CentOS7.4/Debian9.4
Operating Environment Temperature
Operating Environment Humidity
Form Factor Length*Height*Thickness