AI Accelerator Card SC5H

SOPHON SC5H adopts the standard half-height and half-length PCIE card size design and is equipped with a BM1684 chip and a side suction fan. It can be well adapted to complex working conditions and applied in distributed AI computing analysis on the edge side. Its computing capability is 10 times more than that of the Intel E5 CPU; at the same time, the power consumption is over 70% less than other products. Its typical power consumption during overall operation is less than 21W.

Consultancy

Provide 2.2T@FP32, 17.6T@INT8, 35.2T@INT8 (Winograd ON) super AI performance

High performance-consumption ratio for applications with high-performance requirements at the edge

Support multiple precision calculations such as FP32 and INT8

32 channels full HD video hard decoding capability, applicable to high-speed high-frame rate industrial cameras

2-channel HD video hard-coding capability, supporting 4K level semi-real-time encoding output

Video and picture decoding resolution range up to above 8K, suitable for all kinds of ultra-high-definition network cameras

Adapt to local workstation environments and domestic CPU systems such as Phytium and Shenwei.

Can be together used with SC5+ and accelerator cards and GPU cards of other brands to build a heterogeneous computing platform

Wide Application and Scenarios

SC5H is mainly applied in distributed AI computing and analysis scenarios on the edge side, such as traffic, urban management, communities and industrial inspection that require front-AI computing power; it can also be together used with other cards such as SC5 + on the same computing platform.

Easy-to-use, Convenient and Efficient

BMNNSDK (BITMAIN Neural Network SDK) one-stop toolkit provides a series of software tools including the underlying driver environment, compiler and inference deployment tool. The easy-to-use and convenient toolkit covers the model optimization, efficient runtime support and other capabilities required for neural network inference. It provides easy-to-use and efficient full-stack solutions for the development and deployment of deep learning applications. BMNNSDK minimizes the development cycle and cost of algorithms and software. Users can quickly deploy deep learning algorithms on various AI hardware products of SOPHGO to facilitate intelligent applications.

Support mainstream programming framework

More

Performance Parameter

AI Developer Portfolio

AI computing accelerator card

AI computing accelerator card

TPU core architecture

SOPHON

SOPHON

SOPHON

SOPHON

NPU core number

64

-

64

192

AI performance

FP32(FLOPS)

2.2T

-

2.2T

6.6T

INT8(OPS) Winograd OFF

17.6T

-

17.6T

52.8T

INT8(OPS) Winograd ON

35.2T

-

35.2T

105.6T

CPU

ARM 8-core A53 @ 2.3GHz

-

ARM 8-core A53 @ 2.3GHz

3x ARM 8-core A53 @ 2.3GHz

VPU

Video decoding capability

H.264:1080P @960fps
H.265:1080P @960fps

-

H.264:1080P @960fps
H.265:1080P @960fps

H.264:1080P @2880fps
H.265:1080P @2880fps

Video decoding resolution

CIF / D1 / 720P / 1080P / 4K(3840×2160) / 8K(8192×4096)

-


CIF / D1 / 720P / 1080P / 4K(3840×2160) / 8K(8192×4096)

CIF / D1 / 720P / 1080P / 4K(3840×2160) / 8K(8192×4096)

Video encoding capability

H.264:1080P @50fps
H.265:1080P @50fps

-

H.264:1080P @50fps
H.265:1080P @50fps

H.264:1080P @150fps
H.265:1080P @150fps

Video encoding resolution

CIF / D1 / 720P / 1080P / 4K(3840×2160)

-

CIF / D1 / 720P / 1080P / 4K(3840×2160)

CIF / D1 / 720P / 1080P / 4K(3840×2160)

Video transcoding capability (1080P to CIF)

Max. 18 channels

-

Max. 18 channels

Max. 54 channels

JPU

JPEG image decoding capability

480 images / second @ 1080p

-

480 images / second @ 1080p

1440 images / second @ 1080p

Maximum resolution (pixels)

32768×32768

-

32768×32768

32768×32768

System interface

Data link

EP PCIE X8
RC PCIE X8

PCIE X2

PCIE X16

PCIE X8

Operating mode

EP+RC

SOC extension

EP

EP

Physical / power interface

PCIE X16

12VDC Jack

PCIE X16

PCIE X16

RAM

Standard configuration

12GB

-

12GB

36GB

Maximum capacity

16GB

-

16GB

48GB

Power consumption

30W MAX

No load: 6W
With load: 30W

30W MAX

75W MAX

Heat dissipation mode

active

-

active

passive

Working status display

N/A

LED x3 (power / hard disk / status)

LED x1

LED x1

External I/O expansion *

SD-Card

1

-

-

RESET Button

1

-

-

RJ45

2 *1000Base-T

-

-

USB

4

-

-

SATA

1

-

-

4G/LTE

1

-

-

micro USB

1

-

-

working temperature

0℃-55℃

-10℃-55℃

0℃-55℃

Deep learning framework

Caffe / TensorFlow / Pytorch / Mxnet / Darknet / Paddle

Operating system support

Ubuntu / CentOS / Debian

compatibility

Compatible with mainstream x86 architecture and ARM architecture servers

Localization support

Support domestic CPU system such as Feiteng, Shenwei, Zhaoxin, etc.; support domestic Linux operating system such as Kylin, Deepin, etc.; support domestic AI framework Paddle Lite

Length x height x width (including bracket)

200x111.2x19.8mm

206x28.5x59.5mm

169.1x68.9x19mm

169.1x68.9x19.5mm

* All external I/O expansion interfaces in the AI developer portfolio must be used with SC5-IO