SoyNet is an inference optimizing solution for AI models.
-
Accelerate model inference by maximizing the utilization of numerous cores on the GPU without compromising accuracy (2x to 5x compared to Tensorflow).
-
Minimize GPU memory usage (1/5~1/15 level compared to Tensorflow).
※ Performance varies depends on the model and configuration environment.
- can support customer to provide AI applications and AI services in time. (Time to Market)
- can help application developers to easily execute AI projects without additional technical AI knowledge and experience.
- can help customer to reduce H/W (GPU, GPU server) or Cloud Instance cost for the same AI execution. (inference)
- can support customer to respond to real-time environments that require very low latency in AI inference.
- Dedicated engine for inference of deep learning models.
- Supports NVIDIA GPUs
- library files to be easiliy integrated with customer applications dll file (Windows)
- We can provide c++ and python executable files.
├─3party : third party file to run sample code
├─Samples : sample code of AI model on c++ (such as yolov5...)
| ├─model
| | ├─weights
| | | └─ww.py : you can make soynet weight file from your own weight file
| | └─model.cpp : execution file
| └─main.cpp : main on c++
├─SamplesPy : sample code on python
├─SoyNetV5 : include SoyNetV5.sln for running c++ code
├─bin : *.dll file for running soynetV5
├─data : sample data(such as jpg, mp4..) for sample code
├─include : header file for soynetV5
├─lib : lib for release
├─lib_debug : lib for debug
├─mgmt : SoyNet executuion env
│ ├─configs : Model definitions (*.cfg)
│ ├─engines : SoyNet engine files
│ ├─licenses : license file
│ ├─logs : SoyNet log files
│ └─weights : Weight files for SoyNet models (*.weights)
├─output : output direction when build sample code on c++
└─layer_dict_V5.1.0 : dictionary file for soynet layer
engines
: it's made at the first time execution or when you modify the configs file.weights
: You can make .weights file from your own trained weight file(ex. *.pt) on ww.py in weights folder.license file
: Please contact SoyNet if the time has passed.
-
initSoyNet(.cfg, extend_param)
: Created a SoyNet handle. -
feedData(handle, data)
: Put the data into the SoyNet handle. -
inference(handle)
: Start inference. -
getOutput(handle, output)
: Put the inference data into the output. -
freeSoyNet(handle)
: If you're done using a handle, destroy it.※
extend_param
extend_param
contains parameters necessary to define the model, such as input size, engine_serialize, batch_size ...- The parameters required may vary depending on the model.
※
engine_serialize
inextend_param
- This parameter determines whether to build a SoyNet engine or not.
- default is 0.
- If you run SoyNet for the first time or modify config or extended parameters, select one of the following two methods.
- Delete the existing generated bin (engine) file and run it again.
- Run by setting this value to 1 and then change back to 0.
CUDA version that GPU driver supports.
-
CUDA (>= 12.0)
※ You need to use .dll and .so files that match CDUA and TensorRT versions. If you want another version, Please contact SoyNet.
-
Tested GPU architecture : Pascal/Volta/Turing/Ampere/Ada Lovelac (ex: for PC Nvidia GTX 10xx, RTX 20xx/30xx/40xx, etc)
※ Please contact us for specific GPU support.
- OS : Windows 10 / 11