Skip to content

This repo contains model compression(using TensorRT) and documentation of running various deep learning models on NVIDIA Jetson Orin, Nano (aarch64 architectures)

Notifications You must be signed in to change notification settings

build-error/NVIDIA_Jetson_Inference_be

 
 

Repository files navigation

About

This repo contains model compression(using TensorRT) and documentation of running various deep learning models on NVIDIA Jetson Orin, Nano (aarch64 architectures)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Makefile 81.9%
  • C 9.3%
  • C++ 2.2%
  • Roff 2.1%
  • Shell 1.1%
  • CMake 0.8%
  • Other 2.6%