Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix TF export for BottleneckCSP layer #7330

Merged
merged 2 commits into from
Apr 7, 2022

Conversation

nrupatunga
Copy link
Contributor

@nrupatunga nrupatunga commented Apr 7, 2022

The current implementation of BottleneckCSP in common.py uses SiLU activation, whereas the tf.py script for exporting uses ReLU, This PR fixes the same.

🛠️ PR Summary

Made with ❤️ by Ultralytics Actions

🌟 Summary

Improved activation function in TensorFlow models within the YOLOv5 repository.

📊 Key Changes

  • Changed activation function from ReLU (Rectified Linear Unit) with alpha=0.1 to Swish in TensorFlow models.

🎯 Purpose & Impact

  • The purpose of this change is to enhance the performance of the neural network by using the Swish activation function, which has shown better performance in some scenarios compared to ReLU.
  • Users may experience improved model accuracy and training stability.
  • 🔍 Swish is known to be smoother and can help mitigate vanishing gradient problems, leading to potentially better convergence properties during model training.

@glenn-jocher glenn-jocher merged commit b7faeda into ultralytics:master Apr 7, 2022
@glenn-jocher
Copy link
Member

@nrupatunga good catch! PR is merged. Thank you for your contributions to YOLOv5 🚀 and Vision AI ⭐

BjarneKuehl pushed a commit to fhkiel-mlaip/yolov5 that referenced this pull request Aug 26, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants