jameslahm / yolov10x
Last scanned: Nov 5, 2024 at 12:37 PM UTC
Versions
Version | Commit message | Commit author | Last scanned | ||
---|---|---|---|---|---|
initial commit | jameslahm | Never scanned | |||
Push model using huggingface_hub. | jameslahm | Never scanned | |||
Update README.md | jameslahm | Never scanned | |||
Update config.json | jameslahm | Never scanned | |||
Push model using huggingface_hub. | jameslahm | Never scanned | |||
Update README.md | jameslahm | Nov 12, 2024 | |||
Update README.md | jameslahm | Nov 12, 2024 | |||
Add library name (#4) | jameslahm | Oct 19, 2024 | |||
Update config.json | RaeiW | Nov 5, 2024 | |||
Add library name | nielsr | Nov 12, 2024 |
Protect AI's security scanner detects threats in model files
With Protect AI's Guardian you can scan models for threats before ML developers download them for use, and apply policies based on your risk tolerance.
Found a new threat?
Earn bounties and get recognition for your discoveries by submitting them through Huntr.