Falconsai / text_summarization
Last scanned: Nov 12, 2024 at 7:00 PM UTC
Versions
Version | Commit message | Commit author | Last scanned | ||
---|---|---|---|---|---|
Create README.md | mstatt | Nov 12, 2024 | |||
Update README.md | mstatt | Nov 12, 2024 | |||
Upload tokenizer | mstatt | Nov 12, 2024 | |||
Upload T5ForConditionalGeneration | mstatt | Nov 12, 2024 | |||
Upload tokenizer | mstatt | Nov 12, 2024 | |||
Update README.md | mstatt | Nov 12, 2024 | |||
Update README.md | mstatt | Nov 12, 2024 | |||
Update config.json | RealFalconsAI | Nov 4, 2024 | |||
Adding ONNX file of this model (#4) | RealFalconsAI | Nov 8, 2024 | |||
Update README.md | VlastimilLeadspicker | Oct 22, 2024 | |||
Update README.md (#6) | RealFalconsAI | Nov 6, 2024 | |||
Add Core ML conversion | RyanZhuuuuu | Nov 12, 2024 | |||
Adding ONNX file of this model | jparkerweb | Nov 12, 2024 | |||
Adding `safetensors` variant of this model | SFconvertbot | Nov 12, 2024 |
Protect AI's security scanner detects threats in model files
With Protect AI's Guardian you can scan models for threats before ML developers download them for use, and apply policies based on your risk tolerance.
Found a new threat?
Earn bounties and get recognition for your discoveries by submitting them through Huntr.