david

quyet7779
Β·

AI & ML interests

None yet

Recent Activity

liked a Space 6 months ago
Qwen/Qwen2.5-Coder-demo
View all activity

Organizations

clapAI's profile picture

quyet7779's activity

reacted to DawnC's post with πŸ”₯ 15 days ago
view post
Post
4238
I'm excited to introduce VisionScout β€”an interactive vision tool that makes computer vision both accessible and powerful! πŸ‘€πŸ”

What can VisionScout do right now?
πŸ–ΌοΈ Upload any image and detect 80 different object types using YOLOv8.
πŸ”„ Instantly switch between Nano, Medium, and XLarge models depending on your speed vs. accuracy needs.
🎯 Filter specific classes (people, vehicles, animals, etc.) to focus only on what matters to you.
πŸ“Š View detailed statistics about detected objects, confidence levels, and spatial distribution.
🎨 Enjoy a clean, intuitive interface with responsive design and enhanced visualizations.

What's next?
I'm working on exciting updates:
- Support for more models
- Video processing and object tracking across frames
- Faster real-time detection
- Improved mobile responsiveness

The goal is to build a complete but user-friendly vision toolkit for both beginners and advanced users.

Try it yourself! πŸš€
DawnC/VisionScout

I'd love to hear your feedback , what features would you find most useful? Any specific use cases you'd love to see supported?

Give it a try and let me know your thoughts in the comments! Stay tuned for future updates.

#ComputerVision #ObjectDetection #YOLO #MachineLearning #TechForLife
reacted to andito's post with πŸ”₯ 5 months ago
view post
Post
3396
Let's go! We are releasing SmolVLM, a smol 2B VLM built for on-device inference that outperforms all models at similar GPU RAM usage and tokens throughputs.

- SmolVLM generates tokens 7.5 to 16 times faster than Qwen2-VL! 🀯
- Other models at this size crash a laptop, but SmolVLM comfortably generates 17 tokens/sec on a macbook! πŸš€
- SmolVLM can be fine-tuned on a Google collab! Or process millions of documents with a consumer GPU!
- SmolVLM even outperforms larger models in video benchmarks, despite not even being trained on videos!

Check out more!
Demo: HuggingFaceTB/SmolVLM
Blog: https://huggingface.co/blog/smolvlm
Model: HuggingFaceTB/SmolVLM-Instruct
Fine-tuning script: https://github.com/huggingface/smollm/blob/main/finetuning/Smol_VLM_FT.ipynb
reacted to merve's post with πŸ”₯ 6 months ago
view post
Post
5301
OmniVision-968M: a new local VLM for edge devices, fast & small but performant
πŸ’¨ a new vision language model with 9x less image tokens, super efficient
πŸ“– aligned with DPO for reducing hallucinations
⚑️ Apache 2.0 license πŸ”₯

Demo hf.co/spaces/NexaAIDev/omnivlm-dpo-demo
Model https://huggingface.co/NexaAIDev/omnivision-968M
  • 4 replies
Β·
reacted to mrfakename's post with ❀️ 12 months ago