Piotr Skalski's picture

Piotr Skalski PRO

SkalskiP

AI & ML interests

Computer Vision | Multimodality

Recent Activity

Organizations

Hugging Face Fellows's profile picture Gradio-Blocks-Party's profile picture Roboflow's profile picture ZeroGPU Explorers's profile picture Social Post Explorers's profile picture

SkalskiP's activity

reacted to their post with โค๏ธ 6 months ago
view post
Post
YOLO-World: Real-Time, Zero-Shot Object Detection ๐Ÿ”ฅ ๐Ÿ”ฅ ๐Ÿ”ฅ

YOLO-World was designed to solve a limitation of existing zero-shot object detection models: speed. Whereas other state-of-the-art models use Transformers, a powerful but typically slower architecture, YOLO-World uses the faster CNN-based YOLO architecture.

YOLO-World provides three models: small with 13M (re-parametrized 77M), medium with 29M (re-parametrized 92M), and large with 48M (re-parametrized 110M) parameters.

The YOLO-World team benchmarked the model on the LVIS dataset and measured their performance on the V100 without any performance acceleration mechanisms like quantization or TensorRT.

According to the paper, YOLO-World reached 35.4 AP with 52.0 FPS for the L version and 26.2 AP with 74.1 FPS for the S version. While the V100 is a powerful GPU, achieving such high FPS on any device is impressive.

- ๐Ÿ”— YOLO-World arXiv paper: https://lnkd.in/ddRBKCCX
- ๐Ÿ”— my YOLO-World technical report: https://blog.roboflow.com/what-is-yolo-world
- ๐Ÿค— YOLO-World space: SkalskiP/YOLO-World
reacted to clem's post with ๐Ÿš€ 9 months ago
view post
Post
3686
Who said you couldn't build a big business based on open-source AI? Congrats Mistral team: https://huggingface.co/mistralai
reacted to their post with ๐Ÿค— about 1 year ago
view post
Post
YOLO-World: Real-Time, Zero-Shot Object Detection ๐Ÿ”ฅ ๐Ÿ”ฅ ๐Ÿ”ฅ

YOLO-World was designed to solve a limitation of existing zero-shot object detection models: speed. Whereas other state-of-the-art models use Transformers, a powerful but typically slower architecture, YOLO-World uses the faster CNN-based YOLO architecture.

YOLO-World provides three models: small with 13M (re-parametrized 77M), medium with 29M (re-parametrized 92M), and large with 48M (re-parametrized 110M) parameters.

The YOLO-World team benchmarked the model on the LVIS dataset and measured their performance on the V100 without any performance acceleration mechanisms like quantization or TensorRT.

According to the paper, YOLO-World reached 35.4 AP with 52.0 FPS for the L version and 26.2 AP with 74.1 FPS for the S version. While the V100 is a powerful GPU, achieving such high FPS on any device is impressive.

- ๐Ÿ”— YOLO-World arXiv paper: https://lnkd.in/ddRBKCCX
- ๐Ÿ”— my YOLO-World technical report: https://blog.roboflow.com/what-is-yolo-world
- ๐Ÿค— YOLO-World space: SkalskiP/YOLO-World
posted an update about 1 year ago
view post
Post
YOLO-World: Real-Time, Zero-Shot Object Detection ๐Ÿ”ฅ ๐Ÿ”ฅ ๐Ÿ”ฅ

YOLO-World was designed to solve a limitation of existing zero-shot object detection models: speed. Whereas other state-of-the-art models use Transformers, a powerful but typically slower architecture, YOLO-World uses the faster CNN-based YOLO architecture.

YOLO-World provides three models: small with 13M (re-parametrized 77M), medium with 29M (re-parametrized 92M), and large with 48M (re-parametrized 110M) parameters.

The YOLO-World team benchmarked the model on the LVIS dataset and measured their performance on the V100 without any performance acceleration mechanisms like quantization or TensorRT.

According to the paper, YOLO-World reached 35.4 AP with 52.0 FPS for the L version and 26.2 AP with 74.1 FPS for the S version. While the V100 is a powerful GPU, achieving such high FPS on any device is impressive.

- ๐Ÿ”— YOLO-World arXiv paper: https://lnkd.in/ddRBKCCX
- ๐Ÿ”— my YOLO-World technical report: https://blog.roboflow.com/what-is-yolo-world
- ๐Ÿค— YOLO-World space: SkalskiP/YOLO-World
reacted to their post with ๐Ÿ‘๐Ÿคฏโค๏ธ about 1 year ago
view post
Post
Real-Time Vehicle Speed Estimation Tutorial ๐Ÿš—๐Ÿ’จ๐Ÿ’จ๐Ÿ’จ

TL;DR: Watch the tutorial here: https://www.youtube.com/watch?v=uWP6UjDeZvY

Key Steps:
1. Vehicle Detection: Before we jump into speed estimation, we begin by detecting moving vehicles. I demonstrate this using YOLOv8, deployed through the Inference pip package.

2. Tracking with ByteTrack: For effective object tracking, ByteTrack is my tool of choice. It assigns a unique ID to each vehicle, which is essential for accurately monitoring the distance each car travels. This forms the cornerstone of our speed calculation process.

3. Distance Calculation Complexities: Calculating traveled distance can be tricky due to perspective distortion from the camera. A car moving at a constant speed will appear to move a different number of pixels in the image, depending on its distance from the camera.

4. Vehicle Positioning: We can accurately pinpoint each vehicle's position within our monitored area. By representing each vehicle with x and y coordinates in meters, we can compare its current and past positions, paving the way for calculating its speed.

5. We store the position of each car in the last second, calculate the offset, and divide it by the time delta to get the local speed.

- ๐Ÿ”— tutorial: https://www.youtube.com/watch?v=uWP6UjDeZvY
- ๐Ÿ”— code: https://github.com/roboflow/supervision/tree/develop/examples/speed_estimation
  • 2 replies
ยท
posted an update about 1 year ago
view post
Post
Real-Time Vehicle Speed Estimation Tutorial ๐Ÿš—๐Ÿ’จ๐Ÿ’จ๐Ÿ’จ

TL;DR: Watch the tutorial here: https://www.youtube.com/watch?v=uWP6UjDeZvY

Key Steps:
1. Vehicle Detection: Before we jump into speed estimation, we begin by detecting moving vehicles. I demonstrate this using YOLOv8, deployed through the Inference pip package.

2. Tracking with ByteTrack: For effective object tracking, ByteTrack is my tool of choice. It assigns a unique ID to each vehicle, which is essential for accurately monitoring the distance each car travels. This forms the cornerstone of our speed calculation process.

3. Distance Calculation Complexities: Calculating traveled distance can be tricky due to perspective distortion from the camera. A car moving at a constant speed will appear to move a different number of pixels in the image, depending on its distance from the camera.

4. Vehicle Positioning: We can accurately pinpoint each vehicle's position within our monitored area. By representing each vehicle with x and y coordinates in meters, we can compare its current and past positions, paving the way for calculating its speed.

5. We store the position of each car in the last second, calculate the offset, and divide it by the time delta to get the local speed.

- ๐Ÿ”— tutorial: https://www.youtube.com/watch?v=uWP6UjDeZvY
- ๐Ÿ”— code: https://github.com/roboflow/supervision/tree/develop/examples/speed_estimation
  • 2 replies
ยท
reacted to artificialguybr's post with ๐Ÿค— about 1 year ago
view post
Post
Cool feature! Thanks, HF, for allowing me test it.
  • 2 replies
ยท