Vision-Based Vegetable Sorting using UR3e, RealSense D435i & YOLOv8

Vision-Based Vegetable Sorting using UR3e, RealSense D435i & YOLOv8

A ROS2-based perception and manipulation pipeline using UR3e, YOLOv8, RealSense D435i, MoveIt2, and Robotiq 2F-140 for classification, localization, and autonomous pick-and-place.

UR3eROS2PerceptionYOLOv8MoveIt2RealSense D435iRobotiq 2F-140

System Architecture

System Diagram

Build Log & Details

Team

This project was developed collaboratively as part of ROS2 Case Study
at Deggendorf Institute of Technology.

Team Members:

  • Ruchit Bhanushali — ROS2 Integration, Motion Planning, MoveIt2, Simulation
  • Siddharth Ahuja — Perception
  • Sahil Gore — Dataset & YOLOv8 Model
  • Kaung Sett Thu — RealSense Calibration & Localization

Overview

This project implements a vision-based automated vegetable sorting system using:

  • UR3e Collaborative Robotic Arm
  • Robotiq 2F-140 Parallel Gripper
  • Intel RealSense D435i RGB-D Camera
  • YOLOv8 for object detection
  • ROS2 + MoveIt2 for motion planning and control

The system detects carrots, tomatoes, and potatoes, performs 3D localization, computes grasp pose, and executes autonomous pick-and-place.

The goal is to build a fully automated, data-driven sorting pipeline optimized for accuracy, gentle grasping, and industrial throughput.


Problem Statement

  • Manual sorting suffers from:

    • Inconsistent quality & human error
    • Labor dependency and rising costs
    • Limited throughput
    • Product damage
    • Lack of intelligent automation

Our system solves these limitations through robotic manipulation + AI-driven perception.


System Proposal

AI-powered UR3e-based vegetable sorting, capable of:

  • real-time detection
  • stable 3D localization
  • adaptive grasping
  • repeatable automation

Hardware Setup

UR3e Collaborative Arm

  • 6-DOF, force-controlled, reliable for pick-and-place.

Robotiq 2F-140 Gripper

  • Adaptive parallel motion, 140 mm opening for variable vegetable sizes.

Intel RealSense D435i

  • RGB-D sensing + IMU for real-time depth & pose estimation.

ROS2 Workstation

  • Runs YOLOv8 inference, MoveIt2 planning, TF frames, and control nodes.

Software Stack

  • ROS2 Humble — Core middleware
  • YOLOv8 — Real-time vegetable detection
  • MoveIt2 — Grasp generation + trajectory planning
  • RViz2 — Visualization & simulation

Data Flow Pipeline

  • RGB-D Capture (RealSense D435i)
  • YOLOv8 Detection
  • Depth Cloud Cropping
  • Grasp Pose Computation
  • Pick-and-Place Planning
  • UR3e Execution

This creates a fully automated loop from perception to actuation.


Dataset

  • Public Roboflow dataset
  • 683 images
  • Augmented for model robustness
  • Classes: carrot, tomato, potato

Current Status

Completed:

  • YOLOv8 vegetable detection
  • Depth cloud processing
  • UR3e simulation in RViz & MoveIt2
  • Object localization
  • Basic pick-and-place execution

Next Steps:

  • Real-world execution on the UR3e
  • Final calibration & validation
  • Full documentation

References

  1. Spanu, A. et al., 2023. Vision-Based Robotic Sorting System for Agricultural Products. Politecnico di Torino.
    https://webthesis.biblio.polito.it/33164/

  2. Iftikhar, M. et al., 2024. Computer Vision as a Tool to Support Quality Control and Robotic Handling of Fruit: A Case Study.
    https://www.researchgate.net/publication/385205971

  3. Wu, Q. et al., 2023. Vegetable Disease Detection Using an Improved YOLOv8 Algorithm in Greenhouse Plant Environments.
    https://www.researchgate.net/publication/378371592

← Back to Portfolio