Benchmarking Bimanual Cloth Manipulation

Irene Garcia-Camacho*, Martina Lippi*, Michael C. Welle, Hang Yin, Rika Antonova, Anastasiia Varava, Júlia Borràs, Carme Torras, Alessandro Marino, Guillem Alenyà, Danica Kragic

Cloth manipulation is a challenging task that, despite its importance, has received relatively little attention compared to rigid object manipulation. In this paper, we provide three benchmarks for evaluation and comparison of different approaches towards three basic tasks in cloth manipulation: spreading a tablecloth over a table, folding a towel, and dressing. The tasks can be executed on any bimanual robotic platform and the objects involved in the tasks are standardized and easy to acquire. We provide several complexity levels for each task, and describe the quality measures to evaluate task execution. Furthermore, we provide baseline solutions for all the tasks and evaluate them according to the proposed metrics.

Download Paper

Download Bimanual Cloth Manipulation Benchmarks Resources

Benchmark 1: Spreading a tablecloth

This task consists in taking a tablecloth and spreading it on a table.

Tablecloth spreading provides four possible starting configurations using a defined tablecloth

  • [pg2] cloth is pre grasped on two grasping points
  • [pg1] cloth is pre grasped on one grasping point
  • [fd] cloth is folded
  • [cr] cloth is crumpled

Benchmark 2: Folding a towel

This task consists in taking a towel and folding it.

Towel folding provides four possible starting configurations using two different sized towels for up to three folds

  • [pg2] cloth is pre grasped on two grasping points
  • [pg1] cloth is pre grasped on one grasping point
  • [ft] cloth is flat on the table
  • [cr] cloth is crumpled

Benchmark 3: Partial dressing

This task consists in putting a T-shirt over a simple head model.

Partial dressing provides five possible starting configurations using two different sized heads depending on the T-shirt size

  • [pg2] cloth is pre grasped on two grasping points
  • [pg1] cloth is pre grasped on one grasping point
  • [ft] cloth is flat on the table
  • [fd] cloth is folded
  • [cr] cloth is crumpled

Scoring

The scoring considers the following list of performance metrics: success of each phase, execution time, force measures and quality measures

Success of each phase - Benchmark 1: Spreading a tablecloth

MAN

The tablecloth is sucessfully spread if it covers the table top

GR2

Grasping is maintained during the manipulation phase.

GR1

Grasping is maintained during the manipulation phase.

When grasping with starting configuration [fd]: The tablecloth needs to unfold after the first grasp.

Success of each phase - Benchmark 2: Folding a towel

MAN

The towel is successfully folded if one fold has been done

GR2

Grasping is maintained during the manipulation phase.

GR1

Grasping is maintained during the manipulation phase.

When grasping with starting configuration [fd]: The towel needs to unfold after the first grasp.

Success of each phase - Benchmark 3: Partial dressing

MAN

The T-shirt is successfully dressed when the neck hole of it is put over the head and the entirety of the T-shirt lies below the head.

GR2

Grasping is maintained during the manipulation phase.

GR1

Grasping is maintained during the manipulation phase.

When grasping with starting configuration [fd]: The T-shirt needs to unfold after the first grasp.

Task error measures - Benchmark 1: Spreading a tablecloth

These measures of error can only be applied if the task has been finished with success. A 100% rotation error occurs when the tablecloth is rotated π/4 radians. The maximum translation error occurs when one of the hanging parts is zero, meaning the table is almost uncovered

Task error measures - Benchmark 2: Folding a towel

where Af is the final area of the cloth from the top view, and Ai is the initial area of the cloth. Assuming Af will always be smaller than Ai, 100% error occurs when Ai = Af , but also if Af is less than half of Ai, which can only happen if there are wrinkles or extra folds.

Task error measures - Benchmark 3: Partial dressing

The dressing task is binary and is accomplished when the neck hole of the T-shirt is put over the head and the entirety of the T-shirt lies below the head.

The execution time and force measures are identical for each task.

Execution time: The execution time comprises the times needed for all the phases. It starts when the first robot moves.

Force measures: Force measures quantify the interaction between the robots and the environment; they are only acquired during phase [MAN] and minimum, maximum and average norms at the end effectors are considered. Note that, in order to not limit the possibility of using the benchmark, force measures are not mandatory but are highly encouraged especially in dressing tasks, where monitoring of exerted forces on the head represents a key feature

Reporting tables

The final scoring is done in the following reporting tables:

Download all scoring sheets

Resources

Protocol and benchmark documents as well as a list of needed objects can be found here:

Protocols:

Download Task 1: spreading a tablecloth protocol v1.0 (RAL-SI-2020-P19-0832_1-V1.0)

Download Task 2: folding a towel protocol v1.0 (RAL-SI-2020-P19-0832_2-V1.0)

Download Task 3: partial dressing protocol v1.0 (RAL-SI-2020-P19-0832_3-V1.0)

Protocols explanatory videos

Benchmarks:

Download Task 1: spreading a tablecloth benchmark v1.0 (RAL-SI-2020-B19-0832_1-V1.0)

Download Task 2: folding a towel benchmark v1.0 (RAL-SI-2020-B19-0832_2-V1.0)

Download Task 3: Partial dressing benchmark v1.0 (RAL-SI-2020-B19-0832_3-V1.0)

Objects

Table:

Any table in the allowed range. Example is RYDEBÄCK from IKEA

Tablecloth:

Small towel:

IKEA towel HÄREN 30x50 cm or 30x30 cm

Big towel:

IKEA towel HÄREN 50x100 cm or 40x70 cm

T-shirt:

Any T-shirt in the specified range. For example sizes S-XL of the YCB-T-shirt

Big and small head:

Results

To showcase how the presented Benchmark should be used and promote comparison of different methods we describe our own systems tackling the Benchmark.

Benchmark 1: Spreading a tablecloth

The solution for [GR1] to unfold the tablecloth is to grasp the first grasping point and pull the cloth up. The solution for [GR2] is to grasp a point on the edge of the cloth next to the first hand and trace the edge until the corner at the end of the edge is reached. This implies slipping the cloth inside the gripper without loosing it. Each grasp is performed by a different TIAGo robot, and after they have the tablecloth grasped, they move across the table to spread the tablecloth on the table.

Download Benchmark 1: complete scoring sheet

View Benchmark 1: execution videos

Download Benchmark 1: comments on solution

Benchmark 2: folding a towel

For the task of folding a towel, we use the same robotic system as in the previous task and we consider the big towel ( [bt] ) with measures 0.5×1 m. The two mobile manipulators are placed at different sides of the table where the towel is to be folded. We just report the [MAN] phase, as this is the most different phase compared to benchmark 1. Thus, we start with the two corners already grasped, and we then perform the folding motion. For the folding strategy, we only focus on the first fold and we use a Dynamic Movement Primitives (DMP) representation of the motion for each robot learned by demonstration, and execute both trajectories in synchronization. The size of the towel and the localization of the robot with respect to the table are assumed to be known.

Download Benchmark 2: complete scoring sheet

View Benchmark 2: execution videos

Download Benchmark 2: comments on solution

Benchmark 3: Partial dressing

For the dressing task, we propose a human inspired solution based on a vision/force-feedback informed strategy with hand-tuned hyperparameters. Concerning the grasping phases [GR1] and [GR2] , pre- defined grasping poses are selected with initial configurations [pg1] , [cr] , [ft] and [fd] , thus no visual feedback is exploited in these stages. Concerning the manipulation phase [MAN] , the basic idea is to use the visual information to guide the team motion and, at the same time, to perform random wiggling motions which emulate human-like dressing. More specifically, a vision system is designed which splits the top-view circle associated with the head into two halves and measures the free area in each of them. These measures are then exploited to determine in which direction to move the team centroid in such a way that both areas exceed a certain threshold.

Download Benchmark 3: complete scoring sheet

View Benchmark 3: execution videos

Download Benchmark 3: comments on solution

Benchmark 3: Partial dressing second shirt

Download Benchmark 3: complete scoring sheet

View Benchmark 3: execution videos second shirt

Contact

  • Irene Garcia-Camacho*, igarcia@iri.upc.edu, Institut de Robòtica i Informàtica Industrial, Spain
  • Martina Lippi*; lippi@kth.se, mlippi@unisa.it; KTH Royal Institute of Technology, Sweden and University of Salerno, Italy
  • Michael C. Welle; mwelle@kth.se; KTH Royal Institute of Technology, Sweden
  • Hang Yin; hyin@kth.se; KTH Royal Institute of Technology, Sweden
  • Rika Antonova; antonova@kth.se; KTH Royal Institute of Technology, Sweden
  • Anastasiia Varava; varava@kth.se; KTH Royal Institute of Technology, Sweden
  • Júlia Borràs; jborras@iri.upc.edu; Institut de Robòtica i Informàtica Industrial, Spain
  • Carme Torras; torras@iri.upc.edu; Institut de Robòtica i Informàtica Industrial, Spain
  • Alessandro Marino; al.marino@unicas.it; University of Cassino and Southern Lazio, Italy
  • Guillem Alenyà; galenya@iri.upc.edu; Institut de Robòtica i Informàtica Industrial, Spain
  • Danica Kragic; dani@kth.se; KTH Royal Institute of Technology, Sweden