Forum to Develop an Open-Source Ecosystem for Robotic Manipulation
Robotics: Science and Systems (RSS) 2023 Conference Workshop
This half-day workshop was held during the Robotics: Science and Systems (RSS) 2023 conference on July 10, 2023 in Daegu, Republic of Korea.
Workshop Overview
To support and improve the development and deployment of robotic manipulation systems, an open-source ecosystem (OSE) is being scoped to facilitate development and dissemination of open-source assets (i.e., hardware, software, datasets), benchmarking practices, and sharing of results. A distributed, community-driven platform is envisioned where researchers and developers can share and learn about these open-source resources, find tools to easily utilize them, collaborate on developing systematic robot experimentation methodologies, and disseminate their findings effectively. As such, the OSE proposes to address the issues facing the advancement of robot manipulation due to the lack of systematic development and benchmarking methodologies, as well as the unique challenges related to physical assets (both equipment and benchmarking tools), which prevent this community from effectively utilizing such resources compared to other research domains like computer vision.
This workshop is one of several scoping activities conducted this year to support the development of the proposed OSE: two workshops at high-profile robotics conferences (HRI 2023 in March, ICRA 2023 in May), a user survey (with over 100 respondents as of this workshop submission), and multiple deep dive meetings with key stakeholders in the field. These activities will be completed prior to this workshop, which is being proposed as a forum rather than a traditional conference workshop in order to facilitate interactive, guided discussions on our lessons learned from these prior activities and receive detailed feedback from the attendees to guide the development of the OSE.
Key Takeaways
Based on the discussions had at the workshop, a set of key takeaways have been summarized and organized into topics below:
Current landscape: Issues
Ensure the inclusion of simulation assets in ecosystem definition
Ontologies of tasks are needed such that performance thresholds in each can be set
For publications and funding opportunities, novel solutions are favored over replication
ACM Emerging Interest Group (EIG) on Reproducibility and Replicability is working to integrate these practices into ACM journals and conferences https://reproducibility.acm.org/
Developing open-source is thankless and payless; but it can lead to more citations
Different robot platforms cause lack of replication of open-source between labs
Need to start promoting generalization of methods over specific, narrow solutions
Lack of diversity of tasks with benchmarks, yet there is a lot of research in these other application areas
Paper length restrictions can inhibit inclusion of sufficient task description detail
Supplementary material/data sometimes allowed in paper submissions, which could be the right method to include this info
No existing review/qualification process for generating and sharing datasets
Current landscape: Existing efforts to leverage
Datasheets for datasets: https://dl.acm.org/doi/10.1145/3458723 Can this be adopted for use in robotics to document open-source and benchmarking for replication purposes?
HRI conference is starting to incorporate mechanisms similar to those identified through this project, such as standard reporting of study parameters
Advocates found their way onto the organizing committee and helped to make these changes from inside
HRI conference now has short contributions for code, datasets, and replication studies where documentation is reviewed in addition to the paper itself
This practice would be useful in multiple areas of robotics
Robotics competitions are good ways to get the community focused around a common task
Software Bill of Materials (SBOM) practices could be adopted for robotics for software + hardware + the relationships between the two
Recording the packages used, the computing environment it runs in, the hardware of the robot it runs on, etc.
Future solutions: Modular benchmarking software pipelines
Will labs that develop RL solutions, end-to-end solutions, etc., be able to use the pipeline?
We will have to develop it to allow for two distinct type of usage profiles: step-by-step (individual components) or end-to-end (complete systems)
Benchmarking protocols and metrics should apply equally to solutions that use the pipeline and those that do not
Future solutions: Distributed physical benchmarking facilities
Start with a centralized effort conducted at a single facility (e.g., UMass Lowell NERVE Center) to work out the logistics and infrastructure needed for others to participate, then slowly bring on volunteer-based organizations to provide labs/hardware/time
Incentives to participate:
Develop some form of “badging” for participating labs in this network
Agreement on shared authorship of papers for research conducted in a participating lab
When planning benchmarking exercises, consider broad vs. narrow challenge spaces
Similar community-driven effort: Dialogue System Technology Challenge (DSTC) https://dstc11.dstc.community/
If a company that provides resources sponsors a competition or workshop, the reward can be in terms of resources rather than cash (e.g., Amazon sponsored a workshop that awarded some amount of MTurk credits as a prize)
Future solutions: Online community resources
Some existing leaderboards could be used or modeled after, such as the AI leaderboards on https://leaderboard.allenai.org/
LLM model cards include characteristics of a model in a standard format; the proposed ROS Enhancement Proposal (REP) for characterization of benchmarking data in open-source assets could be developed similarly
Future solutions: Working groups and advocacy
Participants must be diverse and decentralized, not heavily influenced by a single entity
Do not set requirements for leveraging open-source and requiring benchmarking comparisons until open-source assets are well documented enough for proper replication and benchmarking methods have been established
Let’s limit the tasks/domains we work in to ensure that we can require/build criteria AND have the best practices to require
Discussion Topics
The forum will be split into two distinct categories: (1) the current landscape of robotic manipulation open-source and benchmarking practices, and (2) proposed new activities to improve robotic manipulation open-source and benchmarking practices. A note taker will record meeting minutes for all topics discussed and salient takeaways will be distilled. Topics for discussion include the following:
Limitations to robotics research and development caused by access to robotic hardware, robot simulation capabilities, lack of relevant comparable benchmarks, and lack of relevant open-source assets.
Current ability of researchers to benchmark their robotic manipulation research to others in the field using existing protocols, or utilize open-source assets in their research and contribute to their development.
Possible activities to improve the current state of open-source and benchmarking for robotic manipulation, including organized repositories of open-source assets and benchmarking results, establishing advisory committees, developing physical distributed benchmarking facilities, and dedicated conference tracks.
Schedule (Tentative)
All times given are in Korean Standard Time (GMT+9)
1:30 Workshop and participant introductions (15 min)
1:45 Presentation and guided discussion on the current landscape (75 min)
3:00 Break (30 min)
3:30 Presentation and guided discussion on proposed new activities (75 min)
4:45 Break (15 min)
5:00 Summarize workshop outputs and discuss next steps (30 min)
5:30 Workshop end
Participation
The workshop will be hybrid, with a focus on in-person participation given the time focus on forum discussions, but a virtual option for remote attendees will be available. Questions will be collected from the online chat and asked by one of the workshop organizers to the speaker. A Slack workspace is being established for pre, during, and post-workshop discussions and coordination, to serve as an open communication platform for the open-source ecosystem.
Organizers
Adam Norton, University of Massachusetts Lowell
Holly Yanco, University of Massachusetts Lowell
Berk Calli, Worcester Polytechnic Institute
Aaron Dollar, Yale University
Contact
Please contact Adam Norton with any questions or comments via e-mail: adam_norton@uml.edu
Funded by the National Science Foundation, Pathways to Enable Open-Source Ecosystems (POSE), Award TI-2229577