no code implementations • 28 Jul 2022 • Zijue Chen, Keenan Granland, Rhys Newbury, Chao Chen
We further validated HOB-CNN against two different types of 2D trees, and HOB-CNN shows generalization across different trees and robustness under different occluded conditions.
no code implementations • 16 Oct 2020 • Keenan Granland, Rhys Newbury, David Ting, Chao Chen
It is shown that in our application, the new Automating-the-Loop method greatly reduces the labeling effort while producing comparable performance to both human-in-the-loop and complete manual labeling methods.
no code implementations • 14 Oct 2020 • Zijue Chen, David Ting, Rhys Newbury, Chao Chen
DeepLabv3 outperforms the other models at Binary accuracy, Mean IoU and Boundary F1 score, but is surpassed by Pix2Pix (without discriminator) and U-Net in Occluded branch recall.
1 code implementation • 2 Jun 2020 • Patrick Rosenberger, Akansel Cosgun, Rhys Newbury, Jun Kwan, Valerio Ortenzi, Peter Corke, Manfred Grafinger
In experiments with 13 objects, the robot was able to successfully take the object from the human in 81. 9% of the trials.
no code implementations • 2 May 2020 • Shray Bansal, Rhys Newbury, Wesley Chan, Akansel Cosgun, Aimee Allen, Dana Kulić, Tom Drummond, Charles Isbell
We compare two robot modes in a shared table pick-and-place task: (1) Task-oriented: the robot only takes actions to further its own task objective and (2) Supportive: the robot sometimes prefers supportive actions to task-oriented ones when they reduce future goal-conflicts.