Project

# Title Team Members TA Documents Sponsor
25 Long-horizon Task Completion with Robotic Arms by Human Instructions
Bingjun Guo
Qi Long
Qingran Wu
Yuxi Chen
design_document1.pdf
final_paper2.pdf
final_paper3.pdf
final_paper4.pdf
proposal1.pdf
Gaoang Wang
# Problem
The use of robotic arms for long-horizon tasks such as assembling, cooking, and packing that involve multi-step operations is growing. The interdependencies among subtasks, shifting environmental conditions, and requirement for constant feedback integration, however, make it extremely difficult to execute such tasks steadily. Current approaches frequently have trouble with skill chaining, task decomposition, and preserving robustness while being executed. For robotic arms to be able to manipulate objects on their own based on real-time feedback and finish long-term tasks, a comprehensive framework that incorporates perception and planning is therefore required.

# Solution Overview
Our solution for enabling a robot to conduct a series of tasks is to combine Perception, Planning and Acting Intelligence as a whole. The robotic arm is our primary entity. Firstly, for perception, it has a e.g. RGB camera on the top to capture the scene, including recognizing the objects using computer vision. Then, for planning, with uploaded captured images and user's instructions, the robot will do the analysis and task planning. Finally, for acting, the plan is reflected as a guide for the robotic arm to move. During the process of acting, the sensors including the rgb-camera on the robotic arm will provide continuous feedback, which will revise the action of the robotic arm in a control system loop. The whole process will loop over these three steps until the series of tasks are completed.

# Solution Components
## Output Subsystem
- A robotic arm (UR3)
- A specially designed grasper
- Exclusive tools for the long-horizon task set (e.g. a screwdriver for assembly tasks)

## Feedback Subsystem
- Visual sensors (e.g. a RGB or RGB-D camera, depending on availability)
- Tactile sensors
- Corresponding circuits preprocessing the perceptional signals

## Planning Subsystem
- A language model to extract semantic information from instructions
- A vision model that preprocess input images
- An agent model that process inputs, plan movements, and carry out tasks according to feedback

# Criteria of Success
- Overall: The robotic arm can successfully complete a certain set of long-horizon tasks (t.b.d according to feasibility) based on human instructions in a zero-shot manner.
- Perception: The system can accurately recognize objects in the scene.
- Planning: The system can generate reasonable multi-step operations.
- Acting: The robotic arm can follow the generated plan and adjust its movement based on real-time feedback to improve accuracy and robustness.
- Safety: The robotic arm can avoid collisions.

Digital Controlled LED Rotating Display System

Guanshujie Fu, Yichi Jin, Keyi Shen, Chentai (Seven) Yuan

Featured Project

# TEAM MEMBERS

Chentai Yuan (chentai2)

Guanshujie Fu (gf9)

Keyi Shen (keyis2)

Yichi Jin (yichij2)

# TITLE OF THE PROJECT

Digital Controlled LED Rotating Display System

# PROBLEM

By visual persistence phenomenon, we can display any images and strings with a rotating LED array. Many devices based on this idea have been developed. However, there are some common issues to be solved. First, the images or strings to be displayed are pre-defined and cannot be changed in a real-time way. Second, the wired connection between some components may limit the rotation behavior, and harm the quality of display. Some economical wireless communication technologies and new ways to connect components can be applied to achieve a better display and real-time image update.

# SOLUTION OVERVIEW

We aim at developing a digital controlled LED rotating display system. A servo motor is controlled to drive the stick with one row of LED to do circular rotation. The connection between LEDs, control circuit, motor and other components should be simple but firm enough to suppose good display and high-speed rotation. Moreover, there is another part to handle users’ input and communicate with the display part via Bluetooth to update images in a real-time and wireless way.

# SOLUTION COMPONENTS

## Subsystem1: Display Subsystem

- LED Array that can display specific patterns.

- Controller and other components that can timely turn the status of LEDs to form aimed patterns.

## Subsystem2: Drive Subsystem

- Servo motor that drive of the LED array to do circular rotation.

- Controller that communicates with the motor to achieve precise rotation and position control.

- An outer shell that has mechanisms to fix the motor and LED array.

## Subsystem3: Logic and Interface Subsystem

- Input peripherals like keyboard to receive users’ input.

- A FPGA board for high-level logics to handle input, give output and communicate with other subsystems.

- Wireless communication protocol like Bluetooth used in communication.

- VGA display hardware offering Graphical User Interface.

# CRITERION OF SUCCESS

- Users can successfully recognize the real-time patterns to be displayed.

- It achieves the precise rotation and position control of motor.

- The motor can drive the LED array and any necessary components to rotate stably and safely.

- The LED array is under real-time control and responds rapidly.

- The communication between components has low latency and enough bandwidth.

# DISTRIBUTION OF WORK

- Chentai Yuan(ME): Mechanisms and servo motor control.

- Guanshujie Fu(CompE): Logic and Interface design and keyboard & VGA display implementation.

- Keyi Shen(EE): Wireless communication and servo motor control.

- Yichi Jin(EE): Circuit design, keyboard & VGA display implementation.