About me

I’m Yicheng, a student passionate about AI and robotics. I hold a Bachelor’s degree in Optoelectronic Information Engineering from Zhejiang University, where I was an active member of the Grasp Lab. This experience allowed me to develop a strong foundation in robotics, focusing on the perception of robotic grasping.

I am pursuing a Master’s degree in Machine Learning at the School of Electrical and Electronic Engineering at Nanyang Technological University to further my academic and research aspirations in robotics. In parallel, I am deepening my practical expertise as an intern at ASTAR SIMTech ARM, where I am engaged in advanced robotics research.


Reach me by:

Email: yichatma@gmail.com

Github: yichatani

Wechat: Ani


Research Interests

My research focuses on the intersection of robotic perception, control, and intelligent manipulation. Specifically, I am interested in:

  • Developing sensorimotor models to integrate perception and control for robotic grasping.
  • Exploring machine learning methods, such as deep learning and reinforcement learning, to enhance robotic decision-making and adaptability.
  • Advancing techniques in robotic grasping and manipulation, including vision-based control and multi-modal sensor fusion, for real-world applications.

Research Experience

ASTAR SIMTech ARM, Singapore

Research Intern
Sep 2024 – Present
Topic: Robotic Manipulation System using Advanced Deep Learning Technique

  • Developed a UR10e grasping system designed to operate in uncertain and dynamic clustered environments. This project involves exploring and leveraging RTDE (Real-Time Data Exchange) as a motion planning tool to enhance system reliability and responsiveness.
  • Enhanced the functionality of AnyGrasp, a state-of-the-art grasp generator, by integrating the Grasp-1Billion dataset with MetaGraspNet. This integration requires preprocessing and aligning the datasets to ensure consistency in their structure and arrangement, enabling seamless application and improved grasping performance.
  • Conducted simulations using NVIDIA Isaac Sim, replicating the grasping system in a simulated environment to evaluate its performance under various scenarios. Currently, I am exploring novel methods to transform task-agnostic grasp detectors into task-oriented systems. Potential approaches include leveraging Diffusion models or Reinforcement Learning (RL).
  • Planning to submit research findings and contributions to RA-L, focusing on innovative techniques and their applications in robotic grasping systems.

    Isaac_sim clutter_grasp


Grasp Lab, Zhejiang University

Graduation Project and Thesis
Sep 2023 - Jun 2024
Topic: Research on static and dynamic grasping of robots for warehousing and logistics applications

  • Constructed a geometric grasping module for grasp generation. The method refers to the GSNet model architecture, using graspness to measure points in the point cloud suitable for grasping and extracting local and global high-dimensional point cloud features. Then, it extracts point-wise grasping degrees and subsequent viewpoint-wise grasping degrees, sets a graspness threshold to filter target point clouds, and generates grasps based on the filtered target point clouds, achieving static grasp generation.
  • Built a dynamic tracking module based on time-based graspness. It refers to AnyGrasp, using multi-threaded high-dimensional feature vectors to represent each grasp in each frame, calculating cosine similarity to measure the similarity between high-dimensional feature vectors, and using this to measure the correspondence of temporal dimensions between grasps in frames, achieving continuous generation of dynamic tracking poses.
  • Established a robot motion control and path planning system. It based on ROS, utilizes ROS’s distributed communication architecture to connect and communicate between nodes. Motion control and path planning mainly use MoveIt API to write robot control modes and path planning, achieving multidimensional control of the robotic arm system.
  • Designed static and dynamic experiments for robot new object grasping. Static experiments include parcel grasping experiments and daily necessities grasping experiments to verify the generalization ability of the constructed robot new object grasping system. Dynamic experiments involve grasping parcels moving on conveyor belts to verify their dynamic grasping capability. Through experimental verification, this article demonstrates that the robot learning-based unknown object grasping system constructed in this article exhibits good performance in both static and dynamic object grasping scenarios.

    dynamic_grasp


Publications

hybrid_figure1 submissioninfo

Construction of Bin-picking System for Logistic Application: A Hybrid Robotic Gripper and Vision-based Grasp Planning

Zhian Su, Yicheng Ma, Haotian Guo, and Huixu Dong

IEEE Robotics and Automation Letters(RA-L), 2024 (Under Review)