Jay Prajapati
Portfolio

A robocist that converts fiber into robots and caffiene into software modules.

Quadruped
Four Legged Robot

In this project, I led the design and development of a quadruped robot, which involved complex mechanical, electrical, and software integration. The goal was to create a robust, four-legged autonomous robot capable of navigating dynamic environments. Using CAD tools like Solidworks and PTC Creo, I designed the robot’s structure, focusing on weight distribution and mechanical strength. The control system relied on ROS and a custom motion planning algorithm that allowed the robot to adjust its gait in real-time based on the environment. I implemented dynamic path planning to ensure obstacle avoidance and smooth navigation. This project won a state-level robotics championship and significantly advanced my understanding of autonomous navigation and robotic kinematics.

Multi-Robot System Software Module

In this project, I developed a software module for a multi-robot system, focusing on coordination and communication between robots. The system was designed to optimize task execution by distributing tasks among multiple robots, ensuring efficient task completion with minimal redundancy. I used ROS to manage inter-robot communication and integrated sensors for real-time data exchange, enabling the robots to adjust their paths based on the actions of other robots in the system. This project demonstrated my ability to build complex systems that require real-time collaboration between autonomous robots.

Dynamic Path Planning and Replanning Using RRT* for Autonomous Robots

This project involved the development of dynamic path-planning algorithms for an autonomous robot to navigate unpredictable environments. Using ROS and Gazebo for simulation, I implemented global path planning using Dijkstra and A* algorithms, while employing local planners for real-time obstacle avoidance. The robot successfully navigated a test environment while dynamically adjusting its path based on sensor input. This project not only refined my understanding of planning algorithms but also helped me master ROS and simulation tools like Rviz and Gazebo.

AI-Driven Real-Time Image Captioning for Enhanced Accessibility

This project aimed to assist visually impaired individuals by developing an AI-driven image captioning system. I trained a deep learning model using the Inception V3 CNN encoder for feature extraction, reaching a 92% accuracy rate in generating descriptive captions from real-time video feeds. To optimize real-time performance, I implemented Block Static Expansion and multi-headed attention vectors, which improved both accuracy and response time. Additionally, I developed Python scripts for seamless integration with mobile devices, enabling visually impaired users to capture and process video in real-time, with automatic caption generation and voice narration. This project highlighted my ability to apply AI solutions for real-world accessibility challenges.

Semantic Segmentation for Real-Time Food Item Recognition

This project revolved around real-time food item segmentation using deep learning techniques. I employed RNN-based ResNet50 and PointRend architectures to achieve an 80% accuracy rate in food item segmentation from live video feeds. Using transfer learning, I was able to improve model performance, reaching 67% test accuracy on a limited dataset of 4,935 training images and 2,135 test images, outperforming a custom model’s 43% accuracy. Fine-tuning pre-trained architectures allowed me to develop a reliable and efficient segmentation system, which was aimed at applications in food identification and inventory management in commercial kitchens.

Autonomous
Pick and Place Robot

This project involved the design and development of an autonomous robot capable of picking and placing color-coded blocks in a designated area. I engineered a real-time color detection system using a Pi Camera, which was integrated with the A* path planning algorithm to ensure accurate retrieval and placement of the blocks. The robot was equipped with IMU sensors (MPU6050) and encoders for precise navigation, and I implemented PID control for stable maneuvering. Obstacle avoidance was achieved through the use of ultrasonic sensors, allowing the robot to navigate dynamically changing environments. The project achieved an 80% success rate in goal placement, demonstrating effective robot localization and task execution.

Miniature Statue Manufacturing

In this project, I led a team to scan and 3D print miniature replicas of an 11ft statue to celebrate the 75th anniversary of my university. We used a 3D scanner to capture the intricate details of the original statue, and I led the scanning process to ensure accuracy and completeness of the data. Using Fusion 360, I performed reverse engineering to create a CAD model from the scanned data. I then used 3D printing technology to produce prototypes with different printing parameters, optimizing for quality, speed, and material usage. The final models were produced and distributed to over 1000 alumni attendees.

Human Tracking Software Module with Real-Time Object Detection

This project focused on building a human tracking system using advanced object detection algorithms in C++. I integrated the YOLOv3 model for real-time detection, which achieved an 84% confidence level and 99 ms inference time on the MS-COCO dataset. I also incorporated a KCF tracker for precise tracking across 164,000 images, enabling reliable performance in dynamic environments. This project demonstrated significant improvement in accuracy and efficiency. To further enhance development, I integrated CI/CD pipelines with Travis CI and Coveralls, which boosted code quality and testing coverage. The module was used in a multi-robot system, where the detection and tracking data fed into autonomous path-planning algorithms for mobile robots.

Perception Project with TurtleBot 3 WafflePi

In this perception-focused project, I worked with the TurtleBot 3 WafflePi to autonomously navigate through a set of waypoints on a flat surface. My key contribution was calibrating the monocular camera using a checkerboard pattern to derive intrinsic and extrinsic parameters. From there, I calculated the homography of the floor, identified the vanishing point, and determined the horizon line, enabling the robot to detect and follow waypoints seamlessly. Additionally, I integrated YOLO-based stop sign detection, allowing the robot to recognize and respond to stop signs in varying orientations. This project sharpened my skills in perception, camera calibration, and robot path planning.

IoT-Based Wastewater Spillage Detection System

I co-invented a patented IoT-based wastewater spillage detection system designed to monitor and alert authorities of any spillage in real-time. The system integrates multiple sensors to detect abnormal spillage patterns, and the data is transmitted via IoT devices to a centralized system for monitoring and action. This project combined sensor networks, data transmission protocols, and environmental monitoring technologies to create a solution that can have a significant impact on environmental safety.

RESTful API Server for Robot Management System

I developed a RESTful API server to manage a collection of OpenRAVE robots using Debian , Docker, and Python. This project aimed to provide a scalable and efficient solution for handling multiple robot configurations stored in JSON format. I utilized Python's Flask framework to implement CRUD operations, allowing for seamless robot data management through endpoints for adding, updating, and deleting robot specifications. Additionally, I containerized the API server using Docker to ensure consistent deployment and compatibility across different environments. The server could efficiently parse JSON requests, validate robot parameters, and provide real-time data response, thereby improving the robotic fleet management workflow. This project deepened my understanding of REST API architecture, Docker containerization, and robotic data handling.

Comparison of 2D Mapping Algorithms

Conducted a comprehensive evaluation of 2D mapping algorithms, including HectorSLAM, Cartographer, and Gmapping. Mapped a 20,000 sqft hostel floor to validate the study, providing actionable insights for optimal mapping algorithm selection.

FRCobot ROS2 Setup

This project involved setting up a robotic control system for a collaborative robot (FRCobot) using ROS2, leveraging its real-time communication capabilities with DDS (Data Distribution Service) for reliable control. I created a ROS2 workspace and developed custom Python nodes for joint control, sensor processing, and safety features. Using URDF models, I defined the robot's kinematics and visualized it in RViz2, ensuring accurate transformations. I also configured MoveIt2 for motion planning, enabling pick-and-place and obstacle avoidance tasks, and integrated Gazebo for simulation to validate control algorithms before hardware deployment.

Autonomous Rover for Indoor and Outdoor Navigation

I designed and developed an autonomous rover capable of navigating both indoor and outdoor environments using SLAM-based and GPS-based navigation, respectively. The rover's indoor navigation was achieved using the Gmapping SLAM algorithm, which combined LIDAR and wheel odometry data for precise localization and mapping of the environment. For outdoor navigation, I integrated a GPS module that allowed the rover to accurately navigate through predefined waypoints in open areas. The project also involved developing a ROS-based control system to seamlessly switch between SLAM and GPS modes, ensuring continuous navigation without manual intervention. I implemented an Extended Kalman Filter (EKF) to fuse sensor data, achieving high accuracy in localization and reducing drift over time. Despite constraints on funding and a tight timeline, I sourced aluminum materials and fabricated custom mounts to house the sensors and electronic components, maintaining a balance between durability and weight. The final prototype demonstrated robust autonomous navigation in a mixed environment, showcasing my proficiency in integrating SLAM, GPS, and ROS, as well as expertise in sensor fusion techniques for practical robotic applications.