Mathematical Modeling and Kinematic Anal

Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

International Journal of Electrical and Computer Engineering (IJECE)

Vol. 13, No. 2, April 2023, pp. 1522~1532


ISSN: 2088-8708, DOI: 10.11591/ijece.v13i2.pp1522-1532  1522

Mathematical modeling and kinematic analysis of 5 degrees of


freedom serial link manipulator for online real-time pick
and place applications

Abhilasha Singh1, Kalaichelvi Venkatesan1, Yuvalakshmi Nagarasan1, Karthikeyan Ramanujam2,


Kumar Karuppusamy3
1
Department of Electrical and Electronics Engineering, Birla Institute of Technology and Science Pilani, Dubai Campus UAE, Dubai,
United Arab Emirates
2
Department of Mechanical Engineering, Birla Institute of Technology and Science Pilani, Dubai Campus UAE, Dubai,
United Arab Emirates
3
Department of Mathematics, Birla Institute of Technology and Science Pilani, Dubai Campus UAE, Dubai, United Arab Emirates

Article Info ABSTRACT


Article history: Modeling and kinematic analysis are crucial jobs in robotics that entail
identifying the position of the robot’s joints in order to accomplish particular
Received Feb 4, 2022 tasks. This article uses an algebraic approach to model the kinematics of a
Revised Sep 13, 2022 serial link, 5 degrees of freedom (DOF) manipulator. The analytical method
Accepted Oct 10, 2022 is compared to an optimization strategy known as sequential least squares
programming (SLSQP). Using an Intel RealSense 3D camera, the colored
object is picked up and placed using vision-based technology, and the pixel
Keywords: location of the object is translated into robot coordinates. The LOBOT
LX15D serial bus servo controller was used to transmit these coordinates to
Algebraic method the robotic arm. Python3 programming language was used throughout the
Kinematics entire analysis. The findings demonstrated that both analytical and optimized
Manipulator inverse kinematic solutions correctly identified colored objects and
Optimization positioned them in their appropriate goal points.
Pick and place operation
This is an open access article under the CC BY-SA license.

Corresponding Author:
Abhilasha Singh
Department of Electrical and Electronics Engineering, Birla Institute of Technology and Science Pilani
P.O. Box 345 055, Dubai Campus, Dubai, United Arab Emirates
Email: p20180906@dubai.bits-pilani.ac.in

1. INTRODUCTION
Robots play a significant role in industrial automation, where they employ machine vision to
recognize and categorize the objects put so they can pick and place them quickly [1]. Vision sensors are
employed to collect visual feedback data on the scattered objects in the surroundings for this task. The
efficient picking of things by the robot requires an understanding of dynamics and kinematics [2]. Robot
dynamics deals with joint force and torque, whereas kinematic analysis determines the position of the robot
joints without taking force into account [3]. The kinematics and inverse kinematics of a 5 degrees of freedom
(DOF) robotic arm are covered in this work. One of the most significant tasks involves selecting the ideal
robot from the available conventional configurations. This requires designing the working positions required
for forward and inverse kinematics modeling [4]. The joint angles are used to determine the position of the
end effector in forward kinematics, whereas the end effector position is used to determine the joint angles in
inverse kinematics. The two methods of inverse kinematics algorithms are closed-form solutions and
numerical approaches [5]. The closed-form solutions are found using algebraic and geometric techniques,
and the Jacobian transpose method [6], pseudoinverse method [7], cyclic coordinate descent methods [8],

Journal homepage: http://ijece.iaescore.com


Int J Elec & Comp Eng ISSN: 2088-8708  1523

optimization algorithms [9], artificial intelligence (AI) methods [10], and singular value decomposition [11]
are some of the widely utilized numerical techniques. Since closed-form solutions require precise initial
values and converge more quickly than numerical ones, an algebraic approach was taken in this study to
calculate joint angles. Gaeid et al. [3] used Denavit-Hartenberg (DH) parameters and matrix transformation
method to perform kinematic modeling of 6 DOF manipulator whereas Sutyasadi and Wicaksono [12] used
hybrid controller of iterative learning and H∞ controller to control the robot joint for trajectory tracking
applications. Quang et al. [13] used radial basis function to derive the inverse dynamics of delta manipulator
and Mashhadany [14] used adaptive neuro fuzzy inference system (ANFIS) controller and fractional order
proportional, integral, derivative (FOPID) controller to obtain optimal trajectory for PUMA 560 manipulator.
Dewi et al. [15] implemented fruit sorting robot for packaging industry using hue saturation value (HSV) and
image processing techniques. In the above discussed works kinematic analysis has been implemented in
servo motors with conventional controllers. Also, most of the works were tested on simulation. Hence in this
paper the main contribution of this work is to implement vision-based control for pick and place operation for
different colors and object size. To obtain joint commands, inverse kinematics solutions for Lewansoul xArm
manipulator made up of serial bus servo motors is derived. The main difference between serial bus servo and
other servo motors is that serial bus servo motors are connected serially. In such motors, when one motor
does not receive power, the other motors does not work. Secondly, attempt has been made to perform real
time experimentation for industrial pick and place operation from live IntelRealSense camera feed. This work
can be commonly used in color-based objects sorting, garbage sorting and picking up the fruits based on their
color, size, and shape.

2. PROPOSED METHOD
The method suggested in this study involves taking screenshots of each frame taken by the Intel
RealSense camera and using HSV morphology and HSV transformation to extract the features of the colored
objects. To acquire the specific channel needed to draw the contours around the objects, the colors are further
masked. Then, camera-to-robot coordinates (𝑥, 𝑦, 𝑧) are determined via homogeneous transformation.
Inverse kinematics solutions utilizing the algebraic approach and sequential least squares programming
(SLSQP) are compared using these coordinates. By positioning the Intel RealSense camera in an eye-to-hand
position, the joint angles that were obtained are used to implement pick and place operations. Finally, these
robot joint commands are transferred to the Lewansoul xArm robotic arm to conduct pick and place of the
relevant colored objects [16]. The xArm-SDK was used throughout the entire procedure and during real-time
experimentation in the Python3 environment. Figure 1 depicts the overall process flow for a vision-based
pick and place operation.

Figure 1. Process flow of vision-based pick and place using Lewansoul xArm manipulator

Mathematical modeling and kinematic analysis of 5 degrees of freedom serial link … (Abhilasha Singh)
1524  ISSN: 2088-8708

3. METHOD
3.1. Kinematic modelling of Lewansoul xArm
In this section, detailed modelling of kinematics and inverse kinematics is discussed for Lewansoul
xArm which is a 5 DOF robotic arm which is used for gripping, sorting, and many other industrial
applications. It is a miniature version of an industrial manipulator that runs with intelligent bus servos
LX-15D servo motor which has built-in position feedback and voltage feedback. The xArm is controlled by
using wireless handle control and through smartphone app [17]. This arm can be operated in robot operating
system (ROS) and Python 3, but the entire work reported in this paper is carried out in python 3.6. The
python wrapper called xArm is installed which has built-in functions to operate the robot. In this paper, DH
method is used which is used for finding the transformation matrix between each joint. Initially, robot links
length and angles between joints are measured and tabulated for xArm. Next, each row in the DH table
corresponds to transformation between each joint. Finally, the overall transformation matrix is obtained
between the end effector to the base of the robot [18]. The basic parameters obtained are 𝜃𝑖 =angle from
𝑥𝑖−1 to 𝑥𝑖 along 𝑧𝑖−1 ; 𝑑𝑖 =distance from the intersection of 𝑧𝑖−1 with 𝑥𝑖 to the origin of (𝑖 − 1) system of axes;
𝑎𝑖 =distance between 𝑧𝑖−1 and 𝑧𝑖 ; 𝛼𝑖 =angle from 𝑧𝑖−1 to 𝑧𝑖 along 𝑥𝑖 .

3.1.1. Forward kinematics of Lewansoul xArm


The forward kinematics involves finding the position of the robot end-effector with joint angles as
inputs [19]. The kinematic diagram of Lewansoul 5 DOF robotic manipulator with individual joint axes is
shown in Figure 2 where Figure 2(a) represents the frame mapping of individual joints and Figure 2(b)
represents the RoboAnalyser model of manipulator which is efficient tool for analyzing the frames. Using the
link length and link biases measurements from the Figure 2, the DH parameters for 5 DOF manipulator is
shown in Table 1. Since the gripper is the 6 th joint, it is not included in the DH parameter calculation.

Table 1. DH parameters of Lewansoul xArm robotic manipulator


𝑎 𝛼 𝑑 𝜃
-0.015 -90 0.035 𝜃1
0.097 180 0 𝜃2 − 90
0.097 180 0 𝜃3
0 90 0 𝜃4 + 90
0 0 0.085 𝜃5

(a) (b)

Figure 2. Kinematic scheme of Lewansoul xArm manipulator (a) frames of each joints and (b) visualization
of frames in RoboAnalyser

Int J Elec & Comp Eng, Vol. 13, No. 2, April 2023: 1522-1532
Int J Elec & Comp Eng ISSN: 2088-8708  1525

The individual transformation matrices of joints are formulated using (1) [14],

𝐶𝜃 −𝑆𝜃𝐶𝛼 𝑆𝜃𝑆𝛼 𝑎𝐶𝜃


𝑖−1 𝑆𝜃 𝐶𝜃𝐶𝛼 −𝐶𝜃𝑆𝛼 𝑎𝑆𝜃
𝑖𝑇 = [ ] (1)
0 𝑆𝛼 𝐶𝛼 𝑑
0 0 0 1

where 𝑖=no. of frames varying from 0 to 4 for xArm; cos(𝑥) = 𝐶𝑥; sin(𝑥) = 𝑆𝑥 𝑥=𝜃 𝑜𝑟 𝛼. The overall
transformation matrix 50𝑇 from the end effector to the base of the robot is obtained.
5
0𝑇 = 54𝑇 × 43𝑇 × 32𝑇 × 21𝑇 × 10𝑇 (2)

−1 0 1 −0.015
5 0 1 0 0
0𝑇 = [ ] (3)
0 0 −1 0.314
0 0 0 1

From (3), the first three rows and columns denote the initial rotation matrices R, and the last column denotes
the initial position of the end effector. Therefore, the home position of the xArm manipulator is (-15,0,314)
mm respectively.

3.1.2. Inverse kinematics of Lewansoul xArm


The inverse kinematics involves finding the individual joint angles of the robot arm using the
position information in cartesian space. In this paper, inverse kinematics of xArm is derived using algebraic
method and it is compared with optimization approach called SLSQP [20]. The analytical approach gives
robust solutions since it involves trigonometric relationships and algebraic equations that determine the
behavior of the degrees of freedom in the manipulator [21]–[23]. Here for simplicity, wrist angles are made
zero and θ2-34 wrist angle relative to ground (WARTG) is taken as a constant value [24]. To start with, the
overall transformation matrix from the end effector to base 05𝑇 is obtained from equations from (3) and is
taken as (4).

𝑛𝑥 𝑜𝑥 𝑎𝑥 𝑝𝑥
5 𝑛𝑦 𝑜𝑦 𝑎𝑦 𝑝𝑦
0𝑇 = 𝑇𝐸 = [
𝑛𝑧 𝑜𝑧 𝑎𝑧 𝑝𝑧
] (4)
0 0 0 1

Take (5).

𝐴1 = 01𝑇 ; 𝐴2 = 12𝑇; 𝐴3 = 23𝑇 ; 𝐴4 = 34𝑇 ; 𝐴5 = 45𝑇 (5)

Consider (6).
𝑛𝑥 𝑜𝑥 𝑎𝑥 𝑝𝑥
𝑛𝑦 𝑜𝑦 𝑎𝑦 𝑝𝑦
𝐴2 𝐴3 𝐴4 𝐴 5 = 𝐴1−1 [ ] (6)
𝑛𝑧 𝑜𝑧 𝑎𝑧 𝑝𝑧
0 0 0 1
Solving LHS of the (6) we get (7).

𝐶1 𝑛𝑥 + 𝑆1 𝑛𝑦 𝐶1 𝑜𝑥 + 𝑆1 𝑜𝑦 𝐶1 𝑎𝑥 + 𝑆1 𝑎𝑦 𝐶1 𝑝𝑥 + 𝑆1 𝑝𝑦
−𝑛𝑧 −𝑜𝑦 −𝑎𝑦 −𝑝𝑧 + 𝑑1
(7)
−𝑆1 𝑛𝑥 + 𝐶1 𝑛𝑦 −𝑆1 𝑜𝑥 + 𝐶1 𝑜𝑦 −𝑆1 𝑎𝑥 + 𝐶1 𝑎𝑦 −𝑆1 𝑝𝑥 + 𝐶1 𝑝𝑦
[ 0 0 0 1 ]

Solving RHS of the (6) we get (8):

𝐶2−34 𝐶5 −𝐶2−34 𝑆5 −𝑆2−34 −𝑆2−34 𝑑5 + 𝑎3 𝐶23 + 𝑎2 𝐶2


𝑆 𝐶 −𝑆2−34 𝑆5 𝐶2−34 −𝐶2−34 𝑑5 − 𝑎3 𝑆23 − 𝑎2 𝑆2
[ 2−34 5 ] (8)
𝑆5 𝐶5 1 0
0 0 0 1

Mathematical modeling and kinematic analysis of 5 degrees of freedom serial link … (Abhilasha Singh)
1526  ISSN: 2088-8708

where 𝜃2−34 is taken as 𝜃2 − 𝜃3 + 𝜃4 . By comparing (7) and (8) we get the values of joint angles which is,

𝑝𝑦
𝜃1 = 𝑡𝑎𝑛 −1 ( ) (9)
𝑝𝑥

where (𝑝𝑥 , 𝑝𝑦 ) is the end-effector position of the robot. Assuming WART configuration 𝜃234 = 0 to obtain 𝜃3 .

𝑆3
𝜃3 = 𝑡𝑎𝑛−1 ( ) (10)
𝐶3

Finally,

𝑆2
𝜃2 = 𝑡𝑎𝑛 −1 ( ) (11)
𝐶2

𝑆5
𝜃5 = 𝑡𝑎𝑛 −1 ( ) (12)
𝐶5

𝜃4 = 𝜃2−34 − 𝜃2 + 𝜃3 (13)

The results obtained from algebraic method is compared with SLSQP which is an iterative method for
constrained nonlinear optimization problem. This algorithm transforms the least squares model into the
quadratic optimization model and the optimal solution can then be obtained by sequential quadratic
programming. The nonlinear programming is of the form (14).

min 𝑓(𝑥) (14)


𝑥

Subject to 𝑏(𝑥) ≥ 0; 𝑐(𝑥) = 0. Then at every iteration 𝑥𝑘 quadratic programming sub problem is given by (15).

min 𝑓(𝑥) + ∇𝑓(𝑥𝑘 )𝑇 𝑑 + 1⁄2 𝑑 𝑇 ∇2 𝑓(𝑥𝑘 )𝑑 (15)


𝑑

Subject to 𝑏(𝑥𝑘 ) + ∇𝑏(𝑥𝑘 )𝑇 𝑑 ≥ 0 ; 𝑐(𝑥𝑘 ) + ∇𝑐(𝑥𝑘 )𝑇 𝑑 = 0.

3.2. Vision-based pick and place operation


In this section, inverse kinematics solutions obtained are used for pick and place operation. In
general, pick and place using vision systems is common in almost all industries, but it still lags accuracy and
efficiency. Hence there is the need to improve the process of picking up objects, possibly using intelligent
techniques [25]. In recent times, there are many works, and improvements made in computer vision
applications. In this paperwork, detailed steps of kinematic design and its application to pick and place using
a 3D camera are discussed and also tested in real-time. This work gives the reader a clear idea about the
mathematics behind the robotic arm and also how to implement computer vision tasks in real-time. The
process flow for color extraction is shown in Figure 3.
Initially, the colored objects are placed in the robot workspace and the frames are captured by Intel
RealSense camera attached to the wooden frame. The camera is placed in eye-to-hand configuration at an
angle of 90° facing the objects [26]. Once the frames are captured HSV conversion takes place to extract the
color of the objects. In this paper, four colors are used-red, green, blue, and yellow. The upper and lower
range of HSV hue range of these colors were specified to mask each color and perform image dilation. Next
contours are created around each color and its bounding box coordinates are extracted along with box
coordinates to find the orientation of the shape. Another main contribution of this work is that without using
a grasping dataset or training deep learning model, a simple and robust solution to find the grasping angle is
implemented in this work. Using the adjacent box coordinates (𝑥1 , 𝑦1 ) and (𝑥2 , 𝑦2 ), the angle of rotation is
found by measuring the length of both opposite sides and adjacent sides [27]. The centroid (𝐶𝑥 , 𝐶𝑦 ) of the
bounding boxes is also obtained which is useful for the robot to understand the position of the object.
The formula for calculating the orientation is given by (16).

𝑙𝑒𝑛𝑔𝑡ℎ 𝑜𝑓 𝑜𝑝𝑝 𝑠𝑖𝑑𝑒 𝑦 − 𝑦1


𝜃 = 𝑡𝑎𝑛−1 ( ) = 𝑡𝑎𝑛−1 ( 2 ⁄𝑥2 − 𝑥1 ) 𝑑𝑒𝑔𝑟𝑒𝑒𝑠 (16)
𝑙𝑒𝑛𝑔𝑡ℎ 𝑜𝑓 𝑎𝑑𝑗𝑎𝑐𝑒𝑛𝑡 𝑠𝑖𝑑𝑒

Int J Elec & Comp Eng, Vol. 13, No. 2, April 2023: 1522-1532
Int J Elec & Comp Eng ISSN: 2088-8708  1527

Using all the above pixel information, 𝑇𝑏𝑐 the transformation between camera frame {𝑐} and robot base {𝑏} is
obtained using the homogeneous transformation technique which converts pixel to world coordinates so that
robot can understand the points [28] as shown in Figure 4. Figure 4(a) illustrates the transformation between
camera and robot base whereas Figure 4(b) explains the calculation of bounding box coordinates in xy plane.

Figure 3. Process flow of color extraction using image processing for pick and place operation

(a) (b)

Figure 4. Homogeneous coordinate transformation (a) transformation of centroid pixel point P from camera
frame to robot base frame and (b) schematic diagram of bounding box for finding the orientation

The general equation for transforming a point from one frame to another frame is,

𝑃𝑐 𝑏 1 𝑅𝑏 𝑑𝑐𝑏 𝑃
[ ] = [𝑅𝑐 + 𝑑0 ] = [ 𝑇𝑐 ][ ] (17)
1 1 0 1 1

In general,

𝑃0 = 𝑇𝑐𝑏 𝑃 𝑎𝑛𝑑 𝑇𝑐𝑏 = 𝑃0 𝑃 −1 (18)

Mathematical modeling and kinematic analysis of 5 degrees of freedom serial link … (Abhilasha Singh)
1528  ISSN: 2088-8708

where 𝑃0 is 4xm robot coordinates with respect to the robot base; 𝑃 is 4xn camera coordinates with respect to
world coordinates; 𝑇𝑐𝑏 is 4x4 transformation matrix between camera and robot base; 𝑅𝑐𝑏 is a 4×4 rotation
matrix between camera and robot base.
In this work for the formation of 𝑃𝑐 matrix, the pixel values of the robot workspace are obtained so
that the entire workspace can be mapped globally. Here the centroid is the input which is converted to robot
coordinates (𝑥, 𝑦, 𝑧). Finally, these cartesian coordinates are converted into joint angles 𝜃1 to 𝜃5 for picking
the object. If the pieces are oriented, the angle is fed directly to the wrist so that it orients itself similar to the
object placed.

4. RESULTS AND DISCUSSION


In this section results obtained from the algebraic and optimization approach are used for
performing pick and place in real-time. The camera captures the colored object kept in the robot workspace
using OpenCV. Once the color is detected it draws corresponding contours over the objects and its centroid is
found out. The 2D workspace bounds of xArm are 50 to 300 mm in x direction (vertical) and -270 to +270
mm in y direction (horizontal) with -120 to +120 degree as joint limits for all the 5 joints. The 2D workspace
plot for xArm is shown in Figure 5. Figure 5(a) shows the reachable area of robot in MATLAB whereas
Figure 5(b). shows the real time workspace setup. The 3D workspace of the robot analyzed in RoboAnalyser
is shown in Figure 5(c). From the figure, it can be inferred that this robot has spherical workspace until it
reaches 435 mm. The real-time experimentation of picking colored objects is shown in Figure 6 where
Figure 6(a) shows the experimental setup with Intel RealSense camera whereas Figures 6(b)-(c) show the live
detection of colored objects using the input from the camera feed. The bounding boxes is drawn around the
identified objects and the centroid and angle of rotation is calculated using the (16). These pixels are
transformed into real world coordinates and finally inverse kinematics is used to find the joint angles for
picking the specific objects in its workspace.

(50,270) mm (345,270) mm
Robot Workspace
Robot

(50,-270) mm (345,-270) mm
(a) (b)

(c)

Figure 5. Illustration of xArm workspace (a) 2D workspace of manipulator (b) real-time experimental
workspace (c) 3D workspace in RoboAnalyser

Int J Elec & Comp Eng, Vol. 13, No. 2, April 2023: 1522-1532
Int J Elec & Comp Eng ISSN: 2088-8708  1529

(a) (b) (c)

Figure 6. Experimental analysis (a) real-time vision-based pick, place (b) live detection of different size, and
(c) colored objects using Intel RealSense camera

The HSV hue ranges of different colors, the orientation of the objects, and centroid are listed in
Table 2. The hue ranges can be adjusted according to the resolution of the camera. These centroids obtained
are in pixels which are further converted to robot coordinates (𝑥, 𝑦, 𝑧) using a homogeneous transformation
matrix as shown below. The robot coordinates are further converted into joint angles and sent to the robot for
picking the object. The coordinates for placing the objects in the workspace were kept constant and it is
chosen as (200, −150) mm. The joint angles for pick and place obtained from the inverse kinematics
algorithm for the yellow-colored object found are listed in Table 3.

0 −0.6145 0 345
0.843 0 0 −270
T= [ ] (19)
0 0 0 0
0 0 0 1

Table 2. Experimental parameters of Lewansoul xArm robotic manipulator


Color HSV range (Upper) HSV range (Lower) Centroid Angle(degrees)
Yellow (30,255,255) (54,255,255) (212,232) 14.47
Red (10,255,255) (0,100,20) (275,180) 6.00
Green (86,255,255) (36,25,25) (354,179) 10.12
Blue (150,255,255) (100,150,0) (453,160) 8.74

Table 3. Comparative analysis of optimal inverse kinematics solutions using analytical method and SLSQP
Color Centroid (Pixels) Transformed Coordinates (mm) SLSQP Algebraic
Yellow (212,232) (202.41,-91.125,0) 𝜃1=24.23 𝜃1=23.56
𝜃2=65.82 𝜃2=64.32
𝜃3=-19.80 𝜃3=-19
𝜃4 =79.53 𝜃4 =80.5
𝜃5=0 𝜃5=0
Red (359,89) (289.99,-33.75,0) 𝜃1=-2.53 𝜃1=-2.03
𝜃2=78.096 𝜃2=77
𝜃3=19.22 𝜃3=19.53
𝜃4 =67.09 𝜃4 =67.5
𝜃5=0 𝜃5=0
Green (188,102) (282.31,-111.37,0) 𝜃1=-22.31 𝜃1=-22.43
𝜃2=63.35 𝜃2=63
𝜃3=-9.23 𝜃3=-9.22
𝜃4 =65.29 𝜃4 =65.34
𝜃5=0 𝜃5=0
Blue (504,93) (287.84,-156.37,0) 𝜃1=-28.52 𝜃1=-28.65
𝜃2=76.56 𝜃2=76.4
𝜃3=-1.985 𝜃3=-1.98
𝜃4 =43.42 𝜃4 =43.44
𝜃5=0 𝜃5=0

Mathematical modeling and kinematic analysis of 5 degrees of freedom serial link … (Abhilasha Singh)
1530  ISSN: 2088-8708

From the Table 3, it can be seen that the centroid of all the colored objects in pixels was converted
into robot coordinates using the (31). Further, these coordinates were converted into joint angles and these
angles were sent to the robot and tested the pick and place in real-time. The gripper angle was kept constant
of about 30^° as it was found to be effective for holding the objects. The robot was able to pick the objects
and place them in the proper goal points efficiently. The same was tested with other colored blocks. The
experimental parameters are illustrated in Figure 7. The joint angles obtained from SLSQP for blue object
with time period of 1,000 ms is shown in Figure 7(a) and triangular angular velocity profile is illustrated in
Figure 7(b). The real-time experimentation of vision-based pick and place was tested and is illustrated in
Figure 8. The real-time experimentation of vision-based pick and place was tested with initial grasping
position as shown in Figure 8(a). The picking of the red object is illustrated in Figure 8(b) whereas the red
object reaching final point is shown in Figure 8(c).

(a) (b)

Figure 7. Graphical illustration of xArm parameters (a) joint angles for blue object and
(b) angular velocity of xArm

(a) (b) (c)

Figure 8. Real-time pick and place of colored objects (a) pick position (b) middle position,
and (c) place position

5. CONCLUSION
Using the 3D camera in real-time, the online experimentation of vision-based pick and place
operation was accomplished. This method can be used in industries to separate or identify the items placed on
a conveyor belt. In agriculture and warehouse applications, they can be used to color-sort vegetables or other
items. Under this article, colorful objects could be reliably identified even in a variety of illumination and
noise situations. The algebraic technique was used to mathematically derive the inverse kinematics. It was
discovered that the algebraic technique and the SLSQP optimization method both produced effective
solutions with only a slight difference in error. In almost every location in the workspace, the coordinate

Int J Elec & Comp Eng, Vol. 13, No. 2, April 2023: 1522-1532
Int J Elec & Comp Eng ISSN: 2088-8708  1531

conversions were exact and correctly positioned the objects. During the real-time experimentation, these
minor errors had very little effect. A more realistic experimental setup for identifying industrial objects that
match the industrial environment will be included in future developments. Deep learning approaches can be
used in conjunction with additional inverse kinematics methods to detect the objects.

ACKNOWLEDGEMENTS
The authors are immensely grateful to the authorities of Birla Institute of Science and Technology
Pilani, Dubai campus for their support throughout this research work.

REFERENCES
[1] M. Javaid, A. Haleem, R. P. Singh, S. Rab, and R. Suman, “Exploring impact and features of machine vision for progressive
industry 4.0 culture,” Sensors International, vol. 3, 2022, doi: 10.1016/j.sintl.2021.100132.
[2] K. Renuka, N. Bhuvanesh, and J. R. Catherine, “Kinematic and dynamic modelling and PID control of three degree-of-freedom
robotic arm,” in Springer Proceedings in Materials, Springer Singapore, 2021, pp. 867–882.
[3] K. S. Gaeid, A. F. Nashee, I. A. Ahmed, and M. H. Dekheel, “Robot control and kinematic analysis with 6DoF manipulator using
direct kinematic method,” Bulletin of Electrical Engineering and Informatics (BEEI), vol. 10, no. 1, pp. 70–78, Feb. 2021, doi:
10.11591/eei.v10i1.2482.
[4] J. Cai, J. Deng, W. Zhang, and W. Zhao, “Modeling method of autonomous robot manipulator based on D-H algorithm,” Mobile
Information Systems, pp. 1–10, Jul. 2021, doi: 10.1155/2021/4448648.
[5] R. Singh, V. Kukshal, and V. S. Yadav, “A review on forward and inverse kinematics of classical serial manipulators,” in Lecture
Notes in Mechanical Engineering, Springer Singapore, 2021, pp. 417–428.
[6] S.-O. Park, M. C. Lee, and J. Kim, “Trajectory planning with collision avoidance for redundant robots using Jacobian and
artificial potential field-based real-time inverse kinematics,” International Journal of Control, Automation and Systems, vol. 18,
no. 8, pp. 2095–2107, Aug. 2020, doi: 10.1007/s12555-019-0076-7.
[7] J. Podobnik, M. Munih, and M. Mihelj, “Magneto-inertial data sensory fusion based on Jacobian weighted-left-pseudoinverse,” in
Advances in Robot Kinematics 2020, Springer International Publishing, 2021, pp. 174–181.
[8] P. Yotchon and Y. Jewajinda, “Combining a differential evolution algorithm with cyclic coordinate descent for inverse kinematics
of manipulator robot,” in 2021 3rd International Conference on Electronics Representation and Algorithm (ICERA), Jul. 2021,
pp. 35–40, doi: 10.1109/ICERA53111.2021.9538798.
[9] H. M. Tuan, F. Sanfilippo, and N. V. Hao, “Modelling and control of a 2-DOF robot arm with elastic joints for safe human-robot
interaction,” Frontiers in Robotics and AI, vol. 8, Aug. 2021, doi: 10.3389/frobt.2021.679304.
[10] X. Wang, X. Liu, L. Chen, and H. Hu, “Deep-learning damped least squares method for inverse kinematics of redundant robots,”
Measurement, vol. 171, Feb. 2021, doi: 10.1016/j.measurement.2020.108821.
[11] N. V. Toan and P. B. Khoi, “A SVD-least-square algorithm for manipulator kinematic calibration based on the product of
exponentials formula,” Journal of Mechanical Science and Technology, vol. 32, no. 11, pp. 5401–5409, Nov. 2018, doi:
10.1007/s12206-018-1038-3.
[12] P. Sutyasadi and M. B. Wicaksono, “Robotic arm joint position control using iterative learning and mixed sensitivity H∞ robust
controller,” Bulletin of Electrical Engineering and Informatics (BEEI), vol. 10, no. 4, pp. 1864–1873, Aug. 2021, doi:
10.11591/eei.v10i4.3059.
[13] N. H. Quang, N. Van Quyen, and N. N. Hien, “Radial basis function neural network control for parallel spatial robot,”
Telecommunication Computing Electronics and Control (TELKOMNIKA), vol. 18, no. 6, Dec. 2020, doi:
10.12928/telkomnika.v18i6.14913.
[14] Y. I. Al Mashhadany, “Virtual reality trajectory of modified PUMA 560 by hybrid intelligent controller,” Bulletin of Electrical
Engineering and Informatics, vol. 9, no. 6, pp. 2261–2269, Dec. 2020, doi: 10.11591/eei.v9i6.2579.
[15] T. Dewi, P. Risma, and Y. Oktarina, “Fruit sorting robot based on color and size for an agricultural product packaging system,”
Bulletin of Electrical Engineering and Informatics, vol. 9, no. 4, pp. 1438–1445, Aug. 2020, doi: 10.11591/eei.v9i4.2353.
[16] Diestra AI, “xArm Lewansoul ROS documentation,” 2021. Accessed: Nov 20, 2021. [Online]. Available: https://xarm-lewansoul-
ros.readthedocs.io/_/downloads/en/latest/pdf/.
[17] D. Kusmenko and K. Schmidt, “Development of an analytical inverse kinematics for a 5 DOF manipulator,” in 2020 21st
International Conference on Research and Education in Mechatronics (REM), Dec. 2020, pp. 1–5, doi:
10.1109/REM49740.2020.9313880.
[18] W. S. Pambudi, E. Alfianto, A. Rachman, and D. P. Hapsari, “Simulation design of trajectory planning robot manipulator,”
Bulletin of Electrical Engineering and Informatics, vol. 8, no. 1, pp. 196–205, Mar. 2019, doi: 10.11591/eei.v8i1.1179.
[19] F. Maric, M. Giamou, A. W. Hall, S. Khoubyarian, I. Petrovic, and J. Kelly, “Riemannian optimization for distance-geometric
inverse kinematics,” IEEE Transactions on Robotics, vol. 38, no. 3, pp. 1703–1722, Jun. 2022, doi: 10.1109/TRO.2021.3123841.
[20] P. Beeson and B. Ames, “TRAC-IK: An open-source library for improved solving of generic inverse kinematics,” in 2015 IEEE-
RAS 15th International Conference on Humanoid Robots (Humanoids), Nov. 2015, pp. 928–935, doi:
10.1109/HUMANOIDS.2015.7363472.
[21] H. F. Durrant-Whyte, “Uncertain geometry in robotics,” IEEE Journal on Robotics and Automation, vol. 4, no. 1, pp. 23–31,
1988, doi: 10.1109/56.768.
[22] A. Steinhauser and J. Swevers, “An efficient iterative learning approach to time-optimal path tracking for industrial robots,” IEEE
Transactions on Industrial Informatics, vol. 14, no. 11, pp. 5200–5207, Nov. 2018, doi: 10.1109/TII.2018.2851963.
[23] L. E. J. Alkurawy, “Recursive least square and control for PUMA robotics,” Indonesian Journal of Electrical Engineering and
Computer Science (IJEECS), vol. 21, no. 2, pp. 1238–1246, Feb. 2021, doi: 10.11591/ijeecs.v21.i2.pp1238-1246.
[24] B. Koyuncu and M. Güzel, “Software development for the kinematic analysis of a Lynx 6 robot arm,” International Journal of
Computer and Information Engineering, vol. 1, no. 6, pp. 1575–1580, 2007.
[25] A. Mathur, C. Bansal, S. Chauhan, and O. Yadav, “A review of pick and place operation using computer vision and ROS,” in
Computational and Experimental Methods in Mechanical Engineering, Springer Singapore, 2022, pp. 411–418.
[26] G. Xu and Y. Yan, “A scene feature based eye-in-hand calibration method for industrial robot,” in Mechatronics and Machine
Vision in Practice 4, Cham: Springer International Publishing, 2021, pp. 179–191.
Mathematical modeling and kinematic analysis of 5 degrees of freedom serial link … (Abhilasha Singh)
1532  ISSN: 2088-8708

[27] M. Ahmad, I. Ahmed, and A. Adnan, “Overhead view person detection using YOLO,” in 2019 IEEE 10th Annual Ubiquitous
Computing, Electronics and Mobile Communication Conference (UEMCON), Oct. 2019, pp. 627–633, doi:
10.1109/UEMCON47517.2019.8992980.
[28] O. Martínez and R. Campa, “Comparing methods using homogeneous transformation matrices for kinematics modeling of robot
manipulators,” in Multibody Mechatronic Systems, Springer International Publishing, 2021, pp. 110–118.

BIOGRAPHIES OF AUTHORS

Abhilasha Singh received the B.E. degree in Electrical and Electronics


engineering from Jerusalem College of Engineering, Chennai, India in 2013 and the M.E.
degree in Instrumentation engineering from MIT, Anna University, Chennai, India, in 2017.
She is currently pursuing a Ph.D. degree in Electrical and Electronics Engineering at Birla
Institute of Technology and Science Pilani, Dubai Campus since 2018. Her current research
area focuses on robotics, computer vision, and intelligent control techniques. She can be
contacted at p20180906@dubai.bits-pilani.ac.in.

Kalaichelvi Venkatesan is an associate professor in the Department of Electrical


and Electronics Engineering at BITS Pilani, Dubai Campus. She is working with Bits Pilani,
Dubai Campus since 2008 and she is having 29 years of teaching experience. She has her
research work published in refereed international journals and many international conferences.
She has also reviewed many papers in international journals and conferences. Her research
area of expertise includes process control, neural networks, fuzzy logic, computer vision, and
control systems. She is currently guiding students in the area of intelligent control techniques
applied to robotics and mechatronics. She can be contacted at kalaichelvi@dubai.bits-
pilani.ac.in.

Yuvalakshmi Nagarasan is a final year B.E. Electrical and Electronics


Engineering student at BITS Pilani-Dubai Campus. She is well-versed in power systems,
design, and development in the fundamentals of electrical engineering and automation. She is
equipped with the knowledge of several software required for robotic manipulation such as
Python and MATLAB. Her research interests include power systems, sustainable energy, and
robotics. She can be contacted at f20180047@dubai.bits-pilani.ac.in.

Karthikeyan Ramanujam is a professor in the Department of Mechanical


Engineering at Bits Pilani, Dubai Campus. He is working with BITS Pilani, Dubai Campus
since 2007 and is having 30 years of teaching experience. He has also guided five Ph.D.
scholars. He has published over 75 papers in international journals and conferences. He has
also reviewed many papers in international journals and conferences. His research area of
expertise includes material processing, manufacturing engineering, mechatronics, and robotics.
He can be contacted at rkarthikeyan@dubai.bits-pilani.ac.in.

Kumar Karuppusamy is a professor in the Department of General Science at


BITS Pilani, Dubai Campus. He is working with Bits Pilani, Dubai Campus since 2004 and is
having 33 years of teaching experience. His research area of expertise includes nonlinear
dynamics, heat transfer and wave theory. He can be contacted at kumar@dubai.bits-
pilani.ac.in.

Int J Elec & Comp Eng, Vol. 13, No. 2, April 2023: 1522-1532

You might also like