By the development of IT systems, human-computer interaction is also developing even faster and newer communication methods become available in human-machine interaction. In this article, the application of a hand gesture controlled human-computer interface is being introduced through the example of a mobile robot. The control of the mobile robot is implemented in a realistic virtual environment that is advantageous regarding the aspect of different tests, parallel examinations, so the purchase of expensive equipment is unnecessary. The usability of the implemented hand gesture control has been evaluated by test subjects. According to the opinion of the testing subjects, the system can be well used, and its application would be recommended on other application fields too.
Socially assistive robotic has become increasingly active and it is present in therapies of people affected for several neurobehavioral conditions, such as Autism Spectrum Disorder (ASD). In fact, robots have played a significant role for positive interaction with children with ASD, by stimulating their social and cognitive skills. This work introduces a mobile socially-assistive robot, which was built for interaction with children with ASD, using non-linear control techniques for this interaction.
In real-time applications, wheeled mobile robots are increasingly used and operated in extreme and diverse conditions traversing challenging surfaces such as a pitted, uneven terrain, natural flat, smooth terrain, as well as wet and dry surfaces. In order to accomplish such tasks, it is critical that the motion control functions without wheel slip and odometry error during the navigation of the two-wheeled mobile robot (WMR). Wheel slip and odometry error are disrupting factors on overall WMR performance in the form of deviation from desired trajectory, navigation, travel time and budgeted energy consumption. The wheeled mobile robot’s ability to operate at peak performance on various work surfaces without wheel slippage and odometry error is directly connected to four main parameters, which are the range of payload distribution, speed, wheel diameter, and wheel width. This paper analyses the effects of those parameters on overall performance and is concerned with determining the ideal range of parameters for optimum performance.
This paper addresses the issue of the autonomous mobile robot (AMR) navigation task based on the hybrid control modes. The novel hybrid control mode, based on multi-sensors information by using the fuzzy approach, has been presented in this research. The system operates in real time, is robust, enables the robot to operate with imprecise knowledge, and takes into account the physical limitations of the environment in which the robot moves, obtaining satisfactory responses for a large number of different situations. An experiment is simulated and carried out with a pioneer mobile robot. From the experimental results, the effectiveness and usefulness of the proposed AMR obstacle avoidance and navigation scheme are confirmed. The experimental results show the feasibility, and the control system has improved the navigation accuracy. The implementation of the controller is robust, has a low execution time, and allows an easy design and tuning of the fuzzy knowledge base.
Robotics brings together several very different engineering areas and skills. There are various types of robot such as humanoid robot, mobile robots, remotely operated vehicles, modern autonomous robots etc. This survey paper advocates the operation of a robotic car (remotely operated vehicle) that is controlled by a mobile phone (communicate on a large scale over a large distance even from different cities). The person makes a call to the mobile phone placed in the car. In the case of a call, if any one of the button is pressed, a tone equivalent to the button pressed is heard at the other end of the call. This tone is known as DTMF (Dual Tone Multiple Frequency). The car recognizes this DTMF tone with the help of the phone stacked in the car. The received tone is processed by the Arduino microcontroller. The microcontroller is programmed to acquire a decision for any given input and outputs its decision to motor drivers in order to drive the motors in the forward direction or backward direction or left or right direction. The mobile phone that makes a call to cell phone stacked in the car act as a remote.
The feedbacks obtained regarding the sense of presence from pilot users operating a Mobile Robotic presence (MRP) system to visit a simulated museum are reported in this paper. The aim is to investigate how much the perception of system’s usefulness and ease of use is affected by operators’ sense of social telepresence (presence) in the remote location. Therefore, scenarios of visiting a museum are simulated and the user operators are supposed to perform some regular tasks inside the remote environment including interaction with local users, navigation and visiting the artworks. Participants were divided into two groups, those who had previous experience of operation and interaction with a MRP system and those who never had experience. Based on the results, both groups provided different feedbacks. Moreover, there was a significant association between user’s sense of presence and their perception of system usefulness and ease of use.
The inspection of underneath vehicle system has been given significant attention by governments after the threat of terrorism become more prevalent. New technologies such as mobile robots and computer vision are led to have more secure environment. This paper proposed that a mobile robot like Aria robot can be used to search and inspect the bombs under parking a lot vehicle. This robot is using fuzzy logic and subsumption algorithms to control the robot that movies underneath the vehicle. An OpenCV library and laser Hokuyo are added to Aria robot to complete the experiment for under vehicle inspection. This experiment was conducted at the indoor environment to demonstrate the efficiency of our methods to search objects and control the robot movements under vehicle. We got excellent results not only by controlling the robot movement but also inspecting object by the robot camera at same time. This success allowed us to know the requirement to construct a new cost effective robot with more functionality.
Weed suppression and weeding are necessary measures for rice cultivation. Weed suppression precedes the process of weeding. It means suppressing the growth of young weeds and creating a weed-less environment. If we suppress the growth of weeds, we can reduce the number of weeds in a paddy field. This would result in a reduction of the weeding work load. In this paper, we will show how we developed a weed suppression robot for the purpose of reducing the weeding work load. The robot has a laser range finder for autonomous mobility and a robot arm for weed suppression. It travels along the rice rows without stepping on and injuring the rice plants in a paddy field. The robot arm applies force to the weed seedlings and thereby suppresses the growth of weeds. This paper will explain the methodology of the autonomous mobile, the experiment in weed suppression, and the method of controlling the robot’s posture on uneven ground.
This paper presents the hardware implemented and validation for a special system to assist the unprofessional users of car with back trailers. The system consists of two platforms; the front car platform (C) and the trailer platform (T). The main objective is to control the Trailer platform using the actuators found in the front platform (c). The mobility of the platform (C) is investigated and inverse and forward kinematics model is obtained for both platforms (C) and (T).The system is simulated using Matlab M-file and the simulation examples results illustrated the system performance. The system is constructed with a hardware setup for the front and trailer platform. The hardware experimental results and the simulated examples outputs showed the validation of the hardware setup.
In this paper we propose a discrete tracking control of nonholonomic mobile robots with two degrees of freedom. The electromechanical model of a mobile robot moving on a horizontal surface without slipping, with two rear wheels controlled by two independent DC electric, and one front roal wheel is considered. We present backstepping design based on the Euler approximate discretetime model of a continuous-time plant. Theoretical considerations are verified by numerical simulation.
This paper presents a hybrid fuzzy logic control strategy for a unicycle trajectory following robot on irregular terrains. In literature, researchers have presented the design of path tracking controllers of mobile robots on non-frictional surface. In this work, the robot is simulated to drive on irregular terrains with contrasting frictional profiles of peat and rough gravel. A hybrid fuzzy logic controller is utilised to stabilise and drive the robot precisely with the predefined trajectory and overcome the frictional impact. The controller gains and scaling factors were optimised using spiral dynamics optimisation algorithm to minimise the mean square error of the linear and angular velocities of the unicycle robot. The robot was simulated on various frictional surfaces and terrains and the controller was able to stabilise the robot with a superior performance that is shown via simulation results.
Now-a-days autonomous mobile robots have found applications in diverse fields. An autonomous robot system must be able to behave in an intelligent manner to deal with complex and changing environment. This work proposes the performance of path planning and navigation of autonomous mobile robot using Gravitational Search Algorithm (GSA), Simulated Annealing (SA) and Particle Swarm optimization (PSO) based intelligent controllers in an unstructured environment. The approach not only finds a valid collision free path but also optimal one. The main aim of the work is to minimize the length of the path and duration of travel from a starting point to a target while moving in an unknown environment with obstacles without collision. Finally, a comparison is made between the three controllers, it is found that the path length and time duration made by the robot using GSA is better than SA and PSO based controllers for the same work.
Localization of mobile robots are important tasks for developing autonomous mobile robots. This paper proposes a method to estimate positions of a mobile robot using a omnidirectional camera on the robot. Landmarks for points of references are set up on a field where the robot works. The omnidirectional camera which can obtain 360 [deg] around images takes photographs of these landmarks. The positions of the robots are estimated from directions of these landmarks that are extracted from the images by image processing. This method can obtain the robot positions without accumulative position errors. Accuracy of the estimated robot positions by the proposed method are evaluated through some experiments. The results show that it can obtain the positions with small standard deviations. Therefore the method has possibilities of more accurate localization by tuning of appropriate offset parameters.
The aim of this research is to design and implement line-tracking mobile robot. The robot must follow a line drawn on the floor with different color, avoids hitting moving object like another moving robot or walking people and achieves color sensing. The control system reacts by controlling each of the motors to keep the tracking sensor over the middle of the line. Proximity sensors used to avoid hitting moving objects that may pass in front of the robot. The programs have been written using micro c instructions, then converted into PIC16F887 ATmega48/88/168 microcontrollers counterparts. Practical simulations show that the walking robot accurately achieves line following action and exactly recognizes the colors and avoids any obstacle in front of it.
Mobile robotics is gaining an increasingly important role in modern society. Several potentially dangerous or laborious tasks for human are assigned to mobile robots, which are increasingly capable. Many of these tasks need to be performed within a specified period, i.e, meet a deadline. Missing the deadline can result in financial and/or material losses. Mechanisms for predicting the missing of deadlines are fundamental because corrective actions can be taken to avoid or minimize the losses resulting from missing the deadline. In this work we propose a simple but reliable deadline missing prediction mechanism for mobile robots through the use of historical data and we use the Pioneer 3-DX robot for experiments and simulations, one of the most popular robots in academia.
Autonomous mobile robots can be found in a wide field of applications. Their types range from household robots over workshop robots to autonomous cars and many more. All of them undergo a number of testing steps during development, production and maintenance. This paper describes an approach to improve testing of robot behavior. It was inspired by the RoboCup @work competition that itself reflects a robotics benchmark for industrial robotics. There, scaled down versions of mobile industrial robots have to navigate through a workshop-like environment or operation area and have to perform tasks of manipulating and transporting work pieces. This paper will introduce an approach of automated vision-based testing of the behavior of the so called youBot robot, which is the most widely used robot platform in the RoboCup @work competition. The proposed system allows automated testing of multiple tries of the robot to perform a specific missions and it allows for the flexibility of the robot, e.g. selecting different paths between two tasks within a mission. The approach is based on a multi-camera setup using, off the shelf cameras and optical markers. It has been applied for test-driven development (TDD) and maintenance-like verification of the robot behavior and performance.
Assistive robotics are playing a vital role in advancing the quality of life for disable people. There exist wide range of systems that can control and guide autonomous mobile robots. The objective of the control system is to guide an autonomous mobile robot using the movement of eyes by means of EOG signal. The EOG signal is acquired using Ag/AgCl electrodes and this signal is processed by a microcontroller unit to calculate the eye gaze direction. Then according to the guidance control strategy, the control commands of the wheelchair are sent. The classification of different eye movements allows us to generate simple code for controlling the wheelchair. This work was aimed towards developing a usable and low-cost assistive robotic wheel chair system for disabled people. To live more independent life, the system can be used by the handicapped people especially those with only eye-motor coordination.
Methods for measuring or estimating ground shape by a laser range finder and a vision sensor (Exteroceptive sensors) have critical weaknesses in terms that these methods need a prior database built to distinguish acquired data as unique surface conditions for driving. Also, ground information by Exteroceptive sensors does not reflect the deflection of ground surface caused by the movement of UGVs. Therefore, this paper proposes a method of recognizing exact and precise ground shape using an Inertial Measurement Unit (IMU) as a proprioceptive sensor. In this paper, firstly this method recognizes the attitude of a robot in real-time using IMU and compensates attitude data of a robot with angle errors through analysis of vehicle dynamics. This method is verified by outdoor driving experiments of a real mobile robot.
We present in this work the performances of a mobile omnidirectional robot through evaluating its management of the redundancy of actuation. Thus we come to the predictive control implemented.
The distribution of the wringer on the robot actions, through the inverse pseudo of Moore-Penrose, corresponds to a « geometric ›› distribution of efforts. We will show that the load on vehicle wheels would not be equi-distributed in terms of wheels configuration and of robot movement.
Thus, the threshold of sliding is not the same for the three wheels of the vehicle. We suggest exploiting the redundancy of actuation to reduce the risk of wheels sliding and to ameliorate, thereby, its accuracy of displacement. This kind of approach was the subject of study for the legged robots.
In this paper, we present a localization of a mobile robot with localization modules which have two ceiling-view cameras in indoor environments. We propose two kinds of localization method. The one is the localization in the local space; we use the line feature and the corner feature between the ceiling and wall. The other is the localization in the large space; we use the natural features such as bulbs, structures on the ceiling. These methods are installed on the embedded module able to mount on the robot. The embedded module has two cameras to be able to localize in both the local space and the large spaces. The experiment is practiced in our indoor test-bed and a government office. The proposed method is proved by the experimental results.