In the ultra-close approaching phase of tethered space robot, a highly stable self-attitude control is essential. However, due to the field of view limitation of cameras, typical point features are difficult to extract, where commonly adopted position-based visual servoing cannot be valid anymore. To provide robot’s relative position and attitude with the target, we propose a monocular visual servoing control method using only the edge lines of satellite brackets. Firstly, real time detection of edge lines is achieved based on image gradient and region growing. Then, we build an edge line based model to estimate the relative position and attitude between the robot and the target. Finally, we design a visual servoing controller combined with PD controller. Experimental results demonstrate that our algorithm can extract edge lines stably and adjust the robot’s attitude to satisfy the grasping requirements.
National Natural Science Foundation of China1127225660805034Fundamental Research Funds for the Central Universities3102016BJJ031. Introduction
The increasing number of space debris has been a serious threat for the safety of space activities. Although many debris removal methods have already been proposed [1–3], such as the electrodynamic tether [4] and foam method [5], the majority of them are limitedly examined on ground and far from being practical. In practice, the traditional “space platform + multi-DOF manipulator arm” system has been the preferred solution; however, it has several disadvantages, such as small operating range, complicated control operation, and being only applicable for cooperative target capturing, whose attitude and movement are stable. Therefore, developing on-orbit service technologies, especially for noncooperative targets, is critical. In order to overcome the problems caused by manipulator, Huang et al. [6, 7] proposed the tethered space robot (TSR) system for noncooperative target capturing and detumbling [8, 9]. TSR is a new type of space robot, which comprises a platform, space tether, and an operational robot, as shown in Figure 1. The operational robot is released by the platform through the space tether to approach the target and accomplish an on-orbit task. Compared with the manipulator arm, TSR has two main merits: a large operational radius (as long as 1,000 m) and the enhanced flexibility.
Diagram of tethered space robot.
The flow of providing an on-orbit service can be divided into three sequential stages [10]:
The platform distinguishes the target from space environment and flies towards it gradually from a far distance.
The platform flies around the target and detects corresponding regions to identify a suitable position to release the operational robot. Then, the operational robot will be launched out and freely fly towards the frontal target. Meanwhile, it keeps detecting the suitable grasping region of the target to guide end-effector in precise manipulation.
The operational robot grasps the target once reaching the appointed position (usually less than 0.15 m) and eliminates the tumbling of the robot-target combination with its own propellants. Then it provides on-orbit services, such as injecting propellants and dragging the target into a graveyard orbit.
In the third stage, the operational robot needs to capture the target spacecraft through the docking system (such as the docking ring) in order to provide on-orbit services. But the docking system usually works for cooperative targets only. When it comes to noncooperative targets, any device or structure mounted on them is unknown in advance. To facilitate the determination of the grasping position, we analyze the common spacecraft structures and select satellite bracket as a suitable location for grasp, which is usually structurally robust and widely present in spacecraft [11].
During the approaching process, it is necessary for TSR to recognize the grasping position and measure their relative positions and attitudes. The calculated information will be fed into the control system to adjust TSR’s self-attitude and guide its approaching path. Though microwave radar and laser radar have been used in relative navigation, we select charge-coupled device (CCD) cameras mounted on the operational robot to provide vision-based measurements considering their low mass and rich view information [12].
Providing the on-orbit service involves many key techniques and one of them is the control of the relative position and attitude between TSR and the target.
Yuga et al. [13] proposed a coordinated control method using tension and thruster. Wang et al. [14, 15] proposed that, through altering the position of the connecting point between the tether and the robot, the tether tension force was able to achieve the orbit and attitude control simultaneously. Mori and Matunaga [16] designed tether tension/rope length controller and realized the attitude control. Experiments showed that this controller could save propellants. But these methods all require that the relative position between the space platform, tethered robot, and the target is known. In order to measure the relative position and orientation between them, visual servoing control problem has been widely studied. Xu et al. [17] proposed a stereo vision-based method to measure the relative position and orientation. The basic idea is to identify the feature points of the target in 3D space. Thienel et al. [18] presented a nonlinear method to estimate the attitude of the target spacecraft and then realized the tracking to it. But it assumed that the target attitude was already estimated using vision method. Hafez et al. [19] proposed a visual servoing control method of the dual-arm space robot. In [20], Dong and Zhu used the extended Kalman filter to estimate pose and motion of noncooperative target in real time. Cai et al. [21] proposed a monocular visual servoing control system for the tethered space robot. It realized the real time tracking of noncooperative target satellite in complex scenes. For these position and orientation measurement methods, the key technology is the detection and recognition of the characteristics of the target spacecraft. Considering the wide existence of circle-shape device, Hough Transform methods [22] are generally adopted to detect space target in autonomous rendezvous and docking. Casonato and Palmerini [23] used Hough Transform to identify basic geometric shapes, such as circles and straight lines, in rendezvous monitoring. The work [24] also used Hough Transform to convert the image from spatial space to parameter space, facilitating the detection of launch bolt holes. In [25], two marks, an external circle and an internal circle, are laid on different planes of the target to indicate the relative attitude and position. They can be easily detected by Hough Transform methods too. In addition, Ramirez et al. [26] proposed a quadrilateral recognition method, which could be used to recognize the spacecraft’s solar panels. Grompone Von Gioi et al. [27] proposed a line segment detection algorithm based on image gradients. It could achieve rapid detection of line segments and had been applied to detect the edge lines of the spacecraft’s solar panels. Wang et al. [28, 29] proposed a new approach to dynamic eye-in-hand visual tracking for the robot manipulator in constrained environments with an adaptive controller.
The above methods generally aim at detecting the characteristics of the target from a relatively far distance, where the point features, linear features, or geometric features of the target could be fully detected [30], as shown in Figures 2(a) and 2(b). But when in extremely close distance (usually less than 1 m), these features are generally difficult to extract and are incomplete due to the limited FOV of cameras. The brackets supporting the solar panels are the element present in the majority of spacecraft, excluding the small cube satellites which often accommodate body-mounted panels. These elements can be easily recognized by cameras using their edge lines (Figure 2(c)). But for the above-mentioned methods, simply using the detected edge line features is inadequate and they tend to fail [31].
Images captured in different approaching distances.
Far distance
Close distance
Extremely close distance
In this paper, we propose a visual servoing control method based on the detection of margin line on the satellite brackets. Firstly, a gradient based edge line detection method is proposed to extract edges of the satellite brackets. Then, we construct the model to acquire the relative position and attitude between TSR and the target. Lastly, we integrate the PD controller to design a new visual servoing controller. It is able to control TSR’s attitude to satisfy the grasping requirements. This method is appropriate for addressing the relative position and attitude problem in extremely close distance when TSR approaches the target satellite.
The remainder of this paper is organized as follows. In Section 2, we introduce the project model of edge lines. The detailed description of the line detection algorithm and servoing control method is presented in Sections 3 and 4. Section 5 illustrates and analyzes the experiments and we conclude our work in Section 6.
2. Projection Model of Edge Lines2.1. Coordinates Definition
Observing system is composed of the operational robot and the satellite bracket. In extremely close stage, TSR moves towards the grasping position and only edge information of the brackets is available. In order to facilitate the relative measurement and future expression, the coordinate systems used in our paper are defined as follows:
The coordinate system of operating robot body OGXGYGZG: the origin OG is located in the centroid of the robot; OGXG axis points to the direction of flight; OGZG axis points to the geocenter; OGYG axis direction is determined by the right-hand rule.
The coordinate system of the capture point OZXZYZZZ: the origin OZ is located at the capture point of the target panel bracket; directions of OZZZ axis and OCZC axis are same; the direction of axis OZXZ is along the grasping pole of panel bracket; axis OZYZ is determined by the right-hand rule.
Camera coordinate system OCXCYCZC: origin OC is at the optical center of the camera, OCZC axis is the optical axis, and OCXC axis and OCYC axis are horizontal axis and vertical axis which are parallel to the image plane.
Image coordinate system ODXDYD: coordinate plane of image coordinate system XDODYD is on the camera’s CCD imaging plane, the origin OD is the intersection of optical axis and imaging plane, and ODXD axis and ODYD axis are parallel to the OCXC and the OCYC axis.
Computer image coordinate system UV: the origin is located in the upper left corner of the digital image, U and V denote the number of columns and rows of a digital image. The relationships of aforementioned coordinate systems are illustrated in Figure 3.
Definition of coordinate systems.
Besides, we choose the widely adopted Pinhole model to represent the camera imaging relationship [32]. Set the coordinate of a space point P at camera coordinate system OCXCYCZC as (xc,yc,zc); then the imaging model of the point P is represented as(1)uv1=αx0u00αyv0001xczcyczc1.
(u,v) is the coordinate of point P in the coordinate system of computer image. (u0,v0) is the main point of the camera. αx=f/dx,αy=f/dy, f is the focal length of the camera, and dx and dy are a pixel’s physical sizes in the horizontal and vertical direction. Generally, dx is equal to dy.
If two points, A(u1,v1) and B(u2,v2), which locate in the computer image coordinate system UV, are considered, then the line AB can be expressed as(2)u=kv+b,k=u1-u2v1-v2,b=v1u2-u1v2v1-v2.
2.2. Camera Imaging Model
Assume the rotation matrix and translation vector from grasping point coordinate system OZXZYZZZ to camera coordinate system as RTC and T, respectively [33]. The matrix transformed from grasping point coordinate system to camera coordinate system is(3)MTC=RTCT0T1.
Therefore, the transformational relationship of the space point P between homogeneous coordinates (xz,yz,zz,1)T in grasping point coordinates OZXZYZZZ and homogeneous coordinates (xc,yc,zc,1)T in the camera coordinate system OCXCYCZC is(4)xcyczc1=MTCxzyzzz1=RTCT0T1xzyzzz1,zcuv1=αx0u000αyv000010RTCT0T1xzyzzz1.
3. Line Segment Detection3.1. Image Gradients
Image gradients reflect the intensity changes between adjacent pixels in the image. A small gradient value corresponds to a region with smooth gray level in the image, and a large gradient value corresponds to a sharp edge region in the image. Pixels along the consistent gradient direction are likely to belong to an edge in the image. Thus, calculating the gradients of images is important for the edge structure detection. For image I, the gradients in pixel x are defined as (5)gxx,y=Ix+1,y+Ix+1,y+1-Ix,y-Ix,y+12,gyx,y=Ix,y+1+Ix+1,y+1-Ix,y-Ix+1,y2,where I(x,y) represents the intensity value of pixel (x,y); gx(x,y) and gy(x,y) are horizontal and vertical gradients of pixel (x,y). Then, we calculate(6)Gx,y=gx2x,y+gy2x,y,θ=arctangyx,ygxx,y,where G(x,y) and θ represent the gradient value and gradient orientation of pixel (x,y), respectively.
To speed up the edge structure detection and exclude the monotonous regions in image, we first preprocess the image by thresholding the gradient value of each pixel. Pixels whose gradient value is less than the predefined threshold are more likely to be from nonedge regions and will be removed directly. Beside, we define pixel consistency τ as the difference between adjacent pixel gradient orientations, which is τ=θ1-θ2. If τ<t, it is considered that the orientations of adjacent pixels are consistent and these pixels should be in the same straight line, where t is an human-defined parameter. The determination of τ is described in Section 3.3.
3.2. Pseudo-Sorting Method
Line support area is composed of pixels with similar gradient directions in the image, as shown in Figure 4. In general, we use the external rectangle, for example, the red color one in Figure 4, to approximate the support area.
Line support areas.
The support area of straight line segment is generated using region growing method [34, 35]. It first selects a small part of pixels as the seeds. Then, it iteratively integrates neighboring pixels with similar orientations and the seed region gradually grows. The detailed region growing strategy is introduced in Section 3.3. The performance of the region growing method is sensitive to the selection of seed pixels. One way to determine seeds is to sort image pixels according to their gradient magnitudes and selects pixels with larger values as the seed pixels. However, the computational efficiency of sorting these pixels is rather low due to the large amount of pixels contained in the image. Even though some fast sorting methods (e.g., bubble sort, quick sort) are used, the real time performance is still hard to obtain. Here, we adopt the pseudo-sorting method to sort the pixels (Figure 5).
Illustration of pseudo-ordering method using a 9-pixel region.
Pseudo-ordering method is generally used to roughly arrange pixels in order according to their gradients. Firstly, a linear interval is generated according to the maximum and minimum values of the image gradients. Each interval represents a range of gradient values. Pixels are placed in the corresponding intervals according to their gradient values. Hence, as the increase of interval number, the pseudo-ordering results will become more accurate while the algorithmic efficiency becomes lower. Then, for pixels in the same interval, they have the same possibility of being selected as seeds.
To determine the number of intervals, we take several experiments and conclude that, for common 8-bit image, setting interval number to be 1024 could achieve a good tradeoff between performance and time complexity.
3.3. Region Growing Method
According to the results of the pseudo-sorting, the pixel point is selected as the seed point of the region growing according to the descending gradient value [36]. First, we define the direction of the line support area as follows: (7)θregion=arctan∑isinθi∑icosθi,where θregion represents the angle of the support area and θi represents the linear direction angle of the pixel i. Set the seed point of region growing as (x,y); then the direction of the support area at the beginning is the direction of the seed point; namely, θregion=θ(x,y).
Let Sx=cos(θregion) and Sy=sin(θregion). For each point p(x,y) in the support area, we visit its 8 neighboring pixels, which are p(x-1,y-1), p(x,y-1), p(x+1,y-1), p(x-1,y), p(x+1,y), p(x-1,y+1), p(x,y+1), and p(x+1,y+1), and calculate the difference between their gradient orientations and θregion. If the difference is less than τ, it is considered that they have the same direction, and the corresponding pixel will be added to the straight line support area. Then, we update the direction of the support area using (8)θregion=arctanSy+sinθP¯Sx+cosθP¯,where P- denotes the selected neighbor pixel of p(x,y). The visited pixels will be removed from the candidate list and not be accessed again. Repeat above steps until no new pixel could be added to the linear support area, and the regional growth algorithm terminates.
During the region growing, the selection of the threshold value τ is key and will influence the final results of the algorithm. Intuitively, when a small τ is adopted, only pixels with almost identical orientations could be included in the line support area. The generated support area may be too short to provide the complete edge information. As τ increases, more nonedge pixels tend to be covered by the support area, which leads to its increase in length and width. This will cause a lot of false positive detection. In order to make a tradeoff between deficient edge information and high false positive rate, we evaluate the performance under different τ and select τ to be 22.5° by trial and error.
The region growing results under different τ are illustrated in Figure 6. Besides 22.5°, two typical thresholds, which are smaller and larger than 22.5°, are selected to provide the visual examples.
Region growing results with different τ. (a) is the original gradient map, including the edge area with noise interference. (b), (c), and (d) are results corresponding to 11.25°, 22.5°, and 45°, respectively.
Origin image
τ=11.25°
τ=22.5°
τ=45°
3.4. Straight Line Determination
According to the aforementioned region growing algorithm, the support area of straight line segment is generated. Then the next step is to produce the corresponding line segments from each support area. First, in order to reduce the computational burden, the support area with extremely few pixels is excluded, which is likely caused by the isolated noise. Excluding areas with few pixels can be regarded as a type of erosion operation in image processing. Both of them aim to eliminate the small regions in image. Specifically, we could use a hollow circle template with small radius and slide it over the image. For areas with few pixels, they will be totally enclosed by the template and hence will be discarded. For support region areas, they will intersect with the circle due to their length and will be retained.
In order to determine the straight line segment, the external rectangle is used to describe its support area. The support area is scored by summing up the gradient values of pixels contained in it. Then the support area centroid (cx,cy) can be calculated by the following formula:(9)cx=∑j∈ReGj∗xj∑j∈ReGj,cy=∑j∈ReGj∗yj∑j∈ReGj,where G(j) is the gradient magnitude value of the pixel j in the support area Re; x(j) and y(j) are coordinates of pixel j. The center of the external rectangle is chosen as the center of the support area, and the long axis of the rectangle is chosen as the direction θregion of the support area, with the short axis vertical to θregion.
Since not all of the straight segment support areas correspond to a straight line segment model, they need to be further judged after obtaining the external rectangle of the support area. In this paper, we develop two criteria to achieve this.
(1) The ratio of long and short axis in the external rectangle should be larger than the set threshold (typically 4 to 7).
Due to the existence of noise, pixels in monotonous region may be wrongly selected as the seeds of region growing method, leading to the support areas shown in Figure 7(b). Besides, the noise is also likely to form some small line segments, which cannot be detected due to the pixel number it contained, as shown in Figure 7(c). It is obvious to see that by thresholding the ratio between long and short axis these support areas can both be eliminated.
Illustration of different support areas including (a) straight line, (b) monotonous region, (c) isolated noise, and (d) low-curvature curve. (b) and (c) could be discarded by criterion 1; (d) could be discarded by criterion 2.
(2) When the direction of the external rectangle is consistent with the direction of the long axis of rectangle (direction deviation is less than τ), the proportion of its pixel number to the total number of pixels contained in the rectangular should be larger than the predefined threshold (typically 50% to 70%).
Determining the straight lines with long-short axis ratio may include some low-curvature curves, such as the circular edge with a larger radius, as shown in Figure 7(d). It can be observed that although the curve shares a large long-short axis ratio, it tends to produce the increase of the support region area. Hence, by thresholding the proportion of the pixel number divided by the support region area, these regions can be eliminated.
The support areas which satisfy these two criteria correspond to the straight line segment model, and the principal axis which comes through the center of the external rectangle is the straight line segment. As shown in Figure 8, the principal axis of the external rectangle, that is, the blue line segment, is the straight segment.
Schematic diagram of the line segment validation.
4. Servoing Controller Design
Through the above algorithm, we can detect the edge lines of the satellite brackets in the image. These operations are all based on the computer image coordinates. Hence, the detected edge lines can be expressed as(10)u=k1v+b1,u=k2v+b2.
According to (10), we design a controller to achieve the control of relative attitude of the robot in the ultra-close approaching stage.
4.1. Edge Lines Analysis
The capturing device of TSR is usually installed in the front plate of the robot and close to the camera. Hence, we set that the capture coordinate system coincides with the camera coordinate system OCXCYCZC, and assume the Euler angles between capture coordinate system OCXCYCZC and target coordinate system OZXZYZZZ as α,β,γ, respectively [37]. It is easy to analyze that, for different relative attitude of the robot and the target, the imaging of the bracket edge lines in the camera image plane is shown in Figure 9.
Relationships of two edge lines under different relative attitudes.
The imaging result in Figure 9(a) shows the final state of an ideal controller. It is requested that the Euler angles β and γ should be zero, and the relative position deviation y along direction Y is zero. The Euler angles α of axis X can be any angle. Figure 9(b) is the case that y is not zero. Figure 9(c) shows that the Euler angle γ is not zero and there exists rotation along axis Z. The edge lines in Figure 9(d) are not parallel (k1≠k2) and it reveals that the robot rotates along axis Y.
4.2. Controller Design
From the above analysis, we can conclude that (1) the relative attitude angle γ can be measured by summing up the slopes of two edge lines; (2) the relative attitude angle β can be measured by subtracting the slopes of two edge lines; (3) the position deviation y along axis Y can be measured using the sum of b1 and b2. Therefore, we refer to the PD controller and design our controller as follows [38]:(11)Mz=-p1k1+k2-p2k˙1+k˙2,My=-p3k1-k2-p4k˙1-k˙2,Fy=-p5b1+b2-2u0-p6b˙1+b˙2,where p1~p6 are the controller parameters and should be adjusted in practice. Mz and My are the control torque of axes Z and Y. Fy is the control force of axis Y. The stability of the proposed controller could be proved by Lyapunov first method [39, 40].
We have to note that the attitude angle α is not controlled by the controller, because only the pole structure of the bracket is imaged in ultra-close distance (Figure 9(a)) and it is resistant to attitude change along axis X. But in practice α should be restricted during the approaching process using additional sensors, for example, the global camera, to keep TSR’s self-attitude stable. The flowchart of our method is shown in Figure 10.
Flowchart of the proposed method.
5. Experiments and Results5.1. Experiment Conditions
In order to verify the effectiveness of the proposed control algorithm, we set up a visual servoing control system, as shown in Figure 11. It mainly consists of visual perception subsystem, control subsystem, movement platform, and the target satellite model. The composition of visual perception subsystem is listed in Table 1. Since this paper mainly focuses on the visual servoing control in ultra-close distance, we adopt a single-pole model to substitute the satellite bracket model. In the experiment, we use the 6-DOF movement platform to simulate TSR’s movement in space. The visual perception subsystem and target model are mounted on the movement platform to control their relative position and attitude.
The composition of visual perception subsystem.
Camera (VS-902H)
CCD: 1/3′′
Resolution: 752 (H) × 582 (V)
Lens (Computar M1614)
Focal length: 5 mm
Field of view: 65.5° (H) × 51.4° (V)
Processor
Pentium Dual-Core E5200 with 2.50 GHz, 2G RAM
Operating system: Windows XP SP3
Programming environment: VS2005 IDE with OpenCV
Visual servoing control system.
Visual perception and movement subsystems
Single-pole satellite model
5.2. Results Analysis
In order to evaluate the edge line detection performance in approaching process, we make a simulation and move TSR from initial position (about 3 m to the target) to grasping position (about 0.2 m to the target). Meanwhile, the single-pole model swings with an initial rate, about 10°/s. Figure 12 qualitatively demonstrates that despite the change of relative position and attitude between TSR and the target model our proposed algorithm is always able to detect the edge lines.
Edge line detection results under the different position and attitude.
Close-range detection
Detection with positive slope
Detection with negative slope
Figure 13 quantitatively reveals the measurement values of slope k and intercept b of the detected edge lines in every frame, respectively. In (a), the initial slope is −0.45, which means TSR’s self-attitude needs to be further adjusted. As TSR approaches the target, slops of the two detected lines gradually converge to about 0, which means TSR reaches the ideal grasping attitude. Meanwhile, intercepts of the detected lines converge to about 400 and 300, respectively. It is easy to find that their sum is rather close to the image width 2u0. It indicates that the two edge lines are horizontally symmetric along the image center and TSR reaches the ideal grasping position.
Measurement value curves of slope k (a) and intercept b (b) of the detected edge lines. All curves converge at about 500th frame.
In Figure 14, we give the time consumption of the proposed method. The average time of detecting one image is about 0.115 seconds. Considering the control period of the controller is 0.25 seconds, our method could satisfy the real time need.
Comparison of time consumption per frame.
5.3. Limitations
As mentioned in Section 4.2, TSR’s rotation around axis x is not taken into consideration in this paper. But in practice this rotation will lead to TSR’s deviation from the approaching path and increase the possibility of collision with the target. Hence in future works the control to attitude angle α should be added to the controller.
Although our proposed method is able to work successfully, the number of frames it needs to offset the relative position and attitude is a bit high. It could be ascribed to the low frame frequency (4 Hz) of visual camera. The time span between adjacent frames leads to the inconsistency between actual measurement values and control parameters. In order to solve this, we consider (1) increasing the frame frequency of cameras as well as further reducing the complexity of our method and (2) introducing the motion prediction mechanism of TSR to offset the attitude change of target, in the future.
6. Conclusions
This paper presents a novel monocular visual servoing control method using only the edge lines. It is designed to measure the relative position and attitude between TSR and the target under nonfull field of view, which is intractable for traditional feature based visual servoing methods.
The novelty of this paper is constructed by the following three aspects: (1) We propose a novel edge line detection method. It is based on gradient growing and can stably extract lines in the image. Also, it meets the real time requirement of controller. (2) We build the model and analyze the relationship between edge line parameters and the relative attitude. (3) We design a PD-based visual servoing controller for ultra-close approaching process, which requires the two detected edge line parameters only.
Experiments on the semiphysical simulation system indicate that our method is invariant to rotation and scale change.
Competing Interests
The authors declare that they have no competing interests.
Acknowledgments
This research is sponsored by the National Natural Science Foundation of China (Grant nos. 11272256, 60805034) and the Fundamental Research Funds for the Central Universities (Grant no. 3102016BJJ03).
ShanM.GuoJ.GillE.Review and comparison of active space debris capturing and removal methods201680183210.1016/j.paerosci.2015.11.0012-s2.0-84959534044HuangP.ZhangF.MengZ.LiuZ.Adaptive control for space debris removal with uncertain kinematics, dynamics and states201612841643010.1016/j.actaastro.2016.07.0432-s2.0-84982170317ZhangF.HuangP.Releasing dynamics and stability control of maneuverable tethered space net2016PP991110.1109/tmech.2016.2628052WilliamsP.Optimal orbital transfer with electrodynamic tether200528236937210.2514/1.120162-s2.0-15744383603PergolaP.RuggieroA.AndrenucciM.SummererL.Low-thrust missions for expanding foam space debris removalProceedings of the 32nd International Electric Propulsion ConferenceSeptember 2011Wiesbaden, GermanyHuangP.WangD.MengZ.ZhangF.LiuZ.Impact dynamic modeling and adaptive target capturing control for tethered space robots with uncertainties20162152260227110.1109/tmech.2016.25694662-s2.0-84983568824HuangP.WangD.MengZ.ZhangF.GuoJ.Adaptive postcapture backstepping control for tumbling tethered space robot–target combination201639115015610.2514/1.g0013092-s2.0-84957680455HuangP.WangM.MengZ.ZhangF.LiuZ.ChangH.Reconfigurable spacecraft attitude takeover control in post-capture of target by space manipulators201635391985200810.1016/j.jfranklin.2016.03.011ZBL1347.931852-s2.0-84964334165HuangP.WangD.ZhangF.MengZ.LiuZ.Postcapture robust nonlinear control for tethered space robot with constraints on actuator and velocity of space tether201610.1002/rnc.3712ChenL.HuangP.CaiJ.MengZ.LiuZ.A non-cooperative target grasping position prediction model for tethered space robot20165857158110.1016/j.ast.2016.09.009BischofB.KersteinL.StarkeJ.GuentherH.FothW.ROGER—robotic geostationary orbit restorer2004109183193FriendR. B.HowardR. T.MotaghediP.Orbital express program summary and mission overviewSPIE Defense and Security SymposiumApril 2008Orlando, Fla, USA10.1117/12.783792YugaN.FumikiS.ShinichiN.Guidance and control of ‘tethered retriever’ with collaborative tension-thruster control for future on-orbit service missionsProceedings of the 8th International Symposium on Artificial Intelligence: Robotics and Automation in Space (i-SAIRAS '05)September 2005Munich, GermanyWangD.HuangP.CaiJ.MengZ.Coordinated control of tethered space robot using mobile tether attachment point in approaching phase20145461077109110.1016/j.asr.2014.05.0162-s2.0-84905510193WangD.HuangP.MengZ.Coordinated stabilization of tumbling targets using tethered space manipulators20155132420243210.1109/TAES.2015.1405302-s2.0-84942876436MoriO.MatunagaS.Formation and attitude control for rotational tethered satellite clusters200744121122010.2514/1.149102-s2.0-33947133750XuW.LiangB.LiC.XuY.Autonomous rendezvous and robotic capturing of non-cooperative target in space201028570571810.1017/S02635747099903972-s2.0-77958455857ThienelJ. K.VanEepoelJ. M.SannerR. M.Accurate state estimation and tracking of a non-cooperative target vehicleProceedings of the AIAA Guidance, Navigation, and Control ConferenceAugust 2006Keystone, Colo, USA551155222-s2.0-33845788727HafezA. H. A.AnuragV. V.ShahS. V.KrishnaK. M.JawaharC. V.Reactionless visual servoing of a dual-arm space robotProceedings of the IEEE International Conference on Robotics and Automation (ICRA '14)June 20144475448010.1109/icra.2014.69075122-s2.0-84929208241DongG.ZhuZ. H.Position-based visual servo control of autonomous robotic manipulators201511529130210.1016/j.actaastro.2015.05.0362-s2.0-84934953557CaiJ.HuangP.ChenL.ZhangB.An efficient circle detector not relying on edge detection2014511123592375YuenH.PrincenJ.IllingworthJ.KittlerJ.Comparative study of hough transform methods for circle finding199081717710.1016/0262-8856(90)90059-e2-s2.0-0025385444CasonatoG.PalmeriniG. B.Visual techniques applied to the ATV/ISS rendez-vous monitoringProceedings of the IEEE Aerospace Conference ProceedingsMarch 2004Big Sky, Mont, USA6136242-s2.0-11244340671AullM.Visual servoing for an autonomous rendezvous and capture system20092313113710.1007/s11370-009-0042-y2-s2.0-70349811557SabatiniM.PalmeriniG. B.GasbarriP.A testbed for visual based navigation and control during space rendezvous operations201511718419610.1016/j.actaastro.2015.07.0262-s2.0-84941552551RamirezV. A.GutierrezS. A. M.YanezR. E. S.Quadrilateral detection using genetic algorithms2011152181193Grompone Von GioiR.JakubowiczJ.MorelJ.-M.RandallG.LSD: a fast line segment detector with a false detection control201032472273210.1109/tpami.2008.3002-s2.0-77649270310WangH.LiuY.-H.ChenW.WangZ.A new approach to dynamic eye-in-hand visual tracking using nonlinear observers201116238739410.1109/TMECH.2009.20399412-s2.0-78951475199WangH.YangB.LiuY.ChenW.LiangX.PfeiferR.Visual servoing of soft robot manipulator in constrained environments with an adaptive controller201610.1109/tmech.2016.2613410Flores-AbadA.MaO.PhamK.UlrichS.A review of space robotics technologies for on-orbit servicing20146812610.1016/j.paerosci.2014.03.0022-s2.0-84899458942LiuJ.CuiN.ShenF.RongS.Dynamics of RObotic GEostationary orbit restorer system during deorbiting20142911364210.1109/MAES.2014.1301972-s2.0-84919904661PetitA.MarchandE.KananiK.Vision-based space autonomous rendezvous: a case studyProceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS '11)September 2011San Francisco, Calif, USAIEEE61962410.1109/iros.2011.60481762-s2.0-84455204824LaroucheB. P.ZhuZ. H.Autonomous robotic capture of non-cooperative target using visual servoing and motion predictive control201437215716710.1007/s10514-014-9383-22-s2.0-84902295709GambottoJ.-P.A new approach to combining region growing and edge detection1993141186987510.1016/0167-8655(93)90150-CZBL0802.681632-s2.0-0027839009HojjatoleslamiS. A.KittlerJ.Region growing: a new approach1998771079108410.1109/83.7011702-s2.0-0032124126ChungK.-L.HuangY.-H.WangJ.-P.ChangT.-C.Mark LiaoH.-Y.Fast randomized algorithm for center-detection20104382659266510.1016/j.patcog.2010.03.010ZBL1207.684402-s2.0-77951258786RostenE.PorterR.DrummondT.Faster and better: a machine learning approach to corner detection201032110511910.1109/tpami.2008.2752-s2.0-84871610129NohmiM.Mission design of a tethered robot satellite “STARS” for orbital experimentProceedings of the IEEE International Conference on Control Applications (CCA '09)July 2009St. Petersburg, Russia1075108010.1109/cca.2009.52811132-s2.0-74049153894BourquardezO.MahonyR.GuenardN.ChaumetteF.HamelT.EckL.Image-based visual servo control of the translation kinematics of a quadrotor aerial vehicle200925374374910.1109/TRO.2008.20114192-s2.0-67649271503MalisE.ChaumetteF.BoudetS.2 1/2 D visual servoing199915223825010.1109/70.7603452-s2.0-0032666619