REAL-TIME POSE ESTIMATION SYSTEM USING INERTIAL AND FEATURE MEASUREMENTS
A hybrid estimator system using visual and inertial sensors for real-time pose tracking on devices with limited processing power using at least one processor, a memory, a storage and communications through a protocol and one or more than one software module for a hybrid estimator, real-time algorithm selection to process different measurements, statistical learning for these characteristics to compute the expected device computing cost of any strategy for allocating measurements to algorithms, and algorithm selection based on the statistical learning module.
Latest The Regents of the University of California Patents:
- Designs and Applications of a Low-Drag, High-Efficiency Microchannel Polymer Heat Exchanger
- METHODS FOR FABRICATING A VERTICAL CAVITY SURFACE EMITTING LASER
- METHODS FOR MAKING AND USING THERAPEUTIC CELLS
- REAL-TIME SINGLES-BASE CARDIO-RESPIRATORY MOTION TRACKING FOR MOTION-FREE PHOTON IMAGING
- AIR DISTRIBUTOR FOR AN ALMOND STOCKPILE HEATED AND AMBIENT AIR DRYER (SHAD)
This application claims the benefit under 35 U.S.C. §119(e) of U.S. Provisional Patent Application Ser. No. 61/824,309, filed on May 16, 2013, the contents of which are incorporated herein by reference in their entirety.
FIELD OF THE INVENTIONThe present invention relates to real-time pose tracking, and more specifically to a hybrid estimator system using visual and inertial sensors for real-time pose tracking on devices with limited processing power.
BACKGROUNDCurrently, there are no adequate solutions to the problem of tracking the 3D-pose of small, resource-constrained systems in unknown environments. Specifically, estimating the motion of miniature devices, similar in size to a mobile phone, is difficult. In contrast to medium and large-scale systems, (e.g. mobile robots, UAVs, autonomous cars), small devices have limited computational capabilities and battery life, factors which make the pose-estimation problem challenging. In the absence of GPS, the types of sensors that can be used for pose estimation in small-scale systems are quite restricted.
A key challenge in designing an estimator for this task is the limited availability of processing power. Feature detection algorithms can track hundreds of features in images of natural scenes, but processing all this data in real time is challenging, particularly for a small, resource-constrained device.
Additional difficulties arise when designing an estimator for any localization task. The computational efficiency of different methods depends strongly on the nature of each particular dataset. For instance, one algorithm may outperform all others when the environment is feature-rich and the vehicle is moving slowly, while a different algorithm may be the fastest in feature-sparse environments under fast motion. This makes algorithm selection a difficult task, for which no general, systematic methods exist to date.
A huge number of methods have been proposed for pose estimation, however, existing methods typically consist of a single approach for processing the feature measurements (e.g., EKF-based SLAM, or sliding-window iterative minimization). As a result, they often do not generalize well to different environments, and are difficult to adapt to the varying availability of computational resources.
Therefore there is a need for a hybrid-estimator system using visual and inertial sensors for real-time pose tracking on devices with limited processing power.
SUMMARYThe present invention solves the problems of the prior art by providing a hybrid estimator system using sensors for real-time pose tracking of moving platforms. The system has a device with a processor, memory, storage and communications through a protocol to execute one or more than one software module. The software modules comprise code executable on the processor for executing the hybrid estimator that uses a plurality of algorithms for processing measurements and selecting the constituent algorithms of the hybrid estimator used to process each measurement.
The system uses a module with non-statutory instructions for learning the statistical characteristics of the measurements and the algorithm selection is based on the learning module. The system gathers information to compute the expected cost of strategies for processing measurements by different algorithms to reduce the computational requirements. These computations are solved by the system in real time. The system solves an optimization problem with an objective function representing the expected computation time to identify the preferred strategy.
The hybrid estimator can then estimate a moving platform's trajectory using inertial measurements and the observations of features by one or more than one sensors or by using inertial measurements and the observations of features by a visual sensor.
The system of claim 1, where the system comprises non-transitory instructions to process each of the available feature measurements, where the feature measurements can be processed by including a description of the features as variables to be estimated in the hybrid estimator. Feature measurements can also be processed by using obtained observations in order to derive constraints, directly used for updating the pose estimates of the moving object.
The system has non-transitory instructions to improve the accuracy of the pose estimates, reduce the computational requirements of the estimator or both improve the accuracy of the pose estimates and reduce the computational requirements of the estimator during system operation to reduce the processing requirements.
The hybrid estimator has non-transitory instructions to determine which of the plurality of methods is to be used for processing each of the feature measurements to reduce the computational requirements of the estimator and to adapt the system to the characteristics of the environment, the trajectory of the moving object, and the availability of processing resources. The hybrid estimator determines the number of features that should be extracted from raw sensor data in order to adjust to the availability of computational resources.
The hybrid estimator constructs linearized approximations of the nonlinear mathematical models that describe the motion and the sensor measurements in order to compute a description of the uncertainty of the estimates, where the linearization points are selected to preserve the system's observability properties. A unique estimate of certain states is used in order to compute the linearization matrices for all measurement equations that involve each of the states. The estimates of one or more than one state is used to compute the linearization matrices and can be modified by equal amounts, to reduce linearization inaccuracy while preserving the system's observability properties.
The hybrid estimator can also implement a hybrid extended Kalman filter. The hybrid extended Kalman filter comprises an extended-Kalman filter algorithm that includes feature states in the state vector, and a sliding-window extended-Kalman-filter algorithm that includes states of the mobile platform in the state vector. The size of the sliding window is selected to reduce the computational cost of the hybrid extended Kalman filter. The hybrid estimator module determines the choice of algorithm to process each individual feature depending on the distribution of the feature track lengths of features. The hybrid estimator also determines the optimal strategy for processing the feature measurements by solving a one-variable optimization problem using the information above and processes all available measurements without loss of localization information.
The plurality of algorithms, employ an extended Kalman filter, must have bounded computational complexity, irrespective of the duration of the trajectory. The hybrid estimator also includes an extended Kalman filter-sliding-window iterative minimization algorithm, where the state vector contains a current IMU state as well as representations of the feature positions. Features that leave the field of view are removed from the state vector leaving only the currently observed ones, to keep the computations bounded. The hybrid estimator has other extended-Kalman-filter algorithms that maintain a sliding window of camera poses in the state vector, and use the feature observations to apply probabilistic constraints between these poses. One of the other extended Kalman filter is a multistate-constraint Kalman filter.
The algorithm used to process feature measurements is selected to have the lowest computational cost. The hybrid filter is a combination of both the extended Kalman filter-sliding-window iterative minimization and the multistate-constraint Kalman filter algorithms. The hybrid filter is a filter whose state vector contains the current IMU state, m camera poses, and sk features, and determines whether a feature will be processed using the multistate-constraint Kalman filter algorithm, or whether it will be included in the state vector and processed using the extended Kalman filter-sliding-window iterative minimization algorithm.
Also provided is method for using a hybrid estimator with visual and inertial sensors for real-time pose tracking on devices with limited processing power. First, a method to be used by the system for processing each of the feature measurements is determined. Then, the system is adapted to the characteristics of the environment, the trajectory of the moving object, and the availability of processing resources. Next, the number of features to be extracted from raw sensor data is determined in order to adjust to the availability of computational resources. Finally, linearized approximations of the nonlinear mathematical models that describe the motion and the sensor measurements are constructed.
Additionally, a method for making a hybrid estimator using visual and inertial sensors for real-time pose tracking on devices with limited processing power. First, a device comprising at least one processor, a memory, a storage and communications through a protocol is provided. Next, one or more than one software module is provided. The software one or more than one module are communicatively coupled to each other. The software modules provide a hybrid estimator, real-time algorithm selection to process different measurements, statistical learning for these characteristics to compute the expected device computing cost of any strategy for allocating measurements to algorithms, and algorithm selection based on the statistical learning module.
The method above also propagates the state vector and covariance matrix using the IMU readings, determines when camera measurements and features are available, augments the state vector with the latest camera pose, determines if the features are to be processed using an multistate-constraint Kalman filter algorithm, computing residuals and measurement Jacobian matrices, and form the residual and Jacobian matrix Hk for features that are included in the state vector, update the state vector and covariance matrix using the residual and Jacobian matrix Hk, and initialize features tracked in all images of the sliding window, updating state management and removing the oldest camera poses from the state vector.
If the features are to be processed using an multistate-constraint Kalman filter algorithm, then the method calculates a residual and Jacobian matrix for each feature to be processed, performs a Mahalanobis gating test, and forms a residual vector and a Jacobian matrix using all features that passed the gating test.
State management is updated by removing sliding-window iterative minimization features that are no longer tracked and changing the anchor pose for sliding-window iterative minimization features anchored at the oldest pose.
In addition, the method can also analyze the computations needed for the hybrid estimator and calculate the number of floating-point operations per update of the hybrid algorithm.
These and other features, aspects and advantages of the present invention will become better understood with regard to the following description, appended claims, and accompanying figures where:
The present invention solves the limitations described above by providing a new paradigm for the design of motion estimation algorithms. Specifically, the system uses a hybrid estimator that incorporates a one or more than one of algorithms with different computational characteristics. The system determines, in real time, the algorithm by which to process different measurements (e.g., different features). Because the preferred choice for each measurement will depend on the characteristics of sensor data, the system can employ statistical learning of these characteristics. In one embodiment, the system gathers statistical information to compute the expected cost of any strategy for allocating measurements to algorithms. To identify the preferred strategy, the system solves, in real time, an optimization problem with an objective function of the expected computation time.
The system uses visual and inertial measurements, because cameras and inertial measurement units (IMUs) are small, lightweight, and inexpensive sensors that can operate in almost any environment. The system estimates a moving platform's trajectory using the inertial measurements and the observations of naturally occurring point features. The system does not assume that a map of the area is available.
The present invention overcomes limitations of the prior art by providing a hybrid estimator system using visual and inertial for real-time pose tracking on devices with limited processing power. The system uses a novel procedure and method for estimating the state (e.g., the position, orientation, velocity) of a moving object (e.g., a vehicle, handheld device, aircraft) by combining data from an inertial sensor and one or more than on additional sensors that provide feature measurements. One key advantage of the method is the use of one or more than one different approaches for the processing of the feature measurements. The method used to process each of the available feature measurements is determined during system operation to improve the accuracy of the pose estimates and/or reduce the computational requirements of the estimator. In one embodiment, the method is executed for pose estimation using measurements from an inertial sensor and from a camera that observes “features” (e.g., points of interest) in the scene. The feature measurements can be processed by (i) explicitly including them as variables to be estimated in the estimator, or (ii) using their observations in order to derive constraints, directly used for updating the pose estimates of the moving object.
There is also provided a method to determine which of the methods should be used by the system for processing each of the feature measurements. This is used to reduce the computational requirements of the estimator and to adapt the system to the characteristics of the environment, the trajectory of the moving object, and the availability of processing resources.
Additionally, in one embodiment the system uses an adaptive approach to determining the number of features that should be extracted from raw sensor data in order to adjust to the availability of computational resources.
Finally, in order to compute an accurate description of the uncertainty of the estimates, the method employs suitably constructed linearized approximations of the nonlinear mathematical models that describe the motion and the sensor measurements. By choosing linearization points that preserve the system's observability properties, the system obtains high-precision, consistent estimates for the state of the moving object, at low computational cost.
Additionally, two other research papers by the inventors “Improving the Accuracy of EKF-Based Visual-Inertial Odometry” and “Vision-aided Inertial Navigation for Resource-constrained Systems” by Mingyang Li and Anastasios Mourikis that are hereby incorporated by reference in their entirety.
All dimensions specified in this disclosure are by way of example only and are not intended to be limiting. Further, the proportions shown in the Figures are not necessarily to scale. As will be understood by those with skill in the art with reference to this disclosure, the actual dimensions and proportions of any system, any device or part of a system or device disclosed in this disclosure will be determined by its intended use.
Systems and methods that implement the embodiments of the various features of the invention will now be described with reference to the drawings. The drawings and the associated descriptions are provided to illustrate embodiments of the invention and not to limit the scope of the invention. Reference in the specification to “one embodiment” or “an embodiment” is intended to indicate that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least an embodiment of the invention. The appearances of the phrase “in one embodiment” or “an embodiment” in various places in the specification are not necessarily all referring to the same embodiment.
Throughout the drawings, reference numbers are re-used to indicate correspondence between referenced elements. In addition, the first digit of each reference number indicates the figure where the element first appears.
As used in this disclosure, except where the context requires otherwise, the term “comprise” and variations of the term, such as “comprising”, “comprises” and “comprised” are not intended to exclude other additives, components, integers or steps.
In the following description, specific details are given to provide a thorough understanding of the embodiments. However, it will be understood by one of ordinary skill in the art that the embodiments may be practiced without these specific details. Well-known circuits, structures and techniques may not be shown in detail in order not to obscure the embodiments. For example, circuits may be shown in block diagrams in order not to obscure the embodiments in unnecessary detail.
Also, it is noted that the embodiments may be described as a process that is depicted as a flowchart, a flow diagram, a structure diagram, or a block diagram. Although a flowchart may describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations may be rearranged. A process is terminated when its operations are completed. A process may correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
Moreover, a storage may represent one or more devices for storing data, including read-only memory (ROM), random access memory (RAM), magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information. The term “machine readable medium” includes, but is not limited to portable or fixed storage devices, optical storage devices, wireless channels and various other mediums capable of storing, containing or carrying instruction(s) and/or data.
Furthermore, embodiments can be implemented by hardware, software, firmware, middleware, microcode, or a combination thereof. When implemented in software, firmware, middleware or microcode, the program code or code segments to perform the necessary tasks can be stored in a machine-readable medium such as a storage medium or other storage(s). One or more than one processor can perform the necessary tasks in series, concurrently, distributed or in parallel. A code segment can represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a class, or a combination of instructions, data structures, or program statements. A code segment can be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, or memory contents. Information, arguments, parameters, data, etc. can be passed, forwarded, or transmitted through a suitable means including memory sharing, message passing, token passing, network transmission, wireless transmission, etc.
The software identified above can also be constructed as a module. The module is a logically self-contained and discrete part of a larger computer program, for example, a subroutine or a co-routine. Modules are typically incorporated into the program through interfaces. The module interface expresses the elements that are provided and required by the module. The elements defined in the interface are detectable by other modules. The implementation contains the executable code that corresponds to the elements declared in the interface. Modules perform logically discrete functions. The module can interact with other modules of the system to achieve its purpose.
In the following description, certain terminology is used to describe certain features of one or more embodiments of the invention.
The term “visual-inertial odometry” refers to using inertial measurements from an IMU and the observations of naturally occurring point features from a visual sensor (e.g., a camera) to estimate the pose of a moving object in an environment without the use of a map.
Referring now to
In one embodiment, the system determines the choice of algorithm to process each individual feature depending on the distribution of the feature track lengths of all features. The distribution is not known in advance, because it depends on the environment, the camera motion, as well as the feature tracker used, and therefore the system learns it from an image sequence. Using this information, the optimal strategy for processing the feature measurements can be computed by solving a one-variable optimization problem. Results demonstrate that the hybrid algorithm outperforms each individual method (EKF-SLAM and sliding-window EKF) by a wide margin. In fact, the hybrid filter allows processing the sensor data at real-time speed on a mobile phone, something that is not possible with either of the individual algorithms. Using the present invention, all the available measurements are processed, and no localization information is lost.
Since real-time performance is necessary, any candidate algorithm for the system must have bounded computational complexity, irrespective of the duration of the trajectory. Within this class, practically all algorithms proposed to date employ either an EKF, or iterative minimization over a window of states. The latter approaches iteratively relinearize the measurement equations to better deal with their nonlinearity, which, however, incurs a high computational cost. By choosing linearization points that preserve the system's observability properties, an EKF-based system can attain better accuracy than iterative-minimization methods, at only a fraction of the computational cost. Specifically, the system can use a unique estimate of certain states, such as, for example, the first available estimate, in order to compute the linearization matrices (i.e., the Jacobians), for all measurement equations that involve each of the states. Optionally, the estimates of one or more than one state can be used to compute the linearization matrices (i.e., the Jacobians) can be modified by equal amounts, to reduce linearization inaccuracy while preserving the system's observability properties.
Within the class of EKF methods, there are two possible formulations of the estimator. Specifically, the system can employ the EKF-SLAM approach, where the state vector contains a current IMU state as well as feature positions. To keep the computations bounded, features that leave the field of view can be removed from the state vector, leaving only the currently observed ones. Other EKF algorithms maintain a sliding window of camera poses in the state vector, and use the feature observations to apply probabilistic constraints between these poses. In a preferred embodiment, the system uses a multistate-constraint Kalman filter (MSCKF) algorithm that uses the all the information provided by the feature measurements.
Both the MSCKF and EKF-SLAM use exactly the same information, and only differ in the way they organize the computations, and in linearization. If the measurement models were linear, both methods would yield exactly the same result, equal to the maximum-a-posteriori (MAP) estimate of the IMU pose. With respect to computational cost, however, the two methods differ substantially. For the EKF-SLAM algorithm the cost at each time-step is cubic in the number of features, since all features in the state vector are observed. In contrast, the MSCKF has computational cost that scales linearly in the number of features, but cubically in the length of the feature tracks. Therefore, if many features are tracked, each in a small number of frames, the MSCKF approach is preferable, but if few features are tracked over long image sequences, EKF-SLAM would result in lower computational cost. The system can integrate both EKF-SLAM and the MSCKF algorithms in a single, hybrid filter due to their complementary nature.
The system uses a modified MSCKF algorithm that attains increased accuracy and consistency by choosing linearization points that ensure the correct observability properties. The MSCKF does not include the feature positions in the state vector, but instead uses the feature measurements to directly impose constraints between the camera poses. The state vector of the MSCKF is:
xkT=[xI
where xIk is the current IMU state, and xCi, i=1 . . . m are the camera poses (positions and orientations) at the times the last m images were recorded. The IMU state is defined as:
xI=[
where q is the unit quaternion describing the rotation from the global frame to the IMU frame, p and v denote the IMU position and velocity, respectively, while bg and ba are the gyroscope and accelerometer biases.
The MSCKF uses the IMU measurements to propagate the current IMU state and the covariance matrix, Pk+1|k, while the feature measurements are used for EKF updates. Let's assume that the i-th feature, that has been observed in l images, has just been lost from tracking (e.g., it went out of the field of view). At this time, the MSCKF method uses all the measurements of the feature to carry out an EKF update. Let the observations of the feature be described by the nonlinear equations zij=h(xCj, fi)+nij, for j=1 . . . l, where fi is the feature position (described by the inverse-depth parameterization) and nij is the noise vector, modeled as zero-mean Gaussian with covariance matrix σ2I2. Using all the measurements we compute a feature position estimate {circumflex over (f)}i, and then compute the residuals {tilde over (z)}ij=zij−h({circumflex over (x)}C
{tilde over (z)}ij≅Hij{tilde over (x)}C
where {tilde over (z)}l and {tilde over (f)}i are the estimation errors of the j-th camera pose and i-th feature respectively, and the matrices Hij and Hf
{tilde over (z)}i≅Hi{tilde over (x)}+Hf
where {tilde over (z)}l and ni are block vectors with elements and {tilde over (z)}ij and nij, respectively, and Hi and Hf
{tilde over (z)}io=VT{tilde over (z)}
Once {tilde over (z)}lo and Hio available, we proceed to carry out a Mahalanobis gating test for the residual {tilde over (z)}io, by computing:
γ=({tilde over (z)}io)T(HioPk+1|k(Hio)T+σ2I)−1{tilde over (z)}io (Eq. 6)
and comparing it against a threshold computed based on the χ2 distribution with 2l−3 degrees of freedom. By stacking together the residuals of all features that pass this gating test, we obtain:
{tilde over (z)}o=Ho{tilde over (x)}+no (Eq. 7)
where {tilde over (z)}o, Ho, and no are block vectors/matrices, with block rows {tilde over (z)}lo, Hlo, and ni for i=1 . . . n, respectively. We can now use the above residual, which only involves the camera poses, for an EKF update. However, if a large number of features are processed at each time instant (the common situation), further computational savings can be achieved. Specifically, instead of using the above residual, we can equivalently compute the thin QR factorization of H
{tilde over (z)}r=QT{tilde over (z)}o=HT{tilde over (x)}+nr (Eq. 8)
where nr is the r×1 noise vector, with covariance matrix σ2Ir. Once the residual and the matrix Hr have been computed, we compute the state correction and update the covariance matrix via the standard EKF equations:
Δx=K{tilde over (z)}r (Eq. 9)
Pk+1|k+1=Pk+1|k−KSKT (Eq. 10)
S=HrPk+1|k(HT)T+σ2Ir (Eq. 11)
K=Pk+1|k(Hr)TS−1 (Eq. 12)
The above-described method of processing feature measurements is optimal, in the sense that no approximations are used, except for the EKF's linearization. This is true, however, only if the sliding window of states, m, contains at least as many states as the longest feature track. If it does not, then the measurements that occurred more than m timesteps in the past cannot be processed. Therefore, to use all the available feature information, the MSCKF must maintain a window of states long enough to include the longest feature tracks.
The computation time of the MSCKF is dominated by the following operations:
1) The Mahalanobis test for each feature, requiring
operations.
2) The computation of the residual and the Jacobian matrix in (Eq. 8), which, by exploiting the structure in Ho, can be implemented in approximately
operations.
3) The computation of the Kalman gain and the update of the covariance matrix, which require O(r3/6+r(15+6m)2) operations. Here 15+6m is the size of the state covariance matrix, and r is the number of rows of Hr. It can be shown that, in general, r (which equals the number of independent constraints for the camera poses) is given by:
r=2(l(1)+l(2)+l(3)−7 (Eq. 13)
where l(1), l(2), l(3) are the three longest feature tracks. Although the computational cost of the MSCKF is linear in the number of features, it is at least quadratic in the size of the sliding window, m. In fact, if the size of the sliding window is chosen to be equal to the longest feature tracks, then r is also O(m), and the overall complexity becomes cubic in m. This demonstrates a shortcoming of the MSCKF for preserving all measurement information: the complexity of the algorithm scales cubically as a function of the longest feature track length.
In real-world datasets the distribution of the feature-track lengths is non-uniform, with many features tracked for short durations, and very few stable features tracked over longer periods. For example,
Alternatively, the system can process feature measurements using the EKF by including the feature measurements in a state vector, and using the observations as in the standard visual-SLAM formulation. As discussed earlier, this approach has computational properties complementary to those of the MSCKF. While the MSCKF is better at processing many features tracked for a few frames, EKF-SLAM is faster when few features are tracked for many frames. The system combines both approaches in a single, “hybrid” estimator. Specifically, the system uses a filter whose state vector at time-step k contains the current IMU state, a sliding window of m camera poses, and sk features:
xk=[xl
This provides the system with the ability to determine whether a feature will be processed using the MSCKF approach, or whether it will be included in the state vector and processed as in EKF-SLAM. By analyzing in detail the computational requirements of each EKF update, it can be shown that when many features are present there is no gain to initializing the state vector if any feature is observed fewer than m times. Thus, in one embodiment of the system the strategy for using the features is as follows: if feature i's track is lost after fewer than m frames (i.e., li<m), then the feature is processed using the MSCKF equations, as described. If a feature is still actively being tracked after m images, it is initialized into the state vector, and used for the described SLAM computation. Therefore, the system can determine the size of the sliding window, m.
At a given time step, the EKF update is carried out using a residual vector constructed by stacking together the residual computed from a number of “MSCKF features” and from a number sk of observations of “SLAM features”:
In the above equation, {tilde over (z)}jm, for j=1 . . . sk are the residuals of the observations of the SLAM features from the latest camera state (state m), and Hjm, for j=1 . . . sk are the associated Jacobian matrices. Each of these residuals is a 2×1 vector, while each Hjm is a 2×(15+6m+3 sk) matrix (here 15+6m+3sk is the size of the state covariance matrix). The residual {tilde over (z)}k and the Jacobian Hk are used for updating the state and covariance matrix.
For initializing new features, the m measurements of a feature are used to triangulate it and to compute its initial covariance and the cross-correlation with other filter states. In our implementation, we use the inverse-depth feature parameterization, due to its superior linearity properties. The latest camera clone, xC
Algorithm 1 Hybrid MSCKF/SLAM Algorithm
Propagation: Propagate the state vector and covariance matrix using the IMU readings. Once camera measurements become available:
Augment the state vector with the latest camera pose.
For features to be processed in the MSCKF (feature tracks of length smaller than m), do the following:
-
- For each feature to be processed, calculate the residual and Jacobian matrix in (Eq. 5).
- Perform the Mahalanobis gating test in (Eq. 6).
- Using all features that passed the gating test, form the residual vector and the Jacobian matrix in (Eq. 8).
For features that are included in the state vector, compute the residuals and measurement Jacobian matrices, and form the residual {tilde over (z)}k and matrix Hk in (Eq. 15).
Update the state vector and covariance matrix, via (Eq. 9)-(Eq. 12), using the residual {tilde over (z)}k and Jacobian matrix Hk.
Initialize features tracked in all m images of the sliding window. State Management:
Remove SLAM features that are no longer tracked, and change the anchor pose for SLAM features anchored at the oldest pose.
Remove the oldest camera pose from the state vector. If no feature is currently tracked for more than mo poses (with mo<m−1), remove the oldest m−mo poses.
The system can select the size of the sliding window, m, to reduce the computational cost of the hybrid MSCKF/SLAM filter. The choice of m can have a profound effect on the time requirements of the algorithm. With a suitable choice, the hybrid method can significantly outperform each of its individual components.
By carefully analyzing the computations needed, the system can calculate the number of floating-point operations per update of the hybrid algorithm:
where the αi's are known constants, n is the number of features used in the MSCKF, r is defined in (13), and l.o.t. stands for lower-order terms. The full expression for the operation count consists of tens of individual terms, whose inclusion would merely complicate the description and not add much insight. The three terms shown above correspond to the computation of the MSCKF residual, the Cholesky factorization of the matrix S in (Eq. 12), and the covariance update equation, respectively. Note that (Eq. 16) also models the probability of failure for the Mahalanobis test.
It is interesting to examine the properties of (Eq. 16). First, note that since r represents the number of constraints provided by the MSCKF features for the poses in the sliding window, it is bounded above by 6m−7: the total number of unknowns in the sliding window is 6m, and the feature measurements cannot provide any information about the global pose or scale, which correspond to 7 degrees of freedom. If many features are available, the system will have r≈6m−7, and thus:
This approximate expression makes it possible to gain intuition about the behavior of the computational cost of the hybrid method: as the size of the sliding window, m, increases, more features will be processed by the MSCKF, and fewer features will be included in the state vector of the filter. Thus, as m increases, the term Σi=1nli3 will increase, but sk will decrease rapidly. These two opposing trends result in the performance curves shown in
In one embodiment, the system determines, in real time, the optimal value of m in order to minimize the runtime of the hybrid EKF estimator. Equation 16 provides the operation count of one filter update, so at first glance, it may appear that the system would need to minimize this equation with respect to m, at each time instant. However, that would be an ill-posed problem. To see why, consider the case where, at time step k, the sliding window has length m=20, and ten features exist that have been continuously tracked in 20 images. At this time, the system can either increase the size of the sliding window, or include the ten features in the state vector. Which is best depends on the future behavior of the feature tracks. If the features end up being tracked for a very large number of frames (>>20), then it would be preferable to include the features in the state vector. If, on the other hand, the features end up being tracked for only 21 frames, it would be preferable to increase m by one.
Clearly, it is impossible to obtain future information about any particular feature track. The system can however, collect statistical information about the properties of the feature tracks, and use this information to minimize the expected cost of the EKF updates. This approach is implemented in one embodiment of the system. Specifically, during the filter's operation, the system collects statistics to learn the probability mass function (pmf) of the feature track lengths, the probability of failure of the Mahalanobis gating test, as well as the pmf of the number of features tracked in the images. Using the learned pmfs, the system computes the average number of operations needed for each EKF update,
The value of m that yields the minimum
It is worth noting that in modern computers the use of flop counts to model computational cost is not always suitable, as performance is affected by several factors including vectorization, cache access patterns, data locality, etc. However, it has been experimentally verified that in the algorithms considered here, and for our implementation, the computed flop counts closely follow the observed runtimes. Specifically,
To test the system simulation data was generated that closely matched real-world datasets, to have as realistic a test environment as possible. A first simulation environment is based on a dataset that consists of a 29.6-km long trajectory through canyons, forested areas, and a town. A ground truth trajectory (position, velocity, orientation) was generated that matches the vehicle's actual trajectory, as computed by a high-precision NS system. Using this trajectory, IMU measurements corrupted with noise and bias characteristics identical to those of the Xsens MTi-G sensor used in the dataset were subsequently generated. Moreover, monocular feature tracks with statistical characteristics similar to those of the features extracted in the actual dataset were also generated. The number of features observed in each image and the feature track distribution change in each part of the trajectory, as in the actual dataset (see
Referring now to
In the second subplot
The timing results presented in
Referring now to
Referring now to
In this dataset, to compensate for the low frame rate of the images, we use lower feature-detection thresholds, which leads to a larger number of features. Specifically, 540 features are tracked in each image on average, and the average track length is 5.06 frames. This substantially increases the overall computational requirements of all algorithms. When run on the mobile phone's processor, the average time per update is 139 msec for the MSCKF, 774 msec for EKF-SLAM, and 77 msec for the hybrid filter with optimally selected thresholds. The trajectory estimates 500 for each of the three methods are plotted on a map of the area where the experiment took place. We observe that the accuracy of the MSKCF and the hybrid filter are similar, and substantially better than that of the EKF-SLAM.
Referring now to
Referring now to
Referring now to
What has been described is a new hybrid estimator system using visual and inertial sensors for real-time pose tracking, overcoming limitations and disadvantages inherent in the related art.
Although the present invention has been described with a degree of particularity, it is understood that the present disclosure has been made by way of example and that other versions are possible. As various changes could be made in the above description without departing from the scope of the invention, it is intended that all matter contained in the above description or shown in the accompanying drawings shall be illustrative and not used in a limiting sense. The spirit and scope of the appended claims should not be limited to the description of the preferred versions contained in this disclosure.
All features disclosed in the specification, including the claims, abstracts, and drawings, and all the steps in any method or process disclosed, may be combined in any combination, except combinations where at least some of such features and/or steps are mutually exclusive. Each feature disclosed in the specification, including the claims, abstract, and drawings, can be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise. Thus, unless expressly stated otherwise, each feature disclosed is one example only of a generic series of equivalent or similar features.
Any element in a claim that does not explicitly state “means” for performing a specified function or “step” for performing a specified function should not be interpreted as a “means” or “step” clause as specified in 35 U.S.C. §112.
Claims
1. A hybrid estimator system using sensors for real-time pose tracking of moving platforms, the system comprising:
- a) a device comprising at least one processor, a memory, a storage and communications through a protocol; and
- b) one or more than one module, communicatively coupled to each other, comprising code executable on the processor for 1) a hybrid estimator, comprising a plurality of algorithms for processing measurements; and 2) selecting the constituent algorithms of the hybrid estimator used to process each measurement.
2. The system of claim 1, where a module comprising non-statutory instructions for learning the statistical characteristics of the measurements is used.
3. The system of claim 1, where the constituent algorithm selection is based on the learning module.
4. The system of claim 1, where the system gathers information to compute the expected cost of strategies for processing measurements by different algorithms to reduce the computational requirements.
5. The system of claim 1, where the system solves, in real time, an optimization problem with an objective function representing the expected computation time to identify the preferred strategy.
6. The system of claim 1, where the hybrid estimator estimates a moving platform's trajectory using inertial measurements and the observations of features by one or more than one sensors.
7. The system of claim 1, where the hybrid estimator estimates a moving platform's trajectory using inertial measurements and the observations of features by a visual sensor.
8. The system of claim 1, where the system comprises non-transitory instructions to process each of the available feature measurements.
9. The system of claim 8, where the feature measurements can be processed by including a description of the features as variables to be estimated in the hybrid estimator.
10. The system of claim 8, where the feature measurements can be processed by using obtained observations in order to derive constraints, directly used for updating the pose estimates of the moving object.
11. The system of claim 1, where the system comprises non-transitory instructions to improve the accuracy of the pose estimates, reduce the computational requirements of the estimator or both improve the accuracy of the pose estimates and reduce the computational requirements of the estimator during system operation to reduce the processing requirements.
12. The system of claim 1, where the hybrid estimator comprises non-transitory instructions to determine which of the plurality of methods is to be used for processing each of the feature measurements to reduce the computational requirements of the estimator and to adapt the system to the characteristics of the environment, the trajectory of the moving object, and the availability of processing resources.
13. The system of claim 1, where the hybrid estimator further comprises instructions for a module that determines the number of features that should be extracted from raw sensor data in order to adjust to the availability of computational resources.
14. The system of claim 1, where the hybrid estimator comprises non-transitory instructions to construct linearized approximations of the nonlinear mathematical models that describe the motion and the sensor measurements in order to compute a description of the uncertainty of the estimates.
15. The system of claim 14, where the linearization points are selected to preserve the system's observability properties.
16. The system of claim 15, where a unique estimate of certain states is used in order to compute the linearization matrices for all measurement equations that involve each of the states.
17. The system of claim 15 where the estimates of one or more than one state is used to compute the linearization matrices and can be modified by equal amounts, to reduce linearization inaccuracy while preserving the system's observability properties.
18. The system of claim 1, where the hybrid estimator comprises non-transitory instructions to implement a hybrid extended Kalman filter.
19. The system of claim 1, where the hybrid extended Kalman filter comprises an extended-Kalman filter algorithm that includes feature states in the state vector, and a sliding-window extended-Kalman-filter algorithm that includes states of the mobile platform in the state vector.
20. The system of claim 19, where the size of the sliding window is selected to reduce the computational cost of the hybrid extended Kalman filter.
21. The system of claim 1, where the hybrid estimator module determines the choice of algorithm to process each individual feature depending on the distribution of the feature track lengths of features.
22. The system of claim 21, where the hybrid estimator module determines the optimal strategy for processing the feature measurements by solving a one-variable optimization problem using the information above.
23. The system of claim 1, where the hybrid estimator module processes all available measurements without loss of localization information.
24. The system of claim 1, where the plurality of algorithms must have bounded computational complexity, irrespective of the duration of the trajectory.
25. The system of claim 24, where the plurality of algorithms employ an extended Kalman filter.
26. The system of 1, where the hybrid estimator includes an extended Kalman filter-sliding-window iterative minimization algorithm, where the state vector contains a current IMU state as well as representations of the feature positions.
27. The system of claim 26, where the features that leave the field of view are removed from the state vector leaving only the currently observed ones, to keep the computations bounded.
28. The system of claim 26, where the hybrid estimator comprises other extended-Kalman-filter algorithms that maintain a sliding window of camera poses in the state vector, and use the feature observations to apply probabilistic constraints between these poses.
29. The system of claim 26, where the hybrid estimator comprises a multistate-constraint Kalman filter.
30. The system of claim 26, where the algorithm used to process feature measurements is selected to have the lowest computational cost.
31. The system of claim 26, where the hybrid filter is a combination of both the extended Kalman filter-sliding-window iterative minimization and the multistate-constraint Kalman filter algorithms.
32. The system of claim 26, where the hybrid filter is a filter whose state vector contains the current IMU state, m camera poses, and sk features, and determines whether a feature will be processed using the multistate-constraint Kalman filter algorithm, or whether it will be included in the state vector and processed using the extended Kalman filter-sliding-window iterative minimization algorithm.
33. A method for using a hybrid estimator using visual and inertial sensors for real-time pose tracking on devices with limited processing power, the method comprising the steps of:
- a) determining a method to be used by the system for processing each of the feature measurements;
- b) adapting the system to the characteristics of the environment, the trajectory of the moving object, and the availability of processing resources;
- c) determining the number of features to be extracted from raw sensor data in order to adjust to the availability of computational resources; and
- d) constructing linearized approximations of the nonlinear mathematical models that describe the motion and the sensor measurements.
34. A method for making a hybrid estimator using visual and inertial sensors for real-time pose tracking on devices with limited processing power, the method comprising the steps of:
- a) providing a device comprising at least one processor, a memory, a storage and communications through a protocol; and
- b) providing one or more than one software module, communicatively coupled to each other, comprising code executable on the processor for: 1) a hybrid estimator; 2) real-time algorithm selection to process different measurements; 3) statistical learning for these characteristics to compute the expected device computing cost of any strategy for allocating measurements to algorithms; and 4) algorithm selection based on the statistical learning module.
35. The method of claim 34 further comprising the steps of:
- a) propagating the state vector and covariance matrix using the IMU readings;
- b) determining when camera measurements and features are available;
- c) augmenting the state vector with the latest camera pose;
- d) determine if the features are to be processed using an multistate-constraint Kalman filter algorithm; where if the features are to be processed the perform the following steps: 1) calculating a residual and Jacobian matrix for each feature to be processed; 2) performing a Mahalanobis gating test; and 3) forming a residual vector and a Jacobian matrix using all features that passed the gating test;
- e) computing residuals and measurement Jacobian matrices, and form the residual {tilde over (z)}k and Jacobian matrix Hk for features that are included in the state vector;
- f) updating the state vector and covariance matrix using the residual {tilde over (z)}k and Jacobian matrix Hk; and
- g) initialize features tracked in all images of the sliding window;
- h) updating state management by performing the following steps: 1) removing sliding-window iterative minimization features that are no longer tracked; and 2) changing the anchor pose for sliding-window iterative minimization features anchored at the oldest pose; and
- i) removing oldest camera poses from the state vector.
36. The method of claim 35 further comprising the steps of:
- a) analyzing the computations needed for the hybrid estimator; and
- b) calculating the number of floating-point operations per update of the hybrid algorithm.
Type: Application
Filed: May 15, 2014
Publication Date: Nov 20, 2014
Applicant: The Regents of the University of California (Oakland, CA)
Inventors: Mingyang Li (Riverside, CA), Anastasios Mouriskis (Riverside, CA)
Application Number: 14/279,123
International Classification: G06K 9/62 (20060101); G06K 9/00 (20060101);