RESUMO
Displacement is critical when it comes to the evaluation of civil structures. Large displacement can be dangerous. There are many methods that can be used to monitor structural displacements, but every method has its benefits and limitations. Lucas-Kanade (LK) optical flow is recognized as a superior computer vision displacement tracking method, but it only applies to small displacement monitoring. An upgraded LK optical flow method is developed in this study and used to detect large displacement motions. One motion controlled by a multiple purpose testing system (MTS) and a free-falling experiment were designed to verify the developed method. The results provided by the upgraded LK optical flow method showed 97 percent accuracy when compared with the movement of the MTS piston. In order to capture the free-falling large displacement, the pyramid and warp optical flow methods are included in the upgraded LK optical flow method and compared with the results of template matching. The warping algorithm with the second derivative Sobel operator provides accurate displacements with 96% average accuracy.
RESUMO
During the radiotherapy treatment of patients with lung cancer, the radiation delivered to healthy tissue around the tumor needs to be minimized, which is difficult because of respiratory motion and the latency of linear accelerator (LINAC) systems. In the proposed study, we first use the Lucas-Kanade pyramidal optical flow algorithm to perform deformable image registration (DIR) of chest computed tomography (CT) scan images of four patients with lung cancer. We then track three internal points close to the lung tumor based on the previously computed deformation field and predict their position with a recurrent neural network (RNN) trained using real-time recurrent learning (RTRL) and gradient clipping. The breathing data is quite regular, sampled at approximately 2.5â¯Hz, and includes artificially added drift in the spine direction. The amplitude of the motion of the tracked points ranged from 12.0â¯mm to 22.7â¯mm. Finally, we propose a simple method for recovering and predicting three-dimensional (3D) tumor images from the tracked points and the initial tumor image, based on a linear correspondence model and the Nadaraya-Watson non-linear regression. The root-mean-square (RMS) error, maximum error and jitter corresponding to the RNN prediction on the test set were smaller than the same performance measures obtained with linear prediction and least mean squares (LMS). In particular, the maximum prediction error associated with the RNN, equal to 1.51â¯mm, is respectively 16.1% and 5.0% lower than the error given by a linear predictor and LMS. The average prediction time per time step with RTRL is equal to 119â¯ms, which is less than the 400â¯ms marker position sampling time. The tumor position in the predicted images appears visually correct, which is confirmed by the high mean cross-correlation between the original and predicted images, equal to 0.955. The standard deviation of the Gaussian kernel and the number of layers in the optical flow algorithm were the parameters having the most significant impact on registration performance. Their optimization led respectively to a 31.3% and 36.2% decrease in the registration error. Using only a single layer proved to be detrimental to the registration quality because tissue motion in the lower part of the lung has a high amplitude relative to the resolution of the CT scan images. The random initialization of the hidden units and the number of these hidden units were found to be the most important factors affecting the performance of the RNN. Increasing the number of hidden units from 15 to 250 led to a 56.3% decrease in the prediction error on the cross-validation data. Similarly, optimizing the standard deviation of the initial Gaussian distribution of the synaptic weights σinitRNN led to a 28.4% decrease in the prediction error on the cross-validation data, with the error minimized for σinitRNN=0.02 with the four patients.
Assuntos
Neoplasias Pulmonares , Redes Neurais de Computação , Algoritmos , Humanos , Pulmão , Neoplasias Pulmonares/diagnóstico por imagem , Neoplasias Pulmonares/radioterapia , Movimento (Física)RESUMO
In the detection of moving objects from vision sources one usually assumes that the scene has been captured by stationary cameras. In case of backing up a vehicle, however, the camera mounted on the vehicle moves according to the vehicle's movement, resulting in ego-motions on the background. This results in mixed motion in the scene, and makes it difficult to distinguish between the target objects and background motions. Without further treatments on the mixed motion, traditional fixed-viewpoint object detection methods will lead to many false-positive detection results. In this paper, we suggest a procedure to be used with the traditional moving object detection methods relaxing the stationary cameras restriction, by introducing additional steps before and after the detection. We also decribe the implementation as a FPGA platform along with the algorithm. The target application of this suggestion is use with a road vehicle's rear-view camera systems.