The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] RED(1933hit)

261-280hit(1933hit)

  • Meeting Tight Security for Multisignatures in the Plain Public Key Model

    Naoto YANAI  

     
    PAPER

      Vol:
    E101-A No:9
      Page(s):
    1484-1493

    Multisignatures are digital signatures for a group consisting of multiple signers where each signer signs common documents via interaction with its co-signers and the data size of the resultant signatures for the group is independent of the number of signers. In this work, we propose a multisignature scheme, whose security can be tightly reduced to the CDH problem in bilinear groups, in the strongest security model where nothing more is required than that each signer has a public key, i.e., the plain public key model. Loosely speaking, our main idea for a tight reduction is to utilize a three-round interaction in a full-domain hash construction. Namely, we surmise that a full-domain hash construction with three-round interaction will become tightly secure under the CDH problem. In addition, we show that the existing scheme by Zhou et al. (ISC 2011) can be improved to a construction with a tight security reduction as an application of our proof framework.

  • An Emotion Similarity Based Severity Prediction of Software Bugs: A Case Study of Open Source Projects

    Geunseok YANG  Tao ZHANG  Byungjeong LEE  

     
    PAPER-Software Engineering

      Pubricized:
    2018/05/02
      Vol:
    E101-D No:8
      Page(s):
    2015-2026

    Many software development teams usually tend to focus on maintenance activities in general. Recently, many studies on bug severity prediction have been proposed to help a bug reporter determine severity. But they do not consider the reporter's expression of emotion appearing in the bug report when they predict the bug severity level. In this paper, we propose a novel approach to severity prediction for reported bugs by using emotion similarity. First, we do not only compute an emotion-word probability vector by using smoothed unigram model (UM), but we also use the new bug report to find similar-emotion bug reports with Kullback-Leibler divergence (KL-divergence). Then, we introduce a new algorithm, Emotion Similarity (ES)-Multinomial, which modifies the original Naïve Bayes Multinomial algorithm. We train the model with emotion bug reports by using ES-Multinomial. Finally, we can predict the bug severity level in the new bug report. To compare the performance in bug severity prediction, we select related studies including Emotion Words-based Dictionary (EWD)-Multinomial, Naïve Bayes Multinomial, and another study as baseline approaches in open source projects (e.g., Eclipse, GNU, JBoss, Mozilla, and WireShark). The results show that our approach outperforms the baselines, and can reflect reporters' emotional expressions during the bug reporting.

  • Predicting Taxi Destination by Regularized RNN with SDZ

    Lei ZHANG  Guoxing ZHANG  Zhizheng LIANG  Qingfu FAN  Yadong LI  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2018/05/02
      Vol:
    E101-D No:8
      Page(s):
    2141-2144

    The traditional Markov prediction methods of the taxi destination rely only on the previous 2 to 3 GPS points. They negelect long-term dependencies within a taxi trajectory. We adopt a Recurrent Neural Network (RNN) to explore the long-term dependencies to predict the taxi destination as the multiple hidden layers of RNN can store these dependencies. However, the hidden layers of RNN are very sensitive to small perturbations to reduce the prediction accuracy when the amount of taxi trajectories is increasing. In order to improve the prediction accuracy of taxi destination and reduce the training time, we embed suprisal-driven zoneout (SDZ) to RNN, hence a taxi destination prediction method by regularized RNN with SDZ (TDPRS). SDZ can not only improve the robustness of TDPRS, but also reduce the training time by adopting partial update of parameters instead of a full update. Experiments with a Porto taxi trajectory data show that TDPRS improves the prediction accuracy by 12% compared to RNN prediction method in literature[4]. At the same time, the prediction time is reduced by 7%.

  • Decentralized Event-Triggered Control of Composite Systems Using M-Matrices

    Kenichi FUKUDA  Toshimitsu USHIO  

     
    PAPER-Systems and Control

      Vol:
    E101-A No:8
      Page(s):
    1156-1161

    A composite system consists of many subsystems, which have interconnections with other subsystems. For such a system, in general, we utilize decentralized control, where each subsystem is controlled by a local controller. On the other hand, event-triggered control is one of useful approaches to reduce the amount of communications between a controller and a plant. In the event-triggered control, an event triggering mechanism (ETM) monitors the information of the plant, and determines the time to transmit the data. In this paper, we propose a design of ETMs for the decentralized event-triggered control of nonlinear composite systems using an M-matrix. We consider the composite system where there is an ETM for each subsystem, and ETMs monitor local states of the corresponding subsystems. Each ETM is designed so that the composite system is stabilized. Moreover, we deal with the case of linear systems. Finally, we perform simulation to show that the proposed triggering rules are useful for decentralized control.

  • Weighted Subtask Controller for Redundant Manipulator Using Auxiliary Positive Function

    Youngjun YOO  Daesung JUNG  Sangchul WON  

     
    PAPER-Systems and Control

      Vol:
    E101-A No:8
      Page(s):
    1162-1171

    We propose a weighted subtask controller and sufficient conditions for boundedness of the controller both velocity and acceleration domain. Prior to designing the subtask controller, a task controller is designed for global asymptotic stability of task space error and subtask error. Although the subtask error converges to zero by the task controller, the boundedness of the subtask controller is also important, therefore its boundedness conditions are presented. The weighted pseudo inverse is introduced to relax the constraints of the null-space of Jacobian. Using the pseudo inverse, we design subtask controller and propose sufficient conditions for boundedness of the auxiliary signal to show the existence of the inverse kinematic solution. The results of experiments using 7-DOF WAM show the effectiveness of the proposed controller.

  • Fast Time-Aware Sparse Trajectories Prediction with Tensor Factorization

    Lei ZHANG  Qingfu FAN  Guoxing ZHANG  Zhizheng LIANG  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2018/04/13
      Vol:
    E101-D No:7
      Page(s):
    1959-1962

    Existing trajectory prediction methods suffer from the “data sparsity” and neglect “time awareness”, which leads to low accuracy. Aiming to the problem, we propose a fast time-aware sparse trajectories prediction with tensor factorization method (TSTP-TF). Firstly, we do trajectory synthesis based on trajectory entropy and put synthesized trajectories into the original trajectory space. It resolves the sparse problem of trajectory data and makes the new trajectory space more reliable. Then, we introduce multidimensional tensor modeling into Markov model to add the time dimension. Tensor factorization is adopted to infer the missing regions transition probabilities to further solve the problem of data sparsity. Due to the scale of the tensor, we design a divide and conquer tensor factorization model to reduce memory consumption and speed up decomposition. Experiments with real dataset show that TSTP-TF improves prediction accuracy generally by as much as 9% and 2% compared to the Baseline algorithm and ESTP-MF algorithm, respectively.

  • Identifying Core Objects for Trace Summarization by Analyzing Reference Relations and Dynamic Properties

    Kunihiro NODA  Takashi KOBAYASHI  Noritoshi ATSUMI  

     
    PAPER

      Pubricized:
    2018/04/20
      Vol:
    E101-D No:7
      Page(s):
    1751-1765

    Behaviors of an object-oriented system can be visualized as reverse-engineered sequence diagrams from execution traces. This approach is a valuable tool for program comprehension tasks. However, owing to the massiveness of information contained in an execution trace, a reverse-engineered sequence diagram is often afflicted by a scalability issue. To address this issue, many trace summarization techniques have been proposed. Most of the previous techniques focused on reducing the vertical size of the diagram. To cope with the scalability issue, decreasing the horizontal size of the diagram is also very important. Nonetheless, few studies have addressed this point; thus, there is a lot of needs for further development of horizontal summarization techniques. We present in this paper a method for identifying core objects for trace summarization by analyzing reference relations and dynamic properties. Visualizing only interactions related to core objects, we can obtain a horizontally compactified reverse-engineered sequence diagram that contains system's key behaviors. To identify core objects, first, we detect and eliminate temporary objects that are trivial for a system by analyzing reference relations and lifetimes of objects. Then, estimating the importance of each non-trivial object based on their dynamic properties, we identify highly important ones (i.e., core objects). We implemented our technique in our tool and evaluated it by using traces from various open-source software systems. The results showed that our technique was much more effective in terms of the horizontal reduction of a reverse-engineered sequence diagram, compared with the state-of-the-art trace summarization technique. The horizontal compression ratio of our technique was 134.6 on average, whereas that of the state-of-the-art technique was 11.5. The runtime overhead imposed by our technique was 167.6% on average. This overhead is relatively small compared with recent scalable dynamic analysis techniques, which shows the practicality of our technique. Overall, our technique can achieve a significant reduction of the horizontal size of a reverse-engineered sequence diagram with a small overhead and is expected to be a valuable tool for program comprehension.

  • Using Scattered X-Rays to Improve the Estimation Accuracy of Attenuation Coefficients: A Fundamental Analysis

    Naohiro TODA  Tetsuya NAKAGAMI  Yoichi YAMAZAKI  Hiroki YOSHIOKA  Shuji KOYAMA  

     
    PAPER-Measurement Technology

      Vol:
    E101-A No:7
      Page(s):
    1101-1114

    In X-ray computed tomography, scattered X-rays are generally removed by using a post-patient collimator located in front of the detector. In this paper, we show that the scattered X-rays have the potential to improve the estimation accuracy of the attenuation coefficient in computed tomography. In order to clarify the problem, we simplified the geometry of the computed tomography into a thin cylinder composed of a homogeneous material so that only one attenuation coefficient needs to be estimated. We then conducted a Monte Carlo numerical experiment on improving the estimation accuracy of attenuation coefficient by measuring the scattered X-rays with several dedicated toroidal detectors around the cylinder in addition to the primary X-rays. We further present a theoretical analysis to explain the experimental results. We employed a model that uses a T-junction (i.e., T-junction model) to divide the photon transport into primary and scattered components. This division is processed with respect to the attenuation coefficient. Using several T-junction models connected in series, we modeled the case of several scatter detectors. The estimation accuracy was evaluated according to the variance of the efficient estimator, i.e., the Cramer-Rao lower bound. We confirmed that the variance decreases as the number of scatter detectors increases, which implies that using scattered X-rays can reduce the irradiation dose for patients.

  • A Relaxed Bit-Write-Reducing and Error-Correcting Code for Non-Volatile Memories

    Tatsuro KOJO  Masashi TAWADA  Masao YANAGISAWA  Nozomu TOGAWA  

     
    LETTER

      Vol:
    E101-A No:7
      Page(s):
    1045-1052

    Non-volatile memories are a promising alternative to memory design but data stored in them still may be destructed due to crosstalk and radiation. The data stored in them can be restored by using error-correcting codes but they require extra bits to correct bit errors. One of the largest problems in non-volatile memories is that they consume ten to hundred times more energy than normal memories in bit-writing. It is quite necessary to reduce writing bits. Recently, a REC code (bit-write-reducing and error-correcting code) is proposed for non-volatile memories which can reduce writing bits and has a capability of error correction. The REC code is generated from a linear systematic error-correcting code but it must include the codeword of all 1's, i.e., 11…1. The codeword bit length must be longer in order to satisfy this condition. In this letter, we propose a method to generate a relaxed REC code which is generated from a relaxed error-correcting code, which does not necessarily include the codeword of all 1's and thus its codeword bit length can be shorter. We prove that the maximum flipping bits of the relaxed REC code is still limited theoretically. Experimental results show that the relaxed REC code efficiently reduce the number of writing bits.

  • User Clustering for Wireless Powered Communication Networks with Non-Orthogonal Multiple Access

    Tianyi XIE  Bin LYU  Zhen YANG  Feng TIAN  

     
    LETTER-Mobile Information Network and Personal Communications

      Vol:
    E101-A No:7
      Page(s):
    1146-1150

    In this letter, we study a wireless powered communication network (WPCN) with non-orthogonal multiple access (NOMA), where the user clustering scheme that groups each two users in a cluster is adopted to guarantee the system performance. The two users in a cluster transmit data simultaneously via NOMA, while time division multiple access (TDMA) is used among clusters. We aim to maximize the system throughput by finding the optimal cluster permutation and the optimal time allocation, which can be obtained by solving the optimization problems corresponding to all cluster permutations. The closed-form solution of each optimization problem is obtained by exploiting its constraint structures. However, the complexity of this exhaustive method is quite high, we further propose a sub-optimal clustering scheme with low complexity. The simulation results demonstrate the superiority of the proposed scheme.

  • HOAH: A Hybrid TCP Throughput Prediction with Autoregressive Model and Hidden Markov Model for Mobile Networks

    Bo WEI  Kenji KANAI  Wataru KAWAKAMI  Jiro KATTO  

     
    PAPER

      Pubricized:
    2018/01/22
      Vol:
    E101-B No:7
      Page(s):
    1612-1624

    Throughput prediction is one of the promising techniques to improve the quality of service (QoS) and quality of experience (QoE) of mobile applications. To address the problem of predicting future throughput distribution accurately during the whole session, which can exhibit large throughput fluctuations in different scenarios (especially scenarios of moving user), we propose a history-based throughput prediction method that utilizes time series analysis and machine learning techniques for mobile network communication. This method is called the Hybrid Prediction with the Autoregressive Model and Hidden Markov Model (HOAH). Different from existing methods, HOAH uses Support Vector Machine (SVM) to classify the throughput transition into two classes, and predicts the transmission control protocol (TCP) throughput by switching between the Autoregressive Model (AR Model) and the Gaussian Mixture Model-Hidden Markov Model (GMM-HMM). We conduct field experiments to evaluate the proposed method in seven different scenarios. The results show that HOAH can predict future throughput effectively and decreases the prediction error by a maximum of 55.95% compared with other methods.

  • Wavelength-Switchable Mid-Infrared Narrowband Thermal Emitters Based on Quantum Wells and Photonic Crystals Open Access

    Takuya INOUE  Menaka DE ZOYSA  Takashi ASANO  Susumu NODA  

     
    INVITED PAPER

      Vol:
    E101-C No:7
      Page(s):
    545-552

    Development of narrowband thermal emitters whose emission wavelengths are dynamically tunable is highly desired for various applications including the sensing of gases and chemical compounds. In this paper, we review our recent demonstration of wavelength-switchable mid-infrared thermal emitters based on multiple quantum wells (MQWs) and photonic crystals (PCs). Through the control of absorptivity by using intersubband transitions in MQWs and optical resonances in PC slabs, we demonstrate novel control of thermal emission, including realization of high-Q (Q>100) thermal emission, dynamic control of thermal emission (∼MHz), and electrical wavelength switching of thermal emission from a single device.

  • On Maximizing the Lifetime of Wireless Sensor Networks in 3D Vegetation-Covered Fields

    Wenjie YU  Xunbo LI  Zhi ZENG  Xiang LI  Jian LIU  

     
    LETTER-Fundamentals of Information Systems

      Pubricized:
    2018/03/01
      Vol:
    E101-D No:6
      Page(s):
    1677-1681

    In this paper, the problem of lifetime extension of wireless sensor networks (WSNs) with redundant sensor nodes deployed in 3D vegetation-covered fields is modeled, which includes building communication models, network model and energy model. Generally, such a problem cannot be solved by a conventional method directly. Here we propose an Artificial Bee Colony (ABC) based optimal grouping algorithm (ABC-OG) to solve it. The main contribution of the algorithm is to find the optimal number of feasible subsets (FSs) of WSN and assign them to work in rotation. It is verified that reasonably grouping sensors into FSs can average the network energy consumption and prolong the lifetime of the network. In order to further verify the effectiveness of ABC-OG, two other algorithms are included for comparison. The experimental results show that the proposed ABC-OG algorithm provides better optimization performance.

  • Block-Adaptive Selection of Recursive and Non-Recursive Type Intra Prediction Modes for Image Coding

    Yuta ISHIDA  Yusuke KAMEDA  Tomokazu ISHIKAWA  Ichiro MATSUDA  Susumu ITOH  

     
    LETTER-Image

      Vol:
    E101-A No:6
      Page(s):
    992-996

    This paper proposes a lossy image coding method for still images. In this method, recursive and non-recursive type intra prediction techniques are adaptively selected on a block-by-block basis. The recursive-type intra prediction technique applies a linear predictor to each pel within a prediction block in a recursive manner, and thus typically produces smooth image values. In this paper, the non-recursive type intra prediction technique is extended from the angular prediction technique adopted in the H.265/HEVC video coding standard to enable interpolative prediction to the maximum possible extent. The experimental results indicate that the proposed method achieves better coding performance than the conventional method that only uses the recursive-type prediction technique.

  • Online Linear Optimization with the Log-Determinant Regularizer

    Ken-ichiro MORIDOMI  Kohei HATANO  Eiji TAKIMOTO  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2018/03/01
      Vol:
    E101-D No:6
      Page(s):
    1511-1520

    We consider online linear optimization over symmetric positive semi-definite matrices, which has various applications including the online collaborative filtering. The problem is formulated as a repeated game between the algorithm and the adversary, where in each round t the algorithm and the adversary choose matrices Xt and Lt, respectively, and then the algorithm suffers a loss given by the Frobenius inner product of Xt and Lt. The goal of the algorithm is to minimize the cumulative loss. We can employ a standard framework called Follow the Regularized Leader (FTRL) for designing algorithms, where we need to choose an appropriate regularization function to obtain a good performance guarantee. We show that the log-determinant regularization works better than other popular regularization functions in the case where the loss matrices Lt are all sparse. Using this property, we show that our algorithm achieves an optimal performance guarantee for the online collaborative filtering. The technical contribution of the paper is to develop a new technique of deriving performance bounds by exploiting the property of strong convexity of the log-determinant with respect to the loss matrices, while in the previous analysis the strong convexity is defined with respect to a norm. Intuitively, skipping the norm analysis results in the improved bound. Moreover, we apply our method to online linear optimization over vectors and show that the FTRL with the Burg entropy regularizer, which is the analogue of the log-determinant regularizer in the vector case, works well.

  • Optimization of Body Biasing for Variable Pipelined Coarse-Grained Reconfigurable Architectures

    Takuya KOJIMA  Naoki ANDO  Hayate OKUHARA  Ng. Anh Vu DOAN  Hideharu AMANO  

     
    PAPER-Computer System

      Pubricized:
    2018/03/09
      Vol:
    E101-D No:6
      Page(s):
    1532-1540

    Variable Pipeline Cool Mega Array (VPCMA) is a low power Coarse Grained Reconfigurable Architecture (CGRA) based on the concept of CMA (Cool Mega Array). It provides a pipeline structure in the PE array that can be configured so as to fit target algorithms and required performance. Also, VPCMA uses the Silicon On Thin Buried oxide (SOTB) technology, a type of Fully Depleted Silicon On Insulator (FDSOI), so it is possible to control its body bias voltage to provide a balance between performance and leakage power. In this paper, we study the optimization of the VPCMA body bias while considering simultaneously its variable pipeline structure. Through evaluations, we can observe that it is possible to achieve an average reduction of energy consumption, for the studied applications, of 17.75% and 10.49% when compared to respectively the zero bias (without body bias control) and the uniform (control of the whole PE array) cases, while respecting performance constraints. Besides, it is observed that, with appropriate body bias control, it is possible to extend the possible performance, hence enabling broader trade-off analyzes between consumption and performance. Considering the dynamic power as well as the static power, more appropriate pipeline structure and body bias voltage can be obtained. In addition, when the control of VDD is integrated, higher performance can be achieved with a steady increase of the power. These promising results show that applying an adequate optimization technique for the body bias control while simultaneously considering pipeline structures can not only enable further power reduction than previous methods, but also allow more trade-off analysis possibilities.

  • A New Read Scheme for High-Density Emerging Memories

    Takashi OHSAWA  

     
    PAPER-Electronic Circuits

      Vol:
    E101-C No:6
      Page(s):
    423-429

    Several new memories are being studied as candidates of future DRAM that seems difficult to be scaled. However, the read signal in these new memories needs to be amplified in a single-end manner with reference signal supplied if they are aimed for being applied to the high-density main memory. This scheme, which is fortunately not necessary in DRAM's 1/2Vdd pre-charge sense amp, can become a serious bottleneck in the new memory development, because the device electrical parameters in these new memory cells are prone to large cell-to-cell variations without exception. Furthermore, the extent to which the parameter fluctuates in data “1” is generally not the same as in data “0”. In these situations, a new sensing scheme is proposed that can minimize the sensing error rate for high-density single-end emerging memories like STT-MRAM, ReRAM and PCRAM. The scheme is based on averaging multiple dummy cell pairs that are written “1” and “0” in a weighted manner according to the fluctuation unbalance between “1” and “0”. A detailed analysis shows that this scheme is effective in designing 128Mb 1T1MTJ STT-MRAM with the results that the required TMR ratio of an MTJ can be relaxed from 130% to 90% for the fluctuation of 6% sigma-to-average ratio of MTJ resistance in a 16 pair-dummy cell averaging case by using this technology when compared with the arithmetic averaging method.

  • Scattering Characteristics of the Human Body in 67-GHz Band

    Ngochao TRAN  Tetsuro IMAI  Koshiro KITAO  Yukihiko OKUMURA  Takehiro NAKAMURA  Hiroshi TOKUDA  Takao MIYAKE  Robin WANG  Zhu WEN  Hajime KITANO  Roger NICHOLS  

     
    PAPER-Antennas and Propagation

      Pubricized:
    2017/12/15
      Vol:
    E101-B No:6
      Page(s):
    1434-1442

    The fifth generation (5G) system using millimeter waves is considered for application to high traffic areas with a dense population of pedestrians. In such an environment, the effects of shadowing and scattering of radio waves by human bodies (HBs) on propagation channels cannot be ignored. In this paper, we clarify based on measurement the characteristics of waves scattered by the HB for typical non-line-of-sight scenarios in street canyon environments. In these scenarios, there are street intersections with pedestrians, and the angles that are formed by the transmission point, HB, and reception point are nearly equal to 90 degrees. We use a wide-band channel sounder for the 67-GHz band with a 1-GHz bandwidth and horn antennas in the measurements. The distance parameter between antennas and the HB is changed in the measurements. Moreover, the direction of the HB is changed from 0 to 360 degrees. The evaluation results show that the radar cross section (RCS) of the HB fluctuates randomly over the range of approximately 20dB. Moreover, the distribution of the RCS of the HB is a Gaussian distribution with a mean value of -9.4dBsm and the standard deviation of 4.2dBsm.

  • Forecasting Service Performance on the Basis of Temporal Information by the Conditional Restricted Boltzmann Machine

    Jiali YOU  Hanxing XUE  Yu ZHUO  Xin ZHANG  Jinlin WANG  

     
    PAPER-Network

      Pubricized:
    2017/11/10
      Vol:
    E101-B No:5
      Page(s):
    1210-1221

    Predicting the service performance of Internet applications is important in service selection, especially for video services. In order to design a predictor for forecasting video service performance in third-party application, two famous service providers in China, Iqiyi and Letv, are monitored and analyzed. The study highlights that the measured performance in the observation period is time-series data, and it has strong autocorrelation, which means it is predictable. In order to combine the temporal information and map the measured data to a proper feature space, the authors propose a predictor based on a Conditional Restricted Boltzmann Machine (CRBM), which can capture the potential temporal relationship of the historical information. Meanwhile, the measured data of different sources are combined to enhance the training process, which can enlarge the training size and avoid the over-fit problem. Experiments show that combining the measured results from different resolutions for a video can raise prediction performance, and the CRBM algorithm shows better prediction ability and more stable performance than the baseline algorithms.

  • Superimposing Thermal-Infrared Data on 3D Structure Reconstructed by RGB Visual Odometry

    Masahiro YAMAGUCHI  Trong Phuc TRUONG  Shohei MORI  Vincent NOZICK  Hideo SAITO  Shoji YACHIDA  Hideaki SATO  

     
    PAPER-Machine Vision and its Applications

      Pubricized:
    2018/02/16
      Vol:
    E101-D No:5
      Page(s):
    1296-1307

    In this paper, we propose a method to generate a three-dimensional (3D) thermal map and RGB + thermal (RGB-T) images of a scene from thermal-infrared and RGB images. The scene images are acquired by moving both a RGB camera and an thermal-infrared camera mounted on a stereo rig. Before capturing the scene with those cameras, we estimate their respective intrinsic parameters and their relative pose. Then, we reconstruct the 3D structures of the scene by using Direct Sparse Odometry (DSO) using the RGB images. In order to superimpose thermal information onto each point generated from DSO, we propose a method for estimating the scale of the point cloud corresponding to the extrinsic parameters between both cameras by matching depth images recovered from the RGB camera and the thermal-infrared camera based on mutual information. We also generate RGB-T images using the 3D structure of the scene and Delaunay triangulation. We do not rely on depth cameras and, therefore, our technique is not limited to scenes within the measurement range of the depth cameras. To demonstrate this technique, we generate 3D thermal maps and RGB-T images for both indoor and outdoor scenes.

261-280hit(1933hit)