Daisuke MAEDA Koki MORIMURA Shintaro NARISADA Kazuhide FUKUSHIMA Takashi NISHIDE
We propose how to homomorphically evaluate arbitrary univariate and bivariate integer functions such as division. A prior work proposed by Okada et al. (WISTP'18) uses polynomial evaluations such that the scheme is still compatible with the SIMD operations in BFV and BGV schemes, and is implemented with the input domain ℤ257. However, the scheme of Okada et al. requires the quadratic numbers of plaintext-ciphertext multiplications and ciphertext-ciphertext additions in the input domain size, and although these operations are more lightweight than the ciphertext-ciphertext multiplication, the quadratic complexity makes handling larger inputs quite inefficient. In this work, first we improve the prior work and also propose a new approach that exploits the packing method to handle the larger input domain size instead of enabling the SIMD operation, thus making it possible to work with the larger input domain size, e.g., ℤ215 in a reasonably efficient way. In addition, we show how to slightly extend the input domain size to ℤ216 with a relatively moderate overhead. Further we show another approach to handling the larger input domain size by using two ciphertexts to encrypt one integer plaintext and applying our techniques for uni/bivariate function evaluation. We implement the prior work of Okada et al., our improved version of Okada et al., and our new scheme in PALISADE with the input domain ℤ215, and confirm that the estimated run-times of the prior work and our improved version of the prior work are still about 117 days and 59 days respectively while our new scheme can be computed in 307 seconds.
Daisuke HIBINO Tomoharu SHIBUYA
Distributed computing is one of the powerful solutions for computational tasks that need the massive size of dataset. Lagrange coded computing (LCC), proposed by Yu et al. [15], realizes private and secure distributed computing under the existence of stragglers, malicious workers, and colluding workers by using an encoding polynomial. Since the encoding polynomial depends on a dataset, it must be updated every arrival of new dataset. Therefore, it is necessary to employ efficient algorithm to construct the encoding polynomial. In this paper, we propose Newton coded computing (NCC) which is based on Newton interpolation to construct the encoding polynomial. Let K, L, and T be the number of data, the length of each data, and the number of colluding workers, respectively. Then, the computational complexity for construction of an encoding polynomial is improved from O(L(K+T)log 2(K+T)log log (K+T)) for LCC to O(L(K+T)log (K+T)) for the proposed method. Furthermore, by applying the proposed method, the computational complexity for updating the encoding polynomial is improved from O(L(K+T)log 2(K+T)log log (K+T)) for LCC to O(L) for the proposed method.
In this letter, a ray tracing (RT) acceleration method based on rank minimization is proposed. RT is a general tool used to simulate wireless communication environments. However, the simulation is time consuming because of the large number of ray calculations. This letter focuses on radio map interpolation as an acceleration approach. In the conventional methods cannot appropriately estimate short-span variation caused by multipath fading. To overcome the shortage of the conventional methods, we adopt rank minimization based interpolation. A computational simulation using commercial RT software revealed that the interpolation accuracy of the proposed method was higher than those of other radio map interpolation methods and that RT simulation can be accelerated approximate five times faster with the missing rate of 0.8.
With the arrival of 5G and the popularity of smart devices, indoor localization technical feasibility has been verified, and its market demands is huge. The channel state information (CSI) extracted from Wi-Fi is physical layer information which is more fine-grained than the received signal strength indication (RSSI). This paper proposes a CSI correction localization algorithm using DenseNet, which is termed CorFi. This method first uses isolation forest to eliminate abnormal CSI, and then constructs a CSI amplitude fingerprint containing time, frequency and antenna pair information. In an offline stage, the densely connected convolutional networks (DenseNet) are trained to establish correspondence between CSI and spatial position, and generalized extended interpolation is applied to construct the interpolated fingerprint database. In an online stage, DenseNet is used for position estimation, and the interpolated fingerprint database and K-nearest neighbor (KNN) are combined to correct the position of the prediction results with low maximum probability. In an indoor corridor environment, the average localization error is 0.536m.
Masahito SHIMAMOTO Yusuke KAMEDA Takayuki HAMAMOTO
We aim at HDR imaging with simple processing while preventing spatial resolution degradation in multiple-exposure-time image sensor where the exposure time is controlled for each pixel. The contributions are the proposal of image interpolation by motion area detection and pixel adaptive weighting method by overexposure and motion blur detection.
Li XU Bing LUO Mingming KONG Bo LI Zheng PEI
This letter proposes a fast superpixel segmentation method based on boundary sampling and interpolation. The basic idea is as follow: instead of labeling local region pixels, we estimate superpixel boundary by interpolating candidate boundary pixel from a down-sampling image segmentation. On the one hand, there exists high spatial redundancy within each local region, which could be discarded. On the other hand, we estimate the labels of candidate boundary pixels via sampling superpixel boundary within corresponding neighbour. Benefiting from the reduction of candidate pixel distance calculation, the proposed method significantly accelerates superpixel segmentation. Experiments on BSD500 benchmark demonstrate that our method needs half the time compared with the state-of-the-arts while almost no accuracy reduction.
Seongwook LEE Young-Jun YOON Seokhyun KANG Jae-Eun LEE Seong-Cheol KIM
In this paper, we propose a received signal interpolation method for enhancing the performance of multiple signal classification (MUSIC) algorithm. In general, the performance of the conventional MUSIC algorithm is very sensitive to signal-to-noise ratio (SNR) of the received signal. When array elements receive the signals with nonuniform SNR values, the resolution performance is degraded compared to elements receiving the signals with uniform SNR values. Hence, we propose a signal calibration technique for improving the resolution of the algorithm. First, based on original signals, rough direction of arrival (DOA) estimation is conducted. In this stage, using frequency-domain received signals, SNR values of each antenna element in the array are estimated. Then, a deteriorated element that has a relatively lower SNR value than those of the other elements is selected by our proposed scheme. Next, the received signal of the selected element is spatially interpolated based on the signals received from the neighboring elements and the DOA information extracted from the rough estimation. Finally, fine DOA estimation is performed again with the calibrated signal. Simulation results show that the angular resolution of the proposed method is better than that of the conventional MUSIC algorithm. Also, we apply the proposed scheme to actual data measured in the testing ground, and it gives us more enhanced DOA estimation result.
Ngoc-Giao PHAM Suk-Hwan LEE Ki-Ryong KWON
Nowadays, vector map content is widely used in the areas of life, science and the military. Due to the fact that vector maps bring great value and that their production process is expensive, a large volume of vector map data is attacked, stolen and illegally distributed by pirates. Thus, vector map data must be encrypted before being stored and transmitted in order to ensure the access and to prevent illegal copying. This paper presents a novel perceptual encryption algorithm for ensuring the secured storage and transmission of vector map data. Polyline data of vector maps are extracted to interpolate a spline curve, which is represented by an interpolating vector, the curvature degree coefficients, and control points. The proposed algorithm is based on encrypting the control points of the spline curve in the frequency domain of discrete cosine transform. Control points are transformed and selectively encrypted in the frequency domain of discrete cosine transform. They are then used in an inverse interpolation to generate the encrypted vector map. Experimental results show that the entire vector map is altered after the encryption process, and the proposed algorithm is very effective for a large dataset of vector maps.
Takuro YAMAGUCHI Masaaki IKEHARA
Image interpolation is one of the image upsampling technologies from a single input image. This technology obtains high resolution images by fitting functions or models. Although image interpolation methods are faster than other upsampling technologies, they tend to cause jaggies and blurs in edge and texture regions. Multi-surface Fitting is one of the image upsampling techniques from multiple input images. This algorithm utilizes multiple local functions and the weighted means of the estimations in each local function. Multi-surface Fitting obtains high quality upsampled images. However, its quality depends on the number of input images. Therefore, this method is used in only limited situations. In this paper, we propose an image interpolation method with both high quality and a low computational cost which can be used in many situations. We adapt the idea of Multi-surface Fitting for the image upsampling problems from a single input image. We also utilize local functions to reduce blurs. To improve the reliability of each local function, we introduce new weights in the estimation of the local functions. Besides, we improve the weights for weighted means to estimate a target pixel. Moreover, we utilize convolutions with small filters instead of the calculation of each local function in order to reduce the computational cost. Experimental results show our method obtains high quality output images without jaggies and blurs in short computational time.
Xingge GUO Liping HUANG Ke GU Leida LI Zhili ZHOU Lu TANG
The quality assessment of screen content images (SCIs) has been attractive recently. Different from natural images, SCI is usually a mixture of picture and text. Traditional quality metrics are mainly designed for natural images, which do not fit well into the SCIs. Motivated by this, this letter presents a simple and effective method to naturalize SCIs, so that the traditional quality models can be applied for SCI quality prediction. Specifically, bicubic interpolation-based up-sampling is proposed to achieve this goal. Extensive experiments and comparisons demonstrate the effectiveness of the proposed method.
Zhixin LIU Dexiu HU Yongjun ZHAO Chengcheng LIU
Considering the obvious bias of the traditional interpolation method, a novel time delay estimation (TDE) interpolation method with sub-sample accuracy is presented in this paper. The proposed method uses a generalized extended approximation method to obtain the objection function. Then the optimized interpolation curve is generated by Second-order Cone programming (SOCP). Finally the optimal TDE can be obtained by interpolation curve. The delay estimate of proposed method is not forced to lie on discrete samples and the sample points need not to be on the interpolation curve. In the condition of the acceptable computation complexity, computer simulation results clearly indicate that the proposed method is less biased and outperforms the other interpolation algorithms in terms of estimation accuracy.
Azril HANIZ Gia Khanh TRAN Ryosuke IWATA Kei SAKAGUCHI Jun-ichi TAKADA Daisuke HAYASHI Toshihiro YAMAGUCHI Shintaro ARATA
Conventional localization techniques such as triangulation and multilateration are not reliable in non-line-of-sight (NLOS) environments such as dense urban areas. Although fingerprint-based localization techniques have been proposed to solve this problem, we may face difficulties because we do not know the parameters of the illegal radio when creating the fingerprint database. This paper proposes a novel technique to localize illegal radios in an urban environment by interpolating the channel impulse responses stored as fingerprints in a database. The proposed interpolation technique consists of interpolation in the bandwidth (delay), frequency and spatial domains. A localization algorithm that minimizes the squared error criterion is employed in this paper, and the proposed technique is evaluated through Monte Carlo simulations using location fingerprints obtained from ray-tracing simulations. Results show that utilizing an interpolated fingerprint database is advantageous in such scenarios.
Orthogonal frequency division multiplexing (OFDM) channel estimation is the key technique used in broadband wireless networks. The Doppler frequency caused by fast mobility environments will cause inter-carrier interference (ICI) and degrade the performance of OFDM systems. Due to the severe ICI, channel estimation becomes a difficult task in higher mobility scenarios. Our aim is to propose a pilot-aided channel estimation method that is robust to high Doppler frequency with low computational complexity and pilot overheads. In this paper, the time duration of each estimate covers multiple consecutive OFDM symbols, named a “window”. A close-form of polynomial channel modeling is derived. The proposed method is initialized to the least squares (LS) estimates of the channels corresponding to the time interval of the pilot symbols within the window. Then, the channel interpolation is performed in the entire window. The results of computer simulations and computation complexity evaluations show that the proposed technique is robust to high Doppler frequency with low computation complexity and low pilot overheads. Compared with the state-of-the-art method and some conventional methods, the new technique proposed here has much lower computational complexity while offering comparable performance.
A Bayer-like White-RGB (W-RGB) color filter array (CFA) was invented for overcoming the weaknesses of commonly used RGB based Bayer CFA. In order to reproduce full-color images from the Bayer-like W-RGB CFA, a demosaicing or a CFA interpolation process which estimates missing color channels of raw mosaiced images from CFA is an essential process for single sensor digital cameras having CFA. In the case of Bayer CFA, numerous demosaicing methods which have remarkable performance were already proposed. In order to take advantage of both remarkable performance of demosaicing method for Bayer CFA and the characteristic of high-sensitive Bayer-like W-RGB CFA, a new method of transforming Bayer-like W-RGB to Bayer pattern is required. Therefore, in this letter, we present a new method of transforming Bayer-like W-RGB pattern to Bayer pattern. The proposed method mainly uses the color difference assumption between different channels which can be applied to practical consumer digital cameras.
Fan FAN Tapan K. SARKAR Changwoo PARK Jinhwan KOH
A new approach to reconstructing antenna far-field patterns from the missing part of the pattern is presented in this paper. The antenna far-field pattern can be reconstructed by utilizing the iterative Hilbert transform, which is based on the relationship between the real and imaginary part of the Hilbert transform. A moving average filter is used to reduce the errors in the restored signal as well as the computation load. Under the constraint of the causality of the current source in space, we could successfully reconstruct the data. Several examples dealing with line source antennas and antenna arrays are simulated to illustrate the applicability of this approach.
Rui SHI Shouyi YIN Leibo LIU Qiongbing LIU Shuang LIANG Shaojun WEI
Video Up-scaling is a hotspot in TV display area; as an important brunch of Video Up-scaling, Texture-Based Video Up-scaling (TBVU) method shows great potential of hardware implementation. Coarse-grained Reconfigurable Architecture (CGRA) is a very promising processor; it is a parallel computing platform which provides high performance of hardware, high flexibility of software, and dynamical reconfiguration ability. In this paper we propose an implementation of TBVU on CGRA. We fully exploit the characters of TBVU and utilize several techniques to reduce memory I/O operation and total execution time. Experimental results show that our work can greatly reduce the I/O operation and the execution time compared with the non-optimized ones. We also compare our work with other platforms and find great advantage in execution time and resource utilization rate.
Decimation and interpolation methods are utilized in image coding for low bit rate image coding. However, the decimation filter (prefilter) and the interpolation filter (postfilter) are irreversible with each other since the prefilter is a wide matrix (a matrix whose number of columns are larger than that of rows) and the postfilter is a tall one (a matrix whose number of rows are larger than that of columns). There will be some distortions in the reconstructed image even without any compression. The method of interpolation-dependent image downsampling (IDID) was used to tackle the problem of producing optimized downsampling images, which led to the optimized prefilter of a given postfilter. We propose integrating the IDID with time-domain lapped transforms (TDLTs) to improve image coding performance.
This paper proposes a so called quasi-linear support vector machine (SVM), which is an SVM with a composite quasi-linear kernel. In the quasi-linear SVM model, the nonlinear separation hyperplane is approximated by multiple local linear models with interpolation. Instead of building multiple local SVM models separately, the quasi-linear SVM realizes the multi local linear model approach in the kernel level. That is, it is built exactly in the same way as a single SVM model, by composing a quasi-linear kernel. A guided partitioning method is proposed to obtain the local partitions for the composition of quasi-linear kernel function. Experiment results on artificial data and benchmark datasets show that the proposed method is effective and improves classification performances.
Sanroku TSUKAMOTO Masaya MIYAHARA Akira MATSUZAWA
A 7bit 1GS/s flash ADC using two bit active interpolation and background offset calibration is proposed and tested. It achieves background calibration using 36 pre-amplifiers with 139 comparators. To cancel the offset, two pre-amplifiers and 12 comparators are set to offline in turn while the others are operating. A two bit active interpolation design and an offset cancellation scheme are implemented in the latch stage. The interpolation and background calibration significantly reduce analog input signal as well as reference voltage load. Fabricated with the 90nm CMOS process, the proposed ADC consumes 95mW under a 1.2V power supply.
In this paper, an efficient method to reduce computational complexity for pedestrian detection is presented. Since trilinear interpolation is not used, the amount of required operations for histogram of oriented gradient (HOG) feature calculation is significantly reduced. By calculating multi-scale HOG features with integral HOG in a two-stage approach, both high detection rate and speed are achieved in the proposed method.