Gou HOUBEN Shu FUJITA Keita TAKAHASHI Toshiaki FUJII
Depth (disparity) estimation from a light field (a set of dense multi-view images) is currently attracting much research interest. This paper focuses on how to handle a noisy light field for disparity estimation, because if left as it is, the noise deteriorates the accuracy of estimated disparity maps. Several researchers have worked on this problem, e.g., by introducing disparity cues that are robust to noise. However, it is not easy to break the trade-off between the accuracy and computational speed. To tackle this trade-off, we have integrated a fast denoising scheme in a fast disparity estimation framework that works in the epipolar plane image (EPI) domain. Specifically, we found that a simple 1-D slanted filter is very effective for reducing noise while preserving the underlying structure in an EPI. Moreover, this simple filtering does not require elaborate parameter configurations in accordance with the target noise level. Experimental results including real-world inputs show that our method can achieve good accuracy with much less computational time compared to some state-of-the-art methods.
Kotchakorn PITUSO Chanon WARISARN Damrongsak TONGSOMPORN
When the track density of two-dimensional magnetic recording (TDMR) systems is increased, intertrack interference (ITI) inevitably grows, resulting in the extreme degradation of an overall system performance. In this work, we present coding, writing, and reading techniques which allow TDMR systems with multi-readers to overcome severe ITI. A rate-5/6 two-dimensional (2D) modulation code is adopted to protect middle-track data from ITI based on cross-track data dependence. Since the rate-5/6 2D modulation code greatly improves the reliability of the middle-track, there is a bit-error rate gap between middle-track and sidetracks. Therefore, we propose the different track width writing technique to optimize the reliability of all three data tracks. In addition, we also evaluate the TDMR system performance using an user areal density capability (UADC) as a main key parameter. Here, an areal density capability (ADC) can be measured by finding the bit-error rate of the system with sweeping track and linear densities. The UADC is then obtained by removing redundancy from the ADC. Simulation results show that a system with our proposed techniques gains the UADC of about 4.66% over the conventional TDMR systems.
Bandhit SUKSIRI Masahiro FUKUMOTO
This paper presents an efficient wideband two-dimensional direction-of-arrival (DOA) estimation for an L-shaped microphone array. We propose a way to construct a wideband sample cross-correlation matrix without any process of DOA preliminary estimation, such as beamforming technique, by exploiting sample cross-correlation matrices of two different frequencies for all frequency bins. Subsequently, wideband DOAs can be estimated by using this wideband matrix along with a scheme of estimating DOA in a narrowband subspace method. Therefore, a contribution of our study is providing an alternative framework for recent narrowband subspace methods to estimating the DOA of wideband sources directly. It means that this framework enables cutting-edge techniques in the existing narrowband subspace methods to implement the wideband direction estimation for reducing the computational complexity and facilitating the estimation algorithm. Theoretical analysis and effectiveness of the proposed method are substantiated through numerical simulations and experiments, which are performed in reverberating environments. The results show that performance of the proposed method performs better than others over a range of signal-to-noise ratio with just a few microphones. All these advantages make the proposed method a powerful tool for navigation systems based on acoustic signal processing.
Xuan WANG Bofeng ZHANG Mingqing HUANG Furong CHANG Zhuocheng ZHOU
When individuals make a purchase from online sources, they may lack first-hand knowledge of the product. In such cases, they will judge the quality of the item by the reviews other consumers have posted. Therefore, it is significant to determine whether comments about a product are credible. Most often, conventional research on comment credibility has employed supervised machine learning methods, which have the disadvantage of needing large quantities of training data. This paper proposes an unsupervised method for judging comment credibility based on the Biterm Sentiment Latent Dirichlet Allocation (BS-LDA) model. Using this approach, first we derived some distributions and calculated each comment's credibility score via them. A comment's credibility was judged based on whether it achieved a threshold score. Our experimental results using comments from Amazon.com demonstrated that the overall performance of our approach can play an important role in determining the credibility of comments in some situation.
Yuan ZHOU Yuichi GOTO Jingde CHENG
Many kinds of questionnaires, testing, and voting are performed in some completely electronic ways to do questions and answers on the Internet as Web applications, i.e. e-questionnaire systems, e-testing systems, and e-voting systems. Because there is no unified communication tool among the stakeholders of e-questionnaire, e-testing, and e-voting systems, until now, all the e-questionnaire, e-testing, and e-voting systems are designed, developed, used, and maintained in various ad hoc ways. As a result, the stakeholders are difficult to communicate to implement the systems, because there is neither an exhaustive requirement list to have a grasp of the overall e-questionnaire, e-testing, and e-voting systems nor a standardized terminology for these systems to avoid ambiguity. A general-purpose specification language to provide a unified description way for specifying various e-questionnaire, e-testing, and e-voting systems can solve the problems such that the stakeholders can refer to and use the complete requirements and standardized terminology for better communications, and can easily and unambiguously specify all the requirements of systems and services of e-questionnaire, e-testing, and e-voting, even can implement the systems. In this paper, we propose the first specification language, named “QSL,” with a standardized, consistent, and exhaustive list of requirements for specifying various e-questionnaire, e-testing, and e-voting systems such that the specifications can be used as the precondition of automatically generating e-questionnaire, e-testing, and e-voting systems. The paper presents our design addressing that QSL can specify all the requirements of various e-questionnaire, e-testing, and e-voting systems in a structured way, evaluates its effectiveness, performs real applications using QSL in case of e-questionnaire, e-testing, and e-voting systems, and shows various QSL applications for providing convenient QSL services to stakeholders.
Tianwen GUO Ping DENG Qiang YU Baoyun WANG
In this letter, we investigate a design of efficient antenna allocation at the full duplex receiver (FDR) in a multi-input multi-output multi-eavesdropper (MIMOME) wiretap channel for physical layer security improvement. Specifically, we propose the allocation which are feasible for the practical scenario with self-interference (SI) taken into account, because the jamming signals from FDR not only confuse the eavesdropper but also inevitably cause SI at the FDR. Due to the nolinear and coupling of the antenna allocation optimization problem, we transform the original problem into an integer programming problem. Then, we derive the optimal solution and the corresponding beamforming matrices in closed-form by means of combining spatial alignment and null-space projection method. Furthermore, we present the feasibility condition and full-protection condition, which offer insight into principles that enable more efficient and effective use of FDR in the wiretap channel for security improvement. From the simulation results, we validate the theoretical analysis and demonstrate the outstanding performance of the proposed antennas allocation at FDR.
Chen CHANG Jianjun CAO Qin FENG Nianfeng WENG Yuling SHANG
Most existing truth discovery approaches are designed for structured data, and cannot meet the strong need to extract trustworthy information from raw text data for its unique characteristics such as multifactorial property of text answers (i.e., an answer may contain multiple key factors) and the diversity of word usages (i.e., different words may have the same semantic meaning). As for text answers, there are no absolute correctness or errors, most answers may be partially correct, which is quite different from the situation of traditional truth discovery. To solve these challenges, we propose an optimization-based text truth discovery model which jointly groups keywords extracted from the answers of the specific question into a set of multiple factors. Then, we select the subset of multiple factors as identified truth set for each question by parallel ant colony synchronization optimization algorithm. After that, the answers to each question can be ranked based on the similarities between factors answer provided and identified truth factors. The experiment results on real dataset show that though text data structures are complex, our model can still find reliable answers compared with retrieval-based and state-of-the-art approaches.
Yingwei FU Kele XU Haibo MI Qiuqiang KONG Dezhi WANG Huaimin WANG Tie HONG
Sound event detection is intended to identify the sound events in audio recordings, which has widespread applications in real life. Recently, convolutional recurrent neural network (CRNN) models have achieved state-of-the-art performance in this task due to their capabilities in learning the representative features. However, the CRNN models are of high complexities with millions of parameters to be trained, which limits their usage for the mobile and embedded devices with limited computation resource. Model distillation is effective to distill the knowledge of a complex model to a smaller one, which can be deployed on the devices with limited computational power. In this letter, we propose a novel multi model-based distillation approach for sound event detection by making use of the knowledge from models of multiple teachers which are complementary in detecting sound events. Extensive experimental results demonstrated that our approach achieves a compression ratio about 50 times. In addition, better performance is obtained for the sound event detection task.
Yancheng CHEN Ning LI Xijian ZHONG Yan GUO
Unmanned aerial vehicle mounted base stations (UAV-BSs) can provide wireless cellular service to ground users in a variety of scenarios. The efficient deployment of such UAV-BSs while optimizing the coverage area is one of the key challenges. We investigate the deployment of UAV-BS to maximize the coverage of ground users, and further analyzes the impact of the deployment of UAV-BS on the fairness of ground users. In this paper, we first calculated the location of the UAV-BS according to the QoS requirements of the ground users, and then the fairness of ground users is taken into account by calculating three different fairness indexes. The performance of two genetic algorithms, namely Standard Genetic Algorithm (SGA) and Multi-Population Genetic Algorithm (MPGA) are compared to solve the optimization problem of UAV-BS deployment. The simulations are presented showing that the performance of the two algorithms, and the fairness performance of the ground users is also given.
Conventional TLB (Translation Lookaside Buffer) coalescing schemes do not fully exploit the contiguity that a memory allocator provides. The conventional schemes accordingly have certain performance overheads due to page table walks. To address this issue, we propose an efficient scheme, called block contiguity translation (BCT), that accommodates the block size information in a page table considering the Buddy algorithm. By fully exploiting the block-level contiguity, we can reduce the page table walks as certain physical memory is allocated in the contiguous way. Additionally, we present unified per-level page sizes to simplify the design and better utilize the contiguity information. Considering the state-of-the-art schemes as references, the comparative analysis and the performance simulations are conducted. Experiments indicate that the proposed scheme can improve the memory system performance with moderate hardware overheads.
To the best of our knowledge, there are a few researches on air-handwriting character-level writer identification only employing acceleration and angular velocity data. In this paper, we propose a deep learning approach to writer identification only using inertial sensor data of air-handwriting. In particular, we separate different representations of degree of freedom (DoF) of air-handwriting to extract local dependency and interrelationship in different CNNs separately. Experiments on a public dataset achieve an average good performance without any extra hand-designed feature extractions.
JianFeng WU HuiBin QIN YongZhu HUA LiHuan SHAO Ji HU ShengYing YANG
This paper proposes a deep neural network (DNN) based framework to address the problem of vector quantization (VQ) for high-dimensional data. The main challenge of applying DNN to VQ is how to reduce the binary coding error of the auto-encoder when the distribution of the coding units is far from binary. To address this problem, three fine-tuning methods have been adopted: 1) adding Gaussian noise to the input of the coding layer, 2) forcing the output of the coding layer to be binary, 3) adding a non-binary penalty term to the loss function. These fine-tuning methods have been extensively evaluated on quantizing speech magnitude spectra. The results demonstrated that each of the methods is useful for improving the coding performance. When implemented for quantizing 968-dimensional speech spectra using only 18-bit, the DNN-based VQ framework achieved an averaged PESQ of about 2.09, which is far beyond the capability of conventional VQ methods.
Sixing YANG Yan GUO Dongping YU Peng QIAN
We research device-free (DF) multi-target tracking scheme in this paper. The existing localization and tracking algorithms are always pay attention to the single target and need to collect a large amount of localization information. In this paper, we exploit the sparse property of multiple target locations to achieve target trace accurately with much less sampling both in the wireless links and the time slots. The proposed approach mainly includes the target localization part and target trace recovery part. In target localization part, by exploiting the inherent sparsity of the target number, Compressive Sensing (CS) is utilized to reduce the wireless links distributed. In the target trace recovery part, we exploit the compressive property of target trace, as well as designing the measurement matrix and the sparse matrix, to reduce the samplings in time domain. Additionally, Kronecker Compressive Sensing (KCS) theory is used to simultaneously recover the multiple traces both of the X label and the Y Label. Finally, simulations show that the proposed approach holds an effective recovery performance.
Bao Trung CHU Kenji HASHIMOTO Hiroyuki SEKI
A program is non-interferent if it leaks no secret information to an observable output. However, non-interference is too strict in many practical cases and quantitative information flow (QIF) has been proposed and studied in depth. Originally, QIF is defined as the average of leakage amount of secret information over all executions of a program. However, a vulnerable program that has executions leaking the whole secret but has the small average leakage could be considered as secure. This counter-intuition raises a need for a new definition of information leakage of a particular run, i.e., dynamic leakage. As discussed in [5], entropy-based definitions do not work well for quantifying information leakage dynamically; Belief-based definition on the other hand is appropriate for deterministic programs, however, it is not appropriate for probabilistic ones.In this paper, we propose new simple notions of dynamic leakage based on entropy which are compatible with existing QIF definitions for deterministic programs, and yet reasonable for probabilistic programs in the sense of [5]. We also investigated the complexity of computing the proposed dynamic leakage for three classes of Boolean programs. We also implemented a tool for QIF calculation using model counting tools for Boolean formulae. Experimental results on popular benchmarks of QIF research show the flexibility of our framework. Finally, we discuss the improvement of performance and scalability of the proposed method as well as an extension to more general cases.
Akihito TAYA Takayuki NISHIO Masahiro MORIKURA Koji YAMAMOTO
In millimeter wave (mmWave) vehicular communications, multi-hop relay disconnection by line-of-sight (LOS) blockage is a critical problem, particularly in the early diffusion phase of mmWave-available vehicles, where not all vehicles have mmWave communication devices. This paper proposes a distributed position control method to establish long relay paths through road side units (RSUs). This is realized by a scheme via which autonomous vehicles change their relative positions to communicate with each other via LOS paths. Even though vehicles with the proposed method do not use all the information of the environment and do not cooperate with each other, they can decide their action (e.g., lane change and overtaking) and form long relays only using information of their surroundings (e.g., surrounding vehicle positions). The decision-making problem is formulated as a Markov decision process such that autonomous vehicles can learn a practical movement strategy for making long relays by a reinforcement learning (RL) algorithm. This paper designs a learning algorithm based on a sophisticated deep reinforcement learning algorithm, asynchronous advantage actor-critic (A3C), which enables vehicles to learn a complex movement strategy quickly through its deep-neural-network architecture and multi-agent-learning mechanism. Once the strategy is well trained, vehicles can move independently to establish long relays and connect to the RSUs via the relays. Simulation results confirm that the proposed method can increase the relay length and coverage even if the traffic conditions and penetration ratio of mmWave communication devices in the learning and operation phases are different.
Guoqiang CHENG Qingquan HUANG Zhi LIN Xiangshuai TAO Jian OUYANG Guodong WU
In this paper, we consider a hybrid satellite terrestrial cooperative network with a multi-antenna relay where the satellite links follows the shadowed-Rician fading and the terrestrial link undergoes the correlated Rayleigh fading. Specifically, two different channel state information (CSI) assumptions are considered: 1) full CSI at the relay; 2) full CSI of satellite-relay link and statistical CSI of relay-destination link at the relay. In addition, selection combining (SC) or maximal ratio combining (MRC) are used at the destination to combine the signals from direct link and relay link. By considering the above four cases, we derived the closed-form expressions for the outage probability (OP) respectively. Furthermore, the asymptotic OP expressions at high signal-to-noise (SNR) are developed to reveal the diversity orders and the array gains of the considered network. Finally, numerical results are provided to validate our analytical expressions as well as the system performance for different cases.
An adaptive bit allocation scheme for zero-forcing (ZF) Tomlinson-Harashima precoding (THP) is proposed. The ZF-THP enables us to achieve feasible bit error rate (BER) performance when appropriate substream permutations are installed at the transmitter. In this study, the number of bits in each substream is adaptively allocated to minimize the average BER in fading environments. Numerical examples are provided to compare the proposed method with eigenbeam space division multiplexing (E-SDM) method.
Luis Rafael MARVAL-PÉREZ Koichi ITO Takafumi AOKI
Access control and surveillance applications like walking-through security gates and immigration control points have a great demand for convenient and accurate biometric recognition in unconstrained scenarios with low user cooperation. The periocular region, which is a relatively new biometric trait, has been attracting much attention for recognition of an individual in such scenarios. This paper proposes a periocular recognition method that combines Phase-Based Correspondence Matching (PB-CM) with a texture enhancement technique. PB-CM has demonstrated high recognition performance in other biometric traits, e.g., face, palmprint and finger-knuckle-print. However, a major limitation for periocular region is that the performance of PB-CM degrades when the periocular skin has poor texture. We address this problem by applying texture enhancement and found out that variance normalization of texture significantly improves the performance of periocular recognition using PB-CM. Experimental evaluation using three public databases demonstrates the advantage of the proposed method compared with conventional methods.
Tsuyoshi SUGIURA Satoshi FURUTA Tadamasa MURAKAMI Koki TANJI Norihisa OTANI Toshihiko YOSHIMASU
This paper presents high efficiency Class-E and compact Doherty power amplifiers (PAs) with novel harmonics termination for handset applications using a GaAs/InGaP heterojunction bipolar transistor (HBT) process. The novel harmonics termination circuit effectively reduces the insertion loss of the matching circuit, allowing a device with a compact size. The Doherty PA uses a lumped-element transformer which consists of metal-insulator-metal (MIM) capacitors on an IC substrate, a bonding-wire inductor and short micro-strip lines on a printed circuit board (PCB). The fabricated Class-E PA exhibits a power added efficiency (PAE) as high as 69.0% at 1.95GHz and as high as 67.6% at 2.535GHz. The fabricated Doherty PA exhibits an average output power of 25.5dBm and a PAE as high as 50.1% under a 10-MHz band width quadrature phase shift keying (QPSK) 6.16-dB peak-to-average-power-ratio (PAPR) LTE signal at 1.95GHz. The fabricated chip size is smaller than 1mm2. The input and output Doherty transformer areas are 0.5mm by 1.0mm and 0.7mm by 0.7mm, respectively.
Envelope tracking (ET) technology provides the potential for achieving high efficiency in power amplifiers (PAs) with high peak-to-average ratio (PAR) signals. Envelope amplifiers with high fidelity, high efficiency, and wide bandwidth are critical components for the widespread application of envelope tracking. This paper presents the design of a linear-assisted switching buck converter for use in an envelope amplifier. To effectively leverage the high efficiency of buck converters and the wide bandwidth capabilities of linear amplifiers, a parallel combination of these two devices is employed in this work. A novel current-sense constant-on-time (COT) controller is proposed to coordinate this hybrid power supply. The combination mainly enables the switching converter to provide the average power required by the PA with high efficiency, while the wideband linear amplifier provides a wide range of dynamic voltages. The technique improves the efficiency of the envelope amplifier, especially for applications requiring high PAR with wider bandwidth signals. Measurement of the envelope amplifier showed an efficiency of approximately 77% with 10 W output power using LTE downlink signals. The overall ET system was demonstrated by using a GaN PA. The measured average power-added efficiency of the amplifier reached above 45% for an LTE modulated signal with 20 MHz bandwidth and PAR of 8.0 dB, at an average output power of 5 W and gain of 10.1 dB. The measured normalized RMS error is below 2.1% with adjacent channel leakage ratio of -48 dBc at an offset frequency of 20 MHz.