Kyota HATTORI Masahiro NAKAGAWA Toshiya MATSUDA Masaru KATAYAMA Katsutoshi KODA
Improvement of conventional networks with an incremental approach is an important design method for the development of the future internet. For this approach, we are developing a future aggregation network based on passive optical network (PON) technology to achieve both cost-effectiveness and high reliability. In this paper, we propose a timeslot (TS) synchronization method for sharing a TS from an optical burst mode transceiver between any route of arbitrary fiber length by changing both the route of the TS transmission and the TS control timing on the optical burst mode transceiver. We show the effectiveness of the proposed method for exchanging TSs in bidirectional bufferless wavelength division multiplexing (WDM) and time division multiplexing (TDM) multi-ring networks under the condition of the occurrence of a link failure through prototype systems. Also, we evaluate the reduction of the required number of optical interfaces in a multi-ring network by applying the proposed method.
Zhanhu HU Wang HU Zhiping WANG
To improve the quality of waveforms and achieve a high input power factor (IPF) for matrix rectifier, a novel quasi sliding mode control (SMC) with adaptive compensation is proposed in this letter. Applying quasi-SMC can effective obviate the disturbances of time delay and spatial lag, and SMC based on continuous function is better than discontinuous function to eliminate the chattering. Furthermore, compared with conventional compensation, an adaptive quasi-SMC compensation without any accurate detection for internal parameters is easier to be implementated, which has shown a superior advance. Theoretical analysis and experiments are carried out to validate the correctness of the novel control scheme.
Hui-Seon GANG Shaikhul Islam CHOWDHURY Chun-Su PARK Goo-Rak KWON Jae-Young PYUN
Video quality generally suffers from packet losses caused by an unreliable channel when video is transmitted over an error-prone wireless channel. This quality degradation is the main reason that a video compression encoder uses error-resilient coding to deal with the high packet-loss probability. The use of adequate error resilience can mitigate the effects of channel errors, but the coding efficiency for bit reduction will be decreased. On the other hand, H.264/AVC uses multiple reference frame (MRF) motion compensation for a higher coding efficiency. However, an increase in the number of reference frames in the H.264/AVC encoder has been recently observed, making the received video quality worse in the presence of transmission errors if the cyclic intra-refresh is used as the error-resilience method. This is because the reference-block selection in the MRF chooses blocks on the basis of the rate distortion optimization, irrespective of the intra-refresh coding. In this paper, a new error-resilient reference selection method is proposed to provide error resilience for MRF based motion compensation. The proposed error-resilient reference selection method achieves an average PSNR enhancement up to 0.5 to 2dB in 10% packet-loss-ratio environments. Therefore, the proposed method can be valuable in most MRF-based interactive video encoding system, which can be used for video broadcasting and mobile video conferencing over an erroneous network.
Abu Nowshed CHY Md Zia ULLAH Masaki AONO
Microblog, especially twitter, has become an integral part of our daily life for searching latest news and events information. Due to the short length characteristics of tweets and frequent use of unconventional abbreviations, content-relevance based search cannot satisfy user's information need. Recent research has shown that considering temporal and contextual aspects in this regard has improved the retrieval performance significantly. In this paper, we focus on microblog retrieval, emphasizing the alleviation of the vocabulary mismatch, and the leverage of the temporal (e.g., recency and burst nature) and contextual characteristics of tweets. To address the temporal and contextual aspect of tweets, we propose new features based on query-tweet time, word embedding, and query-tweet sentiment correlation. We also introduce some popularity features to estimate the importance of a tweet. A three-stage query expansion technique is applied to improve the relevancy of tweets. Moreover, to determine the temporal and sentiment sensitivity of a query, we introduce query type determination techniques. After supervised feature selection, we apply random forest as a feature ranking method to estimate the importance of selected features. Then, we make use of ensemble of learning to rank (L2R) framework to estimate the relevance of query-tweet pair. We conducted experiments on TREC Microblog 2011 and 2012 test collections over the TREC Tweets2011 corpus. Experimental results demonstrate the effectiveness of our method over the baseline and known related works in terms of precision at 30 (P@30), mean average precision (MAP), normalized discounted cumulative gain at 30 (NDCG@30), and R-precision (R-Prec) metrics.
Hideaki KIM Noriko TAKAYA Hiroshi SAWADA
Improvements in information technology have made it easier for industry to communicate with their customers, raising hopes for a scheme that can estimate when customers will want to make purchases. Although a number of models have been developed to estimate the time-varying purchase probability, they are based on very restrictive assumptions such as preceding purchase-event dependence and discrete-time effect of covariates. Our preliminary analysis of real-world data finds that these assumptions are invalid: self-exciting behavior, as well as marketing stimulus and preceding purchase dependence, should be examined as possible factors influencing purchase probability. In this paper, by employing the novel idea of hierarchical time rescaling, we propose a tractable but highly flexible model that can meld various types of intrinsic history dependency and marketing stimuli in a continuous-time setting. By employing the proposed model, which incorporates the three factors, we analyze actual data, and show that our model has the ability to precisely track the temporal dynamics of purchase probability at the level of individuals. It enables us to take effective marketing actions such as advertising and recommendations on timely and individual bases, leading to the construction of a profitable relationship with each customer.
Yuji INAGAKI Yusaku SUGIMORI Eri IOKA Yasuyuki MATSUYA
This paper describes a logarithmic compression ADC using a subranging TDC and the transient response of a comparator. We utilized the settling time of the comparator for a logarithmic compression instead of a logarithmic amplifier. The settling time of the comparator is inversely proportional to the logarithm of an input voltage. In the proposed ADC, an input voltage is converted into a pulse whose width represents the settling time of the comparator. Subsequently, the TDC converts the pulse width into a binary code. The supply voltage of the proposed ADC can be reduced more than a conventional logarithmic ADC because an analog to digital conversion takes place in the time domain. We confirmed through a 0.18-µm CMOS circuit simulation that the proposed ADC achieves a resolution of 11 bits, a sampling rate of 20 MS/s, a dynamic range of 59 dB and a power consumption of 9.8 mW at 1.5 V operation.
Bing DENG Zhengbo SUN Le YANG Dexiu HU
A linear-correction method is developed for source position and velocity estimation using time difference of arrival (TDOA) and frequency difference of arrival (FDOA) measurements. The proposed technique first obtains an initial source location estimate using the first-step processing of an existing algebraic algorithm. It then refines the initial localization result by estimating via weighted least-squares (WLS) optimization and subtracting out its estimation error. The new solution is shown to be able to achieve the Cramer-Rao lower bound (CRLB) accuracy and it has better accuracy over several benchmark methods at relatively high noise levels.
In this work, we study efficient scheduling with network coding in a scalable video coding (SVC) multicast system. Transmission consists of two stages. The original SVC packets are multicasted by the server in the first stage and the lost packets are retransmitted in the second stage. With deadline constraint, the consumer can be only satisfied when the requested packets are received before expiration. Further, the hierarchical encoding architecture of SVC introduces extra decoding delay which poses a challenge for providing acceptable reconstructed video quality. To solve these problems, instantly decodable network coding is applied for reducing the decoding delay, and a novel packet weighted policy is designed to better describe the contribution a packet can make in upgrading the recovered video quality. Finally, an online packet scheduling algorithm based on the maximal weighted clique is proposed to improve the delay, deadline miss ratio and users' experience. Multiple characteristics of SVC packets, such as the packet utility, the slack time and the number of undelivered/wanted packets, are jointly considered. Simulation results prove that the proposed algorithm requires fewer retransmissions and achieves lower deadline miss ratio. Moreover, the algorithm enjoys fine recovery video quality and provides high user satisfaction.
Hidekazu MURATA Makoto MIYAGOSHI Yuji OISHI
The end-to-end packet error rate (PER) performance of a multi-hop cooperative relaying system is discussed in this paper. In this system, the end-to-end PER performance improves with the number of hops under certain conditions. The PER performance of multi-hop cooperative networks is analyzed with the state transition technique. The theoretical analysis reveals that the PER performance can be kept almost constant, or even improved, as the number of hops is increased. Computer simulation results agree closely with the analysis results. Moreover, to confirm this performance characteristic in an actual setup, an in-lab experiment using a fading emulator was conducted. The experimental results confirm the theoretical end-to-end PER performance of this system.
Laksmita RAHADIANTI Wooseong JEONG Fumihiko SAKAUE Jun SATO
In this paper we propose a method for estimating time-to-contact in scattering media. Images taken in scattering media are often unclear and blurry, making it difficult to detect appropriate geometric information from these images for computing the 3 dimensional properties of the scene. Therefore, instead of searching for geometric information, we attempt to use photometric information instead. In our approach, we use the observed image intensity. The method proposed in this paper is able to utilize the effect of scattering media on the resultant image and estimate the time-to-contact toward objects without any prior knowledge of the scene, cameras, and the scattering media. This method is then evaluated using simulated and real images.
Arata KAWAMURA Hiro IGARASHI Youji IIGUNI
Image-to-sound mapping is a technique that transforms an image to a sound signal, which is subsequently treated as a sound spectrogram. In general, the transformed sound differs from a human speech signal. Herein an efficient image-to-sound mapping method, which provides an understandable speech signal without any training, is proposed. To synthesize such a speech signal, the proposed method utilizes a multi-column image and a speech spectral phase that is obtained from a long-time observation of the speech. The original image can be retrieved from the sound spectrogram of the synthesized speech signal. The synthesized speech and the reconstructed image qualities are evaluated using objective tests.
Jana BACKHUS Ichigaku TAKIGAWA Hideyuki IMAI Mineichi KUDO Masanori SUGIMOTO
In this paper, we introduce a self-constructive Normalized Gaussian Network (NGnet) for online learning tasks. In online tasks, data samples are received sequentially, and domain knowledge is often limited. Then, we need to employ learning methods to the NGnet that possess robust performance and dynamically select an accurate model size. We revise a previously proposed localized forgetting approach for the NGnet and adapt some unit manipulation mechanisms to it for dynamic model selection. The mechanisms are improved for more robustness in negative interference prone environments, and a new merge manipulation is considered to deal with model redundancies. The effectiveness of the proposed method is compared with the previous localized forgetting approach and an established learning method for the NGnet. Several experiments are conducted for a function approximation and chaotic time series forecasting task. The proposed approach possesses robust and favorable performance in different learning situations over all testbeds.
Takayoshi SHOUDAI Kazuhide AIKOH Yusuke SUZUKI Satoshi MATSUMOTO Tetsuhiro MIYAHARA Tomoyuki UCHIDA
An efficient means of learning tree-structural features from tree-structured data would enable us to construct effective mining methods for tree-structured data. Here, a pattern representing rich tree-structural features common to tree-structured data and a polynomial time algorithm for learning important tree patterns are necessary for mining knowledge from tree-structured data. As such a tree pattern, we introduce a term tree pattern t such that any edge label of t belongs to a finite alphabet Λ, any internal vertex of t has ordered children and t has a new kind of structured variable, called a height-constrained variable. A height-constrained variable has a pair of integers (i, j) as constraints, and it can be replaced with a tree whose trunk length is at least i and whose height is at most j. This replacement is called height-constrained replacement. A sequence of consecutive height-constrained variables is called a variable-chain. In this paper, we present polynomial time algorithms for solving the membership problem and the minimal language (MINL) problem for term tree patternshaving no variable-chain. The membership problem for term tree patternsis to decide whether or not a given tree can be obtained from a given term tree pattern by applying height-constrained replacements to all height-constrained variables in the term tree pattern. The MINL problem for term tree patternsis to find a term tree pattern t such that the language generated by t is minimal among languages, generated by term tree patterns, which contain all given tree-structured data. Finally, we show that the class, i.e., the set of all term tree patternshaving no variable-chain, is polynomial time inductively inferable from positive data if |Λ| ≥ 2.
Hiroaki TANAKA Ayako KOTANI Katsuyoshi NISHI Yurie IRIBE Koji OGURI
Driving safety related innovations received increasing interest from automotive industry. We performed an experiment to observe what situations are related to the secured feelings drivers feel when they drive, and found out that drivers need to have four to seven seconds to react possible collision when they operate onboard Human Machine Interface (HMI) devices and check display devices. We explored the distance of semantic space to see what factors of HMI interaction lead to the secured feeling in that time period, and extracted 32 types of factors that lead to the secured feelings. Furthermore, in the process of investigating the semantic space distance, the indicators relating to the secured feelings obtained in the prior studies were further determined to be ‘The layout of the operation device is the same as the driver's image' and ‘The driver can use the word he uses every day to give instructions’ in this time period.’, which were more concrete factors of the secured feelings.
The paper studies controllability of an aggregate demand response system, i.e., the amount of the change of the total electric consumption in response to the change of the electric price, for real-time pricing (RTP). In order to quantify the controllability, this paper defines the controllability index as the lowest occurrence probability of the total electric consumption when the best possible the electric price is chosen. Then the paper formulates the problem which finds the consumer group maximizing the controllability index. The controllability problem becomes hard to solve as the number of consumers increases. To give a solution of the controllability problem, the article approximates the controllability index by the generalized central limit theorem. Using the approximated controllability index, the controllability problem can be reduced to a problem for solving nonlinear equations. Since the number of variables of the equations is independent of the number of consumers, an approximate solution of the controllability problem is obtained by numerically solving the equations.
Jianquan LIU Shoji NISHIMURA Takuya ARAKI Yuichi NAKAMURA
Similarity search is an important and fundamental problem, and thus widely used in various fields of computer science including multimedia, computer vision, database, information retrieval, etc. Recently, since loitering behavior often leads to abnormal situations, such as pickpocketing and terrorist attacks, its analysis attracts increasing attention from research communities. In this paper, we present AntiLoiter, a loitering discovery system adopting efficient similarity search on surveillance videos. As we know, most of existing systems for loitering analysis, mainly focus on how to detect or identify loiterers by behavior tracking techniques. However, the difficulties of tracking-based methods are known as that their analysis results are heavily influenced by occlusions, overlaps, and shadows. Moreover, tracking-based methods need to track the human appearance continuously. Therefore, existing methods are not readily applied to real-world surveillance cameras due to the appearance discontinuity of criminal loiterers. To solve this problem, we abandon the tracking method, instead, propose AntiLoiter to efficiently discover loiterers based on their frequent appearance patterns in longtime multiple surveillance videos. In AntiLoiter, we propose a novel data structure Luigi that indexes data using only similarity value returned by a corresponding function (e.g., face matching). Luigi is adopted to perform efficient similarity search to realize loitering discovery. We conducted extensive experiments on both synthetic and real surveillance videos to evaluate the efficiency and efficacy of our approach. The experimental results show that our system can find out loitering candidates correctly and outperforms existing method by 100 times in terms of runtime.
Shadow is an important effect that makes virtual 3D scenes more realistic. In this paper, we propose a fast and correct soft shadow generation method for area lights of various shapes and colors. To conduct efficient as well as accurate visibility tests, we exploit the complexity of shadow and area light color.
Tomohiro KITAGAWA Tetsushi YUGE Shigeru YANAGI
The maintenance of a system on a ship has limitations when the ship is engaged in a voyage because of limited maintenance resources. When a system fails, it is either repaired instantly on ship with probability p or remains unrepaired during the voyage with probability 1-p owing to the lack of maintenance resources. In the latter case, the system is repaired after the voyage. We propose two management policies for the overhaul interval of an IFR system: one manages the overhaul interval by number of voyages and the other manages it by the total voyage time. Our goal is to determine the optimal policy that ensures the required availability of the system and minimizes the expected cost rate.
The Even-Goldreich-Micali framework is a generic method for constructing secure digital signature schemes from weaker signature schemes and one-time signature schemes. Several variations are known due to properties demanded on the underlying building blocks. It is in particular interesting when the underlying signature scheme is a so-called F-signature scheme that admits different message spaces for signing and verification. In this paper we overview these variations in the literature and add a new one to the bucket.
Hideki KAWAGUCHI Thomas WEILAND
The Time Domain Boundary Element Method (TDBEM) has its advantages in the analysis of transient electromagnetic fields (wake fields) induced by a charged particle beam with curved trajectory in a particle accelerator. On the other hand, the TDBEM has disadvantages of huge required memory and computation time compared with those of the Finite Difference Time Domain (FDTD) method or the Finite Integration Technique (FIT). This paper presents a comparison of the FDTD method and 4-D domain decomposition method of the TDBEM based on an initial value problem formulation for the curved trajectory electron beam, and application to a full model simulation of the bunch compressor section of the high-energy particle accelerators.