The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] ATI(18690hit)

4281-4300hit(18690hit)

  • Investigation of Electromagnetic Noise Coupling in a Board with a Digital-RF Mixed IC by Measurement and Analysis

    Kenta TSUKAMOTO  Mizuki IWANAMI  Eiji HANKUI  

     
    PAPER

      Vol:
    E98-B No:7
      Page(s):
    1182-1187

    In this paper the amplitude probability distribution (APD) measurement method is applied to evaluate noise coupling to an antenna on an evaluation board that uses mixed RF and digital signals of an IC. We analytically investigate noise coupling path to the antenna where the correlation coefficient matches the APD curve of the evaluation board. Moreover, in order to verify the analysis results, the noise coupling path in the board is evaluated by measurements involving In-phase/Quadrature (I/Q) signals as well as electromagnetic simulations. As a result, we demonstrate that APD method is effective in evaluating a degree of noise coupling from an IC to multiple antennas on the board, and confirm that the intensity of noise coupling to each antenna is affected greatly by the board layout patterns.

  • A Study on Consistency between MINAVE and MINMAX in SSIM Based Independent Perceptual Video Coding

    Chao WANG  Xuanqin MOU  Lei ZHANG  

     
    LETTER-Image Processing and Video Processing

      Pubricized:
    2015/04/13
      Vol:
    E98-D No:7
      Page(s):
    1417-1421

    In this letter, we study the R-D properties of independent sources based on MSE and SSIM, and compare the bit allocation performance under the MINAVE and MINMAX criteria in video encoding. The results show that MINMAX has similar results in terms of average distortion with MINAVE by using SSIM, which illustrates the consistency between these two criteria in independent perceptual video coding. Further more, MINMAX results in lower quality fluctuation, which shows its advantage for perceptual video coding.

  • New Numerical Target SAR Values and an Optimized Flat Phantom for SAR Validation Tests in the 150MHz Band

    Dong-Geun CHOI  Ki-Hwea KIM  Jaehoon CHOI  

     
    PAPER-Antennas and Propagation

      Vol:
    E98-B No:7
      Page(s):
    1323-1332

    New target specific absorption rate (SAR) values, calculated using a proposed reference dipole antenna and the reference flat phantom, are presented for an SAR validation test at 150MHz. The reference flat phantom recommended by the International Electrotechnical Commission (IEC) standard for 150MHz requires a significant amount of liquid owing to its large size. We conduct a numerical analysis in order to reduce the size of the flat phantom. The optimum size of the flat phantom is 780 (L1) × 540 (W) × 200 (H)mm3, which is approximately a 64% reduction in volume compared to the reference flat phantom. The length of the reference dipole antenna required for the optimized flat phantom (extrapolated from the reference values at 300MHz) becomes 760mm. The calculated and measured return losses (S11) of the antenna at 150MHz are 24.1dB and 22dB, respectively. The calculated and measured results for the return loss of the dipole antenna agree well and satisfy the IEC standard (> 20dB). The target SAR values derived from the numerical analysis are 1.08W/kg for 1g of tissue and 0.77W/kg for 10g of tissue for an SAR validation test at 150MHz.

  • An Avoidance of Local Minimum Stagnation in IIR Filter Design Using PSO

    Yuji NISHIMURA  Kenji SUYAMA  

     
    LETTER-Digital Signal Processing

      Vol:
    E98-A No:7
      Page(s):
    1544-1548

    In this paper, a design method for the infinite impulse response (IIR) filters using the particle swarm optimization (PSO) is developed. It is well-known that the updating in the PSO tends to stagnate around local minimums due to a strong search directivity. Recently, the asynchronous digenetic PSO with nonlinear dissipative term (N-AD-PSO) has been proposed as a purpose for a diverse search. Therefore, it can be expected that the stagnation can be avoided by the N-AD-PSO. However, there is no report that the N-AD-PSO has been applied to any realistic problems. In this paper, the N-AD-PSO is applied for the IIR filter design. Several examples are shown to clarify the effectiveness and the drawback of the proposed method.

  • Method of Spread Spectrum Watermarking Using Quantization Index Modulation for Cropped Images

    Takahiro YAMAMOTO  Masaki KAWAMURA  

     
    PAPER-Data Engineering, Web Information Systems

      Pubricized:
    2015/04/16
      Vol:
    E98-D No:7
      Page(s):
    1306-1315

    We propose a method of spread spectrum digital watermarking with quantization index modulation (QIM) and evaluate the method on the basis of IHC evaluation criteria. The spread spectrum technique can make watermarks robust by using spread codes. Since watermarks can have redundancy, messages can be decoded from a degraded stego-image. Under IHC evaluation criteria, it is necessary to decode the messages without the original image. To do so, we propose a method in which watermarks are generated by using the spread spectrum technique and are embedded by QIM. QIM is an embedding method that can decode without an original image. The IHC evaluation criteria include JPEG compression and cropping as attacks. JPEG compression is lossy compression. Therefore, errors occur in watermarks. Since watermarks in stego-images are out of synchronization due to cropping, the position of embedded watermarks may be unclear. Detecting this position is needed while decoding. Therefore, both error correction and synchronization are required for digital watermarking methods. As countermeasures against cropping, the original image is divided into segments to embed watermarks. Moreover, each segment is divided into 8×8 pixel blocks. A watermark is embedded into a DCT coefficient in a block by QIM. To synchronize in decoding, the proposed method uses the correlation between watermarks and spread codes. After synchronization, watermarks are extracted by QIM, and then, messages are estimated from the watermarks. The proposed method was evaluated on the basis of the IHC evaluation criteria. The PSNR had to be higher than 30 dB. Ten 1920×1080 rectangular regions were cropped from each stego-image, and 200-bit messages were decoded from these regions. Their BERs were calculated to assess the tolerance. As a result, the BERs were less than 1.0%, and the average PSNR was 46.70 dB. Therefore, our method achieved a high image quality when using the IHC evaluation criteria. In addition, the proposed method was also evaluated by using StirMark 4.0. As a result, we found that our method has robustness for not only JPEG compression and cropping but also additional noise and Gaussian filtering. Moreover, the method has an advantage in that detection time is small since the synchronization is processed in 8×8 pixel blocks.

  • Speech Reconstruction from MFCC Based on Nonnegative and Sparse Priors

    Gang MIN  Xiong wei ZHANG  Ji bin YANG  Xia ZOU  Zhi song PAN  

     
    LETTER-Speech and Hearing

      Vol:
    E98-A No:7
      Page(s):
    1540-1543

    In this letter, high quality speech reconstruction approaches from Mel-frequency cepstral coefficients (MFCC) are presented. Taking into account of the nonnegative and sparse properties of the speech power spectrum, an alternating direction method of multipliers (ADMM) based nonnegative l2 norm (NL2) and weighted nonnegative l2 norm (NWL2) minimization approach is proposed to cope with the under-determined nature of the reconstruction problem. The phase spectrum is recovered by the well-known LSE-ISTFTM algorithm. Experimental results demonstrate that the NL2 and NWL2 approach substantially achieves better quality for reconstructed speech than the conventional l2 norm minimization approach, it sounds very close to the original speech when using the high-resolution MFCC, the PESQ score reaches 4.0.

  • New p-ary Sequence Families of Period ${ rac{p^n -1}{2}}$ with Good Correlation Property Using Two Decimated m-Sequences

    Chang-Min CHO  Ji-Youp KIM  Jong-Seon NO  

     
    PAPER-Fundamental Theories for Communications

      Vol:
    E98-B No:7
      Page(s):
    1268-1275

    In this paper, for an odd prime p and i=0,1, we investigate the cross-correlation between two decimated sequences, s(2t+i) and s(dt), where s(t) is a p-ary m-sequence of period pn-1. Here we consider two cases of ${d}$, ${d= rac{(p^m +1)^2}{2} }$ with ${n=2m}$, ${p^m equiv 1 pmod{4}}$ and ${d= rac{(p^m +1)^2}{p^e + 1}}$ with n=2m and odd m/e. The value distribution of the cross-correlation function for each case is completely determined. Also, by using these decimated sequences, two new p-ary sequence families of period ${ rac{p^n -1}{2}}$ with good correlation property are constructed.

  • An Evaluation on Two-Handed and One-Handed Control Methods for Positioning Object in Immersive Virtual Environments

    Noritaka OSAWA  Kikuo ASAI  

     
    PAPER-Human-computer Interaction

      Pubricized:
    2015/03/20
      Vol:
    E98-D No:7
      Page(s):
    1316-1324

    A two-handed distance control method is proposed for precisely and efficiently manipulating a virtual 3D object by hand in an immersive virtual reality environment. The proposed method enhances direct manipulation by hand and is used to precisely control and efficiently adjust the position of an object and the viewpoint using the distance between the two hands. The two-handed method is evaluated and compared with the previously proposed one-handed speed control method, which adjusts the position of an object in accordance with the speed of one hand. The results from experimental evaluation show that two-handed methods, which make position and viewpoint adjustments, are the best among six combinations of control and adjustment methods.

  • A Participating Fine-Grained Cloud Computing Platform with In-Network Guidance

    Kento NISHII  Yosuke TANIGAWA  Hideki TODE  

     
    PAPER-Network

      Vol:
    E98-B No:6
      Page(s):
    1008-1017

    What should be the ultimate form of the cloud computing environment? The solution should have two important features; “Fine-Granularity” and “Participation.” To realize an attractive and feasible solution with these features, we propose a “participating fine-grained cloud computing platform” that a large number of personal or small-company resource suppliers participate in, configure and provide cloud computing on. This enables users to be supplied with smaller units of resources such as computing, memory, content, and applications, in comparison with the traditional Infrastructure as a Service (IaaS). Furthermore, to search for nearby resources efficiently among the many available on the platform, we also propose Resource Breadcrumbs (RBC) as a key technology of our proposed platform to provide in-network guidance capability autonomously for users' queries. RBC allows supplier-nodes to distribute guidance information directed to themselves with dedicated control messages; in addition, the information can be logged along the trail of message from supplier to user. With this distributed information, users can to autonomously locate nearby resources. Distributed management also reduces computational load on the central database and enables a participating fine-grained cloud platform at lower cost.

  • Algorithm for Identifying the Maximum Detour Hinge Vertices of a Permutation Graph

    Hirotoshi HONMA  Yoko NAKAJIMA  Yuta IGARASHI  Shigeru MASUYAMA  

     
    PAPER

      Vol:
    E98-A No:6
      Page(s):
    1161-1167

    A hinge vertex is a vertex in an undirected graph such that there exist two vertices whose removal makes the distance between them longer than before. Identifying hinge vertices in a graph can help detect critical nodes in communication network systems, which is useful for making them more stable. For finding them, an O(n3) time algorithm was developed for a simple graph, and, linear time algorithms were developed for interval and permutation graphs, respectively. Recently, the maximum detour hinge vertex problem is defined by Honma et al. For a hinge vertex u in a graph, the detour degree of u is the largest value of distance between any pair of x and y (x and y are adjacent to u) by removing u. A hinge vertex with the largest detour degree in G is defined as the maximum detour hinge vertex of G. This problem is motivated by practical applications, such as network stabilization with a limited cost, i.e., by enhancing the reliability of the maximum detour hinge vertex, the stability of the network is much improved. We previously developed an O(n2) time algorithm for solving this problem on an interval graph. In this study, we propose an algorithm that identifies the maximum detour hinge vertex on a permutation graph in O(n2) time, where n is the number of vertices in the graph.

  • A Framework for Verifying the Conformance of Design to Its Formal Specifications

    Dieu-Huong VU  Yuki CHIBA  Kenro YATAKE  Toshiaki AOKI  

     
    PAPER-Formal Verification

      Pubricized:
    2015/02/13
      Vol:
    E98-D No:6
      Page(s):
    1137-1149

    Verification of a design with respect to its requirement specification is important to prevent errors before constructing an actual implementation. The existing works focus on verifications where the specifications are described using temporal logics or using the same languages as that used to describe the designs. Our work considers cases where the specifications and the designs are described using different languages. To verify such cases, we propose a framework to check if a design conforms to its specification based on their simulation relation. Specifically, we define the semantics of the specifications and the designs commonly as labelled transition systems (LTSs). We appreciate LTSs since they could interpret information about the system and actions that the system may perform as well as the effect of these actions. Then, we check whether a design conforms to its specification based on the simulation relation of their LTS. In this paper, we present our framework for the verification of reactive systems, and we present the case where the specifications and the designs are described in Event-B and Promela/Spin, respectively. We also present two case studies with the results of several experiments to illustrate the applicability of our framework on practical systems.

  • Multipath Time Delay Estimation Based on Gibbs Sampling under Incoherent Reception Environment

    Sen ZHONG  Wei XIA  Zishu HE  

     
    LETTER-Digital Signal Processing

      Vol:
    E98-A No:6
      Page(s):
    1300-1304

    In the traditional time delay estimation methods, it is usually implicitly assumed that the observed signals are either only direct path propagate or coherently received. In practice, the multipath propagation and incoherent reception always exist simultaneously. In response to this situation, the joint maximum likelihood (ML) estimation of multipath delays and system error is proposed, and the estimation of the number of multipath is considered as well for the specific incoherent signal model. Furthermore, an algorithm based Gibbs sampling is developed to solve the multi-dimensional nonlinear ML estimation. The efficiency of the proposed estimator is demonstrated by simulation results.

  • Multicell Distributed Beamforming Based on Gradient Iteration and Local CSIs

    Zijia HUANG  Xiaoxiang WANG  

     
    PAPER-Antennas and Propagation

      Vol:
    E98-B No:6
      Page(s):
    1058-1064

    In this paper, the multicell distributed beamforming (MDBF) design problem of suppressing intra-cell interference (InCI) and inter-cell interference (ICI) is studied. To start with, in order to decrease the InCI and ICI caused by a user, we propose a gradient-iteration altruistic algorithm to derive the beamforming vectors. The convergence of the proposed iterative algorithm is proved. Second, a metric function is established to restrict the ICI and maximize cell rate. This function depends on only local channel state information (CSI) and does not need additional CSIs. Moreover, an MDBF algorithm with the metric function is proposed. This proposed algorithm utilizes gradient iteration to maximize the metric function to improve sum rate of the cell. Finally, simulation results demonstrate that the proposed algorithm can achieve higher cell rates while offering more advantages to suppress InCI and ICI than the traditional ones.

  • Comparative Study of Open-Loop Transmit Diversity Schemes with Four Antennas in DFT-Precoded OFDMA Using Turbo FDE and Iterative Channel Estimation

    Lianjun DENG  Teruo KAWAMURA  Hidekazu TAOKA  Mamoru SAWAHASHI  

     
    PAPER-Wireless Communication Technologies

      Vol:
    E98-B No:6
      Page(s):
    1065-1077

    This paper presents comprehensive comparisons on the block error rate (BLER) performance of rate-one open-loop (OL) transmit diversity schemes with four antennas for discrete Fourier transform (DFT)-precoded Orthogonal Frequency Division Multiple Access (OFDMA). One candidate scheme employs a quasi-orthogonal (QO) - space-time block code (STBC) in which four-branch minimum mean-square error (MMSE) combining is achieved at the cost of residual inter-code interference (ICI). Another candidate employs a combination of the STBC and selection transmit diversity called time switched transmit diversity (TSTD) (or frequency switched transmit diversity (FSTD)). We apply a turbo frequency domain equalizer (FDE) associated with iterative decision-feedback channel estimation (DFCE) using soft-symbol estimation to reduce channel estimation (CE) error. The turbo FDE includes an ICI canceller to reduce the influence of the residual ICI for the QO-STBC. Based on link-level simulation results, we show that a combination of the STBC and TSTD (or FSTD) is suitable as a four-antenna OL transmit diversity scheme for DFT-precoded OFDMA using the turbo FDE and iterative DFCE.

  • Long-Term Performance Evaluation of Hadoop Jobs in Public and Community Clouds

    Kento AIDA  Omar ABDUL-RAHMAN  Eisaku SAKANE  Kazutaka MOTOYAMA  

     
    PAPER-Computer System

      Pubricized:
    2015/02/25
      Vol:
    E98-D No:6
      Page(s):
    1176-1184

    Cloud computing is a widely used computing platform in business and academic communities. Performance is an important issue when a user runs an application in the cloud. The user may want to estimate the application-execution time beforehand to guarantee the application performance or to choose the most suitable cloud. Moreover, the cloud system architect and the designer need to understand the application performance characteristics, such as the scalability or the utilization of cloud platforms, to improve performance. However, because the application performance in clouds sometime fluctuates, estimation of the application performance is difficult. In this paper, we discuss the performance fluctuation of Hadoop jobs in both a public cloud and a community cloud for one to three months. The experimental results indicate phenomena that we cannot see without long-term experiments and phenomena inherent in Hadoop. The results suggest better ways to estimate Hadoop application performances in clouds. For example, we should be aware of application characteristics (CPU intensive or communication intensive), datacenter characteristics (busy or not), and time frame (time of day and day of the week) to estimate the performance fluctuation due to workload congestion in cloud platforms. Furthermore, we should be aware of performance degradation due to task re-execution in Hadoop applications.

  • An ITI-Mitigating 5/6 Modulation Code for Bit-Patterned Media Recording

    Chanon WARISARN  Autthasith ARRAYANGKOOL  Piya KOVINTAVEWAT  

     
    PAPER-Storage Technology

      Vol:
    E98-C No:6
      Page(s):
    528-533

    In bit-patterned media recording (BPMR), the readback signal is severely corrupted by the inter-symbol interference (ISI) and inter-track interference (ITI), especially at high recording densities, due to small bit and track pitches. One way to alleviate the ITI effect is to encode an input data sequence before recording, so as to avoid some data patterns that easily cause an error at the data detection process. This paper proposes an ITI-mitigating 5/6 modulation code for a multi-track multi-head BPMR system to eliminate the data patterns that lead to severe ITI. Specifically, each of the 5 user bits is converted into a 6-bit codeword in the form of a 3-by-2 data array, based on a look-up table. Experimental results indicate that the system with the proposed coding scheme outperforms that without coding, especially when an areal density is high and/or the position jitter is large.

  • Securely Computing Three-Input Functions with Eight Cards

    Takuya NISHIDA  Yu-ichi HAYASHI  Takaaki MIZUKI  Hideaki SONE  

     
    PAPER

      Vol:
    E98-A No:6
      Page(s):
    1145-1152

    Assume that Alice, Bob, and Carol, each of whom privately holds a one-bit input, want to learn the output of some Boolean function, say the majority function, of their inputs without revealing more of their own secret inputs than necessary. In this paper, we show that such a secure three-input function evaluation can be performed with a deck of real cards; specifically, the three players can learn only the output of the function using eight physical cards — four black and four red cards — with identical backs.

  • Far-Field Pattern Reconstruction Using an Iterative Hilbert Transform

    Fan FAN  Tapan K. SARKAR  Changwoo PARK  Jinhwan KOH  

     
    PAPER-Antennas and Propagation

      Vol:
    E98-B No:6
      Page(s):
    1032-1039

    A new approach to reconstructing antenna far-field patterns from the missing part of the pattern is presented in this paper. The antenna far-field pattern can be reconstructed by utilizing the iterative Hilbert transform, which is based on the relationship between the real and imaginary part of the Hilbert transform. A moving average filter is used to reduce the errors in the restored signal as well as the computation load. Under the constraint of the causality of the current source in space, we could successfully reconstruct the data. Several examples dealing with line source antennas and antenna arrays are simulated to illustrate the applicability of this approach.

  • Variable Data-Flow Graph for Lightweight Program Slicing and Visualization

    Yu KASHIMA  Takashi ISHIO  Shogo ETSUDA  Katsuro INOUE  

     
    PAPER-Software Engineering

      Pubricized:
    2015/03/17
      Vol:
    E98-D No:6
      Page(s):
    1194-1205

    To understand the behavior of a program, developers often need to read source code fragments in various modules. System-dependence-graph-based (SDG) program slicing is a good candidate for supporting the investigation of data-flow paths among modules, as SDG is capable of showing the data-dependence of focused program elements. However, this technique has two problems. First, constructing SDG requires heavyweight analysis, so SDG is not suitable for daily uses. Second, the results of SDG-based program slicing are difficult to visualize, as they contain many vertices. In this research, we proposed variable data-flow graphs (VDFG) for use in program slicing techniques. In contrast to SDG, VDFG is created by lightweight analysis because several approximations are used. Furthermore, we propose using the fractal value to visualize VDFG-based program slice in order to reduce the graph complexity for visualization purposes. We performed three experiments that demonstrate the accuracy of VDFG program slicing with fractal value, the size of a visualized program slice, and effectiveness of our tool for source code reading.

  • Optimization Methods for Nop-Shadows Typestate Analysis

    Chengsong WANG  Xiaoguang MAO  Yan LEI  Peng ZHANG  

     
    PAPER-Dependable Computing

      Pubricized:
    2015/02/23
      Vol:
    E98-D No:6
      Page(s):
    1213-1227

    In recent years, hybrid typestate analysis has been proposed to eliminate unnecessary monitoring instrumentations for runtime monitors at compile-time. Nop-shadows Analysis (NSA) is one of these hybrid typestate analyses. Before generating residual monitors, NSA performs the data-flow analysis which is intra-procedural flow-sensitive and partially context-sensitive to improve runtime performance. Although NSA is precise, there are some cases on which it has little effects. In this paper, we propose three optimizations to further improve the precision of NSA. The first two optimizations try to filter interferential states of objects when determining whether a monitoring instrumentation is necessary. The third optimization refines the inter-procedural data-flow analysis induced by method invocations. We have integrated our optimizations into Clara and conducted extensive experiments on the DaCapo benchmark. The experimental results demonstrate that our first two optimizations can further remove unnecessary instrumentations after the original NSA in more than half of the cases, without a significant overhead. In addition, all the instrumentations can be removed for two cases, which implies the program satisfy the typestate property and is free of runtime monitoring. It comes as a surprise to us that the third optimization can only be effective on 8.7% cases. Finally, we analyze the experimental results and discuss the reasons why our optimizations fail to further eliminate unnecessary instrumentations in some special situations.

4281-4300hit(18690hit)