The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] Z(5900hit)

601-620hit(5900hit)

  • Throughput Maximization of UAV-Enabled Wireless Network in the Presence of Jammers: Joint Trajectory and Communication Design

    Yang WU  Weiwei YANG  Di ZHANG  Xiaoli SUN  

     
    PAPER

      Pubricized:
    2019/04/26
      Vol:
    E102-B No:10
      Page(s):
    1983-1990

    Unmanned aerial vehicle (UAV) communication has drawn rising interest recently with the distinctive gains brought by its inherent mobility. In this paper, we investigate the throughput maximization problem in UAV-enabled uplink communication, where multiple ground nodes communicate with a UAV while a group of ground jammers send jamming signals to jam the communications between UAV and the ground nodes. In contrast to the previous works that only considering UAV's transmit power allocation and two-dimension (2D) trajectory design, the ground nodes' transmit power allocation and scheduling along with the UAV's three-dimensional (3D) trajectory design are jointly optimized. The formulated throughput maximization problem is a mixed-integer non-convex programme that hard to be solved in general. Thus, we propose an iterative algorithm to make the problem trackable by applying the block coordinate descent and successive convex optimization techniques. Simulation results show that our proposed algorithm outperforms the benchmark methods that improving the throughput of the system significantly.

  • Effects of Software Modifications and Development After an Organizational Change on Software Metrics Value Open Access

    Ryo ISHIZUKA  Naohiko TSUDA  Hironori WASHIZAKI  Yoshiaki FUKAZAWA  Shunsuke SUGIMURA  Yuichiro YASUDA  

     
    LETTER-Software Quality Management

      Pubricized:
    2019/06/13
      Vol:
    E102-D No:9
      Page(s):
    1693-1695

    Deterioration of software quality developed by multiple organizations has become a serious problem. To predict software degradation after an organizational change, this paper investigates the influence of quality deterioration on software metrics by analyzing three software projects. To detect factors indicating a low evolvability, we focus on the relationships between the change in software metric values and refactoring tendencies. Refactoring after an organization change impacts the quality.

  • A Fully-Connected Ising Model Embedding Method and Its Evaluation for CMOS Annealing Machines

    Daisuke OKU  Kotaro TERADA  Masato HAYASHI  Masanao YAMAOKA  Shu TANAKA  Nozomu TOGAWA  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2019/06/10
      Vol:
    E102-D No:9
      Page(s):
    1696-1706

    Combinatorial optimization problems with a large solution space are difficult to solve just using von Neumann computers. Ising machines or annealing machines have been developed to tackle these problems as a promising Non-von Neumann computer. In order to use these annealing machines, every combinatorial optimization problem is mapped onto the physical Ising model, which consists of spins, interactions between them, and their external magnetic fields. Then the annealing machines operate so as to search the ground state of the physical Ising model, which corresponds to the optimal solution of the original combinatorial optimization problem. A combinatorial optimization problem can be firstly described by an ideal fully-connected Ising model but it is very hard to embed it onto the physical Ising model topology of a particular annealing machine, which causes one of the largest issues in annealing machines. In this paper, we propose a fully-connected Ising model embedding method targeting for CMOS annealing machine. The key idea is that the proposed method replicates every logical spin in a fully-connected Ising model and embeds each logical spin onto the physical spins with the same chain length. Experimental results through an actual combinatorial problem show that the proposed method obtains spin embeddings superior to the conventional de facto standard method, in terms of the embedding time and the probability of obtaining a feasible solution.

  • Cross-VM Cache Timing Attacks on Virtualized Network Functions

    Youngjoo SHIN  

     
    LETTER-Information Network

      Pubricized:
    2019/05/27
      Vol:
    E102-D No:9
      Page(s):
    1874-1877

    Network function virtualization (NFV) achieves the flexibility of network service provisioning by using virtualization technology. However, NFV is exposed to a serious security threat known as cross-VM cache timing attacks. In this letter, we look into real security impacts on network virtualization. Specifically, we present two kinds of practical cache timing attacks on virtualized firewalls and routers. We also propose some countermeasures to mitigate such attacks on virtualized network functions.

  • A Method for Smartphone Theft Prevention When the Owner Dozes Off Open Access

    Kouhei NAGATA  Yoshiaki SEKI  

     
    LETTER-Physical Security

      Pubricized:
    2019/06/04
      Vol:
    E102-D No:9
      Page(s):
    1686-1688

    We propose a method for preventing smartphone theft when the owner dozes off. The owner of the smartphone wears a wristwatch type device that has an acceleration sensor and a vibration mode. This device detects when the owner dozes off. When the acceleration sensor in the smartphone detects an accident while dozing, the device vibrates. We implemented this function and tested its usefulness.

  • TFIDF-FL: Localizing Faults Using Term Frequency-Inverse Document Frequency and Deep Learning

    Zhuo ZHANG  Yan LEI  Jianjun XU  Xiaoguang MAO  Xi CHANG  

     
    LETTER-Software Engineering

      Pubricized:
    2019/05/27
      Vol:
    E102-D No:9
      Page(s):
    1860-1864

    Existing fault localization based on neural networks utilize the information of whether a statement is executed or not executed to identify suspicious statements potentially responsible for a failure. However, the information just shows the binary execution states of a statement, and cannot show how important a statement is in executions. Consequently, it may degrade fault localization effectiveness. To address this issue, this paper proposes TFIDF-FL by using term frequency-inverse document frequency to identify a high or low degree of the influence of a statement in an execution. Our empirical results on 8 real-world programs show that TFIDF-FL significantly improves fault localization effectiveness.

  • Two-Level Named Packet Forwarding for Enhancing the Performance of Virtualized ICN Router

    Kazuaki UEDA  Kenji YOKOTA  Jun KURIHARA  Atsushi TAGAMI  

     
    PAPER

      Pubricized:
    2019/03/22
      Vol:
    E102-B No:9
      Page(s):
    1813-1821

    Information-Centric Networking (ICN) can offer rich functionalities to the network, e.g, in-network caching, and name-based forwarding. Incremental deployment of ICN is a key challenge that enable smooth migration from current IP network to ICN. We can say that Network Function Virtualization (NFV) must be one of the key technologies to achieve this deployment because of its flexibility to support new network functions. However, when we consider the ICN deployment with NFV, there exist two performance issues, processing delay of name-based forwarding and computational overhead of virtual machine. In this paper we proposed a NFV infrastructure-assisted ICN packet forwarding by integrating the name look-up to the Open vSwitch. The contributions are twofold: 1) First, we provide the novel name look-up scheme that can forward ICN packets without costly longest prefix match searching. 2) Second, we design the ICN packet forwarding scheme that integrates the partial name look-up into the virtualization infrastructure to mitigate computation overhead.

  • Analysis of Eye Movement and Critical Fusion Frequency Responses to Different Movie Types Open Access

    Takahide OTOMO  Shinya MOCHIDUKI  Eriko ISHII  Yuko HOSHINO  Mitsuho YAMADA  

     
    LETTER

      Vol:
    E102-A No:9
      Page(s):
    1254-1258

    We can enjoy various video contents such as movies in several ways. In this report, we show the effects of content differences on physiological parameters such as eye movements and CFF. This time we confirmed the difference in responses that after watching a movie. In addition, a consistent change that can infer that due to a movie was also indicated. Our results showed that content differences affect the parameters. This suggests the possibility that the influence of movie contents on the viewer can be evaluated by physiological parameters.

  • Generalized Shogi, Chess, and Xiangqi are Constant-Time Testable

    Hiro ITO  Atsuki NAGAO  Teagun PARK  

     
    PAPER-Puzzles

      Vol:
    E102-A No:9
      Page(s):
    1126-1133

    We present constant-time testing algorithms for generalized shogi (Japanese chess), chess, and xiangqi (Chinese chess). These problems are known or believed to be EXPTIME-complete. A testing algorithm (or a tester) for a property accepts an input if it has the property, and rejects it with high probability if it is far from having the property (e.g., at least 2/3) by reading only a constant part of the input. A property is said to be testable if a tester exists. Given any position on a ⌊√n⌋×⌊√n⌋ board with O(n) pieces, the generalized shogi, chess, and xiangqi problem are problems determining the property that “the player who moves first has a winning strategy.” We propose that this property is testable for shogi, chess, and xiangqi. The shogi tester and xiangqi tester have a one-sided-error, but surprisingly, the chess tester has no-error. Over the last decade, many problems have been revealed to be testable, but most of such problems belong to NP. This is the first result on the constant-time testability of EXPTIME-complete problems.

  • On Computational Complexity of Pipe Puzzles

    Takumu SHIRAYAMA  Takuto SHIGEMURA  Yota OTACHI  Shuichi MIYAZAKI  Ryuhei UEHARA  

     
    PAPER-Puzzles

      Vol:
    E102-A No:9
      Page(s):
    1134-1141

    In this paper, we investigate computational complexity of pipe puzzles. A pipe puzzle is a kind of tiling puzzle; the input is a set of cards, and a part of a pipe is drawn on each card. For a given set of cards, we arrange them and connect the pipes. We have to connect all pipes without creating any local loop. While ordinary tiling puzzles, like jigsaw puzzles, ask to arrange the tiles with local consistency, pipe puzzles ask to join all pipes. We first show that the pipe puzzle is NP-complete in general even if the goal shape is quite restricted. We also investigate restricted cases and show some polynomial-time algorithms.

  • Enhancing Multipath TCP Initialization with SYN Duplication

    Kien NGUYEN  Mirza Golam KIBRIA  Kentaro ISHIZU  Fumihide KOJIMA  

     
    PAPER-Network

      Pubricized:
    2019/03/18
      Vol:
    E102-B No:9
      Page(s):
    1904-1913

    A Multipath TCP (MPTCP) connection uses multiple subflows (i.e., TCP flows), each of which traverses over a wireless link, enabling throughput and resilience enhancements in mobile wireless networks. However, to achieve the benefits, the subflows are necessarily initialized (i.e., must complete TCP handshakes) and sequentially attached to the MPTCP connection. In the standard (MPTCPST), MPTCP initialization raises several problems. First, the TCP handshake of opening subflow is generally associated with a predetermined network. That leads to degraded MPTCP performance when the network does not have the lowest latency among available ones. Second, the first subflow's initialization needs to be successful before the next subflow can commence its attempt to achieve initialization. Therefore, the resilience of multiple paths fails when the first initialization fails. This paper proposes a novel method for MPTCP initialization, namely MPTCPSD (i.e., MPTCP with SYN duplication), which can solve the problems. MPTCPSD duplicates the first SYN and attempts to establish TCP handshakes for all subflows simultaneously, hence inherently improves the loss-resiliency. The subflow that achieves initialization first, is selected as the first subflow, consequently solving the first problem. We have implemented and extensively evaluated MPTCPSD in comparison to MPTCPST. In an emulated network, the evaluation results show that MPTCPSD has better performance that MPTCPST with the scenarios of medium and short flows. Moreover, MPTCPSD outperforms MPTCPST in the case that the opening subflow fails. Moreover, a real network evaluation proves that MPTCPSD efficiently selects the lowest delay network among three ones for the first subflow regardless of the preconfigured default network. Additionally, we propose and implement a security feature for MPTCPSD, that prevents the malicious subflow from being established by a third party.

  • Reducing CPU Power Consumption with Device Utilization-Aware DVFS for Low-Latency SSDs

    Satoshi IMAMURA  Eiji YOSHIDA  Kazuichi OE  

     
    PAPER-Computer System

      Pubricized:
    2019/06/18
      Vol:
    E102-D No:9
      Page(s):
    1740-1749

    Emerging solid state drives (SSDs) based on a next-generation memory technology have been recently released in market. In this work, we call them low-latency SSDs because the device latency of them is an order of magnitude lower than that of conventional NAND flash SSDs. Although low-latency SSDs can drastically reduce an I/O latency perceived by an application, the overhead of OS processing included in the I/O latency has become noticeable because of the very low device latency. Since the OS processing is executed on a CPU core, its operating frequency should be maximized for reducing the OS overhead. However, a higher core frequency causes the higher CPU power consumption during I/O accesses to low-latency SSDs. Therefore, we propose the device utilization-aware DVFS (DU-DVFS) technique that periodically monitors the utilization of a target block device and applies dynamic voltage and frequency scaling (DVFS) to CPU cores executing I/O-intensive processes only when the block device is fully utilized. In this case, DU-DVFS can reduce the CPU power consumption without hurting performance because the delay of OS processing incurred by decreasing the core frequency can be hidden. Our evaluation with 28 I/O-intensive workloads on a real server containing an Intel® Optane™ SSD demonstrates that DU-DVFS reduces the CPU power consumption by 41.4% on average (up to 53.8%) with a negligible performance degradation, compared to a standard DVFS governor on Linux. Moreover, the evaluation with multiprogrammed workloads composed of I/O-intensive and non-I/O-intensive programs shows that DU-DVFS is also effective for them because it can apply DVFS only to CPU cores executing I/O-intensive processes.

  • A Knowledge Representation Based User-Driven Ontology Summarization Method

    Yuehang DING  Hongtao YU  Jianpeng ZHANG  Huanruo LI  Yunjie GU  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2019/05/30
      Vol:
    E102-D No:9
      Page(s):
    1870-1873

    As the superstructure of knowledge graph, ontology has been widely applied in knowledge engineering. However, it becomes increasingly difficult to be practiced and comprehended due to the growing data size and complexity of schemas. Hence, ontology summarization surfaced to enhance the comprehension and application of ontology. Existing summarization methods mainly focus on ontology's topology without taking semantic information into consideration, while human understand information based on semantics. Thus, we proposed a novel algorithm to integrate semantic information and topological information, which enables ontology to be more understandable. In our work, semantic and topological information are represented by concept vectors, a set of high-dimensional vectors. Distances between concept vectors represent concepts' similarity and we selected important concepts following these two criteria: 1) the distances from important concepts to normal concepts should be as short as possible, which indicates that important concepts could summarize normal concepts well; 2) the distances from an important concept to the others should be as long as possible which ensures that important concepts are not similar to each other. K-means++ is adopted to select important concepts. Lastly, we performed extensive evaluations to compare our algorithm with existing ones. The evaluations prove that our approach performs better than the others in most of the cases.

  • Card-Based Physical Zero-Knowledge Proof for Kakuro

    Daiki MIYAHARA  Tatsuya SASAKI  Takaaki MIZUKI  Hideaki SONE  

     
    PAPER-Cryptography and Information Security

      Vol:
    E102-A No:9
      Page(s):
    1072-1078

    Kakuro is a popular logic puzzle, in which a player fills in all empty squares with digits from 1 to 9 so that the sum of digits in each (horizontal or vertical) line is equal to a given number, called a clue, and digits in each line are all different. In 2016, Bultel, Dreier, Dumas, and Lafourcade proposed a physical zero-knowledge proof protocol for Kakuro using a deck of cards; their proposed protocol enables a prover to convince a verifier that the prover knows the solution of a Kakuro puzzle without revealing any information about the solution. One possible drawback of their protocol would be that the protocol is not perfectly extractable, implying that a prover who does not know the solution can convince a verifier with a small probability; therefore, one has to repeat the protocol to make such an error become negligible. In this paper, to overcome this, we design zero-knowledge proof protocols for Kakuro having perfect extractability property. Our improvement relies on the ideas behind the copy protocols in the field of card-based cryptography. By executing our protocols with a real deck of physical playing cards, humans can practically perform an efficient zero-knowledge proof of knowledge for Kakuro.

  • Calibration of Turntable Based 3D Scanning Systems

    Duhu MAN  Mark W. JONES  Danrong LI  Honglong ZHANG  Zhan SONG  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2019/05/30
      Vol:
    E102-D No:9
      Page(s):
    1833-1841

    The consistent alignment of point clouds obtained from multiple scanning positions is a crucial step for many 3D modeling systems. This is especially true for environment modeling. In order to observe the full scene, a common approach is to rotate the scanning device around a rotation axis using a turntable. The final alignment of each frame data can be computed from the position and orientation of the rotation axis. However, in practice, the precise mounting of scanning devices is impossible. It is hard to locate the vertical support of the turntable and rotation axis on a common line, particularly for lower cost consumer hardware. Therefore the calibration of the rotation axis of the turntable is an important step for the 3D reconstruction. In this paper we propose a novel calibration method for the rotation axis of the turntable. With the proposed rotation axis calibration method, multiple 3D profiles of the target scene can be aligned precisely. In the experiments, three different evaluation approaches are used to evaluate the calibration accuracy of the rotation axis. The experimental results show that the proposed rotation axis calibration method can achieve a high accuracy.

  • Computational Complexity of Herugolf and Makaro

    Chuzo IWAMOTO  Masato HARUISHI  Tatsuaki IBUSUKI  

     
    PAPER-Puzzles

      Vol:
    E102-A No:9
      Page(s):
    1118-1125

    Herugolf and Makaro are Nikoli's pencil puzzles. We study the computational complexity of Herugolf and Makaro puzzles. It is shown that deciding whether a given instance of each puzzle has a solution is NP-complete.

  • Development of a Novel Accurate Analysis System Regarding Information Processing within the Gazing Point Open Access

    Tsuyoshi KUSHIMA  Miyuki SUGANUMA  Shinya MOCHIDUKI  Mitsuho YAMADA  

     
    PAPER

      Vol:
    E102-A No:9
      Page(s):
    1205-1216

    Over the last 10 years, tablets have spread to the point where we can now read electronic books (e-books) like paper books. There is a long history of studies of eye movement during reading. Remarkable results have been reported for reading experiments in which displayed letters are changed in conjunction with eye movement during reading. However, these studies were conducted in the 1970s, and it is difficult to judge the detailed descriptions of the experimental techniques and whether the display time was correctly controlled when changing letters. Here, we propose an experimental system to control the display information exactly, as well as the display time, and inspect the results of past reading research, with the aim of being at the forefront of reading research in the e-book era.

  • Prediction of the Helmholtz-Kohlrausch Effect for Natural Images Using a Correction Function

    Yuki HAYAMI  Daiki TAKASU  Hisakazu AOYANAGI  Hiroaki TAKAMATSU  Yoshifumi SHIMODAIRA  Gosuke OHASHI  

     
    PAPER

      Vol:
    E102-A No:9
      Page(s):
    1217-1224

    The human visual system exhibits a characteristic known as the Helmholtz-Kohlrausch (H-K) effect: even if the hue and the lightness retain the same values, the actual lightness (perceived lightness) changes with changes in the color saturation. Quantification of this effect is expected to be useful for the future development and evaluation of high-quality displays. We have been studying the H-K effect in natural images projected by LED projectors, which play important roles in practical uses. To verify the effectiveness of the determinations of the H-K effect for natural images, we have performed a subjective-evaluation experiment by method of adjustment for natural images and compared the experimental values with values calculated from extended form of Nayatani's equation to apply to natural images. In general, we found a high correlation between the two, although there was a low correlation for some images. Therefore, we obtained a correction function derived from the subjective evaluation experiment value of 108 color (hue: 12 × saturation: 3 × lightness: 3) patterns and have applied it to estimate the equation H-K effect.

  • Design and Evaluation of Information Bottleneck LDPC Decoders for Digital Signal Processors Open Access

    Jan LEWANDOWSKY  Gerhard BAUCH  Matthias TSCHAUNER  Peter OPPERMANN  

     
    INVITED PAPER

      Pubricized:
    2019/02/20
      Vol:
    E102-B No:8
      Page(s):
    1363-1370

    Receiver implementations with very low quantization resolution will play an important role in 5G, as high precision quantization and signal processing are costly in terms of computational resources and chip area. Therefore, low resolution receivers with quasi optimum performance will be required to meet complexity and latency constraints. The Information Bottleneck method allows for a novel, information centric approach to design such receivers. The method was originally introduced by Naftali Tishby et al. and mostly used in the machine learning field so far. Interestingly, it can also be applied to build surprisingly good digital communication receivers which work fundamentally different than state-of-the-art receivers. Instead of minimizing the quantization error, receiver components with maximum preservation of relevant information for a given bit width can be designed. All signal processing in the resulting receivers is performed using only simple lookup operations. In this paper, we first provide a brief introduction to the design of receiver components with the Information Bottleneck method. We keep referring to decoding of low-density parity-check codes as a practical example. The focus of the paper lies on practical decoder implementations on a digital signal processor which illustrate the potential of the proposed technique. An Information Bottleneck decoder with 4bit message passing decoding is found to outperform 8bit implementations of the well-known min-sum decoder in terms of bit error rate and to perform extremely close to an 8bit belief propagation decoder, while offering considerably higher net decoding throughput than both conventional decoders.

  • Pre-Training of DNN-Based Speech Synthesis Based on Bidirectional Conversion between Text and Speech

    Kentaro SONE  Toru NAKASHIKA  

     
    PAPER-Speech and Hearing

      Pubricized:
    2019/05/15
      Vol:
    E102-D No:8
      Page(s):
    1546-1553

    Conventional approaches to statistical parametric speech synthesis use context-dependent hidden Markov models (HMMs) clustered using decision trees to generate speech parameters from linguistic features. However, decision trees are not always appropriate to model complex context dependencies of linguistic features efficiently. An alternative scheme that replaces decision trees with deep neural networks (DNNs) was presented as a possible way to overcome the difficulty. By training the network to represent high-dimensional feedforward dependencies from linguistic features to acoustic features, DNN-based speech synthesis systems convert a text into a speech. To improved the naturalness of the synthesized speech, this paper presents a novel pre-training method for DNN-based statistical parametric speech synthesis systems. In our method, a deep relational model (DRM), which represents a joint probability of two visible variables, is applied to describe the joint distribution of acoustic and linguistic features. As with DNNs, a DRM consists several hidden layers and two visible layers. Although DNNs represent feedforward dependencies from one visible variables (inputs) to other visible variables (outputs), a DRM has an ability to represent the bidirectional dependencies between two visible variables. During the maximum-likelihood (ML) -based training, the model optimizes its parameters (connection weights between two adjacent layers, and biases) of a deep architecture considering the bidirectional conversion between 1) acoustic features given linguistic features, and 2) linguistic features given acoustic features generated from itself. Owing to considering whether the generated acoustic features are recognizable, our method can obtain reasonable parameters for speech synthesis. Experimental results in a speech synthesis task show that pre-trained DNN-based systems using our proposed method outperformed randomly-initialized DNN-based systems, especially when the amount of training data is limited. Additionally, speaker-dependent speech recognition experimental results also show that our method outperformed DNN-based systems, by setting the initial parameters of our method are the same as that in the synthesis experiments.

601-620hit(5900hit)