The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] MPO(945hit)

61-80hit(945hit)

  • Applying K-SVD Dictionary Learning for EEG Compressed Sensing Framework with Outlier Detection and Independent Component Analysis Open Access

    Kotaro NAGAI  Daisuke KANEMOTO  Makoto OHKI  

     
    LETTER-Biometrics

      Pubricized:
    2021/03/01
      Vol:
    E104-A No:9
      Page(s):
    1375-1378

    This letter reports on the effectiveness of applying the K-singular value decomposition (SVD) dictionary learning to the electroencephalogram (EEG) compressed sensing framework with outlier detection and independent component analysis. Using the K-SVD dictionary matrix with our design parameter optimization, for example, at compression ratio of four, we improved the normalized mean square error value by 31.4% compared with that of the discrete cosine transform dictionary for CHB-MIT Scalp EEG Database.

  • Classification Functions for Handwritten Digit Recognition

    Tsutomu SASAO  Yuto HORIKAWA  Yukihiro IGUCHI  

     
    PAPER-Logic Design

      Pubricized:
    2021/04/01
      Vol:
    E104-D No:8
      Page(s):
    1076-1082

    A classification function maps a set of vectors into several classes. A machine learning problem is treated as a design problem for partially defined classification functions. To realize classification functions for MNIST hand written digits, three different architectures are considered: Single-unit realization, 45-unit realization, and 45-unit ×r realization. The 45-unit realization consists of 45 ternary classifiers, 10 counters, and a max selector. Test accuracy of these architectures are compared using MNIST data set.

  • Creation of Temporal Model for Prioritized Transmission in Predictive Spatial-Monitoring Using Machine Learning Open Access

    Keiichiro SATO  Ryoichi SHINKUMA  Takehiro SATO  Eiji OKI  Takanori IWAI  Takeo ONISHI  Takahiro NOBUKIYO  Dai KANETOMO  Kozo SATODA  

     
    PAPER-Network

      Pubricized:
    2021/02/01
      Vol:
    E104-B No:8
      Page(s):
    951-960

    Predictive spatial-monitoring, which predicts spatial information such as road traffic, has attracted much attention in the context of smart cities. Machine learning enables predictive spatial-monitoring by using a large amount of aggregated sensor data. Since the capacity of mobile networks is strictly limited, serious transmission delays occur when loads of communication traffic are heavy. If some of the data used for predictive spatial-monitoring do not arrive on time, prediction accuracy degrades because the prediction has to be done using only the received data, which implies that data for prediction are ‘delay-sensitive’. A utility-based allocation technique has suggested modeling of temporal characteristics of such delay-sensitive data for prioritized transmission. However, no study has addressed temporal model for prioritized transmission in predictive spatial-monitoring. Therefore, this paper proposes a scheme that enables the creation of a temporal model for predictive spatial-monitoring. The scheme is roughly composed of two steps: the first involves creating training data from original time-series data and a machine learning model that can use the data, while the second step involves modeling a temporal model using feature selection in the learning model. Feature selection enables the estimation of the importance of data in terms of how much the data contribute to prediction accuracy from the machine learning model. This paper considers road-traffic prediction as a scenario and shows that the temporal models created with the proposed scheme can handle real spatial datasets. A numerical study demonstrated how our temporal model works effectively in prioritized transmission for predictive spatial-monitoring in terms of prediction accuracy.

  • 4K 120fps HEVC Encoder with Multi-Chip Configuration Open Access

    Yuya OMORI  Ken NAKAMURA  Takayuki ONISHI  Daisuke KOBAYASHI  Tatsuya OSAWA  Hiroe IWASAKI  

     
    PAPER

      Pubricized:
    2021/02/04
      Vol:
    E104-B No:7
      Page(s):
    749-759

    This paper describes a novel 4K 120fps (frames per second) real-time HEVC (High Efficiency Video Coding) encoder for high-frame-rate video encoding and transmission. Motion portrayal problems such as motion blur and jerkiness may occur in video scenes containing fast-moving objects or quick camera panning. A high-frame-rate solves such problems and provides a more immersive viewing experience that can express even the fast-moving scenes without discomfort. It can also be used in remote operation for scenes with high motion, such as VAR (Video Assistant Referee) systems in sports. Real-time encoding of high-frame-rate videos with low latency and temporal scalability is required for providing such high-frame-rate video services. The proposed encoder achieves full 4K/120fps real-time encoding, which is twice the current 4K service frame rate of 60fps, by multichip configuration with two encoder LSI. Exchange of reference picture data near a spatially divided slice boundary provides cross-chip motion estimation, and maintains the coding efficiency. The encoder supports temporal-scalable coding mode, in which it output stream with temporal scalability transmitted over one or two transmission paths. The encoder also supports the other mode, low-delay coding mode, in which it achieves 21.8msec low-latency processing through motion vector restriction. Evaluation of the proposed encoder's multichip configuration shows that the BD-bitrate (the average rate of bitrate increase), compared to simple slice division without inter-chip transfer, is -2.86% at minimum and -2.41% on average in temporal-scalable coding mode. The proposed encoder system will open the door to the next generation of high-frame-rate UHDTV (ultra-high-definition television) services.

  • Video Smoke Removal from a Single Image Sequence Open Access

    Shiori YAMAGUCHI  Keita HIRAI  Takahiko HORIUCHI  

     
    PAPER

      Pubricized:
    2021/01/07
      Vol:
    E104-A No:6
      Page(s):
    876-886

    In this study, we present a novel method for removing smoke from videos based on a single image sequence. Smoke is a significant artifact in images or videos because it can reduce the visibility in disaster scenes. Our proposed method for removing smoke involves two main processes: (1) the development of a smoke imaging model and (2) smoke removal using spatio-temporal pixel compensation. First, we model the optical phenomena in natural scenes including smoke, which is called a smoke imaging model. Our smoke imaging model is developed by extending conventional haze imaging models. We then remove the smoke from a video in a frame-by-frame manner based on the smoke imaging model. Next, we refine the appearance of the smoke-free video by spatio-temporal pixel compensation, where we align the smoke-free frames using the corresponding pixels. To obtain the corresponding pixels, we use SIFT and color features with distance constraints. Finally, in order to obtain a clear video, we refine the pixel values based on the spatio-temporal weightings of the corresponding pixels in the smoke-free frames. We used simulated and actual smoke videos in our validation experiments. The experimental results demonstrated that our method can obtain effective smoke removal results from dynamic scenes. We also quantitatively assessed our method based on a temporal coherence measure.

  • A Low-Complexity QR Decomposition with Novel Modified RVD for MIMO Systems

    Lu SUN  Bin WU  Tianchun YE  

     
    LETTER-Digital Signal Processing

      Pubricized:
    2020/11/02
      Vol:
    E104-A No:5
      Page(s):
    814-817

    In this letter, a two-stage QR decomposition scheme based on Givens rotation with novel modified real-value decomposition (RVD) is presented. With the modified RVD applied to the result from complex Givens rotation at first stage, the number of non-zero terms needed to be eliminated by real Givens rotation at second stage decreases greatly and the computational complexity is thereby reduced significantly compared to the decomposition scheme with the conventional RVD. Besides, the proposed scheme is suitable for the hardware design of QR decomposition. Evaluation shows that the proposed QR decomposition scheme is superior to the related works in terms of computational complexity.

  • Evaluation of Temporal Characteristics of Olfactory Displays with Different Structures Open Access

    Masaaki ISEKI  Takamichi NAKAMOTO  

     
    PAPER-Human Communications

      Pubricized:
    2020/09/29
      Vol:
    E104-A No:4
      Page(s):
    744-750

    An olfactory display is a device to present smells. Temporal characteristics of three types of olfactory displays such as one based upon high-speed switching of solenoid valves, desktop-type one based on SAW atomizer and wearable-type one based on SAW atomizer were evaluated using three odorants with different volatilities. The sensory test revealed that the olfactory displays based on SAW atomizer had the presentation speeds faster than that of solenoid valves switching. Especially, the wearable one had an excellent temporal characteristic. These results largely depend on the difference in the odor delivery method. The data obtained in this study provides basic knowledge when we make olfactory contents.

  • Approximate Simultaneous Diagonalization of Matrices via Structured Low-Rank Approximation

    Riku AKEMA  Masao YAMAGISHI  Isao YAMADA  

     
    PAPER-Digital Signal Processing

      Pubricized:
    2020/10/15
      Vol:
    E104-A No:4
      Page(s):
    680-690

    Approximate Simultaneous Diagonalization (ASD) is a problem to find a common similarity transformation which approximately diagonalizes a given square-matrix tuple. Many data science problems have been reduced into ASD through ingenious modelling. For ASD, the so-called Jacobi-like methods have been extensively used. However, the methods have no guarantee to suppress the magnitude of off-diagonal entries of the transformed tuple even if the given tuple has an exact common diagonalizer, i.e., the given tuple is simultaneously diagonalizable. In this paper, to establish an alternative powerful strategy for ASD, we present a novel two-step strategy, called Approximate-Then-Diagonalize-Simultaneously (ATDS) algorithm. The ATDS algorithm decomposes ASD into (Step 1) finding a simultaneously diagonalizable tuple near the given one; and (Step 2) finding a common similarity transformation which diagonalizes exactly the tuple obtained in Step 1. The proposed approach to Step 1 is realized by solving a Structured Low-Rank Approximation (SLRA) with Cadzow's algorithm. In Step 2, by exploiting the idea in the constructive proof regarding the conditions for the exact simultaneous diagonalizability, we obtain an exact common diagonalizer of the obtained tuple in Step 1 as a solution for the original ASD. Unlike the Jacobi-like methods, the ATDS algorithm has a guarantee to find an exact common diagonalizer if the given tuple happens to be simultaneously diagonalizable. Numerical experiments show that the ATDS algorithm achieves better performance than the Jacobi-like methods.

  • Design and VLSI Implementation of a Sorted MMSE QR Decomposition for 4×4 MIMO Detectors

    Lu SUN  Bin WU  Tianchun YE  

     
    LETTER-VLSI Design Technology and CAD

      Pubricized:
    2020/10/12
      Vol:
    E104-A No:4
      Page(s):
    762-767

    In this letter, a low latency, high throughput and hardware efficient sorted MMSE QR decomposition (MMSE-SQRD) for multiple-input multiple-output (MIMO) systems is presented. In contrast to the method of extending the complex matrix to real model and thereafter applying real-valued QR decomposition (QRD), we develop a highly parallel decomposition scheme based on coordinate rotation digital computer (CORDIC) which performs the QRD in complex domain directly and then converting the complex result to its real counterpart. The proposed scheme can greatly improve the processing parallelism and curtail the nullification and sorting procedures. Besides, we also design the corresponding pipelined hardware architecture of the MMSE-SQRD based on highly parallel Givens rotation structure with CORDIC algorithm for 4×4 MIMO detectors. The proposed MMSE-SQRD is implemented in SMIC 55nm CMOS technology achieving up to 50M QRD/s throughput and a latency of 59 clock cycles with only 218 kilo-gates (KG). Compared to the previous works, the proposed design achieves the highest normalized throughput efficiency and lowest processing latency.

  • Transmission Control Method for Data Retention Taking into Account the Low Vehicle Density Environments

    Ichiro GOTO  Daiki NOBAYASHI  Kazuya TSUKAMOTO  Takeshi IKENAGA  Myung LEE  

     
    LETTER-Information Network

      Pubricized:
    2021/01/05
      Vol:
    E104-D No:4
      Page(s):
    508-512

    With the development and spread of Internet of Things (IoT) technology, various kinds of data are now being generated from IoT devices. Some data generated from IoT devices depend on geographical location and time, and we refer to them as spatio-temporal data (STD). Since the “locally produced and consumed” paradigm of STD use is effective for location-dependent applications, the authors have previously proposed a vehicle-based STD retention system. However, in low vehicle density environments, the data retention becomes difficult due to the decrease in the number of data transmissions in this method. In this paper, we propose a new data transmission control method for data retention in the low vehicle density environments.

  • Robust Blind Watermarking Algorithm Based on Contourlet Transform with Singular Value Decomposition

    Lei SONG  Xue-Cheng SUN  Zhe-Ming LU  

     
    LETTER-Cryptography and Information Security

      Pubricized:
    2020/09/11
      Vol:
    E104-A No:3
      Page(s):
    640-643

    In this Letter, we propose a blind and robust multiple watermarking scheme using Contourlet transform and singular value decomposition (SVD). The host image is first decomposed by Contourlet transform. Singular values of Contourlet coefficient blocks are adopted to embed watermark information, and a fast calculation method is proposed to avoid the heavy computation of SVD. The watermark is embedded in both low and high frequency Contourlet coefficients to increase the robustness against various attacks. Moreover, the proposed scheme intrinsically exploits the characteristics of human visual system and thus can ensure the invisibility of the watermark. Simulation results show that the proposed scheme outperforms other related methods in terms of both robustness and execution time.

  • Envy-Free Resource Sharing on a Temporal Network Using a Minimum Cost Circulation Problem

    Ryo HASE  Mitsue IMAHORI  Norihiko SHINOMIYA  

     
    PAPER

      Vol:
    E104-A No:2
      Page(s):
    462-473

    The relationships between producers and consumers have changed radically by the recent growth of sharing economy. Promoting resource sharing can contribute to finding a solution to environmental issues (e.g. reducing food waste, consuming surplus electricity, and so on). Although prosumers have both roles as consumers and suppliers, matching between suppliers and consumers should be determined when the prosumers share resources. Especially, it is important to achieve envy-freeness that is a metric indicating how the number of prosumers feeling unfairness is kept small since the capacity of prosumers to supply resources is limited. Changing resource capacity and demand will make the situation more complex. This paper proposes a resource sharing model based on a temporal network and flows to realize envy-free resource sharing among prosumers. Experimental results demonstrate the deviation of envy among prosumers can be reduced by setting appropriate weights in a flow network.

  • Iterative Carrier Frequency Offset Estimation with Independent Component Analysis in BLE Systems

    Masahiro TAKIGAWA  Takumi TAKAHASHI  Shinsuke IBI  Seiichi SAMPEI  

     
    PAPER-Wireless Communication Technologies

      Pubricized:
    2020/07/14
      Vol:
    E104-B No:1
      Page(s):
    88-98

    This paper proposes iterative carrier frequency offset (CFO) compensation for spatially multiplexed Bluetooth Low Energy (BLE) signals using independent component analysis (ICA). We apply spatial division multiple access (SDMA) to BLE system to deal with massive number of connection requests of BLE devices expected in the future. According to specifications, each BLE peripheral device is assumed to have CFO of up to 150 [kHz] due to hardware impairments. ICA can resolve spatially multiplexed signals even if they include independent CFO. After the ICA separation, the proposed scheme compensates for the CFO. However, the length of the BLE packet preamble is not long enough to obtain accurate CFO estimates. In order to accurately conduct the CFO compensation using the equivalent of a long pilot signal, preamble and a part of estimated data in the previous process are utilized. In addition, we reveal the fact that the independent CFO of each peripheral improves the capability of ICA blind separation. The results confirm that the proposed scheme can effectively compensate for CFO in the range of up to 150[kHz], which is defined as the acceptable value in the BLE specification.

  • Quantitative Evaluation of Software Component Behavior Discovery Approach

    Cong LIU  

     
    LETTER

      Pubricized:
    2020/05/21
      Vol:
    E104-D No:1
      Page(s):
    117-120

    During the execution of software systems, their execution data can be recorded. By fully exploiting these data, software practitioners can discover behavioral models describing the actual execution of the underlying software system. The recorded unstructured software execution data may be too complex, spanning over several days, etc. Applying existing discovery techniques results in spaghetti-like models with no clear structure and no valuable information for comprehension. Starting from the observation that a software system is composed of a set of logical components, Liu et al. propose to decompose the software behavior discovery problem into smaller independent ones by discovering a behavioral model per component in [1]. However, the effectiveness of the proposed approach is not fully evaluated and compared with existing approaches. In this paper, we evaluate the quality (in terms of understandability/complexity) of discovered component behavior models in a quantitative manner. Based on evaluation, we show that this approach can reduce the complexity of the discovered model and gives a better understanding.

  • PCA-LDA Based Color Quantization Method Taking Account of Saliency

    Yoshiaki UEDA  Seiichi KOJIMA  Noriaki SUETAKE  

     
    LETTER-Image

      Vol:
    E103-A No:12
      Page(s):
    1613-1617

    In this letter, we propose a color quantization method based on saliency. In the proposed method, the salient colors are selected as representative colors preferentially by using saliency as weights. Through experiments, we verify the effectiveness of the proposed method.

  • A Simple Depth-Key-Based Image Composition Considering Object Movement in Depth Direction

    Mami NAGOYA  Tomoaki KIMURA  Hiroyuki TSUJI  

     
    LETTER-Computer Graphics

      Vol:
    E103-A No:12
      Page(s):
    1603-1608

    A simple depth-key-based image composition is proposed, which uses two still images with depth information, background and foreground object. The proposed method can place the object at various locations in the background considering the depth in the 3D world coordinate system. The main feature is that a simple algorithm is provided, which enables us to achieve the depthward movement within the camera plane, without being aware of the 3D world coordinate system. Two algorithms are proposed (P-OMDD and O-OMDD), which are based on the pin-hole camera model. As an advantage, camera calibration is not required before applying the algorithm in these methods. Since a single image is used for the object representation, each of the proposed methods has its limitations in terms of fidelity of the composite image. P-OMDD faithfully reproduces the angle at which the object is seen, but the pixels of the hidden surface are missing. On the contrary, O-OMDD can avoid the hidden surface problem, but the angle of the object is fixed, wherever it moves. It is verified through several experiments that, when using O-OMDD, subjectively natural composite images can be obtained under any object movement, in terms of size and position in the camera plane. Future tasks include improving the change in illumination due to positional changes and the partial loss of objects due to noise in depth images.

  • Compressed Sensing Framework Applying Independent Component Analysis after Undersampling for Reconstructing Electroencephalogram Signals Open Access

    Daisuke KANEMOTO  Shun KATSUMATA  Masao AIHARA  Makoto OHKI  

     
    PAPER-Biometrics

      Pubricized:
    2020/06/22
      Vol:
    E103-A No:12
      Page(s):
    1647-1654

    This paper proposes a novel compressed sensing (CS) framework for reconstructing electroencephalogram (EEG) signals. A feature of this framework is the application of independent component analysis (ICA) to remove the interference from artifacts after undersampling in a data processing unit. Therefore, we can remove the ICA processing block from the sensing unit. In this framework, we used a random undersampling measurement matrix to suppress the Gaussian. The developed framework, in which the discrete cosine transform basis and orthogonal matching pursuit were used, was evaluated using raw EEG signals with a pseudo-model of an eye-blink artifact. The normalized mean square error (NMSE) and correlation coefficient (CC), obtained as the average of 2,000 results, were compared to quantitatively demonstrate the effectiveness of the proposed framework. The evaluation results of the NMSE and CC showed that the proposed framework could remove the interference from the artifacts under a high compression ratio.

  • Loss Function Considering Multiple Attributes of a Temporal Sequence for Feed-Forward Neural Networks

    Noriyuki MATSUNAGA  Yamato OHTANI  Tatsuya HIRAHARA  

     
    PAPER-Speech and Hearing

      Pubricized:
    2020/08/31
      Vol:
    E103-D No:12
      Page(s):
    2659-2672

    Deep neural network (DNN)-based speech synthesis became popular in recent years and is expected to soon be widely used in embedded devices and environments with limited computing resources. The key intention of these systems in poor computing environments is to reduce the computational cost of generating speech parameter sequences while maintaining voice quality. However, reducing computational costs is challenging for two primary conventional DNN-based methods used for modeling speech parameter sequences. In feed-forward neural networks (FFNNs) with maximum likelihood parameter generation (MLPG), the MLPG reconstructs the temporal structure of the speech parameter sequences ignored by FFNNs but requires additional computational cost according to the sequence length. In recurrent neural networks, the recursive structure allows for the generation of speech parameter sequences while considering temporal structures without the MLPG, but increases the computational cost compared to FFNNs. We propose a new approach for DNNs to acquire parameters captured from the temporal structure by backpropagating the errors of multiple attributes of the temporal sequence via the loss function. This method enables FFNNs to generate speech parameter sequences by considering their temporal structure without the MLPG. We generated the fundamental frequency sequence and the mel-cepstrum sequence with our proposed method and conventional methods, and then synthesized and subjectively evaluated the speeches from these sequences. The proposed method enables even FFNNs that work on a frame-by-frame basis to generate speech parameter sequences by considering the temporal structure and to generate sequences perceptually superior to those from the conventional methods.

  • An MMT-Based Hierarchical Transmission Module for 4K/120fps Temporally Scalable Video

    Yasuhiro MOCHIDA  Takayuki NAKACHI  Takahiro YAMAGUCHI  

     
    PAPER

      Pubricized:
    2020/06/22
      Vol:
    E103-D No:10
      Page(s):
    2059-2066

    High frame rate (HFR) video is attracting strong interest since it is considered as a next step toward providing Ultra-High Definition video service. For instance, the Association of Radio Industries and Businesses (ARIB) standard, the latest broadcasting standard in Japan, defines a 120 fps broadcasting format. The standard stipulates temporally scalable coding and hierarchical transmission by MPEG Media Transport (MMT), in which the base layer and the enhancement layer are transmitted over different paths for flexible distribution. We have developed the first ever MMT transmitter/receiver module for 4K/120fps temporally scalable video. The module is equipped with a newly proposed encapsulation method of temporally scalable bitstreams with correct boundaries. It is also designed to be tolerant to severe network constraints, including packet loss, arrival timing offset, and delay jitter. We conducted a hierarchical transmission experiment for 4K/120fps temporally scalable video. The experiment demonstrated that the MMT module was successfully fabricated and capable of dealing with severe network constraints. Consequently, the module has excellent potential as a means to support HFR video distribution in various network situations.

  • Block Randomized Singular Value Decomposition on GPUs

    Yuechao LU  Yasuyuki MATSUSHITA  Fumihiko INO  

     
    PAPER-Dependable Computing

      Pubricized:
    2020/06/08
      Vol:
    E103-D No:9
      Page(s):
    1949-1959

    Fast computation of singular value decomposition (SVD) is of great interest in various machine learning tasks. Recently, SVD methods based on randomized linear algebra have shown significant speedup in this regime. For processing large-scale data, computing systems with accelerators like GPUs have become the mainstream approach. In those systems, access to the input data dominates the overall process time; therefore, it is needed to design an out-of-core algorithm to dispatch the computation into accelerators. This paper proposes an accurate two-pass randomized SVD, named block randomized SVD (BRSVD), designed for matrices with a slow-decay singular spectrum that is often observed in image data. BRSVD fully utilizes the power of modern computing system architectures and efficiently processes large-scale data in a parallel and out-of-core fashion. Our experiments show that BRSVD effectively moves the performance bottleneck from data transfer to computation, so that outperforms existing randomized SVD methods in terms of speed with retaining similar accuracy.

61-80hit(945hit)