The search functionality is under construction.
The search functionality is under construction.

Keyword Search Result

[Keyword] SPAR(322hit)

1-20hit(322hit)

  • Reinforced Voxel-RCNN: An Efficient 3D Object Detection Method Based on Feature Aggregation Open Access

    Jia-ji JIANG  Hai-bin WAN  Hong-min SUN  Tuan-fa QIN  Zheng-qiang WANG  

     
    PAPER-Image Recognition, Computer Vision

      Pubricized:
    2024/04/24
      Vol:
    E107-D No:9
      Page(s):
    1228-1238

    In this paper, the Towards High Performance Voxel-based 3D Object Detection (Voxel-RCNN) three-dimensional (3D) point cloud object detection model is used as the benchmark network. Aiming at the problems existing in the current mainstream 3D point cloud voxelization methods, such as the backbone and the lack of feature expression ability under the bird’s-eye view (BEV), a high-performance voxel-based 3D object detection network (Reinforced Voxel-RCNN) is proposed. Firstly, a 3D feature extraction module based on the integration of inverted residual convolutional network and weight normalization is designed on the 3D backbone. This module can not only well retain more point cloud feature information, enhance the information interaction between convolutional layers, but also improve the feature extraction ability of the backbone network. Secondly, a spatial feature-semantic fusion module based on spatial and channel attention is proposed from a BEV perspective. The mixed use of channel features and semantic features further improves the network’s ability to express point cloud features. In the comparison of experimental results on the public dataset KITTI, the experimental results of this paper are better than many voxel-based methods. Compared with the baseline network, the 3D average accuracy and BEV average accuracy on the three categories of Car, Cyclist, and Pedestrians are improved. Among them, in the 3D average accuracy, the improvement rate of Car category is 0.23%, Cyclist is 0.78%, and Pedestrians is 2.08%. In the context of BEV average accuracy, enhancements are observed: 0.32% for the Car category, 0.99% for Cyclist, and 2.38% for Pedestrians. The findings demonstrate that the algorithm enhancement introduced in this study effectively enhances the accuracy of target category detection.

  • Artifact Removal Using Attention Guided Local-Global Dual-Stream Network for Sparse-View CT Reconstruction Open Access

    Chang SUN  Yitong LIU  Hongwen YANG  

     
    LETTER-Biological Engineering

      Pubricized:
    2024/03/29
      Vol:
    E107-D No:8
      Page(s):
    1105-1109

    Sparse-view CT reconstruction has gained significant attention due to the growing concerns about radiation safety. Although recent deep learning-based image domain reconstruction methods have achieved encouraging performance over iterative methods, effectively capturing intricate details and organ structures while suppressing noise remains challenging. This study presents a novel dual-stream encoder-decoder-based reconstruction network that combines global path reconstruction from the entire image with local path reconstruction from image patches. These two branches interact through an attention module, which enhances visual quality and preserves image details by learning correlations between image features and patch features. Visual and numerical results show that the proposed method has superior reconstruction capabilities to state-of-the-art 180-, 90-, and 45-view CT reconstruction methods.

  • Lower Bounds for the Thickness and the Total Number of Edge Crossings of Euclidean Minimum Weight Laman Graphs and (2,2)-Tight Graphs Open Access

    Yuki KAWAKAMI  Shun TAKAHASHI  Kazuhisa SETO  Takashi HORIYAMA  Yuki KOBAYASHI  Yuya HIGASHIKAWA  Naoki KATOH  

     
    PAPER-Fundamentals of Information Systems

      Pubricized:
    2024/02/16
      Vol:
    E107-D No:6
      Page(s):
    732-740

    We explore the maximum total number of edge crossings and the maximum geometric thickness of the Euclidean minimum-weight (k, ℓ)-tight graph on a planar point set P. In this paper, we show that (10/7-ε)|P| and (11/6-ε)|P| are lower bounds for the maximum total number of edge crossings for any ε > 0 in cases (k,ℓ)=(2,3) and (2,2), respectively. We also show that the lower bound for the maximum geometric thickness is 3 for both cases. In the proofs, we apply the method of arranging isomorphic units regularly. While the method is developed for the proof in case (k,ℓ)=(2,3), it also works for different ℓ.

  • Technology Remapping Approach Using Multi-Gate Reconfigurable Cells for Post-Mask Functional ECO

    Tomohiro NISHIGUCHI  Nobutaka KUROKI  Masahiro NUMA  

     
    PAPER-VLSI Design Technology and CAD

      Pubricized:
    2023/10/10
      Vol:
    E107-A No:3
      Page(s):
    592-599

    This paper proposes multi-gate reconfigurable (RECON) cells and a technology remapping approach using them as spare cells for post-mask functional engineering change orders (ECOs). With the rapid increase in circuit complexity, ECOs often occur in the post-mask stage of LSI designs. To deal with post-mask ECOs at a low cost, only the metal layers are redesigned by making functional changes using spare cells. For this purpose, 2T/4T/6T-RECON cells were proposed as reconfigurable spare cells. However, conventional RECON cells are used to implement single functions, which may result in unused transistors in the cells. In addition, the number of 2T/4T/6T-RECON spare cells used for post-mask ECOs varies greatly depending on the circuit to be implemented and the type of ECO that occurs. Therefore, functional ECOs may fail due to a lack of certain types of RECON cells, even if other types of RECON cells remain. To solve this problem, we propose multi-gate RECON cells that implement multiple functions in a single RECON cell while retaining the layouts of conventional 4T/6T-RECON base cells, and a technology remapping approach using them. The proposed approach not only reduces the number of used spare cells for modifications but also allows the flexible use of spare cells to fix them with less increase in wire length and delay. Experimental results have confirmed that the functional ECO success ratio is increased by 4.8pt on average and the total number of used spare cells is reduced by 5.6% on average. It has also been confirmed that the increase in wire length is reduced by 17.4% on average and the decrease in slack is suppressed by 21.6% on average.

  • Introduction to Compressed Sensing with Python Open Access

    Masaaki NAGAHARA  

     
    INVITED PAPER-Fundamental Theories for Communications

      Pubricized:
    2023/08/15
      Vol:
    E107-B No:1
      Page(s):
    126-138

    Compressed sensing is a rapidly growing research field in signal and image processing, machine learning, statistics, and systems control. In this survey paper, we provide a review of the theoretical foundations of compressed sensing and present state-of-the-art algorithms for solving the corresponding optimization problems. Additionally, we discuss several practical applications of compressed sensing, such as group testing, sparse system identification, and sparse feedback gain design, and demonstrate their effectiveness through Python programs. This survey paper aims to contribute to the advancement of compressed sensing research and its practical applications in various scientific disciplines.

  • Network Traffic Anomaly Detection: A Revisiting to Gaussian Process and Sparse Representation

    Yitu WANG  Takayuki NAKACHI  

     
    PAPER-Communication Theory and Signals

      Pubricized:
    2023/06/27
      Vol:
    E107-A No:1
      Page(s):
    125-133

    Seen from the Internet Service Provider (ISP) side, network traffic monitoring is an indispensable part during network service provisioning, which facilitates maintaining the security and reliability of the communication networks. Among the numerous traffic conditions, we should pay extra attention to traffic anomaly, which significantly affects the network performance. With the advancement of Machine Learning (ML), data-driven traffic anomaly detection algorithms have established high reputation due to the high accuracy and generality. However, they are faced with challenges on inefficient traffic feature extraction and high computational complexity, especially when taking the evolving property of traffic process into consideration. In this paper, we proposed an online learning framework for traffic anomaly detection by embracing Gaussian Process (GP) and Sparse Representation (SR) in two steps: 1). To extract traffic features from past records, and better understand these features, we adopt GP with a special kernel, i.e., mixture of Gaussian in the spectral domain, which makes it possible to more accurately model the network traffic for improving the performance of traffic anomaly detection. 2). To combat noise and modeling error, observing the inherent self-similarity and periodicity properties of network traffic, we manually design a feature vector, based on which SR is adopted to perform robust binary classification. Finally, we demonstrate the superiority of the proposed framework in terms of detection accuracy through simulation.

  • Sparse Reconstruction and Resolution Improvement of Synthetic Aperture Radar with Low Computational Complexity Using Deconvolution ISTA

    Masanori GOCHO  

     
    PAPER

      Pubricized:
    2023/07/27
      Vol:
    E106-B No:12
      Page(s):
    1363-1371

    Synthetic aperture radar (SAR) is a device for observing the ground surface and is one of the important technologies in the field of microwave remote sensing. In SAR observation, a platform equipped with a small-aperture antenna flies in a straight line and continuously radiates pulse waves to the ground during the flight. After that, by synthesizing the series of observation data obtained during the flight, one realize high-resolution ground surface observation. In SAR observation, there are two spatial resolutions defined in the range and azimuth directions and they are limited by the bandwidth of the SAR system. The purpose of this study is to improve the resolution of SAR by sparse reconstruction. In particular, we aim to improve the resolution of SAR without changing the frequency parameters. In this paper, we propose to improve the resolution of SAR using the deconvolution iterative shrinkage-thresholding algorithm (ISTA) and verify the proposed method by carrying out an experimental analysis using an actual SAR dataset. Experimental results show that the proposed method can improve the resolution of SAR with low computational complexity.

  • A Unified Design of Generalized Moreau Enhancement Matrix for Sparsity Aware LiGME Models

    Yang CHEN  Masao YAMAGISHI  Isao YAMADA  

     
    PAPER-Digital Signal Processing

      Pubricized:
    2023/02/14
      Vol:
    E106-A No:8
      Page(s):
    1025-1036

    In this paper, we propose a unified algebraic design of the generalized Moreau enhancement matrix (GME matrix) for the Linearly involved Generalized-Moreau-Enhanced (LiGME) model. The LiGME model has been established as a framework to construct linearly involved nonconvex regularizers for sparsity (or low-rank) aware estimation, where the design of GME matrix is a key to guarantee the overall convexity of the model. The proposed design is applicable to general linear operators involved in the regularizer of the LiGME model, and does not require any eigendecomposition or iterative computation. We also present an application of the LiGME model with the proposed GME matrix to a group sparsity aware least squares estimation problem. Numerical experiments demonstrate the effectiveness of the proposed GME matrix in the LiGME model.

  • Segmentation of Optic Disc and Optic Cup Based on Two-Layer Level Set with Sparse Shape Prior Constraint in Fundus Images

    Siqi WANG  Ming XU  Xiaosheng YU  Chengdong WU  

     
    LETTER-Computer Graphics

      Pubricized:
    2023/01/16
      Vol:
    E106-A No:7
      Page(s):
    1020-1024

    Glaucoma is a common high-incidence eye disease. The detection of the optic cup and optic disc in fundus images is one of the important steps in the clinical diagnosis of glaucoma. However, the fundus images are generally intensity inhomogeneity, and complex organizational structure, and are disturbed by blood vessels and lesions. In order to extract the optic disc and optic cup regions more accurately, we propose a segmentation method of the optic disc and optic cup in fundus image based on distance regularized two-layer level with sparse shape prior constraint. The experimental results show that our method can segment the optic disc and optic cup region more accurately and obtain satisfactory results.

  • L0-Norm Based Adaptive Equalization with PMSER Criterion for Underwater Acoustic Communications

    Tian FANG  Feng LIU  Conggai LI  Fangjiong CHEN  Yanli XU  

     
    LETTER-Communication Theory and Signals

      Pubricized:
    2022/12/06
      Vol:
    E106-A No:6
      Page(s):
    947-951

    Underwater acoustic channels (UWA) are usually sparse, which can be exploited for adaptive equalization to improve the system performance. For the shallow UWA channels, based on the proportional minimum symbol error rate (PMSER) criterion, the adaptive equalization framework requires the sparsity selection. Since the sparsity of the L0 norm is stronger than that of the L1, we choose it to achieve better convergence. However, because the L0 norm leads to NP-hard problems, it is difficult to find an efficient solution. In order to solve this problem, we choose the Gaussian function to approximate the L0 norm. Simulation results show that the proposed scheme obtains better performance than the L1 based counterpart.

  • Group Sparse Reduced Rank Tensor Regression for Micro-Expression Recognition

    Sunan LI  Yuan ZONG  Cheng LU  Chuangan TANG  Yan ZHAO  

     
    LETTER-Human-computer Interaction

      Pubricized:
    2023/01/05
      Vol:
    E106-D No:4
      Page(s):
    575-578

    To overcome the challenge in micro-expression recognition that it only emerge in several small facial regions with low intensity, some researchers proposed facial region partition mechanisms and introduced group sparse learning methods for feature selection. However, such methods have some shortcomings, including the complexity of region division and insufficient utilization of critical facial regions. To address these problems, we propose a novel Group Sparse Reduced Rank Tensor Regression (GSRRTR) to transform the fearure matrix into a tensor by laying blocks and features in different dimensions. So we can process grids and texture features separately and avoid interference between grids and features. Furthermore, with the use of Tucker decomposition, the feature tensor can be decomposed into a product of core tensor and a set of matrix so that the number of parameters and the computational complexity of the scheme will decreased. To evaluate the performance of the proposed micro-expression recognition method, extensive experiments are conducted on two micro expression databases: CASME2 and SMIC. The experimental results show that the proposed method achieves comparable recognition rate with less parameters than state-of-the-art methods.

  • Deep Learning of Damped AMP Decoding Networks for Sparse Superposition Codes via Annealing

    Toshihiro YOSHIDA  Keigo TAKEUCHI  

     
    PAPER-Communication Theory and Signals

      Pubricized:
    2022/07/22
      Vol:
    E106-A No:3
      Page(s):
    414-421

    This paper addresses short-length sparse superposition codes (SSCs) over the additive white Gaussian noise channel. Damped approximate message-passing (AMP) is used to decode short SSCs with zero-mean independent and identically distributed Gaussian dictionaries. To design damping factors in AMP via deep learning, this paper constructs deep-unfolded damped AMP decoding networks. An annealing method for deep learning is proposed for designing nearly optimal damping factors with high probability. In annealing, damping factors are first optimized via deep learning in the low signal-to-noise ratio (SNR) regime. Then, the obtained damping factors are set to the initial values in stochastic gradient descent, which optimizes damping factors for slightly larger SNR. Repeating this annealing process designs damping factors in the high SNR regime. Numerical simulations show that annealing mitigates fluctuation in learned damping factors and outperforms exhaustive search based on an iteration-independent damping factor.

  • Present Status and Prospect of Graphene Interconnect Applications

    Kazuyoshi UENO  

     
    PAPER

      Pubricized:
    2022/04/21
      Vol:
    E105-C No:10
      Page(s):
    572-577

    Graphene has been expected as an alternative material for copper interconnects in which resistance increases and reliability deteriorates in nanoscale. While the principle advantages are verified by simulations and experiments, they have not been put into practical use due to the immaturity of the manufacturing process leading to mass production. On the other hand, recent steady progress in the fabrication process has increased the possibility of practical application. In this paper, I will review the recent advances and the latest prospects for conductor applications of graphene centered on interconnects. The possibility of further application utilizing the unique characteristics of graphene is discussed.

  • Bridging between Soft and Hard Thresholding by Scaling

    Katsuyuki HAGIWARA  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2022/06/09
      Vol:
    E105-D No:9
      Page(s):
    1529-1536

    This study considered an extension of a sparse regularization method with scaling, especially in thresholding methods that are simple and typical examples of sparse modeling. In this study, in the setting of a non-parametric orthogonal regression problem, we developed and analyzed a thresholding method in which soft thresholding estimators are independently expanded by empirical scaling values. The scaling values have a common hyper-parameter that is an order of expansion of an ideal scaling value to achieve hard thresholding. We simply refer to this estimator as a scaled soft thresholding estimator. The scaled soft thresholding method is a bridge method between soft and hard thresholding methods. This new estimator is indeed consistent with an adaptive LASSO estimator in the orthogonal case; i.e., it is thus an another derivation of an adaptive LASSO estimator. It is a general method that includes soft thresholding and non-negative garrote as special cases. We subsequently derived the degree of freedom of the scaled soft thresholding in calculating the Stein's unbiased risk estimate. We found that it is decomposed into the degree of freedom of soft thresholding and the remainder term connecting to the hard thresholding. As the degree of freedom reflects the degree of over-fitting, this implies that the scaled soft thresholding has an another source of over-fitting in addition to the number of un-removed components. The theoretical result was verified by a simple numerical example. In this process, we also focused on the non-monotonicity in the above remainder term of the degree of freedom and found that, in a sparse and large sample setting, it is mainly caused by useless components that are not related to the target function.

  • Spectral Reflectance Reconstruction Based on BP Neural Network and the Improved Sparrow Search Algorithm

    Lu ZHANG  Chengqun WANG  Mengyuan FANG  Weiqiang XU  

     
    LETTER-Neural Networks and Bioengineering

      Pubricized:
    2022/01/24
      Vol:
    E105-A No:8
      Page(s):
    1175-1179

    To solve the problem of metamerism in the color reproduction process, various spectral reflectance reconstruction methods combined with neural network have been proposed in recent years. However, these methods are generally sensitive to initial values and can easily converge to local optimal solutions, especially on small data sets. In this paper, we propose a spectral reflectance reconstruction algorithm based on the Back Propagation Neural Network (BPNN) and an improved Sparrow Search Algorithm (SSA). In this algorithm, to solve the problem that BPNN is sensitive to initial values, we propose to use SSA to initialize BPNN, and we use the sine chaotic mapping to further improve the stability of the algorithm. In the experiment, we tested the proposed algorithm on the X-Rite ColorChecker Classic Mini Chart which contains 24 colors, the results show that the proposed algorithm has significantly better performance compared to other algorithms and moreover it can meet the needs of spectral reflectance reconstruction on small data sets. Code is avaible at https://github.com/LuraZhang/spectral-reflectance-reconsctuction.

  • SeCAM: Tightly Accelerate the Image Explanation via Region-Based Segmentation

    Phong X. NGUYEN  Hung Q. CAO  Khang V. T. NGUYEN  Hung NGUYEN  Takehisa YAIRI  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2022/05/11
      Vol:
    E105-D No:8
      Page(s):
    1401-1417

    In recent years, there has been an increasing trend of applying artificial intelligence in many different fields, which has a profound and direct impact on human life. Consequently, this raises the need to understand the principles of model making predictions. Since most current high-precision models are black boxes, neither the AI scientist nor the end-user profoundly understands what is happening inside these models. Therefore, many algorithms are studied to explain AI models, especially those in the image classification problem in computer vision such as LIME, CAM, GradCAM. However, these algorithms still have limitations, such as LIME's long execution time and CAM's confusing interpretation of concreteness and clarity. Therefore, in this paper, we will propose a new method called Segmentation - Class Activation Mapping (SeCAM)/ This method combines the advantages of these algorithms above while at simultaneously overcoming their disadvantages. We tested this algorithm with various models, including ResNet50, InceptionV3, and VGG16 from ImageNet Large Scale Visual Recognition Challenge (ILSVRC) data set. Outstanding results were achieved when the algorithm has met all the requirements for a specific explanation in a remarkably short space of time.

  • A Large-Scale SCMA Codebook Optimization and Codeword Allocation Method

    Shiqing QIAN  Wenping GE  Yongxing ZHANG  Pengju ZHANG  

     
    PAPER-Fundamental Theories for Communications

      Pubricized:
    2021/12/24
      Vol:
    E105-B No:7
      Page(s):
    788-796

    Sparse code division multiple access (SCMA) is a non-orthogonal multiple access (NOMA) technology that can improve frequency band utilization and allow many users to share quite a few resource elements (REs). This paper uses the modulation of lattice theory to develop a systematic construction procedure for the design of SCMA codebooks under Gaussian channel environments that can achieve near-optimal designs, especially for cases that consider large-scale SCMA parameters. However, under the condition of large-scale SCMA parameters, the mother constellation (MC) points will overlap, which can be solved by the method of the partial dimensions transformation (PDT). More importantly, we consider the upper bounded error probability of the signal transmission in the AWGN channels, and design a codeword allocation method to reduce the inter symbol interference (ISI) on the same RE. Simulation results show that under different codebook sizes and different overload rates, using two different message passing algorithms (MPA) to verify, the codebook proposed in this paper has a bit error rate (BER) significantly better than the reference codebooks, moreover the convergence time does not exceed that of the reference codebooks.

  • Performance Evaluation of Bluetooth Low Energy Positioning Systems When Using Sparse Training Data

    Tetsuya MANABE  Kosuke OMURA  

     
    PAPER

      Pubricized:
    2021/11/01
      Vol:
    E105-A No:5
      Page(s):
    778-786

    This paper evaluates the bluetooth low energy (BLE) positioning systems using the sparse-training data through the comparison experiments. The sparse-training data is extracted from the database including enough data for realizing the highly accurate and precise positioning. First, we define the sparse-training data, i.e., the data collection time and the number of smartphones, directions, beacons, and reference points, on BLE positioning systems. Next, the positioning performance evaluation experiments are conducted in two indoor environments, that is, an indoor corridor as a one-dimensionally spread environment and a hall as a twodimensionally spread environment. The algorithms for comparison are the conventional fingerprint algorithm and the hybrid algorithm (the authors already proposed, and combined the proximity algorithm and the fingerprint algorithm). Based on the results, we confirm that the hybrid algorithm performs well in many cases even when using sparse-training data. Consequently, the robustness of the hybrid algorithm, that the authors already proposed for the sparse-training data, is shown.

  • Fully Connected Imaging Network for Near-Field Synthetic Aperture Interferometric Radiometer

    Zhimin GUO  Jianfei CHEN  Sheng ZHANG  

     
    LETTER-Artificial Intelligence, Data Mining

      Pubricized:
    2022/02/09
      Vol:
    E105-D No:5
      Page(s):
    1120-1124

    Millimeter wave synthetic aperture interferometric radiometers (SAIR) are very powerful instruments, which can effectively realize high-precision imaging detection. However due to the existence of interference factor and complex near-field error, the imaging effect of near-field SAIR is usually not ideal. To achieve better imaging results, a new fully connected imaging network (FCIN) is proposed for near-field SAIR. In FCIN, the fully connected network is first used to reconstruct the image domain directly from the visibility function, and then the residual dense network is used for image denoising and enhancement. The simulation results show that the proposed FCIN method has high imaging accuracy and shorten imaging time.

  • Dual Self-Guided Attention with Sparse Question Networks for Visual Question Answering

    Xiang SHEN  Dezhi HAN  Chin-Chen CHANG  Liang ZONG  

     
    PAPER-Natural Language Processing

      Pubricized:
    2022/01/06
      Vol:
    E105-D No:4
      Page(s):
    785-796

    Visual Question Answering (VQA) is multi-task research that requires simultaneous processing of vision and text. Recent research on the VQA models employ a co-attention mechanism to build a model between the context and the image. However, the features of questions and the modeling of the image region force irrelevant information to be calculated in the model, thus affecting the performance. This paper proposes a novel dual self-guided attention with sparse question networks (DSSQN) to address this issue. The aim is to avoid having irrelevant information calculated into the model when modeling the internal dependencies on both the question and image. Simultaneously, it overcomes the coarse interaction between sparse question features and image features. First, the sparse question self-attention (SQSA) unit in the encoder calculates the feature with the highest weight. From the self-attention learning of question words, the question features of larger weights are reserved. Secondly, sparse question features are utilized to guide the focus on image features to obtain fine-grained image features, and to also prevent irrelevant information from being calculated into the model. A dual self-guided attention (DSGA) unit is designed to improve modal interaction between questions and images. Third, the sparse question self-attention of the parameter δ is optimized to select these question-related object regions. Our experiments with VQA 2.0 benchmark datasets demonstrate that DSSQN outperforms the state-of-the-art methods. For example, the accuracy of our proposed model on the test-dev and test-std is 71.03% and 71.37%, respectively. In addition, we show through visualization results that our model can pay more attention to important features than other advanced models. At the same time, we also hope that it can promote the development of VQA in the field of artificial intelligence (AI).

1-20hit(322hit)