The search functionality is under construction.

Author Search Result

[Author] Ruiyang HUANG(3hit)

1-3hit
  • Super-Node Based Detection of Redundant Ontology Relations

    Yuehang DING  Hongtao YU  Jianpeng ZHANG  Yunjie GU  Ruiyang HUANG  Shize KANG  

     
    LETTER-Data Engineering, Web Information Systems

      Pubricized:
    2019/04/18
      Vol:
    E102-D No:7
      Page(s):
    1400-1403

    Redundant relations refer to explicit relations which can also be deducted implicitly. Although there exist several ontology redundancy elimination methods, they all do not take equivalent relations into consideration. Actually, real ontologies usually contain equivalent relations; their redundancies cannot be completely detected by existing algorithms. Aiming at solving this problem, this paper proposes a super-node based ontology redundancy elimination algorithm. The algorithm consists of super-node transformation and transitive redundancy elimination. During the super-node transformation process, nodes equivalent to each other are transferred into a super-node. Then by deleting the overlapped edges, redundancies relating to equivalent relations are eliminated. During the transitive redundancy elimination process, redundant relations are eliminated by comparing concept nodes' direct and indirect neighbors. Most notably, we proposed a theorem to validate real ontology's irredundancy. Our algorithm outperforms others on both real ontologies and synthetic dynamic ontologies.

  • Network Embedding with Deep Metric Learning

    Xiaotao CHENG  Lixin JI  Ruiyang HUANG  Ruifei CUI  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2018/12/26
      Vol:
    E102-D No:3
      Page(s):
    568-578

    Network embedding has attracted an increasing amount of attention in recent years due to its wide-ranging applications in graph mining tasks such as vertex classification, community detection, and network visualization. Network embedding is an important method to learn low-dimensional representations of vertices in networks, aiming to capture and preserve the network structure. Almost all the existing network embedding methods adopt the so-called Skip-gram model in Word2vec. However, as a bag-of-words model, the skip-gram model mainly utilized the local structure information. The lack of information metrics for vertices in global network leads to the mix of vertices with different labels in the new embedding space. To solve this problem, in this paper we propose a Network Representation Learning method with Deep Metric Learning, namely DML-NRL. By setting the initialized anchor vertices and adding the similarity measure in the training progress, the distance information between different labels of vertices in the network is integrated into the vertex representation, which improves the accuracy of network embedding algorithm effectively. We compare our method with baselines by applying them to the tasks of multi-label classification and data visualization of vertices. The experimental results show that our method outperforms the baselines in all three datasets, and the method has proved to be effective and robust.

  • Chinese Named Entity Recognition Method Based on Dictionary Semantic Knowledge Enhancement

    Tianbin WANG  Ruiyang HUANG  Nan HU  Huansha WANG  Guanghan CHU  

     
    PAPER-Artificial Intelligence, Data Mining

      Pubricized:
    2023/02/15
      Vol:
    E106-D No:5
      Page(s):
    1010-1017

    Chinese Named Entity Recognition is the fundamental technology in the field of the Chinese Natural Language Process. It is extensively adopted into information extraction, intelligent question answering, and knowledge graph. Nevertheless, due to the diversity and complexity of Chinese, most Chinese NER methods fail to sufficiently capture the character granularity semantics, which affects the performance of the Chinese NER. In this work, we propose DSKE-Chinese NER: Chinese Named Entity Recognition based on Dictionary Semantic Knowledge Enhancement. We novelly integrate the semantic information of character granularity into the vector space of characters and acquire the vector representation containing semantic information by the attention mechanism. In addition, we verify the appropriate number of semantic layers through the comparative experiment. Experiments on public Chinese datasets such as Weibo, Resume and MSRA show that the model outperforms character-based LSTM baselines.