Multi-modal Knowledge Graph Link Prediction Based on Dual Fusion and Graph Attention Networks
Knowledge graph link prediction aims to predict missing facts within the knowledge graph,addressing the issue of knowledge graph incompleteness.However,existing knowledge graph link prediction methods exhibit certain limitations.Traditional methods are re-stricted to utilizing a single data modality,thereby missing the opportunity to fully leverage the wealth of information provided by diverse data modalities.Moreover,in the context of graph neural networks,entities and relationships are frequently treated as independent elements,often overlooking the varying significance of entity relationships within distinct neighborhoods.To address these problems,a multi-modal knowledge graph link prediction model based on dual-fusion graph attention network is proposed.Firstly,three modalities of image,text and attribute were incorporated.To ensure consistency and synergy among data modal features,a dual-fusion mechanism that combines early and late fusion strategies was devised for multi-modal data amalgamation.To strengthen the fusion of entity relationships and neighborhood relationships in the knowledge graph,consideration is given to the diversity of entities and relationships.The fusion of entity and relationship representations is followed by aggregation through the graph attention network,thereby enhancing the feature representation of the entities.By conducting simulation experiments on four public datasets,specifically FB15K-237,WN18RR,DB15K,and YAGO15K,the results demonstrate the strong performance of the proposed multi-modal knowledge graph link prediction method.