• Complex
  • Title
  • Keyword
  • Abstract
  • Scholars
  • Journal
  • ISSN
  • Conference
搜索
High Impact Results & Cited Count Trend for Year Keyword Cloud and Partner Relationship

Query:

学者姓名:杨金福

Refining:

Source

Submit Unfold

Co-Author

Submit Unfold

Clean All

Sort by:
Default
  • Default
  • Title
  • Year
  • WOS Cited Count
  • Impact factor
  • Ascending
  • Descending
< Page ,Total 16 >
Sparse-view planar 3D reconstruction method based on hierarchical token pooling Transformer SCIE
期刊论文 | 2025 , 174 | APPLIED SOFT COMPUTING
Abstract&Keyword Cite

Abstract :

Sparse-view planar 3D reconstruction aims to recover scene information from limited camera frames, which poses a fundamental problem in computer vision. Although previous methods have made significant improvements in this field, they have not adequately considered the multi-scale properties of the surrounding environment, thus limiting the reconstruction performance. Additionally, the conventional feed-forward network in the vanilla Transformer is constructed using fully connected layers, lacking the ability to capture local information from image features. To address these two problems, this paper proposes a sparse-view planar 3D reconstruction method based on hierarchical token pooling Transformer (i.e. HTP-Formers). Specifically, we utilize average pooling layers with various ratios in Transformer model to capture multi-scale features. Subsequently, we propose a depth-wise convolution based inverted residual feed-forward network to enhance local information extraction performance at negligible computational cost. To demonstrate the effectiveness of HTP-Formers on planar 3D reconstruction tasks, we thoroughly evaluate the proposed model on Matterport3D public dataset. Especially, HTP-Formers improves performance by 6.1% and 18.3% in translational and rotational errors, respectively, outperforming most existing planar 3D reconstruction methods in terms of planar correspondence inference and relative camera pose estimation.

Keyword :

Feed-forward network Feed-forward network Planar 3D reconstruction Planar 3D reconstruction Hierarchical token pooling Hierarchical token pooling Depth-wise convolution Depth-wise convolution

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Zhang, Jiahui , Yang, Jinfu , Fu, Fuji et al. Sparse-view planar 3D reconstruction method based on hierarchical token pooling Transformer [J]. | APPLIED SOFT COMPUTING , 2025 , 174 .
MLA Zhang, Jiahui et al. "Sparse-view planar 3D reconstruction method based on hierarchical token pooling Transformer" . | APPLIED SOFT COMPUTING 174 (2025) .
APA Zhang, Jiahui , Yang, Jinfu , Fu, Fuji , Ma, Jiaqi . Sparse-view planar 3D reconstruction method based on hierarchical token pooling Transformer . | APPLIED SOFT COMPUTING , 2025 , 174 .
Export to NoteExpress RIS BibTex
Deep support vector data description based on correntropy for few-shot anomaly detection SCIE
期刊论文 | 2025 , 160 | DIGITAL SIGNAL PROCESSING
Abstract&Keyword Cite

Abstract :

Few-shot anomaly detection aims to identify samples that differ from or are abnormal compared to normal samples using a limited number of training samples. Training deep one-class classifiers requires a large number of normal samples, and it may not even require any anomaly samples. Therefore, deep one-class classification can address the issue of class imbalance in few-shot anomaly detection. However, existing methods based on deep one-class classification lack sufficient utilization of the non-linea r relationships and local feature differences among input samples in high-dimensional feature spaces when addressing class imbalance issues. Moreover, they are sensitive to noise and anomaly samples due to the use of the Euclidean distance loss function. To address these limitations, we propose a Deep support vector data description based on Correntropy for Few-Shot Anomaly Detection (DC-FSAD). Specifically, we introduce an improved loss function that replaces the Euclidean distance loss function in deep one-class classification with correntropy. By utilizing correntropy as a measure of similarity, the new loss function can better capture the non-linear relationships among input samples in high-dimensional feature spaces and fully exploit the local feature differences among samples. Additionally, the width parameter of correntropy can be adaptively adjusted to enhance robustness to noise and anomaly samples. By formulating a new optimization problem and leveraging semi-quadratic optimization techniques, our method achieves a tighter hyper-spherical boundary for accurately describing the distribution of normal samples. Experimental results demonstrate that our proposed method outperforms state-of-the-art methods on benchmark datasets.

Keyword :

Deep support vector data description Deep support vector data description Correntropy Correntropy Few-shot anomaly detection Few-shot anomaly detection

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Shang, Qingzhen , Yang, Jinfu , Fu, Fuji et al. Deep support vector data description based on correntropy for few-shot anomaly detection [J]. | DIGITAL SIGNAL PROCESSING , 2025 , 160 .
MLA Shang, Qingzhen et al. "Deep support vector data description based on correntropy for few-shot anomaly detection" . | DIGITAL SIGNAL PROCESSING 160 (2025) .
APA Shang, Qingzhen , Yang, Jinfu , Fu, Fuji , Ma, Jiaqi . Deep support vector data description based on correntropy for few-shot anomaly detection . | DIGITAL SIGNAL PROCESSING , 2025 , 160 .
Export to NoteExpress RIS BibTex
Consistency knowledge distillation based on similarity attribute graph guidance SCIE
期刊论文 | 2025 , 269 | EXPERT SYSTEMS WITH APPLICATIONS
Abstract&Keyword Cite

Abstract :

Knowledge Distillation (KD) transfers knowledge from a larger and well-trained teacher network to a smaller and trainable student network. Existing KD methods are divided into response-based KD and feature-based KD according to the type of knowledge. However, these two types of methods typically only consider the discrepancy between the outputs (logits and features, etc.) of the teacher network and the student network during distillation, disregarding the uniqueness of the distribution of outputs in the feature space. Optimizing the distribution of outputs in the feature space while achieving consistent knowledge transfer is crucial for enhancing the learning of the student network. Inspired by this, we propose a framework, Consistency Knowledge Distillation based on Similarity Attribute Graph Guidance (CKD-SAG2), which extracts attribute knowledge from the similarity attribute graph and employs consistency distillation to transfer knowledge. Specifically, we design a Similarity Attribute Graph Building (SimAGB) module, which constructs similarity attribute graphs based on sample embeddings, class centers, and their relevance at the feature level to explore the unique distribution of sample embeddings. Then, we propose a Consistency Knowledge Distillation (ConKD) module that obtains attribute knowledge based on the similarity attribute graph and introduces center loss and attribute loss to achieve consistency transfer of knowledge between the teacher network and the student network. Detailed experiments are conducted on three visual tasks (image classification, object detection and semantic segmentation), which fully demonstrate the effectiveness and superiority of the proposed framework.

Keyword :

Consistency loss Consistency loss Similarity attribute graph Similarity attribute graph Attribute knowledge Attribute knowledge Knowledge distillation Knowledge distillation Feature distribution Feature distribution

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Ma, Jiaqi , Yang, Jinfu , Fu, Fuji et al. Consistency knowledge distillation based on similarity attribute graph guidance [J]. | EXPERT SYSTEMS WITH APPLICATIONS , 2025 , 269 .
MLA Ma, Jiaqi et al. "Consistency knowledge distillation based on similarity attribute graph guidance" . | EXPERT SYSTEMS WITH APPLICATIONS 269 (2025) .
APA Ma, Jiaqi , Yang, Jinfu , Fu, Fuji , Zhang, Jiahui . Consistency knowledge distillation based on similarity attribute graph guidance . | EXPERT SYSTEMS WITH APPLICATIONS , 2025 , 269 .
Export to NoteExpress RIS BibTex
An effective two-stage channel pruning method based on two-dimensional information entropy SCIE
期刊论文 | 2024 , 54 (17-18) , 8491-8504 | APPLIED INTELLIGENCE
Abstract&Keyword Cite

Abstract :

Channel pruning can reduce the number of neural network parameters and computational cost by eliminating redundant channels, its main purpose is to adapt to resource constrained devices. Evaluation-based global pruning and network search-based pruning are two common methods of channel pruning. However, the network architecture pruned by the global mask is often not optimal, while the method that directly searches for the optimal architecture will introduce a large number of hyperparameters, which greatly increases the training cost. In this paper, we propose a novel Two-dimensional information Entropy based Channel Pruning method (TECP). The pruning process consists of two steps. First, a global mask pruning scheme is employed to obtained a pre-pruning model. Then, the two-dimensional information entropy is calculated by using feature maps of dense network to adjust the pre-pruning model adaptively to get a compact network. Moreover, the entropy values are used to determine the minimum number of reserved channels per layer based on to avoid the imbalance of network architecture and the layer collapse caused by global pruning. Extensive experiments with a variety of networks on several datasets clearly demonstrate the effectiveness of our proposed TECP method. For example, results show that on CIFAR-10, the compressed model achieves comparable accuracy to the original model, but with a significantly lower number of parameters (44.29% for ResNet-20 and 46.79% for VGG-16). This is beneficial for industrial deployment. And experimental results also show that TECP method obtain the better performance compared with state-of-the-art method.

Keyword :

Network architecture Network architecture Two-dimensional information entropy Two-dimensional information entropy Channel pruning Channel pruning Feature map Feature map

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Xu, Yifei , Yang, Jinfu , Wang, Runshi et al. An effective two-stage channel pruning method based on two-dimensional information entropy [J]. | APPLIED INTELLIGENCE , 2024 , 54 (17-18) : 8491-8504 .
MLA Xu, Yifei et al. "An effective two-stage channel pruning method based on two-dimensional information entropy" . | APPLIED INTELLIGENCE 54 . 17-18 (2024) : 8491-8504 .
APA Xu, Yifei , Yang, Jinfu , Wang, Runshi , Li, Haoqing . An effective two-stage channel pruning method based on two-dimensional information entropy . | APPLIED INTELLIGENCE , 2024 , 54 (17-18) , 8491-8504 .
Export to NoteExpress RIS BibTex
Few-shot classification based on manifold metric learning SCIE
期刊论文 | 2024 , 33 (1) | JOURNAL OF ELECTRONIC IMAGING
Abstract&Keyword Cite

Abstract :

Few-shot classification aims to classify samples with a limited quantity of labeled training data, and it can be widely applied in practical scenarios such as wastewater treatment plants and healthcare. Compared with traditional methods, existing deep metric-based algorithms have excelled in few-shot classification tasks, but some issues need to be further investigated. While current standard convolutional networks can extract expressive depth features, they do not fully exploit the relationships among input sample attributes. Two problems are included here: (1) how to extract more expressive features and transform them into attributes, and (2) how to obtain the optimal combination of sample class attributes. This paper proposes a few-shot classification method based on manifold metric learning (MML) with feature space embedded in symmetric positive definite (SPD) manifolds to overcome the above limitations. First, significant features are extracted using the proposed joint dynamic convolution module. Second, the definition and properties of Riemannian popular strictly convex geodesics are used to minimize the proposed MML loss function and obtain the optimal attribute correlation matrix A. We theoretically prove that the MML is popularly strictly convex in the SPD and obtain the global optimal solution in the closed space. Extensive experimental results on popular datasets show that our proposed approach outperforms other state-of-the-art methods.

Keyword :

dynamic convolution dynamic convolution metric learning metric learning few-shot classification few-shot classification symmetric positive definite manifold symmetric positive definite manifold

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Shang, Qingzhen , Yang, Jinfu , Ma, Jiaqi et al. Few-shot classification based on manifold metric learning [J]. | JOURNAL OF ELECTRONIC IMAGING , 2024 , 33 (1) .
MLA Shang, Qingzhen et al. "Few-shot classification based on manifold metric learning" . | JOURNAL OF ELECTRONIC IMAGING 33 . 1 (2024) .
APA Shang, Qingzhen , Yang, Jinfu , Ma, Jiaqi , Zhang, Jiahui . Few-shot classification based on manifold metric learning . | JOURNAL OF ELECTRONIC IMAGING , 2024 , 33 (1) .
Export to NoteExpress RIS BibTex
Monocular visual-inertial odometry leveraging point-line features with structural constraints SCIE
期刊论文 | 2023 , 40 (2) , 647-661 | VISUAL COMPUTER
WoS CC Cited Count: 3
Abstract&Keyword Cite

Abstract :

Structural geometry constraints, such as perpendicularity, parallelism and coplanarity, are widely existing in man-made scene, especially in Manhattan scene. By fully exploiting these structural properties, we propose a monocular visual-inertial odometry (VIO) using point and line features with structural constraints. First, a coarse-to-fine vanishing points estimation method with line segment consistency verification is presented to classify lines into structural and non-structural lines accurately with less computation cost. Then, to get precise estimation of camera pose and the position of 3D landmarks, a cost function which combines structural line constraints with feature reprojection residual and inertial measurement unit residual is minimized under a sliding window framework. For geometric representation of lines, Plucker coordinates and orthonormal representation are utilized for 3D line transformation and non-linear optimization respectively. Sufficient evaluations are conducted using two public datasets to verify that the proposed system can effectively enhance the localization accuracy and robustness than other existing state-of-the-art VIO systems with acceptable time consumption.

Keyword :

Structural line Structural line Vanishing point Vanishing point Structural constraints Structural constraints Visual-inertial odometry Visual-inertial odometry

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Zhang, Jiahui , Yang, Jinfu , Ma, Jiaqi . Monocular visual-inertial odometry leveraging point-line features with structural constraints [J]. | VISUAL COMPUTER , 2023 , 40 (2) : 647-661 .
MLA Zhang, Jiahui et al. "Monocular visual-inertial odometry leveraging point-line features with structural constraints" . | VISUAL COMPUTER 40 . 2 (2023) : 647-661 .
APA Zhang, Jiahui , Yang, Jinfu , Ma, Jiaqi . Monocular visual-inertial odometry leveraging point-line features with structural constraints . | VISUAL COMPUTER , 2023 , 40 (2) , 647-661 .
Export to NoteExpress RIS BibTex
Dense Face Network: A Dense Face Detector Based on Global Context and Visual Attention Mechanism
期刊论文 | 2022 , 19 (3) , 247-256 | MACHINE INTELLIGENCE RESEARCH
Abstract&Keyword Cite

Abstract :

Face detection has achieved tremendous strides thanks to convolutional neural networks. However, dense face detection remains an open challenge due to large face scale variation, tiny faces, and serious occlusion. This paper presents a robust, dense face detector using global context and visual attention mechanisms which can significantly improve detection accuracy. Specifically, a global context fusion module with top-down feedback is proposed to improve the ability to identify tiny faces. Moreover, a visual attention mechanism is employed to solve the problem of occlusion. Experimental results on the public face datasets WIDER FACE and FDDB demonstrate the effectiveness of the proposed method.

Keyword :

Face detection Face detection global context global context deep learning deep learning computer vision computer vision attention mechanism attention mechanism

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Song, Lin , Yang, Jin-Fu , Shang, Qing-Zhen et al. Dense Face Network: A Dense Face Detector Based on Global Context and Visual Attention Mechanism [J]. | MACHINE INTELLIGENCE RESEARCH , 2022 , 19 (3) : 247-256 .
MLA Song, Lin et al. "Dense Face Network: A Dense Face Detector Based on Global Context and Visual Attention Mechanism" . | MACHINE INTELLIGENCE RESEARCH 19 . 3 (2022) : 247-256 .
APA Song, Lin , Yang, Jin-Fu , Shang, Qing-Zhen , Li, Ming-Ai . Dense Face Network: A Dense Face Detector Based on Global Context and Visual Attention Mechanism . | MACHINE INTELLIGENCE RESEARCH , 2022 , 19 (3) , 247-256 .
Export to NoteExpress RIS BibTex
Cross-modal Video Moment Retrieval Based on Enhancing Significant Features
期刊论文 | 2022 , 44 (12) , 4395-4404 | JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY
Abstract&Keyword Cite

Abstract :

With the continuous development of video acquisition equipment and technology, the number of videos has grown rapidly. It is a challenging task in video retrieval to find target video moments accurately in massive videos. Cross-modal video moment retrieval is to find a moment matching the query from the video database. Existing works focus mostly on matching the text with the moment, while ignoring the context content in the adjacent moment. As a result, there exists the problem of insufficient expression of feature relation. In this paper, a novel moment retrieval network is proposed, which highlights the significant features through residual channel attention. At the same time, a temporal adjacent network is designed to capture the context information of the adjacent moment. Experimental results show that the proposed method achieves better performance than the mainstream candidate matching based and video- text features relation based methods.

Keyword :

Temporal adjacent network Temporal adjacent network Feature relationship Feature relationship Cross-modal video moment retrieval Cross-modal video moment retrieval Residual channel attention Residual channel attention

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Yang Jinfu , Liu Yubin , Song Lin et al. Cross-modal Video Moment Retrieval Based on Enhancing Significant Features [J]. | JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY , 2022 , 44 (12) : 4395-4404 .
MLA Yang Jinfu et al. "Cross-modal Video Moment Retrieval Based on Enhancing Significant Features" . | JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY 44 . 12 (2022) : 4395-4404 .
APA Yang Jinfu , Liu Yubin , Song Lin , Yan Xue . Cross-modal Video Moment Retrieval Based on Enhancing Significant Features . | JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY , 2022 , 44 (12) , 4395-4404 .
Export to NoteExpress RIS BibTex
Eliminating short-term dynamic elements for robust visual simultaneous localization and mapping using a coarse-to-fine strategy SCIE
期刊论文 | 2022 , 31 (5) | JOURNAL OF ELECTRONIC IMAGING
WoS CC Cited Count: 1
Abstract&Keyword Cite

Abstract :

Visual simultaneous localization and mapping (VSLAM) is one of the foremost principal technologies for intelligent robots to implement environment perception. Many research works have focused on proposing comprehensive and integrated systems based on the static environment assumption. However, the elements whose motion status changes frequently, namely short-term dynamic elements, can significantly affect the system performance. Therefore, it is extremely momentous to cope with short-term dynamic elements to make the VSLAM system more adaptable to dynamic scenes. This paper proposes a coarse-to-fine elimination strategy for short-term dynamic elements based on motion status check (MSC) and feature points update (FPU). First, an object detection module is designed to obtain semantic information and screen out the potential short-term dynamic elements. And then an MSC module is proposed to judge the true status of these elements and thus ultimately determine whether to eliminate them. In addition, an FPU module is introduced to update the extracted feature points according to calculating the dynamic region factor to improve the robustness of VSLAM system. Quantitative and qualitative experiments on two challenging public datasets are performed. The results demonstrate that our method effectively eliminates the influence of short-term dynamic elements and outperforms other state-of-the-art methods. (c) 2022 SPIE and IS&T

Keyword :

visual simultaneous localization and mapping visual simultaneous localization and mapping motion status check motion status check short-term dynamic elements short-term dynamic elements feature points update feature points update coarse-to-fine coarse-to-fine

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Fu, Fuji , Yang, Jinfu , Zhang, Jiahui et al. Eliminating short-term dynamic elements for robust visual simultaneous localization and mapping using a coarse-to-fine strategy [J]. | JOURNAL OF ELECTRONIC IMAGING , 2022 , 31 (5) .
MLA Fu, Fuji et al. "Eliminating short-term dynamic elements for robust visual simultaneous localization and mapping using a coarse-to-fine strategy" . | JOURNAL OF ELECTRONIC IMAGING 31 . 5 (2022) .
APA Fu, Fuji , Yang, Jinfu , Zhang, Jiahui , Ma, Jiaqi . Eliminating short-term dynamic elements for robust visual simultaneous localization and mapping using a coarse-to-fine strategy . | JOURNAL OF ELECTRONIC IMAGING , 2022 , 31 (5) .
Export to NoteExpress RIS BibTex
Multimodal based attention-pyramid for predicting pedestrian trajectory SCIE
期刊论文 | 2022 , 31 (5) | JOURNAL OF ELECTRONIC IMAGING
WoS CC Cited Count: 1
Abstract&Keyword Cite

Abstract :

The goal of pedestrian trajectory prediction is to predict the future trajectory according to the historical one of pedestrians. Multimodal information in the historical trajectory is conducive to perception and positioning, especially visual information and position coordinates. However, most of the current algorithms ignore the significance of multimodal information in the historical trajectory. We describe pedestrian trajectory prediction as a multimodal problem, in which historical trajectory is divided into an image and coordinate information. Specifically, we apply fully connected long short-term memory (FC-LSTM) and convolutional LSTM (ConvLSTM) to receive and process location coordinates and visual information respectively, and then fuse the information by a multimodal fusion module. Then, the attention pyramid social interaction module is built based on information fusion, to reason complex spatial and social relations between target and neighbors adaptively. The proposed approach is validated on different experimental verification tasks on which it can get better performance in terms of accuracy than other counterparts. (c) 2022 SPIE and IS&T

Keyword :

trajectory prediction trajectory prediction recurrent neural network recurrent neural network multimodal fusion multimodal fusion attention mechanism attention mechanism

Cite:

Copy from the list or Export to your reference management。

GB/T 7714 Yan, Xue , Yang, Jinfu , Liu, Yubin et al. Multimodal based attention-pyramid for predicting pedestrian trajectory [J]. | JOURNAL OF ELECTRONIC IMAGING , 2022 , 31 (5) .
MLA Yan, Xue et al. "Multimodal based attention-pyramid for predicting pedestrian trajectory" . | JOURNAL OF ELECTRONIC IMAGING 31 . 5 (2022) .
APA Yan, Xue , Yang, Jinfu , Liu, Yubin , Song, Lin . Multimodal based attention-pyramid for predicting pedestrian trajectory . | JOURNAL OF ELECTRONIC IMAGING , 2022 , 31 (5) .
Export to NoteExpress RIS BibTex
10| 20| 50 per page
< Page ,Total 16 >

Export

Results:

Selected

to

Format:
Online/Total:437/9312227
Address:BJUT Library(100 Pingleyuan,Chaoyang District,Beijing 100124, China Post Code:100124) Contact Us:010-67392185
Copyright:BJUT Library Technical Support:Beijing Aegean Software Co., Ltd.