no code implementations • 25 Apr 2024 • Zhijie Rao, Jingcai Guo, Xiaocheng Lu, Jingming Liang, Jie Zhang, Haozhao Wang, Kang Wei, Xiaofeng Cao
Zero-shot learning has consistently yielded remarkable progress via modeling nuanced one-to-one visual-attribute correlation.
no code implementations • 23 Apr 2024 • Bo Lin, Yingjing Xu, Xuanwen Bao, Zhou Zhao, Zuyong Zhang, Zhouyang Wang, Jie Zhang, Shuiguang Deng, Jianwei Yin
With the continuous advancement of vision language models (VLMs) technology, remarkable research achievements have emerged in the dermatology field, the fourth most prevalent human disease category.
1 code implementation • 23 Apr 2024 • Xuanhua He, Quande Liu, Shengju Qian, Xin Wang, Tao Hu, Ke Cao, Keyu Yan, Man Zhou, Jie Zhang
Based on this pipeline, a random face reference training method is further devised to precisely capture the ID-relevant embeddings from reference images, thus improving the fidelity and generalization capacity of our model for ID-specific video generation.
no code implementations • 9 Apr 2024 • Mahmut S. Gokmen, Cody Bumgardner, Jie Zhang, Ge Wang, Jin Chen
The results show that the polynomial noise distribution outperforms the model trained with log-normal noise distribution, yielding a 33. 54 FID score after 100, 000 training steps with constant discretization steps.
no code implementations • 8 Apr 2024 • Jie Zhang, Jun Li, Long Shi, Zhe Wang, Shi Jin, Wen Chen, H. Vincent Poor
By leveraging the power of DT models learned over extensive datasets, the proposed architecture is expected to achieve rapid convergence with many fewer training epochs and higher performance in a new context, e. g., similar tasks with different state and action spaces, compared with DRL.
no code implementations • 2 Apr 2024 • Jiachen Ma, Anda Cao, Zhiqing Xiao, Jie Zhang, Chao Ye, Junbo Zhao
The fast advance of the image generation community has attracted attention worldwide.
no code implementations • 25 Mar 2024 • Jie Zhang
The ability to form memories is a basic feature of learning and accumulating knowledge.
1 code implementation • 25 Mar 2024 • Yirong Zeng, Xiao Ding, Yi Zhao, Xiangyu Li, Jie Zhang, Chao Yao, Ting Liu, Bing Qin
Furthermore, we construct RU22Fact, a novel multilingual explainable fact-checking dataset on the Russia-Ukraine conflict in 2022 of 16K samples, each containing real-world claims, optimized evidence, and referenced explanation.
no code implementations • 25 Mar 2024 • Ziyan Wang, Yingpeng Du, Zhu Sun, Haoyan Chua, Kaidong Feng, Wenya Wang, Jie Zhang
However, the former methods struggle with optimal prompts to elicit the correct reasoning of LLMs due to the lack of task-specific feedback, leading to unsatisfactory recommendations.
no code implementations • 11 Mar 2024 • Sikai Bai, Jie Zhang, Shuaicheng Li, Song Guo, Jingcai Guo, Jun Hou, Tao Han, Xiaocheng Lu
Federated learning (FL) has emerged as a powerful paradigm for learning from decentralized data, and federated domain generalization further considers the test dataset (target domain) is absent from the decentralized training data (source domains).
no code implementations • 8 Mar 2024 • Zichong Meng, Jie Zhang, Changdi Yang, Zheng Zhan, Pu Zhao, Yanzhi Wang
On top of that, Exemplar-free Class Incremental Learning is even more challenging due to forbidden access to previous task data.
1 code implementation • 29 Feb 2024 • Chen Qian, Jie Zhang, Wei Yao, Dongrui Liu, Zhenfei Yin, Yu Qiao, Yong liu, Jing Shao
This research provides an initial exploration of trustworthiness modeling during LLM pre-training, seeking to unveil new insights and spur further developments in the field.
no code implementations • 29 Feb 2024 • Jie Zhang, Xubing Yang, Rui Jiang, Wei Shao, Li Zhang
While the direct application of SAM to remote sensing image segmentation tasks does not yield satisfactory results, we propose RSAM-Seg, which stands for Remote Sensing SAM with Semantic Segmentation, as a tailored modification of SAM for the remote sensing field and eliminates the need for manual intervention to provide prompts.
no code implementations • 27 Feb 2024 • Yanghao Su, Jie Zhang, Ting Xu, Tianwei Zhang, Weiming Zhang, Nenghai Yu
To address it, in this paper, we begin by presenting an intriguing observation: the decision boundary of the backdoored model exhibits a greater degree of closeness than that of the clean model.
no code implementations • 21 Feb 2024 • Xiao-Yang Liu, Jie Zhang, Guoxuan Wang, Weiqing Tong, Anwar Walid
However, the resulting model still consumes a large amount of GPU memory.
1 code implementation • 19 Feb 2024 • Xuanhua He, Ke Cao, Keyu Yan, Rui Li, Chengjun Xie, Jie Zhang, Man Zhou
To the best of our knowledge, this work is the first attempt in exploring the potential of the Mamba model and establishes a new frontier in the pan-sharpening techniques.
1 code implementation • 19 Feb 2024 • Marcus de Carvalho, Mahardhika Pratama, Jie Zhang, Chua Haoyan, Edward Yapp
In this work, we introduce a novel approach called Cross-Domain Continual Learning (CDCL) that addresses the limitations of being limited to single supervised domains.
no code implementations • 19 Feb 2024 • Zhihao Wen, Jie Zhang, Yuan Fang
Fine-tuning all parameters of large language models (LLMs) necessitates substantial computational power and extended time.
no code implementations • 14 Feb 2024 • Yingpeng Du, Ziyan Wang, Zhu Sun, Haoyan Chua, Hongzhi Liu, Zhonghai Wu, Yining Ma, Jie Zhang, Youchen Sun
To adapt text-based LLMs with structured graphs, We use the LLM as an aggregator in graph processing, allowing it to understand graph-based information step by step.
1 code implementation • 4 Feb 2024 • Jiacheng Chen, Zeyuan Ma, Hongshu Guo, Yining Ma, Jie Zhang, Yue-Jiao Gong
Recent Meta-learning for Black-Box Optimization (MetaBBO) methods harness neural networks to meta-learn configurations of traditional black-box optimizers.
1 code implementation • 2 Feb 2024 • Guanlin Li, Shuai Yang, Jie Zhang, Tianwei Zhang
With the development of generative models, the quality of generated content keeps increasing.
1 code implementation • 30 Jan 2024 • Yijie Lin, Jie Zhang, Zhenyu Huang, Jia Liu, Zujie Wen, Xi Peng
Existing video-language studies mainly focus on learning short video clips, leaving long-term temporal dependencies rarely explored due to over-high computational cost of modeling long videos.
Ranked #1 on Zero-Shot Video Retrieval on YouCook2
Action Segmentation Long Video Retrieval (Background Removed) +2
no code implementations • 26 Jan 2024 • Chaochao Lu, Chen Qian, Guodong Zheng, Hongxing Fan, Hongzhi Gao, Jie Zhang, Jing Shao, Jingyi Deng, Jinlan Fu, Kexin Huang, Kunchang Li, Lijun Li, LiMin Wang, Lu Sheng, Meiqi Chen, Ming Zhang, Qibing Ren, Sirui Chen, Tao Gui, Wanli Ouyang, Yali Wang, Yan Teng, Yaru Wang, Yi Wang, Yinan He, Yingchun Wang, Yixu Wang, Yongting Zhang, Yu Qiao, Yujiong Shen, Yurong Mou, Yuxi Chen, Zaibin Zhang, Zhelun Shi, Zhenfei Yin, Zhipin Wang
Multi-modal Large Language Models (MLLMs) have shown impressive abilities in generating reasonable responses with respect to multi-modal contents.
no code implementations • 22 Jan 2024 • Shihao Chen, Liping Chen, Jie Zhang, KongAik Lee, ZhenHua Ling, LiRong Dai
For validation, we employ the open-source pre-trained YourTTS model for speech generation and protect the target speaker's speech in the white-box scenario.
1 code implementation • 19 Jan 2024 • Xiangshuo Qiao, Xianxin Li, Xiaozhe Qu, Jie Zhang, Yang Liu, Yu Luo, Cihang Jin, Jin Ma
Differently, video covers in short video search scenarios are presented as user-originated contents that provide important visual summaries of videos.
Ranked #1 on Image Retrieval on CBVS
no code implementations • 17 Jan 2024 • Xingming Long, Shiguang Shan, Jie Zhang
In this paper, we conduct an Anomalous cue Guided FAS (AG-FAS) method, which leverages real faces for improving model generalization via a De-spoofing Face Generator (DFG).
no code implementations • 15 Jan 2024 • Jie Zhang, Zhifan Wan, Lanqing Hu, Stephen Lin, Shuzhe Wu, Shiguang Shan
Considering the close connection between action recognition and human pose estimation, we design a Collaboratively Self-supervised Video Representation (CSVR) learning framework specific to action recognition by jointly considering generative pose prediction and discriminative context matching as pretext tasks.
1 code implementation • 9 Jan 2024 • Sibo Wang, Jie Zhang, Zheng Yuan, Shiguang Shan
Specifically, PMG-AFT minimizes the distance between the features of adversarial examples in the target model and those in the pre-trained model, aiming to preserve the generalization features already captured by the pre-trained model.
no code implementations • 8 Jan 2024 • Zhongjiang He, Zihan Wang, Xinzhang Liu, Shixuan Liu, Yitong Yao, Yuyao Huang, Xuelong Li, Yongxiang Li, Zhonghao Che, Zhaoxi Zhang, Yan Wang, Xin Wang, Luwen Pu, Huinan Xu, Ruiyu Fang, Yu Zhao, Jie Zhang, Xiaomeng Huang, Zhilong Lu, Jiaxin Peng, Wenjun Zheng, Shiquan Wang, Bingkai Yang, Xuewei he, Zhuoru Jiang, Qiyi Xie, Yanhan Zhang, Zhongqiu Li, Lingling Shi, Weiwei Fu, Yin Zhang, Zilu Huang, Sishi Xiong, Yuxiang Zhang, Chao Wang, Shuangyong Song
Subsequently, the model undergoes fine-tuning to align with human preferences, following a detailed methodology that we describe.
1 code implementation • 7 Jan 2024 • Qiushi Zhu, Jie Zhang, Yu Gu, Yuchen Hu, LiRong Dai
Considering that visual information helps to improve speech recognition performance in noisy scenes, in this work we propose a multichannel multi-modal speech self-supervised learning framework AV-wav2vec2, which utilizes video and multichannel audio data as inputs.
Audio-Visual Speech Recognition Automatic Speech Recognition +7
1 code implementation • 4 Jan 2024 • Xuanhua He, Keyu Yan, Rui Li, Chengjun Xie, Jie Zhang, Man Zhou
Pan-sharpening involves reconstructing missing high-frequency information in multi-spectral images with low spatial resolution, using a higher-resolution panchromatic image as guidance.
1 code implementation • 4 Jan 2024 • Xuanhua He, Tao Hu, Guoli Wang, Zejin Wang, Run Wang, Qian Zhang, Keyu Yan, Ziyi Chen, Rui Li, Chenjun Xie, Jie Zhang, Man Zhou
However, current methods often ignore the difference between cell phone RAW images and DSLR camera RGB images, a difference that goes beyond the color matrix and extends to spatial structure due to resolution variations.
no code implementations • 3 Jan 2024 • Zheng Yuan, Jie Zhang, Yude Wang, Shiguang Shan, Xilin Chen
The attention mechanism has been proven effective on various visual tasks in recent years.
no code implementations • 3 Jan 2024 • Zheng Yuan, Jie Zhang, Shiguang Shan
In recent years, the Vision Transformer (ViT) model has gradually become mainstream in various computer vision tasks, and the robustness of the model has received increasing attention.
1 code implementation • 17 Dec 2023 • Yi Xie, Jie Zhang, Shiqian Zhao, Tianwei Zhang, Xiaofeng Chen
While deep learning models have shown significant performance across various domains, their deployment needs extensive resources and advanced computing infrastructure.
no code implementations • 15 Dec 2023 • Jingcai Guo, Qihua Zhou, Ruibing Li, Xiaocheng Lu, Ziming Liu, Junyang Chen, Xin Xie, Jie Zhang
Then, to facilitate the generalization of local linearities, we construct a maximal margin geometry on the learned features by enforcing low-rank constraints on intra-class samples and high-rank constraints on inter-class samples, resulting in orthogonal subspaces for different classes and each subspace lies on a compact manifold.
no code implementations • 14 Dec 2023 • Yuan Sun, Xuan Wang, Yunfan Zhang, Jie Zhang, Caigui Jiang, Yu Guo, Fei Wang
We present a method named iComMa to address the 6D camera pose estimation problem in computer vision.
no code implementations • 12 Dec 2023 • Kangneng Zhou, Daiheng Gao, Xuan Wang, Jie Zhang, Peng Zhang, Xusen Sun, Longhao Zhang, Shiqi Yang, Bang Zhang, Liefeng Bo, Yaxing Wang
To address this limitation, we propose \textbf{MaTe3D}: mask-guided text-based 3D-aware portrait editing.
no code implementations • 12 Dec 2023 • Jiawei Sun, Bin Zhao, Dong Wang, Zhigang Wang, Jie Zhang, Nektarios Koukourakis, Juergen W. Czarske, Xuelong Li
Quantitative phase imaging (QPI) through multi-core fibers (MCFs) has been an emerging in vivo label-free endoscopic imaging modality with minimal invasiveness.
1 code implementation • 11 Dec 2023 • Jiyan He, Weitao Feng, Yaosen Min, Jingwei Yi, Kunsheng Tang, Shuai Li, Jie Zhang, Kejiang Chen, Wenbo Zhou, Xing Xie, Weiming Zhang, Nenghai Yu, Shuxin Zheng
In this study, we aim to raise awareness of the dangers of AI misuse in science, and call for responsible AI development and use in this domain.
1 code implementation • 10 Dec 2023 • Xiaojian Yuan, Kejiang Chen, Wen Huang, Jie Zhang, Weiming Zhang, Nenghai Yu
In response to these identified gaps, we introduce a novel Data-Free Hard-Label Robustness Stealing (DFHL-RS) attack in this paper, which enables the stealing of both model accuracy and robustness by simply querying hard labels of the target model without the help of any natural data.
no code implementations • 4 Dec 2023 • Guanlin Li, Naishan Zheng, Man Zhou, Jie Zhang, Tianwei Zhang
However, these works lack analysis of adversarial information or perturbation, which cannot reveal the mystery of adversarial examples and lose proper interpretation.
no code implementations • 25 Nov 2023 • Ruibin Li, Jingcai Guo, Song Guo, Qihua Zhou, Jie Zhang
Specifically, we find that the very last few steps of the denoising (i. e., generation) process strongly correspond to the stylistic information of images, and based on this, we propose to augment the latent features of both the foreground and background images with Gaussians for a direct denoising-based harmonization.
no code implementations • 23 Nov 2023 • Zhijie Rao, Jingcai Guo, Xiaocheng Lu, Qihua Zhou, Jie Zhang, Kang Wei, Chenxin Li, Song Guo
In this paper, we propose a simple yet effective Attribute-Aware Representation Rectification framework for GZSL, dubbed $\mathbf{(AR)^{2}}$, to adaptively rectify the feature extractor to learn novel features while keeping original valuable features.
no code implementations • 22 Nov 2023 • Jie Zhang, Qing-Tian Xu, Zhen-Hua Ling
In this work, we therefore propose a novel end-to-end brain-assisted speech enhancement network (BASEN), which incorporates the listeners' EEG signals and adopts a temporal convolutional network together with a convolutional multi-layer cross attention module to fuse EEG-audio features.
no code implementations • 18 Nov 2023 • Jiayang Liu, Siyu Zhu, Siyuan Liang, Jie Zhang, Han Fang, Weiming Zhang, Ee-Chien Chang
Various techniques have emerged to enhance the transferability of adversarial attacks for the black-box scenario.
no code implementations • 24 Oct 2023 • Zhiling Zhang, Jie Zhang, Kui Zhang, Wenbo Zhou, Weiming Zhang, Nenghai Yu
To address these concerns, researchers are actively exploring the concept of ``unlearnable examples", by adding imperceptible perturbation to data in the model training stage, which aims to prevent the model from learning discriminate features of the target face.
no code implementations • 24 Oct 2023 • Caixin Wang, Jie Zhang, Matthew A. Wilson, Ralph Etienne-Cummings
By combining the versatility of pixel-wise sampling patterns with the strength of deep neural networks at decoding complex scenes, our method greatly enhances the vision system's adaptability and performance in dynamic conditions.
no code implementations • 18 Oct 2023 • Zengguang Hao, Jie Zhang, Binxia Xu, Yafang Wang, Gerard de Melo, Xiaolong Li
Intent detection and identification from multi-turn dialogue has become a widely explored technique in conversational agents, for example, voice assistants and intelligent customer services.
1 code implementation • 16 Oct 2023 • Jianhao Yuan, Jie Zhang, Shuyang Sun, Philip Torr, Bo Zhao
Synthetic training data has gained prominence in numerous learning tasks and scenarios, offering advantages such as dataset augmentation, generalization evaluation, and privacy preservation.
no code implementations • 11 Oct 2023 • Jie Zhang, Yongshan Zhang, Yicong Zhou
To aggregate the multiview information, a fully-convolutional SED with a U-shape in spectral dimension is introduced to extract a multiview feature map.
no code implementations • 8 Oct 2023 • Md Selim, Jie Zhang, Faraneh Fathi, Michael A. Brooks, Ge Wang, Guoqiang Yu, Jin Chen
Finally, the decoder uses the transformed latent representation to generate a standardized CT image, providing a more consistent basis for downstream analysis.
1 code implementation • IEEE Transactions on Neural Networks and Learning Systems 2023 • Wenhui Huang, Cong Zhang, Jingda Wu, Xiangkun He, Jie Zhang, Chen Lv.
Stochastic exploration is the key to the success of the Deep Q-network (DQN) algorithm.
no code implementations • 27 Sep 2023 • Guanlin Li, Yifei Chen, Jie Zhang, Jiwei Li, Shangwei Guo, Tianwei Zhang
We propose Warfare, a unified methodology to achieve both attacks in a holistic way.
no code implementations • 25 Sep 2023 • Haokun Song, Rui Lin, Andrea Sgambelluri, Filippo Cugini, Yajie Li, Jie Zhang, Paolo Monti
We propose a cluster-based method to detect and locate eavesdropping events in optical line systems characterized by small power losses.
no code implementations • 11 Sep 2023 • Haotian Wang, Yuxuan Xi, Hang Chen, Jun Du, Yan Song, Qing Wang, Hengshun Zhou, Chenxi Wang, Jiefeng Ma, Pengfei Hu, Ya Jiang, Shi Cheng, Jie Zhang, Yuzhe Weng
Three different structures based on attention-guided feature gathering (AFG) are designed for deep feature fusion.
no code implementations • 7 Sep 2023 • Zhendong Liu, Jie Zhang, Qiangqiang He, Chongjun Wang
In the realm of visual recognition, data augmentation stands out as a pivotal technique to amplify model robustness.
no code implementations • 3 Sep 2023 • Dong Huang, Qingwen Bu, Jie Zhang, Xiaofei Xie, Junjie Chen, Heming Cui
To mitigate bias for code generation models, we evaluate five bias mitigation prompt strategies, i. e., utilizing bias testing results to refine the code (zero-shot), one-, few-shot, and two Chain-of-Thought (CoT) prompts.
no code implementations • 28 Aug 2023 • Qiushi Zhu, Yu Gu, Rilin Chen, Chao Weng, Yuchen Hu, LiRong Dai, Jie Zhang
Noise-robust TTS models are often trained using the enhanced speech, which thus suffer from speech distortion and background noise that affect the quality of the synthesized speech.
no code implementations • 21 Aug 2023 • Changzhen Li, Jie Zhang, Yang Wei, Zhilong Ji, Jinfeng Bai, Shiguang Shan
Vision Transformers have achieved great success in computer visions, delivering exceptional performance across various tasks.
no code implementations • 21 Aug 2023 • Yutong Wu, Jie Zhang, Florian Kerschbaum, Tianwei Zhang
Users can easily download the word embedding from public websites like Civitai and add it to their own stable diffusion model without fine-tuning for personalization.
no code implementations • 19 Aug 2023 • Jie Zhang, Pengcheng Shi, Zaiwang Gu, Yiyang Zhou, Zhi Wang
In this paper, we present Semantic-Human, a novel method that achieves both photorealistic details and viewpoint-consistent human parsing for the neural rendering of humans.
no code implementations • 18 Aug 2023 • Pengcheng Shi, Jie Zhang, Haozhe Cheng, Junyang Wang, Yiyang Zhou, Chenlin Zhao, Jihua Zhu
Specifically, we propose a plug-and-play Overlap Bias Matching Module (OBMM) comprising two integral components, overlap sampling module and bias prediction module.
no code implementations • 7 Aug 2023 • Zhenhao Jiang, Biao Zeng, Hao Feng, Jin Liu, Jie Zhang, Jia Jia, Ning Hu
In order to address the problem of pagination trigger mechanism, we propose a completely new module in the pipeline of recommender system named Mobile Supply.
no code implementations • 31 Jul 2023 • Yuzheng Wang, Zhaoyu Chen, Jie Zhang, Dingkang Yang, Zuhao Ge, Yang Liu, Siao Liu, Yunquan Sun, Wenqiang Zhang, Lizhe Qi
Then, we introduce a low-noise representation to alleviate the domain shifts and build a structured relationship of multiple data examples to exploit data knowledge.
1 code implementation • 27 Jul 2023 • Lingdong Kong, Yaru Niu, Shaoyuan Xie, Hanjiang Hu, Lai Xing Ng, Benoit R. Cottereau, Ding Zhao, Liangjun Zhang, Hesheng Wang, Wei Tsang Ooi, Ruijie Zhu, Ziyang Song, Li Liu, Tianzhu Zhang, Jun Yu, Mohan Jing, Pengwei Li, Xiaohua Qi, Cheng Jin, Yingfeng Chen, Jie Hou, Jie Zhang, Zhen Kan, Qiang Ling, Liang Peng, Minglei Li, Di Xu, Changpeng Yang, Yuanqi Yao, Gang Wu, Jian Kuai, Xianming Liu, Junjun Jiang, Jiamian Huang, Baojun Li, Jiale Chen, Shuang Zhang, Sun Ao, Zhenyu Li, Runze Chen, Haiyong Luo, Fang Zhao, Jingze Yu
In this paper, we summarize the winning solutions from the RoboDepth Challenge -- an academic competition designed to facilitate and advance robust OoD depth estimation.
1 code implementation • ICCV 2023 • Dongyao Zhu, Bowen Lei, Jie Zhang, Yanbo Fang, Ruqi Zhang, Yiqun Xie, Dongkuan Xu
Neural networks trained on distilled data often produce over-confident output and require correction by calibration methods.
no code implementations • 20 Jul 2023 • Yingpeng Du, Di Luo, Rui Yan, Hongzhi Liu, Yang song, HengShu Zhu, Jie Zhang
However, directly leveraging LLMs to enhance recommendation results is not a one-size-fits-all solution, as LLMs may suffer from fabricated generation and few-shot problems, which degrade the quality of resume completion.
1 code implementation • 19 Jul 2023 • Yu-chen Fan, Yitong Ji, Jie Zhang, Aixin Sun
First, there are significant differences in user interactions at the different stages when a user interacts with the MovieLens platform.
no code implementations • 18 Jul 2023 • Zhenhao Jiang, Biao Zeng, Hao Feng, Jin Liu, Jicong Fan, Jie Zhang, Jia Jia, Ning Hu, Xingyu Chen, Xuguang Lan
We propose a novel Entire Space Multi-Task Model for Post-Click Conversion Rate via Parameter Constraint (ESMC) and two alternatives: Entire Space Multi-Task Model with Siamese Network (ESMS) and Entire Space Multi-Task Model in Global Domain (ESMG) to address the PSC issue.
no code implementations • 15 Jul 2023 • Wenxin Xu, Hexin Jiang, Xuefeng Liang, Ying Zhou, Yin Zhao, Jie Zhang
In this work, we propose Utopia Label Distribution Approximation (ULDA) for time-series data, which makes the training label distribution closer to real-world but unknown (utopia) label distribution.
no code implementations • 11 Jul 2023 • Sikai Bai, Shuaicheng Li, Weiming Zhuang, Jie Zhang, Song Guo, Kunlin Yang, Jun Hou, Shuai Zhang, Junyu Gao, Shuai Yi
Theoretically, we show the convergence guarantee of the dual regulators.
no code implementations • 7 Jul 2023 • Min Yu, Jie Zhang, Arndt Joedicke, Tom Reddyhoff
Overall, the proposed method is applicable to general lubricated interfaces for the identification of equivalent circuit models, which in turn facilitates in-situ tribo-contacts with electric impedance measurement of oil film thickness.
no code implementations • 6 Jul 2023 • Jie Zhang, Masanori Suganuma, Takayuki Okatani
They consider an unsupervised setting, specifically the one-class setting, in which we assume the availability of a set of normal (\textit{i. e.}, anomaly-free) images for training.
no code implementations • 6 Jul 2023 • Jie Zhang, Masanori Suganuma, Takayuki Okatani
The local student, which is used in previous studies mainly focuses on structural anomaly detection while the global student pays attention to logical anomalies.
Ranked #13 on Anomaly Detection on MVTec LOCO AD
1 code implementation • 28 Jun 2023 • Jie Zhang, Xiaohua Qi, Bo Zhao
Existing federated learning solutions focus on transmitting features, parameters or gadients between clients and server, which suffer from serious low-efficiency and privacy-leakage problems.
no code implementations • 8 Jun 2023 • Quanjie Wang, Jie Zhang, Vladimir Chernysh, Xiangjun Liu
However, due to mode conversion and inelastic scattering, we found a portion of high-frequency TA phonons, which are higher than the cut-off frequency and cannot transmit across the ideal sharp interface, can partially transmit across the amorphous interlayer, which introduces additional thermal transport channels through the interface and has positive effect on interfacial thermal conductance.
no code implementations • 2 Jun 2023 • Leijie Wu, Song Guo, Junxiao Wang, Zicong Hong, Jie Zhang, Jingren Zhou
As Federated Learning (FL) has gained increasing attention, it has become widely acknowledged that straightforwardly applying stochastic gradient descent (SGD) on the overall framework when learning over a sequence of tasks results in the phenomenon known as ``catastrophic forgetting''.
no code implementations • 1 Jun 2023 • Ruibin Li, Qihua Zhou, Song Guo, Jie Zhang, Jingcai Guo, Xinyang Jiang, Yifei Shen, Zhenhua Han
Diffusion-based Generative Models (DGMs) have achieved unparalleled performance in synthesizing high-quality visual content, opening up the opportunity to improve image super-resolution (SR) tasks.
1 code implementation • 31 May 2023 • Jianan Zhou, Yaoxin Wu, Wen Song, Zhiguang Cao, Jie Zhang
Learning heuristics for vehicle routing problems (VRPs) has gained much attention due to the less reliance on hand-crafted rules.
1 code implementation • 29 May 2023 • Tao Feng, Jie Zhang, Peizheng Wang, Zhijie Wang
The expenses involved in training state-of-the-art deep hashing retrieval models have witnessed an increase due to the adoption of more sophisticated models and large-scale datasets.
2 code implementations • 23 May 2023 • Yang Qi, Zhichao Zhu, Yiming Wei, Lu Cao, Zhigang Wang, Jie Zhang, Wenlian Lu, Jianfeng Feng
To account for the propagation of correlated neural variability, we derive from first principles a moment embedding for spiking neural network (SNN).
no code implementations • 23 May 2023 • Siyuan Pan, Linna Zhang, Jie Zhang, Xiaoshuang Li, Liang Hou, Xiaobing Tu
Structured pruning can simplify network architecture and improve inference speed.
no code implementations • 23 May 2023 • Qiushi Zhu, Xiaoying Zhao, Jie Zhang, Yu Gu, Chao Weng, Yuchen Hu
Recently, many efforts have been made to explore how the brain processes speech using electroencephalographic (EEG) signals, where deep learning-based approaches were shown to be applicable in this field.
no code implementations • 21 May 2023 • Mohan Shi, Yuchun Shu, Lingyun Zuo, Qian Chen, Shiliang Zhang, Jie Zhang, Li-Rong Dai
For speech interaction, voice activity detection (VAD) is often used as a front-end.
no code implementations • 21 May 2023 • Mohan Shi, Zhihao Du, Qian Chen, Fan Yu, Yangze Li, Shiliang Zhang, Jie Zhang, Li-Rong Dai
In addition, a two-pass decoding strategy is further proposed to fully leverage the contextual modeling ability resulting in a better recognition performance.
no code implementations • 18 May 2023 • Yihao Huang, Felix Juefei-Xu, Qing Guo, Jie Zhang, Yutong Wu, Ming Hu, Tianlin Li, Geguang Pu, Yang Liu
Although recent personalization methods have democratized high-resolution image synthesis by enabling swift concept acquisition with minimal examples and lightweight computation, they also present an exploitable avenue for high accessible backdoor attacks.
1 code implementation • 17 May 2023 • Jie Zhang, Qing-Tian Xu, Qiu-Shi Zhu, Zhen-Hua Ling
In this paper, we thus propose a novel time-domain brain-assisted SE network (BASEN) incorporating electroencephalography (EEG) signals recorded from the listener for extracting the target speaker from monaural speech mixtures.
1 code implementation • 14 May 2023 • Xi Yang, Kejiang Chen, Weiming Zhang, Chang Liu, Yuang Qi, Jie Zhang, Han Fang, Nenghai Yu
To allow third-parties to autonomously inject watermarks into generated text, we develop a watermarking framework for black-box language model usage scenarios.
no code implementations • 10 May 2023 • Jie Zhang, Wei Ma, Qiang Hu, Shangqing Liu, Xiaofei Xie, Yves Le Traon, Yang Liu
Furthermore, the perturbation of adversarial examples introduced by RNNS is smaller compared to the baselines in terms of the number of replaced variables and the change in variable length.
no code implementations • 8 May 2023 • Chaoya Jiang, Wei Ye, Haiyang Xu, Miang yan, Shikun Zhang, Jie Zhang, Fei Huang
Cross-modal contrastive learning in vision language pretraining (VLP) faces the challenge of (partial) false negatives.
no code implementations • 5 May 2023 • Yitong Ji, Aixin Sun, Jie Zhang
Then we blend the historical and new preferences in the form of node embeddings in the new graph, through a Disentanglement Module.
no code implementations • 1 May 2023 • Jie Zhang, Xiaosong Ma, Song Guo, Wenchao Xu
Federated Semi-supervised Learning (FedSSL) has emerged as a new paradigm for allowing distributed clients to collaboratively train a machine learning model over scarce labeled data and abundant unlabeled data.
no code implementations • 11 Apr 2023 • Xiaoqing Huang, Andersen Ang, Kun Huang, Jie Zhang, Yijie Wang
We study estimation of piecewise smooth signals over a graph.
no code implementations • 10 Apr 2023 • Jie Zhang, Minghui Nie, Junjie Cao, Jian Liu, Ligang Liu
Comprehensive experiments demonstrate that the two proposed unsupervised methods are noticeably superior to some supervised deep normal estimators on the most common synthetic dataset.
1 code implementation • 9 Apr 2023 • Zhongqi Wang, Jie Zhang, Zhilong Ji, Jinfeng Bai, Shiguang Shan
While the style aggregator module is to generate paintings of a style corresponding to a reference image.
no code implementations • 15 Mar 2023 • Liang Shi, Jie Zhang, Shiguang Shan
In this study, we propose Real Face Foundation Representation Learning (RFFR), which aims to learn a general representation from large-scale real face datasets and detect potential artifacts outside the distribution of RFFR.
1 code implementation • ICCV 2023 • Jie Zhang, Chen Chen, Weiming Zhuang, LingJuan Lv
This paper focuses on an under-explored yet important problem: Federated Class-Continual Learning (FCCL), where new classes are dynamically added in federated learning.
1 code implementation • 4 Mar 2023 • Yaoxin Wu, Jianan Zhou, Yunwen Xia, Xianli Zhang, Zhiguang Cao, Jie Zhang
Airport ground handling (AGH) offers necessary operations to flights during their turnarounds and is of great importance to the efficiency of airport management and the economics of aviation.
1 code implementation • 27 Feb 2023 • Jianan Zhou, Yaoxin Wu, Zhiguang Cao, Wen Song, Jie Zhang, Zhenghua Chen
Dispatching vehicle fleets to serve flights is a key task in airport ground handling (AGH).
1 code implementation • 20 Feb 2023 • Xiaojian Yuan, Kejiang Chen, Jie Zhang, Weiming Zhang, Nenghai Yu, Yang Zhang
At first, a top-n selection strategy is proposed to provide pseudo-labels for public data, and use pseudo-labels to guide the training of the cGAN.
no code implementations • 19 Feb 2023 • Jie Zhang, Bo Li, Chen Chen, Lingjuan Lyu, Shuang Wu, Shouhong Ding, Chao Wu
In this work, we propose a novel algorithm called Decision Boundary based Federated Adversarial Training (DBFAT), which consists of two components (local re-weighting and global regularization) to improve both accuracy and robustness of FL systems.
no code implementations • 16 Feb 2023 • Xiao-Ying Zhao, Qiu-Shi Zhu, Jie Zhang
With advances in deep learning, neural network based speech enhancement (SE) has developed rapidly in the last decade.
no code implementations • 9 Feb 2023 • Yingchun Wang, Jingcai Guo, Jie Zhang, Song Guo, Weizhan Zhang, Qinghua Zheng
Federated learning (FL) is an emerging technique that trains massive and geographically distributed edge data while maintaining privacy.
no code implementations • 1 Feb 2023 • Ziji Shi, Le Jiang, Ang Wang, Jie Zhang, Xianyan Jia, Yong Li, Chencan Wu, Jialin Li, Wei Lin
However, finding a suitable model parallel schedule for an arbitrary neural network is a non-trivial task due to the exploding search space.
no code implementations • 20 Jan 2023 • Md Selim, Jie Zhang, Michael A. Brooks, Ge Wang, Jin Chen
This work addresses the issue of CT image harmonization using a new diffusion-based model, named DiffusionCT, to standardize CT images acquired from different vendors and protocols.
no code implementations • ICCV 2023 • Xuanhua He, Keyu Yan, Rui Li, Chengjun Xie, Jie Zhang, Man Zhou
To this end, we first revisit the degradation process of pan-sharpening in Fourier space, and then devise a Pyramid Dual Domain Injection Pan-sharpening Network upon the above observation by fully exploring and exploiting the distinguished information in both the spatial and frequency domains.
no code implementations • CVPR 2023 • Jie Zhang, Yongshan Zhang, Yicong Zhou
Using QSSN as the building block, we propose an end-to-end quantum-inspired spectral-spatial pyramid network (QSSPN) for HSI feature extraction and classification.
no code implementations • 20 Dec 2022 • Wei Ma, Shangqing Liu, Mengjie Zhao, Xiaofei Xie, Wenhan Wang, Qiang Hu, Jie Zhang, Yang Liu
These structures are fundamental to understanding code.
no code implementations • 19 Dec 2022 • Yingchun Wang, Jingcai Guo, Song Guo, Weizhan Zhang, Jie Zhang
Recent studies show that even highly biased dense networks contain an unbiased substructure that can achieve better out-of-distribution (OOD) generalization than the original model.
2 code implementations • CVPR 2023 • Lei Zhang, Jie Zhang, Bowen Lei, Subhabrata Mukherjee, Xiang Pan, Bo Zhao, Caiwen Ding, Yao Li, Dongkuan Xu
Dataset Distillation (DD), a newly emerging field, aims at generating much smaller but efficient synthetic training datasets from large ones.
no code implementations • 7 Dec 2022 • Yingchun Wang, Song Guo, Jingcai Guo, Weizhan Zhang, Yida Xu, Jie Zhang, Yi Liu
Extensive experiments based on small Cifar-10 and large-scaled ImageNet demonstrate that our method can obtain sparser networks with great generalization performance while providing quantified reliability for the pruned model.
no code implementations • 29 Nov 2022 • Kui Zhang, Hang Zhou, Jie Zhang, Qidong Huang, Weiming Zhang, Nenghai Yu
Deep 3D point cloud models are sensitive to adversarial attacks, which poses threats to safety-critical applications such as autonomous driving.
no code implementations • 22 Nov 2022 • Jie Zhang, Yihui Zhao, Tianzhe Bao, Zhenhong Li, Kun Qian, Alejandro F. Frangi, Sheng Quan Xie, Zhi-Qiang Zhang
The salient advantages of the proposed framework are twofold: 1) For the generic model, physics-based domain knowledge is embedded into the loss function of the data-driven model as soft constraints to penalise/regularise the data-driven model.
no code implementations • 22 Nov 2022 • Jing Sun, Shuo Chen, Cong Zhang, Yining Ma, Jie Zhang
To address this issue, we introduce Distributional Opponent-aided Multi-agent Actor-Critic (DOMAC), the first speculative opponent modelling algorithm that relies solely on local information (i. e., the controlled agent's observations, actions, and rewards).
no code implementations • 21 Nov 2022 • Xueyang Tang, Song Guo, Jie Zhang
Recently, data heterogeneity among the training datasets on the local clients (a. k. a., Non-IID data) has attracted intense interest in Federated Learning (FL), and many personalized federated learning methods have been proposed to handle it.
Out-of-Distribution Generalization Personalized Federated Learning
no code implementations • 21 Nov 2022 • Qiushi Zhu, Long Zhou, Ziqiang Zhang, Shujie Liu, Binxing Jiao, Jie Zhang, LiRong Dai, Daxin Jiang, Jinyu Li, Furu Wei
Although speech is a simple and effective way for humans to communicate with the outside world, a more realistic speech interaction contains multimodal information, e. g., vision, text.
1 code implementation • 20 Nov 2022 • Cong Zhang, Zhiguang Cao, Wen Song, Yaoxin Wu, Jie Zhang
Recent studies in using deep reinforcement learning (DRL) to solve Job-shop scheduling problems (JSSP) focus on construction heuristics.
no code implementations • 15 Nov 2022 • Jinyu Chen, Wenchao Xu, Song Guo, Junxiao Wang, Jie Zhang, Haozhao Wang
Federated Learning (FL) is an emerging paradigm that enables distributed users to collaboratively and iteratively train machine learning models without sharing their private data.
no code implementations • 14 Nov 2022 • Yi Liu, Song Guo, Jie Zhang, Qihua Zhou, Yingchun Wang, Xiaohan Zhao
We prove that FedFoA is a model-agnostic training framework and can be easily compatible with state-of-the-art unsupervised FL methods.
1 code implementation • 13 Nov 2022 • Kaixin Wang, Cheng Long, Da Yan, Jie Zhang, H. V. Jagadish
Specifically, we propose a weighted sampling algorithm called WSD for estimating the subgraph count in a fully dynamic graph stream, which samples the edges based on their weights that indicate their importance and reflect their properties.
no code implementations • 13 Nov 2022 • Leijie Wu, Song Guo, Yaohong Ding, Junxiao Wang, Wenchao Xu, Richard Yida Xu, Jie Zhang
In contrast, visual data exhibits a fundamentally different structure: Its basic unit (pixel) is a natural low-level representation with significant redundancies in the neighbourhood, which poses obvious challenges to the interpretability of MSA mechanism in ViT.
no code implementations • 1 Nov 2022 • Mohan Shi, Jie Zhang, Zhihao Du, Fan Yu, Qian Chen, Shiliang Zhang, Li-Rong Dai
Speaker-attributed automatic speech recognition (SA-ASR) in multi-party meeting scenarios is one of the most valuable and challenging ASR task.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
1 code implementation • 27 Oct 2022 • Qiu-Shi Zhu, Long Zhou, Jie Zhang, Shu-Jie Liu, Yu-Chen Hu, Li-Rong Dai
Self-supervised pre-training methods based on contrastive learning or regression tasks can utilize more unlabeled data to improve the performance of automatic speech recognition (ASR).
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
no code implementations • 5 Oct 2022 • Mohammad Divband Soorati, Enrico H. Gerding, Enrico Marchioni, Pavel Naumov, Timothy J. Norman, Sarvapali D. Ramchurn, Bahar Rastegari, Adam Sobey, Sebastian Stein, Danesh Tarpore, Vahid Yazdanpanah, Jie Zhang
The Agents, Interaction and Complexity research group at the University of Southampton has a long track record of research in multiagent systems (MAS).
no code implementations • 28 Sep 2022 • Xiao-Ying Zhao, Qiu-Shi Zhu, Jie Zhang
Specifically, the encoder and bottleneck layer of the DEMUCS model are initialized using the self-supervised pretrained WavLM model, the convolution in the encoder is replaced by causal convolution, and the transformer encoder in the bottleneck layer is based on causal attention mask.
1 code implementation • 13 Sep 2022 • Rongkai Zhang, Cong Zhang, Zhiguang Cao, Wen Song, Puay Siew Tan, Jie Zhang, Bihan Wen, Justin Dauwels
We propose a manager-worker framework based on deep reinforcement learning to tackle a hard yet nontrivial variant of Travelling Salesman Problem (TSP), \ie~multiple-vehicle TSP with time window and rejections (mTSPTWR), where customers who cannot be served before the deadline are subject to rejections.
1 code implementation • 5 Sep 2022 • Marcus de Carvalho, Mahardhika Pratama, Jie Zhang, Yajuan San
Catastrophic forgetting has been a significant problem hindering the deployment of deep learning algorithms in the continual learning setting.
1 code implementation • 1 Sep 2022 • Zhu Sun, Yu Lei, Lu Zhang, Chen Li, Yew-Soon Ong, Jie Zhang
Being equipped with three modules (i. e., global user behavior encoder, local multi-channel encoder, and region-aware weighting strategy), MCMG is capable of capturing both fine- and coarse-grained sequential regularities as well as exploring the dynamic impact of multi-channel by differentiating the region check-in patterns.
2 code implementations • 1 Sep 2022 • Jie Zhang, Zhiqi Li, Bo Li, Jianghe Xu, Shuang Wu, Shouhong Ding, Chao Wu
Extensive experiments on federated datasets and real-world datasets demonstrate that FedLC leads to a more accurate global model and much improved performance.
no code implementations • 21 Aug 2022 • Jingcai Guo, Song Guo, Jie Zhang, Ziming Liu
Concretely, we maintain an edge-agnostic hidden model in the cloud server to estimate a less-accurate while direction-aware inversion of the global model.
no code implementations • 19 Aug 2022 • Changzhen Li, Jie Zhang, Shuzhe Wu, Xin Jin, Shiguang Shan
Recently action recognition has received more and more attention for its comprehensive and practical applications in intelligent surveillance and human-computer interaction.
no code implementations • 15 Aug 2022 • Chia Hong Tseng, Jie Zhang, Min-Te Sun, Kazuya Sakai, Wei-Shinn Ku
To better utilize the lane information, the lanes which are in opposite direction to target agent are not likely to be taken by the target agent and are consequently filtered out.
no code implementations • 5 Aug 2022 • Sai Xu, Yanan Du, Jiliang Zhang, Jie Zhang
This letter proposes to employ intelligent reflecting surface (IRS) as an information media to display a microwave quick response (QR) code for Internet-of-Things applications.
1 code implementation • Pattern Recognition 2022 • Yude Wang, Jie Zhang, Meina Kan, Shiguang Shan
Firstly, we introduce a class-aware cross entropy (CCE) loss for network training.
no code implementations • 20 Jul 2022 • Sai Xu, Yanan Du, Jiliang Zhang, Jiangzhou Wang, Jie Zhang
This paper proposes to leverage intelligent reflecting surface (IRS) backscatter to realize radio-frequency-chain-free uplink-transmissions (RFCF-UT).
no code implementations • 4 Jul 2022 • Jie Zhang, Yihui Zhao, Fergus Shone, Zhenhong Li, Alejandro F. Frangi, Shengquan Xie, Zhiqiang Zhang
At the same time, the physics law between muscle forces and joint kinematics is used the soft constraint.
no code implementations • 1 Jul 2022 • Haonan Hu, Yan Jiang, Jiliang Zhang, Yanan Zheng, Qianbin Chen, Jie Zhang
The fog-radio-access-network (F-RAN) has been proposed to address the strict latency requirements, which offloads computation tasks generated in user equipments (UEs) to the edge to reduce the processing latency.
2 code implementations • 22 Jun 2022 • Zhu Sun, Hui Fang, Jie Yang, Xinghua Qu, Hongyang Liu, Di Yu, Yew-Soon Ong, Jie Zhang
Recently, one critical issue looms large in the field of recommender systems -- there are no effective benchmarks for rigorous evaluation -- which consequently leads to unreproducible evaluation and unfair comparison.
1 code implementation • 20 Jun 2022 • Wenhui Huang, Cong Zhang, Jingda Wu, Xiangkun He, Jie Zhang, Chen Lv
We theoretically prove that the policy improvement theorem holds for the preference-guided $\epsilon$-greedy policy and experimentally show that the inferred action preference distribution aligns with the landscape of corresponding Q-values.
no code implementations • 26 May 2022 • Qiu-Shi Zhu, Jie Zhang, Zi-Qiang Zhang, Li-Rong Dai
Speech enhancement (SE) is usually required as a front end to improve the speech quality in noisy environments, while the enhanced speech might not be optimal for automatic speech recognition (ASR) systems due to speech distortion.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
1 code implementation • 23 May 2022 • Jie Zhang, Chen Chen, Lingjuan Lyu
Knowledge Distillation (KD) is a typical method for training a lightweight student model with the help of a well-trained teacher model.
no code implementations • 19 May 2022 • Xiaodong Sun, Huijiong Yang, Nan Wu, T. C. Scott, Jie Zhang, Wanzhou Zhang
In order to obtain a physical phase diagram, the snake model with an artificial neural network is applied in an unsupervised learning way by the authors of [Phys. Rev. Lett.
no code implementations • 25 Apr 2022 • Yu Qian, Jian Cao, Xiaoshuang Li, Jie Zhang, Hufei Li, Jue Chen
To address this challenge, we propose a novel method that first linearly over-parameterizes the compact layers in pruned networks to enlarge the number of fine-tuning parameters and then re-parameterizes them to the original layers after fine-tuning.
no code implementations • 14 Apr 2022 • Feijie Wu, Shiqi He, Song Guo, Zhihao Qu, Haozhao Wang, Weihua Zhuang, Jie Zhang
Traditional one-bit compressed stochastic gradient descent can not be directly employed in multi-hop all-reduce, a widely adopted distributed training paradigm in network-intensive high-performance computing systems such as public clouds.
no code implementations • 14 Apr 2022 • Yansong Gao, Jie Zhang
That is, mechanism K is pointwise better than mechanism P. Next, for each task $j$, when machines' execution costs $t_i^j$ are independent and identically drawn from a task-specific distribution $F^j(t)$, we show that the average-case approximation ratio of mechanism K converges to a constant.
no code implementations • 13 Apr 2022 • Songjiang Yang, Zitian Zhang, Jiliang Zhang, Xiaoli Chu, Jie Zhang
Based on the designed detectors, we propose an adaptive modulation scheme to maximize the average transmission rate under imperfect CSI by optimizing the data transmission time subject to the maximum tolerable BEP.
1 code implementation • 12 Apr 2022 • Yitong Ji, Aixin Sun, Jie Zhang, Chenliang Li
Our study offers a different perspective to understand recommender accuracy, and our findings could trigger a revisit of recommender model design.
1 code implementation • 11 Apr 2022 • Yuanxing Zhang, Langshi Chen, Siran Yang, Man Yuan, Huimin Yi, Jie Zhang, Jiamang Wang, Jianbo Dong, Yunlong Xu, Yue Song, Yong Li, Di Zhang, Wei Lin, Lin Qu, Bo Zheng
However, we observe that GPU devices in training recommender systems are underutilized, and they cannot attain an expected throughput improvement as what it has achieved in CV and NLP areas.
no code implementations • 5 Apr 2022 • Ye-Qian Du, Jie Zhang, Qiu-Shi Zhu, Li-Rong Dai, Ming-Hui Wu, Xin Fang, Zhou-Wang Yang
Unpaired data has shown to be beneficial for low-resource automatic speech recognition~(ASR), which can be involved in the design of hybrid models with multi-task training or language model dependent pre-training.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 26 Mar 2022 • Jie Zhang, Jun Li, Yijin Zhang, Qingqing Wu, Xiongwei Wu, Feng Shu, Shi Jin, Wen Chen
Intelligent reflecting surface (IRS) is envisioned to be widely applied in future wireless networks.
Multi-agent Reinforcement Learning reinforcement-learning +1
no code implementations • 13 Mar 2022 • Jiaxin Wu, Xin Chen, Sobhan Badakhshan, Jie Zhang, Pingfeng Wang
Establishing cleaner energy generation therefore improving the sustainability of the power system is a crucial task in this century, and one of the key strategies being pursued is to shift the dependence on fossil fuel to renewable technologies such as wind, solar, and nuclear.
no code implementations • 15 Feb 2022 • Zi-Qiang Zhang, Jie Zhang, Jian-Shu Zhang, Ming-Hui Wu, Xin Fang, Li-Rong Dai
The proposed approach explores both the complementarity of audio-visual modalities and long-term context dependency using a transformer-based fusion module and a flexible masking strategy.
1 code implementation • 15 Feb 2022 • Yuan Jiang, Yaoxin Wu, Zhiguang Cao, Jie Zhang
Recent deep models for solving routing problems always assume a single distribution of nodes for training, which severely impairs their cross-distribution generalization ability.
no code implementations • 5 Feb 2022 • Leijie Wu, Song Guo, Yaohong Ding, Yufeng Zhan, Jie Zhang
Facing the challenge of statistical diversity in client local data distribution, personalized federated learning (PFL) has become a growing research hotspot.
1 code implementation • 28 Jan 2022 • Jie Zhang, Lei Zhang, Gang Li, Chao Wu
Adversarial examples are inputs for machine learning models that have been designed by attackers to cause the model to make mistakes.
no code implementations • 22 Jan 2022 • Qiu-Shi Zhu, Jie Zhang, Zi-Qiang Zhang, Ming-Hui Wu, Xin Fang, Li-Rong Dai
In this work, we therefore first analyze the noise robustness of wav2vec2. 0 via experiments.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 22 Jan 2022 • Xing-Yu Chen, Qiu-Shi Zhu, Jie Zhang, Li-Rong Dai
By using the acoustic signals to train the network, respectively, we can build individual models for three tasks, whose parameters are averaged to obtain an average model, which is then used as the initialization for the BiLSTM model training of each task.
no code implementations • 10 Jan 2022 • Zhenyuan Zhang, Tao Shen, Jie Zhang, Chao Wu
This technique mitigates the user heterogeneity problem and better protects user privacy.
no code implementations • CVPR 2022 • Mingjie He, Jie Zhang, Shiguang Shan, Xilin Chen
In this paper, we propose to enhance face recognition with a bypass of self-supervised 3D reconstruction, which enforces the neural backbone to focus on the identity-related depth and albedo information while neglects the identity-irrelevant pose and illumination information.
1 code implementation • CVPR 2022 • Jie Zhang, Bo Li, Jianghe Xu, Shuang Wu, Shouhong Ding, Lei Zhang, Chao Wu
The proposed method can efficiently imitate the target model through a small number of queries and achieve high attack success rate.
1 code implementation • 23 Dec 2021 • Jie Zhang, Chen Chen, Bo Li, Lingjuan Lyu, Shuang Wu, Shouhong Ding, Chunhua Shen, Chao Wu
One-shot Federated Learning (FL) has recently emerged as a promising approach, which allows the central server to learn a model in a single communication round.
1 code implementation • 19 Dec 2021 • Qidong Huang, Jie Zhang, Wenbo Zhou, WeimingZhang, Nenghai Yu
To this end, we first imitate the target manipulation model with a surrogate model, and then devise a poison perturbation generator to obtain the desired venom.
1 code implementation • 17 Dec 2021 • Feijie Wu, Song Guo, Haozhao Wang, Zhihao Qu, Haobo Zhang, Jie Zhang, Ziming Liu
In the setting of federated optimization, where a global model is aggregated periodically, step asynchronism occurs when participants conduct model training by efficiently utilizing their computational resources.
no code implementations • 16 Dec 2021 • Jie Zhang, Ke-Jia Chen, Jingqiang Chen
Sequential recommendation based on multi-interest framework models the user's recent interaction sequence into multiple different interest vectors, since a single low-dimensional vector cannot fully represent the diversity of user interests.
no code implementations • 15 Dec 2021 • Xi Yang, Jie Zhang, Kejiang Chen, Weiming Zhang, Zehua Ma, Feng Wang, Nenghai Yu
Tracing text provenance can help claim the ownership of text content or identify the malicious users who distribute misleading content like machine-generated fake news.
no code implementations • 30 Nov 2021 • Xuming Ran, Jie Zhang, Ziyuan Ye, Haiyan Wu, Qi Xu, Huihui Zhou, Quanying Liu
In this study, we propose an integrated framework called Deep Autoencoder with Neural Response (DAE-NR), which incorporates information from ANN and the visual cortex to achieve better image reconstruction performance and higher neural representation similarity between biological and artificial neurons.
no code implementations • 27 Nov 2021 • Zheng Yuan, Jie Zhang, Zhaoyan Jiang, Liangliang Li, Shiguang Shan
Instead of using the sign function, we propose to directly utilize the exact gradient direction with a scaling factor for generating adversarial perturbations, which improves the attack success rates of adversarial examples even with fewer perturbations.
2 code implementations • 27 Nov 2021 • Zheng Yuan, Jie Zhang, Shiguang Shan
Adversarial attacks provide a good way to study the robustness of deep learning models.
no code implementations • 18 Nov 2021 • Jie Zhang, Robert B. Fisher
We define a motion divergence measure using 3D lip landmarks to quantify the interframe dynamics of a 3D speaking lip.
1 code implementation • 8 Nov 2021 • Danni Peng, Sinno Jialin Pan, Jie Zhang, AnXiang Zeng
Recommender Systems (RSs) in real-world applications often deal with billions of user interactions daily.
1 code implementation • NeurIPS 2021 • Jie Zhang, Song Guo, Xiaosong Ma, Haozhao Wang, Wencao Xu, Feijie Wu
To deal with such model constraints, we exploit the potentials of heterogeneous model settings and propose a novel training framework to employ personalized models for different clients.
1 code implementation • NeurIPS 2021 • Yaoxin Wu, Wen Song, Zhiguang Cao, Jie Zhang
We then design a neural network to learn policies for each variable in parallel, trained by a customized actor-critic algorithm.
no code implementations • 28 Oct 2021 • Congqing He, Jie Zhang, Xiangyu Zhu, Huan Liu, Yukun Huang
To this end, we introduce a fresh perspective to revisit the relational event-cause extraction task and propose a novel sequence tagging framework, instead of extracting event types and events-causes separately.
no code implementations • 24 Oct 2021 • Kafeng Wang, Haoyi Xiong, Jie Zhang, Hongyang Chen, Dejing Dou, Cheng-Zhong Xu
Extensive experiment based on real-word field deployment (on the highways in Shenzhen, China) shows that SenseMag significantly outperforms the existing methods in both classification accuracy and the granularity of vehicle types (i. e., 7 types by SenseMag versus 4 types by the existing work in comparisons).
no code implementations • 19 Oct 2021 • Md Selim, Jie Zhang, Baowei Fei, Guo-Qiang Zhang, Gary Yeeming Ge, Jin Chen
We propose a novel deep learning approach called CVH-CT for harmonizing CT images captured using scanners from different vendors.
1 code implementation • NeurIPS 2021 • Liang Xin, Wen Song, Zhiguang Cao, Jie Zhang
We present NeuroLKH, a novel algorithm that combines deep learning with the strong traditional heuristic Lin-Kernighan-Helsgaun (LKH) for solving Traveling Salesman Problem.
1 code implementation • 14 Oct 2021 • Jie Zhang, Bo Hui, Po-Wei Harn, Min-Te Sun, Wei-Shinn Ku
We test our model on several graph datasets including directed homogeneous and heterogeneous graphs.
no code implementations • 8 Oct 2021 • Junyang Lin, An Yang, Jinze Bai, Chang Zhou, Le Jiang, Xianyan Jia, Ang Wang, Jie Zhang, Yong Li, Wei Lin, Jingren Zhou, Hongxia Yang
Recent expeditious developments in deep learning algorithms, distributed training, and even hardware design for large models have enabled training extreme-scale models, say GPT-3 and Switch Transformer possessing hundreds of billions or even trillions of parameters.
1 code implementation • 6 Oct 2021 • Jingwen Li, Yining Ma, Ruize Gao, Zhiguang Cao, Andrew Lim, Wen Song, Jie Zhang
To solve those problems, we propose a DRL method based on the attention mechanism with a vehicle selection decoder accounting for the heterogeneous fleet constraint and a node selection decoder accounting for the route construction, which learns to construct a solution by automatically selecting both a vehicle and a node for this vehicle at each step.
no code implementations • 6 Oct 2021 • Jingwen Li, Liang Xin, Zhiguang Cao, Andrew Lim, Wen Song, Jie Zhang
In particular, the heterogeneous attention mechanism specifically prescribes attentions for each role of the nodes while taking into account the precedence constraint, i. e., the pickup node must precede the pairing delivery node.
1 code implementation • 4 Oct 2021 • Marcus de Carvalho, Mahardhika Pratama, Jie Zhang, Edward Yapp
We consider the problem of online unsupervised cross-domain adaptation, where two independent but related data streams with different feature spaces -- a fully labeled source stream and an unlabeled target stream -- are learned together.
no code implementations • 29 Sep 2021 • Liang Xin, Wen Song, Zhiguang Cao, Jie Zhang
Recent studies show that deep neural networks can be trained to learn good heuristics for various Combinatorial Optimization Problems (COPs).
no code implementations • ICLR 2022 • Yaoxin Wu, Wen Song, Zhiguang Cao, Jie Zhang
Many practical combinatorial optimization problems under uncertainty can be modeled as stochastic integer programs (SIPs), which are extremely challenging to solve due to the high complexity.
1 code implementation • 23 Sep 2021 • Xun Gao, Yin Zhao, Jie Zhang, Longjun Cai
We expect the ERATO as well as our proposed SMTA to open up a new way for PERR task in video understanding and further improve the research of multi-modal fusion methodology.
1 code implementation • 30 Aug 2021 • Gui-Song Xia, Jian Ding, Ming Qian, Nan Xue, Jiaming Han, Xiang Bai, Michael Ying Yang, Shengyang Li, Serge Belongie, Jiebo Luo, Mihai Datcu, Marcello Pelillo, Liangpei Zhang, Qiang Zhou, Chao-hui Yu, Kaixuan Hu, Yingjia Bu, Wenming Tan, Zhe Yang, Wei Li, Shang Liu, Jiaxuan Zhao, Tianzhi Ma, Zi-han Gao, Lingqi Wang, Yi Zuo, Licheng Jiao, Chang Meng, Hao Wang, Jiahao Wang, Yiming Hui, Zhuojun Dong, Jie Zhang, Qianyue Bao, Zixiao Zhang, Fang Liu
This report summarizes the results of Learning to Understand Aerial Images (LUAI) 2021 challenge held on ICCV 2021, which focuses on object detection and semantic segmentation in aerial images.
no code implementations • Applied Energy 2021 • Hannah O. Kargbo, Jie Zhang, Anh N. Phan
The developed neural network model was then applied for optimising operating conditions of the two-stage gasification for high carbon conversion, high hydrogen yield and low carbon dioxide in nitrogen and carbon dioxide environments.
1 code implementation • ICCV 2021 • Zheng Yuan, Jie Zhang, Yunpei Jia, Chuanqi Tan, Tao Xue, Shiguang Shan
In recent years, research on adversarial attacks has become a hot spot.
no code implementations • 5 Aug 2021 • Jie Zhang, Dongdong Chen, Jing Liao, Han Fang, Zehua Ma, Weiming Zhang, Gang Hua, Nenghai Yu
However, little attention has been devoted to the protection of DNNs in image processing tasks.
1 code implementation • 5 Aug 2021 • Jie Zhang, Dongdong Chen, Qidong Huang, Jing Liao, Weiming Zhang, Huamin Feng, Gang Hua, Nenghai Yu
As the image structure can keep its semantic meaning during the data transformation, such trigger pattern is inherently robust to data transformations.
no code implementations • 20 Jul 2021 • Mingjie He, Jie Zhang, Shiguang Shan, Xiao Liu, Zhongqin Wu, Xilin Chen
Furthermore, by randomly dropping out several feature channels, our method can well simulate the occlusion of larger area.
no code implementations • 19 Jul 2021 • Jie Zhang, Alexandra Brintrup, Anisoara Calinescu, Edward Kosasih, Angira Sharma
This paper explains what is 'twined' in supply chain digital twin and how to 'twin' them to handle the spatio-temporal dynamic issue.
no code implementations • 14 Jul 2021 • Songjiang Yang, Zitian Zhang, Jiliang Zhang, Jie Zhang
Our contributions of this paper lie in: i) modeling the wobbling process of a hovering RW UAV; ii) developing an analytical model to derive the channel temporal autocorrelation function (ACF) for the millimeter-wave RW UAV A2G link in a closed-form expression; and iii) investigating how RW UAV wobbling impacts the Doppler effect on the millimeter-wave RW UAV A2G link.
2 code implementations • 7 Jul 2021 • Xin Zhou, Aixin Sun, Yong liu, Jie Zhang, Chunyan Miao
Collaborative filtering (CF) is widely used to learn informative latent representations of users and items from observed interactions.
no code implementations • 3 Jul 2021 • Md Selim, Jie Zhang, Baowei Fei, Guo-Qiang Zhang, Jin Chen
While remarkable advances have been made in Computed Tomography (CT), capturing CT images with non-standardized protocols causes low reproducibility regarding radiomic features, forming a barrier on CT image analysis in a large scale.
no code implementations • 2 Jul 2021 • Zhiyuan Wang, Haoyi Xiong, Jie Zhang, Sijia Yang, Mehdi Boukhechba, Laura E. Barnes, Daqing Zhang, Dejing Dou
Mobile Sensing Apps have been widely used as a practical approach to collect behavioral and health-related information from individuals and provide timely intervention to promote health and well-beings, such as mental health and chronic cares.
no code implementations • 29 Jun 2021 • Fadi Boutros, Naser Damer, Jan Niklas Kolf, Kiran Raja, Florian Kirchbuchner, Raghavendra Ramachandra, Arjan Kuijper, Pengcheng Fang, Chao Zhang, Fei Wang, David Montero, Naiara Aginako, Basilio Sierra, Marcos Nieto, Mustafa Ekrem Erakin, Ugur Demir, Hazim Kemal, Ekenel, Asaki Kataoka, Kohei Ichikawa, Shizuma Kubo, Jie Zhang, Mingjie He, Dan Han, Shiguang Shan, Klemen Grm, Vitomir Štruc, Sachith Seneviratne, Nuran Kasthuriarachchi, Sanka Rasnayaka, Pedro C. Neto, Ana F. Sequeira, Joao Ribeiro Pinto, Mohsen Saffari, Jaime S. Cardoso
These teams successfully submitted 18 valid solutions.
no code implementations • 31 May 2021 • An Yang, Junyang Lin, Rui Men, Chang Zhou, Le Jiang, Xianyan Jia, Ang Wang, Jie Zhang, Jiamang Wang, Yong Li, Di Zhang, Wei Lin, Lin Qu, Jingren Zhou, Hongxia Yang
Mixture-of-Experts (MoE) models can achieve promising results with outrageous large amount of parameters but constant computation cost, and thus it has become a trend in model scaling.
2 code implementations • 13 May 2021 • Liangyu Chen, Xin Lu, Jie Zhang, Xiaojie Chu, Chengpeng Chen
Specifically, we present a novel block: Half Instance Normalization Block (HIN Block), to boost the performance of image restoration networks.
Ranked #3 on Single Image Deraining on Test2800
no code implementations • 9 Apr 2021 • Xiquan Guan, Huamin Feng, Weiming Zhang, Hang Zhou, Jie Zhang, Nenghai Yu
Specifically, we present the reversible watermarking problem of deep convolutional neural networks and utilize the pruning theory of model compression technology to construct a host sequence used for embedding watermarking information by histogram shift.