摘要:
Multi-Unmanned Aerial Vehicle (UAV)-supported Mobile Edge Computing (MEC) can meet the computational requirements of tasks with high complexity and latency sensitivity to compensate for the lack of computational resources and coverage. In this paper, a multi-user and multi-UAV MEC networks is built as a two-tier UAV system in a task-intensive region where base stations are insufficient, with a centralized top-center UAV and a set of distributed bottom-UAVs providing computing services. The total energy consumption of the system is minimized by jointly optimizing the task offloading decision, 3D deployment of two-tier UAVs, the elevation angle of the bottom UAV, the number of UAVs, and computational resource allocation. To this end, an algorithm based on Differential Evolution and greedy algorithm with the objective of minimizing Energy Consumption (DEEC) is proposed in this paper. The algorithm uses a two-tier optimization framework where the upper tier uses a population optimization algorithm to solve for the location and elevation angle of the bottom UAV and the number of UAVs based on the actual ground equipment and the lower tier uses clustering and greedy algorithms to solve for the position of the top UAV, the offloading decision of the task, and the allocation of computational resources based on the results of the upper layer. The simulation results show that the algorithm effectively reduces the total energy consumption of the system while satisfying the task computation success rate and time delay.
Multi-Unmanned Aerial Vehicle (UAV)-supported Mobile Edge Computing (MEC) can meet the computational requirements of tasks with high complexity and latency sensitivity to compensate for the lack of computational resources and coverage. In this paper, a multi-user and multi-UAV MEC networks is built as a two-tier UAV system in a task-intensive region where base stations are insufficient, with a centralized top-center UAV and a set of distributed bottom-UAVs providing computing services. The total energy consumption of the system is minimized by jointly optimizing the task offloading decision, 3D deployment of two-tier UAVs, the elevation angle of the bottom UAV, the number of UAVs, and computational resource allocation. To this end, an algorithm based on Differential Evolution and greedy algorithm with the objective of minimizing Energy Consumption (DEEC) is proposed in this paper. The algorithm uses a two-tier optimization framework where the upper tier uses a population optimization algorithm to solve for the location and elevation angle of the bottom UAV and the number of UAVs based on the actual ground equipment and the lower tier uses clustering and greedy algorithms to solve for the position of the top UAV, the offloading decision of the task, and the allocation of computational resources based on the results of the upper layer. The simulation results show that the algorithm effectively reduces the total energy consumption of the system while satisfying the task computation success rate and time delay.
期刊:
IEEE Transactions on Circuits and Systems for Video Technology,2025年:1-1 ISSN:1051-8215
作者机构:
[Jiaohua Qin] College of Computer Science and Information Technology, Central South University of Forestry and Technology, Changsha, China;[Zhangdong Wang; Zhihuang Liu; Tongqing Zhou; Zhiping Cai] College of Computer Science and Technology, National University of Defense Technology, Changsha, China;[Yuanjing Luo] College of Computer Science and Technology, National University of Defense Technology, Changsha, China<&wdkj&>College of Computer Science and Information Technology, Central South University of Forestry and Technology, Changsha, China
摘要:
Cloud environments enhance diffusion model efficiency but introduce privacy risks, including intellectual property theft and data breaches. As AI-generated images gain recognition as copyright-protected works, ensuring their security and intellectual property protection in cloud environments has become a pressing challenge. This paper addresses privacy protection in diffusion model inference under cloud environments, identifying two key characteristics—denoising-encryption antagonism and stepwise generative nature—that create challenges such as incompatibility with traditional encryption, incomplete input parameter representation, and inseparability of the generative process. We propose PPIDM (Privacy-Preserving Inference for Diffusion Models), a framework that balances efficiency and privacy by retaining lightweight text encoding and image decoding on the client while offloading computationally intensive U-Net layers to multiple non-colluding cloud servers. Clientside aggregation reduces computational overhead and enhances security. Experiments show PPIDM offloads 67% of Stable Diffusion computations to the cloud, reduces image leakage by 75%, and maintains high output quality (PSNR = 36.9, FID = 4.56), comparable to standard outputs. PPIDM offers a secure and efficient solution for cloud-based diffusion model inference.
期刊:
International Journal of Autonomous and Adaptive Communications Systems,2025年18(1):45-66 ISSN:1754-8632
作者机构:
[Zhilin Chen; Jiaohua Qin] College of Computer Science and Information Technology, Central South University of Forestry and Technology, Changsha, 410004, China
关键词:
RDH;reversible data hiding;prediction error;Huffman coding;encrypted images;extended run-length coding
摘要:
To reduce prediction errors and create more room for embedding data, the paper proposes a reversible data hiding (RDH) in encrypted images scheme based on histogram shifting and prediction error block coding. Firstly, the histogram of the prediction error image is shifted according to the signs of prediction errors. Next, the prediction error plane is partitioned into uniformly sized blocks, and these blocks are labelled as three types: an all-zero block, a block containing only one 1, and a block containing more than one 1. These three types of blocks are compressed using labelling, binary tree coding, and Huffman coding, respectively. To better compress the label map, an improved extended run-length coding is proposed. Finally, the image is secured by encryption and the secret data is hidden within it. The experimental results indicate a significant improvement in the embedding rate of the scheme compared to other schemes.
作者机构:
[Sheng, Guo; She, Kang; Shan, Zhengping; Liu, Exian] College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China;[Peng, Yuchen] School of Engineering, Westlake University, Hangzhou, Zhejiang 310024, China;[Liu, Jianjun] The Key Laboratory for Micro/Nano-Optoelectronic Devices of Ministry of Education, School of Physics and Electronics, Hunan University, Changsha 410082, China
通讯机构:
[Exian Liu] C;College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China
摘要:
Topological photonic crystals (TPCs) are considered excellent optical sensor platform due to the robustness of light propagation, immunity to interference, and ease of integration. This paper theoretically proposes a topological valley photonic crystal (VPC) sensor based on second-order corner states and, to our knowledge, realizes for the first time topologically protected dual-parameter detection for refractive index (RI) and temperature. Two highly localized opological corner states (TCSs) are generated by splicing two types of VPCs with broken inversion symmetry. Owing to the very weak coupling between the two TCSs, they respond independently to RI and temperature variations, respectively. Results show that the RI sensitivity is around 653 nm/RIU over a wide range from 1.26 to 1.35, and the temperature sensitivity reaches 235 pm/°C over a broad range from 0 to 100 °C. Fascinatingly, the resonance peak's quality factor (Q) surpasses the order of 10(5) due to the high localization of TCS. Moreover, the TPC sensor maintains high-performance robustness by deliberately introducing 12 types of defects. This design has great potential in engineering compact, easy-integrated, and dual-parameter sensors in biomedical, chemical, and pharmaceutical fields.
通讯机构:
[Li, T ] H;Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411201, Hunan, Peoples R China.
关键词:
Short texts typically refer to texts of relatively short length;generally not exceeding 160 characters. Short text classification (Chakraborty & Singh;2022) is a form of text mining technology and a significant research direction in natural language processing (NLP) (Hirschberg & Manning;2015). With the rapid development of the internet;a large volume of information in the form of short texts such as news tags;reviews;instant messages;and tweets has emerged on online platforms. These short texts contain a wealth of helpful information that contributes to establishing services in various fields;such as news classification;social media (Kateb & Kalita;sentiment analysis (Balomenos et al.;e-commerce;and spam filtering. As a result;with the growing depth of related research;short text classification has emerged as a critical challenge that urgently needs to be addressed in information retrieval. However;short text classification faces several challenges that complicate the task. First;short texts tend to be brief;with sparse semantic features and weak contextual connections;often requiring additional feature information to assist in classification (Wu;2023). For example;Hua et al. (2024) integrated text;entities;and words into a heterogeneous graph convolutional network;representing features through a word graph and enhancing word features with BiLSTM;thereby enabling document category prediction. However;due to length limitations;GNNs typically do not incorporate additional information when processing short texts;treating each short text as a single node in the graph. This leads to insufficient feature information;which in turn affects performance. Therefore;enhancing feature information and semantic relevance becomes crucial. Second;in rapidly changing text data environments (such as social media and news);expert knowledge required to organize and label training documents properly is often limited;expensive;and sometimes even unavailable. In some cases;training text classifiers using well-organized and accurately labeled documents is prohibitively costly;making unsupervised learning a crucial solution. Unsupervised tasks allow for fast classification without labeled data;avoiding the time-consuming and error-prone process of manual annotation. For example;Cameron-Steinke (2019) employed a novel text expansion approach in an unsupervised manner;whereby new words are generated directly for each input sentence. The newly generated keywords are formed within the hidden states of a pre-trained language model;which are then used to create expanded pseudo-headlines. However;it has limitations in terms of the accuracy and redundancy of the expanded content. In addition;Mishra (2024) compared and analyzed unsupervised text classification techniques;especially Latent Dirichlet Allocation (LDA) and BERTopic;and applied these topic modeling techniques to very short documents. However;since their evaluation datasets are based on specific scenarios;the generalization of these methods is weak. Therefore;how to effectively classify large amounts of unlabeled data using unsupervised methods is a research-worthy problem. To address the aforementioned issues;we propose a novel approach called UEDE: The Unsupervised Short Text Classification Method based on GCN Encoder-Decoder and Local Enhancement. This method uses unsupervised learning to cleverly integrate the GCN encoder–decoder and local enhancement technology under the autoencoder framework to achieve accurate short-text classification. Specifically;we first design the MRFasGCN modeling approach within the autoencoder framework;learning the hierarchical representations of node attributes and the network’s deep structure. We model and infer the relationships between nodes by integrating these deep representations with network structural information;thereby capturing text features and semantic information more effectively. Then;we introduce a dual decoder suited for unsupervised learning;which alleviates the issue of short text feature sparsity from multiple perspectives by reconstructing the topology and node attributes. Furthermore;we further enhance the latent semantic module locally by constructing an information fusion graph;ensuring that each pair of connected nodes shares a similar module distribution and complementary features. This improves the model’s ability to understand the local context of the text while addressing the overfitting issues caused by short text feature sparsity;significantly boosting the model’s performance and classification accuracy. The main contributions of this article can be summarized as follows: • We propose an unsupervised learning-based short text classification method;utilizing multiple convolutional layers as encoders to learn short text attribute networks’ deep representations and structural information. This approach facilitates module partitioning and inference;more effectively capturing text features and semantic information. • We design a dual decoder structure that;through unsupervised learning;separately reconstructs the topology and node attributes while accounting for the heterogeneity of node degrees. This approach compensates for insufficient short text feature information from multiple perspectives. • We introduce a local enhancement method that enables nodes to share more common neighbors and similar node attributes;strengthening the connections between related nodes. This improves the model’s understanding of the text’s local context while alleviating the overfitting issues caused by short text feature sparsity. • We perform comprehensive experiments on real-world datasets encompassing news articles;concise comments;and search snippets to assess the efficacy of our model in comparison to fourteen baseline approaches. The experimental findings unequivocally establish that our model surpasses the current state-of-the-art baseline methods on the benchmark datasets. We propose an unsupervised learning-based short text classification method;utilizing multiple convolutional layers as encoders to learn short text attribute networks’ deep representations and structural information. This approach facilitates module partitioning and inference;more effectively capturing text features and semantic information. We design a dual decoder structure that;through unsupervised learning;separately reconstructs the topology and node attributes while accounting for the heterogeneity of node degrees. This approach compensates for insufficient short text feature information from multiple perspectives. We introduce a local enhancement method that enables nodes to share more common neighbors and similar node attributes;strengthening the connections between related nodes. This improves the model’s understanding of the text’s local context while alleviating the overfitting issues caused by short text feature sparsity. We perform comprehensive experiments on real-world datasets encompassing news articles;concise comments;and search snippets to assess the efficacy of our model in comparison to fourteen baseline approaches. The experimental findings unequivocally establish that our model surpasses the current state-of-the-art baseline methods on the benchmark datasets. The remainder of the paper is structured as follows: Section 2 reviews previous work. Section 3 describes our proposed method and model;including the unsupervised short text classification model and local enhancement method. In Section 4;we perform comprehensive experiments on the datasets and analyze the outcomes. Lastly;Section 5 concludes the paper;offering insights into the future research directions.
摘要:
Like all fields of data science, short text classification seeks to achieve high-quality results with limited data. Although supervised learning methods have made notable progress in this area, they require much-labeled data to achieve adequate accuracy. However, in many practical applications, labeled data is scarce, and manual labeling is not only time-consuming and labor-intensive but also expensive and may require specialized expertise. Therefore, this paper addresses the challenge of insufficient labeled data through unsupervised methods while ensuring the effective extraction of semantic features from the text. Building on this objective, we propose a novel unsupervised short text classification method within the framework of autoencoders. Specifically, we first design the MRFasGCN encoder and derive the relationships between nodes in its hidden layers, thereby enhancing the capture of text features and semantic information. Furthermore, we construct a dual-node-based decoder that reconstructs the topology and node attributes unsupervised. This approach compensates for feature deficiencies from multiple perspectives, alleviating the issue of insufficient features in short texts. Finally, we propose a localized enhancement method that integrates node features and topology, strengthening the connections between relevant nodes. This improves the model’s understanding of the text’s local context while mitigating the overfitting issues caused by feature sparsity in short texts. Extensive experimental results demonstrate the pronounced superiority of our proposed UEDE model over existing methods on the dataset, validating its effectiveness in short-text classification. Our code is submitted in https://github.com/w123yy/UEDE .
Like all fields of data science, short text classification seeks to achieve high-quality results with limited data. Although supervised learning methods have made notable progress in this area, they require much-labeled data to achieve adequate accuracy. However, in many practical applications, labeled data is scarce, and manual labeling is not only time-consuming and labor-intensive but also expensive and may require specialized expertise. Therefore, this paper addresses the challenge of insufficient labeled data through unsupervised methods while ensuring the effective extraction of semantic features from the text. Building on this objective, we propose a novel unsupervised short text classification method within the framework of autoencoders. Specifically, we first design the MRFasGCN encoder and derive the relationships between nodes in its hidden layers, thereby enhancing the capture of text features and semantic information. Furthermore, we construct a dual-node-based decoder that reconstructs the topology and node attributes unsupervised. This approach compensates for feature deficiencies from multiple perspectives, alleviating the issue of insufficient features in short texts. Finally, we propose a localized enhancement method that integrates node features and topology, strengthening the connections between relevant nodes. This improves the model’s understanding of the text’s local context while mitigating the overfitting issues caused by feature sparsity in short texts. Extensive experimental results demonstrate the pronounced superiority of our proposed UEDE model over existing methods on the dataset, validating its effectiveness in short-text classification. Our code is submitted in https://github.com/w123yy/UEDE .
会议论文集名称:
Communications in Computer and Information Science
关键词:
Image manipulation detection;frequency information
摘要:
The goal of Image Manipulation Detection (IMD) is to identify and locate manipulated regions within images. Recent approaches have primarily designed sophisticated neural networks to capture highf-requency information for IMD tasks. However, these methods often overemphasize high-frequency information while overlooking the important role of low-frequency information in IMD tasks. To address this issue, we propose a Triple-Branch Frequency-Aware Network (TFNet), which includes an Information Separation Module (ISM), a Main Steam Branch (MSB), a Low-Frequency Learning Branch (LFL), a High-Frequency Learning Branch (HFL), and an Adaptive Aggregate Module (AAM) within a unified framework. Specifically, TFNet initially employs FSM to separate the manipulated image into RGB, high-frequency, and low-frequency components. Then, MSB, LFB, and HFB take the above components as input to learn the distinct features. Furthermore, the HFB is supervised with the boundary information to encourage the network to focus on the high-frequency information. Finally, the outputs of MSB, LFB, and HFB are sent to the ABM to adaptively aggregate features learned from the MSB, LFB, and HFB. Experiments on the CASIA, NIST, and Coverage datasets demonstrate the effectiveness of our TFNet.
期刊:
JOURNAL OF SUPERCOMPUTING,2025年81(1):1-21 ISSN:0920-8542
通讯作者:
Xuyu Xiang
作者机构:
[Peng, Jianting; Xiang, Xuyu; Qin, Jiaohua; Tan, Yun] College of Computer Science and Information Technology, Central South University of Forestry and Technology, Changsha, China
通讯机构:
[Xuyu Xiang] C;College of Computer Science and Information Technology, Central South University of Forestry and Technology, Changsha, China
摘要:
Cross-modal retrieval can break through the limitations of modalities and carry out information retrieval across data of different modalities to meet the needs of users in obtaining multi-modal correlation retrieval. Cloud computing has the advantages of high efficiency and low cost, but data security hinders its development. While cloud computing offers high efficiency and cost-effectiveness, concerns surrounding data security impede its full potential. Privacy-preserving cross-modal retrieval emerges as a viable solution, catering to users’ demands for efficient retrieval while safeguarding data confidentiality. However, a major challenge still exists in this field: how to bridge the inherent semantic gap within heterogeneous and chaotic information. To address this challenge, this paper proposes dual-branch networks for privacy-preserving cross-modal retrieval in cloud computing. Firstly, a dual-branch feature extraction network of encrypted image-text is constructed, enhancing the extraction of meaningful features from encrypted data. Secondly, a cross-modal alignment method is designed to eliminate the heterogeneous gap between different modalities through the alignment within and between modalities. Finally, to fully exploit the storage and computing advantages of cloud computing, both encrypted data and the cross-modal feature extractor are deployed to the cloud. Leveraging the dynamic update capabilities of cloud-stored encrypted data enables continuous model refinement, enhancing retrieval accuracy while reducing the storage and computational burdens on data owners. Extensive experiments conducted on the publicly available benchmark image-text dataset Wikipedia indicate that, compared to existing methods, our approach achieves improvements of 5.4%, 1%, 1.6%, and 20.1% in the four metrics of image-to-text (i2t), text-to-image (t2i), image-to-all (i2all), and text-to-all (t2all), respectively.
摘要:
With the continuous development of deep learning (DL), the task of multimodal dialog emotion recognition (MDER) has recently received extensive research attention, which is also an essential branch of DL. The MDER aims to identify the emotional information contained in different modalities, e.g., text, video, and audio, and in different dialog scenes. However, the existing research has focused on modeling contextual semantic information and dialog relations between speakers while ignoring the impact of event relations on emotion. To tackle the above issues, we propose a novel dialog and event relation-aware graph convolutional neural network (DER-GCN) for multimodal emotion recognition method. It models dialog relations between speakers and captures latent event relations information. Specifically, we construct a weighted multirelationship graph to simultaneously capture the dependencies between speakers and event relations in a dialog. Moreover, we also introduce a self-supervised masked graph autoencoder (SMGAE) to improve the fusion representation ability of features and structures. Next, we design a new multiple information Transformer (MIT) to capture the correlation between different relations, which can provide a better fuse of the multivariate information between relations. Finally, we propose a loss optimization strategy based on contrastive learning to enhance the representation learning ability of minority class features. We conduct extensive experiments on the benchmark datasets, Interactive Emotional Dyadic Motion Capture (IEMOCAP) and Multimodal EmotionLines Dataset (MELD), which verify the effectiveness of the DER-GCN model. The results demonstrate that our model significantly improves both the average accuracy and the $F1$ value of emotion recognition. Our code is publicly available at https://github.com/yuntaoshou/DER-GCN.
摘要:
The popularity of smart devices has made it more convenient for users to take screen shots, but it has also made it easier to take clandestine shots, resulting in compromised and untraceable information. Therefore, this paper introduces a screen-shooting robust watermarking algorithm based on the pattern complexity just noticeable difference (PC-JND) model. This approach involves the utilisation of local binary patterns (LBP) for block filtering based on texture complexity in the original image. Stable feature blocks are selected and processed using the speeded-up robust features (SURF) algorithm to extract key feature points, defining them as the watermark embedding regions. Finally, the watermark is embedded in the integer wavelet domain's HH sub-band, guided by the PC-JND model. Experimental results demonstrate that this algorithm not only significantly improves the visual quality of images in a shorter embedding time but also exhibits enhanced robustness against screen captures from various angles.
通讯机构:
[Liu, EX ] C;Cent South Univ Forestry & Technol, Coll Comp & Informat Engn, Changsha 410000, Peoples R China.
关键词:
Sensors;Photonic band gap;Couplings;Lattices;Refractive index;Biosensors;Band structures;Refractive index (RI) sensor;topological corner state (TCS);topological photonic crystals (TPC);ultrahigh quality factor
摘要:
Owing to the flexible light control and the immunity to defects and impurities, topological photonic crystals (TPC) are considered an excellent platform in the engineering of optoelectronic devices. Herein, four TPC refractive index (RI) sensors are theoretically proposed based on the su-Schrieffer-Heeger (SSH) model. Zero-dimensional topological corner states (TCS) and 1-D topological edge states (TES) are demonstrated by the band structure and the Zak phases. Four TPC sensors are based on single topological microcavity (TCS formed by two TES and four TES), trivial corner states (TRCS)-TES coupling, and TCS-TES coupling. The TCS-based TPC sensors exhibit an ultrahigh quality factor (Q) of $\gt 10^{{6}}$ . Moreover, TPC sensor based on TCS-TES coupling remain a topological robustness with >99.4% sensitivity quality of reproduction by artificially introducing eight-type defects. Notably, the design based on TCS-TES coupling presents a sensitivity of 413.76 nm/RIU and a Q value of $2.8\times 10^{{6}}$ in a broad RI range of 1.00–1.50. This work not only provides a new way of 2-D photonic crystals (PC) sensor applications but also offers a guidance for the design of TPC sensors.
关键词:
RDH;reversible data hiding;prediction error;Huffman coding;encrypted images;extended run-length coding
摘要:
To reduce prediction errors and create more room for embedding data, the paper proposes a reversible data hiding (RDH) in encrypted images scheme based on histogram shifting and prediction error block coding. Firstly, the histogram of the prediction error image is shifted according to the signs of prediction errors. Next, the prediction error plane is partitioned into uniformly sized blocks, and these blocks are labelled as three types: an all-zero block, a block containing only one 1, and a block containing more than one 1. These three types of blocks are compressed using labelling, binary tree coding, and Huffman coding, respectively. To better compress the label map, an improved extended run-length coding is proposed. Finally, the image is secured by encryption and the secret data is hidden within it. The experimental results indicate a significant improvement in the embedding rate of the scheme compared to other schemes.
作者机构:
[Sun, Haiwen; Tang, Chunling] Cent South Univ Forestry & Technol, Coll Econ & Management, Changsha 410004, Hunan, Peoples R China.;[Gao, Ping] Univ Manchester, Global Dev Inst, Oxford Rd, Manchester M13 9PL, England.;[Zhou, Guoxiong] Cent South Univ Forestry & Technol, Coll Comp & Informat Engn, Changsha 410004, Hunan, Peoples R China.
通讯机构:
[Tang, CL ] C;Cent South Univ Forestry & Technol, Coll Econ & Management, Changsha 410004, Hunan, Peoples R China.
关键词:
digital economy;energy consumption;carbon emissions;non-linear relationships;spatial spillover effect
摘要:
The intensifying challenge of global climate change has made accelerating energy conservation and emission reduction an urgent global imperative. As one of the world’s largest carbon emitters, China plays a pivotal role in global decarbonization efforts. The digital economy, emerging as a key driver of China’s economic transformation, provides novel pathways for advancing carbon reduction. This paper employs kernel density estimation and ArcGIS 10.8 to analyze the spatiotemporal dynamics of the digital economy and urban carbon emissions in China. Using panel data from 278 prefecture-level cities, the study applies fixed-effects models, mediation effect models, and spatial Durbin models to explore the mechanisms and spatial impacts of the digital economy on carbon reduction. The findings reveal that: (1) the development of the digital economy exerts a significant “inverted U-shaped” influence on urban carbon emission reduction; (2) energy consumption intensity is the critical mechanism underlying the nonlinear relationship between the digital economy and carbon emissions; (3) the digital economy's impact on carbon emissions exhibits spatial spillover effects, following a similar “inverted U-shaped” trajectory. These results contribute valuable empirical insights into the dual objectives of digital economic growth and carbon emission reduction, offering policymakers guidance on leveraging digitalization to achieve sustainable and coordinated regional development.
The intensifying challenge of global climate change has made accelerating energy conservation and emission reduction an urgent global imperative. As one of the world’s largest carbon emitters, China plays a pivotal role in global decarbonization efforts. The digital economy, emerging as a key driver of China’s economic transformation, provides novel pathways for advancing carbon reduction. This paper employs kernel density estimation and ArcGIS 10.8 to analyze the spatiotemporal dynamics of the digital economy and urban carbon emissions in China. Using panel data from 278 prefecture-level cities, the study applies fixed-effects models, mediation effect models, and spatial Durbin models to explore the mechanisms and spatial impacts of the digital economy on carbon reduction. The findings reveal that: (1) the development of the digital economy exerts a significant “inverted U-shaped” influence on urban carbon emission reduction; (2) energy consumption intensity is the critical mechanism underlying the nonlinear relationship between the digital economy and carbon emissions; (3) the digital economy's impact on carbon emissions exhibits spatial spillover effects, following a similar “inverted U-shaped” trajectory. These results contribute valuable empirical insights into the dual objectives of digital economic growth and carbon emission reduction, offering policymakers guidance on leveraging digitalization to achieve sustainable and coordinated regional development.
通讯机构:
[Zhang, LT ] C;Cent South Univ Forestry & Technol, Sch Comp & Informat Engn, Changsha 410000, Peoples R China.
关键词:
Temperature prediction;Hybrid model;CNN-BiLSTM;RandomForest
摘要:
Temperature fluctuations have profound impacts on both human society and the natural environment. However, the diversity of geographical temperature data and the non-linearity and complexity of meteorological phenomena present significant challenges to accurate prediction. Previous studies in temperature prediction have faced certain limitations, such as inadequate consideration of correlations between multiple meteorological factors in some models or limited modeling capability for non-linear and spatiotemporal relationships. To address these shortcomings, we propose a hybrid model based on CNN-BiLSTM and RandomForest for temperature prediction. Relevant meteorological indicators were selected as model inputs using a combined Pearson's correlation and mutual information analysis. Subsequently, utilizing a sequence-to-sequence framework, the CNN-BiLSTM model effectively explores spatiotemporal relationships and feature representations in time-series data as the encoder for sequence modeling. This aids in learning feature representations and sequence relationships in time series data, thus generating abstract representations for each sample in the hidden or output layers, covering key information learned by the model. Feature representations extracted from the CNN-BiLSTM model are aggregated into a feature matrix. Each sample corresponds to a row of the feature vector, with each column representing a specific feature extracted from the CNN-BiLSTM model. Finally, RandomForest, as an ensemble learning method, is capable of handling complex non-linear relationships during both model training and prediction stages, exhibiting good robustness and predictive capability. Utilizing the feature matrix as input data, it is fed into the RandomForest decoder model along with the corresponding target variables, further enhancing the prediction accuracy and stability of the model to obtain the final temperature prediction results. Simulated experiments were conducted using meteorological data from Changsha, Hunan Province, from 2017 to 2021. A significant number of experiments have demonstrated that, compared to the current leading Dliner method, the hybrid model exhibits higher prediction accuracy and stability in capturing daily temperature trends. The mean absolute error and mean squared error were reduced by 35.6 and 57.5%\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\%$$\end{document}, respectively.
摘要:
Tomatoes are an important global crop, and automating the segmentation of leaf diseases is essential for agricultural security. Effective segmentation of these diseases is vital for timely intervention, which can significantly enhance crop yield and reduce pesticide usage. However, challenges such as background interference, tiny diseases, and blurred disease edges pose immense obstacles to the segmentation of tomato leaf diseases. To address these issues effectively, we propose a fusion adversarial segmentation network for tomato disease segmentation named FATDNet. Firstly, to eliminate background interference effectively, we introduce a dual-path fusion adversarial algorithm (DFAA). This algorithm employs parallel dual-path convolution to extract features of leaf disease regions, merging and adversarially processing complex background noise and disease features. Secondly, to enhance the feature extraction of small lesion regions, we employ a multi-dimensional attention mechanism (MDAM). This mechanism allocates weights in both horizontal and vertical directions, subsequently calculating weights in different channels of the feature map. This enhances the dispersion of semantic information through the adoption of diverse weight calculation strategies. Furthermore, to improve the model’s ability to extract features at the edges of leaf diseases, we introduce a Gaussian weighted edge segmentation module (GWESM). This module calculates weight distribution through a Gaussian-weighted function, guiding the model to highlight features at different scales and reduce information loss caused by pooling operations. To demonstrate the superiority of FATDNet, we conduct comparative experiments using a self-built dataset and a public dataset. Experimental results show that FATDNet outperforms nine state-of-the-art segmentation networks. This validates that FATDNet provides a reliable solution for the automated segmentation of tomato leaf diseases.
Tomatoes are an important global crop, and automating the segmentation of leaf diseases is essential for agricultural security. Effective segmentation of these diseases is vital for timely intervention, which can significantly enhance crop yield and reduce pesticide usage. However, challenges such as background interference, tiny diseases, and blurred disease edges pose immense obstacles to the segmentation of tomato leaf diseases. To address these issues effectively, we propose a fusion adversarial segmentation network for tomato disease segmentation named FATDNet. Firstly, to eliminate background interference effectively, we introduce a dual-path fusion adversarial algorithm (DFAA). This algorithm employs parallel dual-path convolution to extract features of leaf disease regions, merging and adversarially processing complex background noise and disease features. Secondly, to enhance the feature extraction of small lesion regions, we employ a multi-dimensional attention mechanism (MDAM). This mechanism allocates weights in both horizontal and vertical directions, subsequently calculating weights in different channels of the feature map. This enhances the dispersion of semantic information through the adoption of diverse weight calculation strategies. Furthermore, to improve the model’s ability to extract features at the edges of leaf diseases, we introduce a Gaussian weighted edge segmentation module (GWESM). This module calculates weight distribution through a Gaussian-weighted function, guiding the model to highlight features at different scales and reduce information loss caused by pooling operations. To demonstrate the superiority of FATDNet, we conduct comparative experiments using a self-built dataset and a public dataset. Experimental results show that FATDNet outperforms nine state-of-the-art segmentation networks. This validates that FATDNet provides a reliable solution for the automated segmentation of tomato leaf diseases.
摘要:
Electrolytic double-layer gate insulation transistors (EDLGITs) have gained significant attention recently due to their features for achieving high-density charge accumulation and low operating voltage. This is primarily attributed to the large capacitance of the electrolyte insulation in these devices. This article proposes a physical model for the capacitance of the electrolytic double-layer gate insulation (EDLGI) and the drain current of amorphous InGaZnO (a-IGZO) EDLGITs based on the channel surface potential. It is demonstrated that the EDLGI capacitance, which comprises the compact-layer and diffuse-layer capacitances, is not only influenced by the characteristics of the electrolyte but also by the gate voltage. Furthermore, the frequency-dependent equivalent capacitance of the electrolyte insulation ( ${C} _{\text {eq}}$ ) under ac gate voltage in a-IGZO EDLGITs is analyzed. The drain current and ${C} _{\text {eq}}$ calculated using our model show good consistency with the reported experimental data.
作者:
Fan Yang;Tan Zhu*;Jing Huang;Zhilin Huang;Guoqi Xie
期刊:
COMPUTER SPEECH AND LANGUAGE,2025年:101818 ISSN:0885-2308
通讯作者:
Tan Zhu
作者机构:
[Fan Yang; Tan Zhu; Zhilin Huang] College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha, Hunan, 41004, PR China;[Jing Huang] School of Computer Science and Engineering, Hunan University of Science and Technology, Xiangtan, Hunan, 411201, PR China;[Guoqi Xie] Key Laboratory for Embedded and Network Computing of Hunan Province, College of Computer Science and Electronic Engineering, Hunan University, Changsha, Hunan 410082, PR China
通讯机构:
[Tan Zhu] C;College of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha, Hunan, 41004, PR China
摘要:
Text classification is an important topic in natural language processing. In recent years, both graph kernel methods and deep learning methods have been widely employed in text classification tasks. However, previous graph kernel algorithms focused too much on the graph structure itself, such as the shortest path subgraph,while focusing limited attention to the information of the text itself.Previous deep learning methods have often resulted in substantial utilization of computational resources. Therefore,we propose a new graph kernel algorithm to address the disadvantages. First,we extract the textual information of the document using the term weighting scheme. Second,we collect the structural information on the document graph. Third, graph kernel is used for similarity measurement for text classification. We compared eight baseline methods on three experimental datasets, including traditional deep learning methods and graph-based classification methods, and tested our algorithm on multiple indicators. The experimental results demonstrate that our algorithm outperforms other baseline methods in terms of accuracy. Furthermore, it achieves a minimum reduction of 69% in memory consumption and a minimum decrease of 23% in runtime. Furthermore, as we decrease the percentage of training data, our algorithm continues to achieve superior results compared to other deep learning methods. The excellent experimental results show that our algorithm can improve the efficiency of text classification tasks and reduce the occupation of computer resources under the premise of ensuring high accuracy.
Text classification is an important topic in natural language processing. In recent years, both graph kernel methods and deep learning methods have been widely employed in text classification tasks. However, previous graph kernel algorithms focused too much on the graph structure itself, such as the shortest path subgraph,while focusing limited attention to the information of the text itself.Previous deep learning methods have often resulted in substantial utilization of computational resources. Therefore,we propose a new graph kernel algorithm to address the disadvantages. First,we extract the textual information of the document using the term weighting scheme. Second,we collect the structural information on the document graph. Third, graph kernel is used for similarity measurement for text classification.
We compared eight baseline methods on three experimental datasets, including traditional deep learning methods and graph-based classification methods, and tested our algorithm on multiple indicators. The experimental results demonstrate that our algorithm outperforms other baseline methods in terms of accuracy. Furthermore, it achieves a minimum reduction of 69% in memory consumption and a minimum decrease of 23% in runtime. Furthermore, as we decrease the percentage of training data, our algorithm continues to achieve superior results compared to other deep learning methods. The excellent experimental results show that our algorithm can improve the efficiency of text classification tasks and reduce the occupation of computer resources under the premise of ensuring high accuracy.
作者机构:
[Wang, Jiangang; Peng, Ting; Lin, Yanhui] Department of Health Management Center, The Third Xiangya Hospital, Central South University, Changsha, Hunan, China;[Fu, Min] Ophthalmology Department of Zhujiang Hospital Affiliated to Southern Medical University, Guangdong, China;[Lin, Ren] Department of Health Management Center, Renmin Hospital of Wuhan University, Wuhan, China;[Hong, Min] Department of Health Management & Institute of Health Management, Sichuan Provincial People's Hospital, School of Medicine, University of Electronic Science and Technology of China, Chengdu, China;[Zheng, Zhian] School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha, China
摘要:
BACKGROUND: Metabolic dysfunction-associated fatty liver disease and steatotic liver disease (SLD) have emerged as one of the fastest-growing liver complications over recent decades; this study evaluated trends in SLD prevalence among Chinese adults and its variation across demographic groups. METHODS: A multicenter, retrospective study analyzed health examination data from 600,536 Chinese adults across 10 provinces between 2015 and 2023. Participants (≥18y) with complete data on age, sex, weight, waist circumference, ultrasound, and biochemical tests were included. SLD was defined as hepatic steatosis with at least 1 metabolic disorder. Age-adjusted SLD prevalence was calculated using direct standardization, and trends were assessed through Joinpoint regression. Logistic regression models examined associations between SLD and demographic/metabolic factors. RESULTS: Of 501,017 participants (mean age: 43.39±12.89y, 54.8% male), the age-standardized SLD prevalence rose from 36.24% in 2015 to 48.11% in 2023 (estimated annual percentage change=3.38, p=0.002). Metabolic dysfunction-associated SLD (MASLD) prevalence also increased from 30.01% to 43.17% (estimated annual percentage change=4.48, p<0.001). Subgroup analysis revealed significant increases in individuals aged 18-40, and higher prevalence in older adults, with the highest risk in those aged 41-64. Males had a higher risk than females (53.0% vs. 18.6%, adjusted prevalence ratio=1.97). Participants from North China had a higher risk of both SLD and MASLD compared with South China. Laborers had a lower risk of SLD compared with farmers, while administrative personnel had a higher risk of MASLD. CONCLUSIONS: The prevalence of SLD and MASLD significantly increased from 2015 to 2023, with notable variations across demographic groups, highlighting the need for targeted public health strategies to address the growing burden of SLD in China.
期刊:
Biomedical Signal Processing and Control,2025年107:107815 ISSN:1746-8094
通讯作者:
Genhua Liu<&wdkj&>Guoxiong Zhou
作者机构:
["Yang, Yixin; Sun, Lixiang; Tang, Zhiwen; Liu, Genhua; Zhou, Guoxiong; Li, Lin] School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China;[Cai, Weiwei] School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi 214122, China;[Li, Liujun] Department of Civil, Architectural and Environmental Engineering, Missouri University of Science & Technology, Rolla 65401, USA;[Chen, Lin] Zhuzhou Sany Aier Eye Hospital, Zhuzhou 412002, China;[Hu, Linan"] Department of Radiology, Zhuzhou Central Hospital, Zhuzhou 412001, China
通讯机构:
[Genhua Liu; Guoxiong Zhou] S;School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha 410004, China
摘要:
In 2013, an estimated 64 million people between the ages of 40 and 80 were suffering from eye disease. By 2020, that number had climbed to 76 million. It is predicted that by 2040, there will be a staggering 111.8 million glaucoma patients worldwide. Segmentation of blood vessels in retinal images can be used to investigate many diseases, and the complexity of the blood vessels and the variable conditions inside the retina pose a high challenge for accurate segmentation. Therefore, a competing fusion segmentation network (TAOD −CFNet) with a trumpet-like attention mechanism and optic disc gradient adjustment algorithm for retinal blood vessel segmentation. First, an optic disc gradient adjustment algorithm (ODGA) is proposed, which designs dual threshold weights for accurate localization and optimization of optic disc areas. Then, a competing fusion block (CFB) is proposed to improve the feature dissimilarity between the arteriovenous vascular sensitive area and the interference area. Finally, a Trumpet Attention Mechanism (TAM) is proposed to enhance the edge features of fine and peripheral blood vessels. TAOD-CFNet outperforms ten SOTA methods in ten-fold cross-validation, with IOU, F1-Score, Dice, Jaccard, ACC and MCC metrics reaching 83.28%, 89.41%, 84.28%, 80.35%, 96.94% and 88.81%. To demonstrate the generalization performance of the model, TAOD-CFNet outperforms ten SOTA image segmentation methods on six retinal image datasets (DRIVE, CHASEDB1, STARE, HRF, IOSTAR, and LES). The experimental results proved that TAOD-CFNet possesses better segmentation performance and can assist clinicians in determining the condition of retinopathy patients.
In 2013, an estimated 64 million people between the ages of 40 and 80 were suffering from eye disease. By 2020, that number had climbed to 76 million. It is predicted that by 2040, there will be a staggering 111.8 million glaucoma patients worldwide. Segmentation of blood vessels in retinal images can be used to investigate many diseases, and the complexity of the blood vessels and the variable conditions inside the retina pose a high challenge for accurate segmentation. Therefore, a competing fusion segmentation network (TAOD −CFNet) with a trumpet-like attention mechanism and optic disc gradient adjustment algorithm for retinal blood vessel segmentation. First, an optic disc gradient adjustment algorithm (ODGA) is proposed, which designs dual threshold weights for accurate localization and optimization of optic disc areas. Then, a competing fusion block (CFB) is proposed to improve the feature dissimilarity between the arteriovenous vascular sensitive area and the interference area. Finally, a Trumpet Attention Mechanism (TAM) is proposed to enhance the edge features of fine and peripheral blood vessels. TAOD-CFNet outperforms ten SOTA methods in ten-fold cross-validation, with IOU, F1-Score, Dice, Jaccard, ACC and MCC metrics reaching 83.28%, 89.41%, 84.28%, 80.35%, 96.94% and 88.81%. To demonstrate the generalization performance of the model, TAOD-CFNet outperforms ten SOTA image segmentation methods on six retinal image datasets (DRIVE, CHASEDB1, STARE, HRF, IOSTAR, and LES). The experimental results proved that TAOD-CFNet possesses better segmentation performance and can assist clinicians in determining the condition of retinopathy patients.
作者:
Lin Tan;Songtao Guo;Zhufang Kuang;Pengzhan Zhou;Mingyan Li
期刊:
IEEE Transactions on Wireless Communications,2025年:1-1 ISSN:1536-1276
作者机构:
[Lin Tan; Songtao Guo; Pengzhan Zhou; Mingyan Li] Key Laboratory of Dependable Service Computing in Cyber-Physical-Society (Ministry of Education), and College of Computer Science, Chongqing University, Chongqing, China;[Zhufang Kuang] School of Computer and Information Engineering, Central South University of Forestry and Technology, Changsha, China
摘要:
Low Earth Orbit (LEO) satellite networks hold great promise in the field of wireless communication due to their global coverage. However, the long communication distances and massive data computations present significant challenges for current satellite networks. To overcome these barriers, we propose SkyLink, a universal Integrated Ground-Air-Space Collaborative Edge Computing system that leverages horizontal collaboration among aerial platforms (AirXs) as well as vertical collaboration among Ground-Air-Space. We propose a bi-level optimization framework based on a Multi-Agent Twin Delayed Deep Deterministic policy gradient (MATD3) with Hybrid Action Space and constructe a latent representation space for each agent to allow the agent to learn the latent policy. This enabling each AirX to act as an agent and autonomously optimize its hybrid action decisions to improve system efficiency in real-time based on the dynamic network environment, a capability not achievable by conventional DRL methods. This includes continuous optimization variables such as AirX deployment (location changes) and resource allocation, as well as discrete optimization variables for collaborative task offloading decisions. Extensive experiments against state-of-the-art algorithms (e.g., MADDPG, QMIX) demonstrate that the proposed system improves energy efficiency by 27.2% and task completion rate by 6.8% compared to traditional Integrated Ground-Air-Space (IG) Network.