Abstract:Named entity recognition is an important foundational tool in application fields such as machine translation and question answering systems, and has always been a research hotspot in the field of natural language processing. Firstly, introduce the definition of named entity recognition, and organize the commonly used implementation tools, datasets, and evaluation criteria in named entity recognition tasks; Then, based on their development history, existing named entity recognition methods are summarized and divided into traditional methods and deep learning methods; Next, summarize the key ideas, advantages, and disadvantages of various methods for named entity recognition, provide the main process of deep learning based named entity methods, and summarize them in the order of the process; Finally, looking forward to the future research directions of named entity recognition, providing ideas for subsequent research.
Keywords:natural language processing;named entity recognition;machine learning;deep learning
Abstract:Since its proposal in 2014, generative adversarial networks have greatly promoted the progress of image generation research. Through the mutual game between two neural networks, it gradually improves the ability to distinguish between real images and generate images, as well as the ability to generate realistic images, ultimately achieving a Nash equilibrium between the two parties. Briefly introduce the generation of adversarial networks and sort out their application methods in the field of image generation around the issue of generating images containing targets. They are divided into five categories: direct method, iterative method, hierarchical method, decoupling method, and 3D modeling method. Focus on the research progress of generative adversarial networks in generating images containing targets, and prospect the development direction of image generation methods containing objects, in order to provide reference for relevant researchers in image generation research.
Abstract:With the rise of deep learning technologies, significant advancements have been achieved in the field of natural language processing(NLP), particularly in the domain of dialogue systems. This paper begins by providing an overview of the fundamental processes involved in dialogue systems. Subsequently, it comprehensively reviews deep learning-based techniques for dialogue systems, encompassing three key categories: convolutional neural network(CNN), recurrent neural network(RNN), and attention mechanism(AM). The paper introduces the principles of these models, and then provides an in-depth analysis and comparison of the applications, characteristics, and advantages and disadvantages of various basic models and their derivative models in dialogue tasks form the perspectives of information extraction, dialogue state tracking, and dialogue generation. Finally, this paper enumerates persisting challenges within dialogue tasks, and proposes feasible solutions.
Keywords:deep learning;natural language processing;attention mechanism;dialogue system;neural network
Abstract:The rapid development of social media platforms not only greatly enhances the accessibility of information, but also accelerates the spread of fake news. The explosive growth of fake news not only damages social stability, but also erodes public trust in the media. In the field of natural language processing, fake news detection is a crucial and challenging task. To this end, first provide a definition of fake news and analyze its characteristics in depth; Then, analyze and evaluate the existing fake news detection methods from four perspectives: news content, social context, communication networks, and hybrid methods. Introduce the performance of relevant models, commonly used datasets, and evaluation indicators; Finally, summarize and analyze the current problems in fake news detection research, and propose possible future research directions.
Keywords:social network;fake news;natural language processing;early detection;explainability
Abstract:Aiming at the problem that most recommendation algorithms in real recommendation scenarios ignore the timeliness factors of dynamic changes in user preferences, resulting in limited model performance, a recommendation fusion model based on LSTM and deep matrix factorization (Long Short-Term Memory Fusion Deep Matrix Factorization, LFDMF) is proposed. The model uses generalized matrix factorization to learn nonlinear low-order features between users and items, uses multi-layer perceptron to learn nonlinear high-order features between users and items, and obtains users ' long-term dynamic preferences. LSTM 's strong fitting ability to time series is used to obtain users ' short-term dynamic preferences. In order to verify the effectiveness and feasibility of the LFDMF model, comparative experiments are carried out on the public datasets MovieLens-1M and Pinterest. The simulation results show that the HR @ 10 and NDCG @ 10 indexes of the LFDMF model are improved by 0.103 4 and 0.132 2,0.118 1 and 0.101 8, respectively, compared with the traditional MF algorithm. Compared with the DMF model, it is improved by 0.022 8 and 0.032 3, 0.016 9 and 0.013 5, respectively,the recommendation performance is significantly improved.
Abstract:Chinese ancient poetry has the characteristics of neat structure, rich emotions, and appropriate length, making it suitable as a natural carrier for generative information hiding. However, due to its implicit semantics, profound meaning, and diverse styles, the automatic generation of poetry hiding is extremely challenging. With the emergence of big language generation models, the quality of poetry generation has been greatly improved, but there have been no reports on applying big language models to implicit writing in poetry generation. To this end, the BERT sentiment analysis model was introduced and combined with the Attention mechanism to design a hidden writing algorithm and model for Seq2Seq to automatically generate poetry. The corresponding hardware system and GUI interface were constructed based on PN40. Under the guidance of theme words and the constraints of rhythm/emotion/mutual information, multi-mode implicit poetry was generated and quickly output on the system. The experimental results show that the proposed model generates implicit poems with clear themes and consistent emotions. The BLEU evaluation value is as high as 44.3%, and the average accuracy of sentiment analysis is above 85%, greatly enhancing the perception and statistical concealment of implicit poems and accelerating the application of generative information steganography.
Keywords:poetry;large language model;sentiment analysis;concealability;information steganography
Abstract:Knowledge graph question answering is one of the hot research areas in the field of natural language processing. Existing methods face two main challenges: difficulty in understanding complex natural language questions and limited semantic interpretation of entity representations. To address these challenges, a knowledge graph question answering method called DEK-KGQA is proposed, which integrates external knowledge. First, a joint graph is constructed by combining the question knowledge graph subgraph and the QA context. Then, the relevance scores of nodes in the joint graph are calculated using pre-trained language models. Finally, external knowledge is introduced to enhance information interaction and reasoning ability during the question answering process. Experimental validation is conducted on the CommonsenseQA dataset, comparing the proposed method with existing methods. The results demonstrate that the proposed method achieves better performance in commonsense question answering tasks, validating its effectiveness. In addition, ablation experiments are conducted to evaluate the impact of each component on the overall performance.
Keywords:knowledge graph question answering;QA context;pre-trained language model;external knowledge
Abstract:In the complex distribution network environment, there are problems of insufficient accuracy and real-time calculation of line loss. This paper proposes an improved TCN-BiGRU distribution line loss prediction method based on recurrent neural network autoencoder. The TCN neural network model, which is good at processing time series, was selected as the backbone feature extraction network, and the BiGRU unit was integrated into the TCN to effectively solve the problem of gradient vanishing. On this basis, combined with the recurrent neural network autoencoder, the line loss outliers were classified and labeled unsupervised, and the causes of the line loss rate anomalies were predicted through the softmax loss function and the corresponding loss reduction measures were formulated. The experimental results show that compared with the traditional distribution network line loss prediction method, the root mean square error of the proposed line loss prediction method is reduced by 0.036 99 and 0.004 02 compared with the traditional EMD-LSTM and PSO-CNN algorithms, respectively, and the accuracy of line loss cause analysis is improved by 1.500% and 5.841% compared with the ResNet50 and DBN-DNN algorithms, respectively. It provides a scientific reference for energy saving and loss reduction in the distribution network after the distributed power generation is connected and the realization of the dual carbon goal of the power grid.
Keywords:recurrent neural network autoencoder;TCN-BiGRU line loss prediction algorithm;smart grid;analysis of the causes of abnormal line loss;prediction of line loss in the substation area
Abstract:In order to monitor and analyze the constraining effect of meteorological factors on the outbreak of blue-green algae blooms, establish the response relationship between blue-green algae blooms and meteorological factors, and predict changes in blue-green algae biomass and distribution. Quantitatively evaluate the importance and contribution rate of meteorological factors through random forest algorithm, select meteorological factors, and construct a blue-green algae bloom prediction model using long short-term memory network. Selecting the annual values of normalized vegetation index in Dianchi Lake from 2000 to 2021 and meteorological data from Kunming Station during the same period as the research objects, this study explores the feasibility of using the long short-term memory network RF-LSTM based on the random forest algorithm to predict blue-green algae blooms, with normalized vegetation index as the indicator of changes in blue-green algae blooms. The results showed that compared with the single structure long short-term memory network model and the single algorithm random forest model, the RF-LSTM model's annual value prediction passed the 0.01 significance test, and the simulation accuracy reached 90.9%. The random forest algorithm is beneficial for understanding the relationship between blue-green algae blooms and meteorological factors, selecting meteorological factors with predictive performance, and thus improving the predictive ability of long short-term memory network models.
Keywords:cyanobacterial blooms;meteorological factor;random forest;long and short term memory network;Dianchi Lake
Abstract:A speech recognition approach has been proposed to address the problem of reduced recognition accuracy and poorer generalization performance due to insufficient training data in low-resource conditions. This method leverages convolutional neural networks to extract feature information. It combines the attention mechanism with delayed time-delay neural networks, referred to as ADTDNN, enhancing the model's ability to capture key information in sequences within low-resource environments. The approach employs linking temporal classification to streamline the recognition process of the model. Additionally, a Transformer is utilized as the language model. Experimental results on the Aishell-1 dataset demonstrate that the ADTDNN-based speech recognition model in low-resource settings reduces word error rates by 3.7% and 1% compared to mainstream end-to-end models like LAS and Transformer, respectively.
Abstract:Open-domain question answering (OpenQA) is a challenging task in natural language processing,the conventional machine learning and deep learning techniques are commonly used to retrieve many candidate document fragments related to the question from the raw corpus for answer extraction. However, the candidate document fragments retrieved by current methods tend to include considerable noise and irrelevant information to the question, and the previous OpenQA model falls short in accurately responding to questions that necessitate multiple document fragments as correlative evidence. Therefore, this paper proposes an open-domain question answering method based on refinement and fusion of multiple documents(RFMD). Specifically, RFMD designs a Transformer-based document refinement module during the retrieval stage to reduce noise information in the candidate documents. In the reading comprehension stage, RFMD employs a text generation-focused question answering module. By constructing a global attention mechanism across document fragments, it integrates information from multiple relevant document fragments to accurately answer questions that require multiple document fragments as supporting evidence. RFMD achieves EM scores of 45.8% and 63.4% on the NaturalQuestions and TriviaQA datasets respectively, verifying the effectiveness and superiority of the model in OpenQA tasks.
Abstract:To solve the intelligent psychological evaluation task for high-risk industry employees in the power grid before work, a multimodal psychological evaluation algorithm is proposed for employees in the power grid industry, which includes expressions, sounds, and walking posture. Firstly, construct a dataset of employees in the power grid industry, extracting facial RGB image sequences, audio ComparE feature sets, and human skeleton keypoint sequences from videos; Secondly, residual networks and bidirectional long short-term memory networks are used to extract facial visual features, audio features are extracted in time windows, and gait features are extracted in spatiotemporal graph convolutional networks, respectively, to obtain the optimal single modal models; Finally, a deep learning training method for polarity loss function and a multimodal fusion algorithm based on attention mechanism are proposed to obtain the optimal multimodal psychological state evaluation model by fusing the output features of a single modal model. The experiment shows that multimodal fusion can significantly improve the accuracy of psychological evaluation compared with single-mode system, and the accuracy rate of the four classification tasks of psychological labels reaches 65.66%. Compared with the model based on facial expression, voice, and gait, the effect of multimodal fusion is increased by 18.04%, 21.22%, and 13.28% respectively.
Keywords:multimodality;deep learning;attention mechanism;psychological evaluation;power grid industry
Abstract:Due to harsh operating conditions and high load requirements, rotating machinery faults can lead to high maintenance costs and unnecessary downtime. It is necessary to develop an efficient and accurate rotating machinery fault online diagnosis and prediction system to help enterprises quickly identify faults, predict future events, and optimize maintenance plans. The construction state matrix represents the vibration signal generated by rotating machinery, that is, a continuous time series is divided into several Windows, and each window is converted into an image. Image features are extracted and processed by a series codec of a specific structure to classify vibration patterns in a training data set. The reliability and effectiveness of the online fault diagnosis and prediction system for rotating machinery based on self-attention codec structure are verified by simulation experiments. The state characteristic database of rotating machinery can accurately diagnose and predict the faults of rotating machinery. The system can help businesses optimize maintenance schedules and reduce downtime and maintenance costs.
Abstract:To solve the problems of low accuracy and severe lack of corpus in named entity recognition tasks in the education field, a named entity recognition model WBBAC that integrates word information and self attention is proposed. This model utilizes a BERT pre trained language model to enhance the semantic representation of word vectors and introduces word frequency information into them. The word vectors are concatenated with each other as inputs to a bidirectional long short-term memory network, which further searches for internal connections within the sequence through a self attention layer. Finally, the optimal sequence is obtained through CRF decoding. Create a computer composition principle dataset based on the characteristics of the course text and annotate it. Conduct experiments on the Resume dataset and the computer composition principle dataset, and the F1 values of the WBBAC model are 95.65% and 73.94%, respectively. The experimental results show that compared with the baseline model, the WBBAC model has a higher F1 value, effectively solving the problem of insufficient annotated data in named entity recognition tasks in the education field.
Abstract:Aiming at the problem that current short text classification only uses classification labels as the basis for judging classification results, and ignores the semantic information contained in the classified label text, a short text classification method based on label aware attention is proposed, which is based on a large-scale pre trained language model. This method represents text data in distributed vector form through large-scale pre trained language models to obtain richer semantic information; At the same time, incorporating classification label information into the text data training process, using attention mechanisms to make the text data perceive and classify the most relevant information; Using CNN networks and max pooling layers to extract local word level vector features, in order to better address semantic issues such as double negation and comparative negation in English texts; Using residual connections to fuse sentence level vectors with word level vectors effectively alleviates the problem of text information decay. Tests were conducted on three common English datasets, R8, R52, and MR, and the experimental results showed that the proposed method achieved accuracies of 98.51% and 97.10% on the R8 and R52 datasets, respectively, which are better than DeBERTa and BertGCN.
Keywords:short text classification;CNN;label awareness;attention;pre-trained
Abstract:Oversampling is a commonly used method to solve the problem of imbalanced class distribution in a dataset by synthesizing new samples of the same class. A PSON algorithm based on neighborhood concept is proposed to address the issue of imbalanced sample distribution in the dataset. This algorithm defines the influence of each minority class sample and oversamples the minority class samples based on different influences to obtain a balanced dataset. Classification tests were conducted on datasets obtained from 8 oversampling algorithms on 50 datasets. The Wilcoxon symbol rank test was used to compare 7 classification performance indicators, and the results showed that the use of PSON algorithm significantly improved classification accuracy.
Abstract:A mathematical model for vehicle path planning (DCVRPTW) was established to optimize the delivery path of agricultural inputs supply chain orders, taking into account constraints such as the comprehensive range, maximum load capacity, and time window of new energy trucks. The model comprehensively optimizes the fixed and transportation costs of vehicles, and proposes a swarm intelligence optimization algorithm framework based on deep reinforcement learning (DRL-SIA). An intelligent agent is a decision-maker who selects the best action from the action pool based on the environmental state as input to change the environment and obtain environmental rewards. The DRL-SIA algorithm combines trained agents with swarm intelligence algorithms to replace the original algorithm for decision selection, thereby improving optimization speed and accuracy. The experiment shows that the optimal solution of the proposed algorithm is superior to other algorithms in all cases, verifying that the algorithm can effectively reduce logistics transportation costs in the agricultural material supply chain.
Abstract:Aiming at a variety of small picking problems of parts, in the multi-person collaborative picking mode, the task allocation is unreasonable, the picking time varies greatly, and the picking link is easy to timeout. A multi-person collaborative picking model aiming at the shortest picking time is constructed, and the improved K-means algorithm and genetic algorithm are used to solve the model. Aiming at the shortcomings of the traditional K-means algorithm clustering results, the number of picking points contained in each cluster varies greatly. The picking time of each cluster is used as an index to transform the cluster where the picking points belong. The genetic algorithm is used to perform path planning and picking time calculation on the clustering results to obtain the optimal clustering results. Taking the parts picking process of a security equipment manufacturing enterprise as the research object, the effectiveness of the algorithm is verified by comparing with the picking time obtained by simple batching.
Abstract:With the continuous development of our country's economy, wine, as a delightful medium of life, has entered the public dining table. However, the quality inspection of wine is still mainly based on the tasting of wine tasters, which can no longer meet the needs of the large-scale and intelligent development of the food industry. Therefore, based on the support vector machine algorithm, the physicochemical indicators of wine were modeled, and the Box plot method was implemented using R language to handle outliers. At the same time, the support vector machine parameters of the RBF kernel were optimized, resulting in a wine quality detection model with an accuracy of 96.46%. This provides an effective approach for wine quality control.
Abstract:Ensuring security and efficiency when intelligent agents perform tasks in complex environments is a major challenge. Traditional reinforcement learning methods use model free reinforcement learning to solve intelligent decision-making problems, constantly trial and error to find the optimal strategy using a large amount of data, ignoring the training cost and security risks of the agent, and therefore cannot effectively ensure the safety of decision-making. To this end, safety constraints are added to the actions of intelligent agents in the model predictive control framework, and a safety reinforcement learning algorithm is designed to obtain the safest action control sequence. At the same time, in response to the problems of high computational complexity and low efficiency in the cross entropy method, as well as the problem of falling into local optima in the gradient optimization method, a combination of robust cross entropy and gradient optimization methods is used to optimize the action control sequence to improve algorithm safety and solving efficiency. The experiment shows that the proposed method can effectively improve the convergence speed compared to the robust cross entropy method, and has the best safety performance compared to other optimization algorithms without sacrificing much performance.
Abstract:In the era of big data, the scale of mobile terminal users continues to expand, and the Internet of everything brings great convenience to people. At the same time, there is also the problem of geographic dispersion of a large amount of data, which brings great challenges to the QoS of user service. In this paper, a task unloading model based on the three-layer service architecture of the mobile edge computing platform is first built. Combined with the actual application scenario of the MEC platform, the deep reinforcement learning algorithm is improved by using the same policy experience playback and entropy regularization, and the task unloading strategy of the MEC platform is optimized. Experiments are designed to compare and analyze the three indexes of energy consumption, delay and network usage of the three traditional algorithms and the improved algorithm, and verify that the improved algorithm has better performance in reducing energy consumption, delay and network usage.
Abstract:Since the tobacco retail stores in cities are dense, traditional path planning algorithms for solving the optimal supervision path will consume a lot of time, and cannot guarantee the effect within the specified time. In addition, existing methods seldom consider the network characteristics and the explainability of the candidate subset.This study proposes a graph attention-based node selection and path optimization algorithm (GA-SGPO), which iteratively selects the optimal coordinate node subset and performs calculation on the subset to reduce computation time. In addition, the structural similarity between nodes is calculated to reduce the sparsity of training samples.The experimental data includes the coordinates of 40,000 retail stores in Dongguan City. The experimental results show that the GA-SGPO model ensures the solution accuracy while the solution time is reduced by an average of 48%.The GA-SGPO can significantly save computational time and is closer to practical application scenarios. The attention mechanism and node similarity calculation can provide visualization basis for optimal node selection.
Abstract:Aiming at the problems of sample imbalance and category overlap in personal credit risk assessment, an improved active generative oversampling model is proposed. Firstly, based on the auxiliary classifier generative adversarial network (ACGAN) framework, Wasserstein distance is introduced to improve the true false discrimination loss function, and gradient penalty is added to prevent pattern collapse; Secondly, Focal loss is used instead of traditional cross entropy loss to enhance the ability to identify difficult samples; Finally, the proposed model is used to oversample imbalanced data to improve classifier performance. Experiments on real credit data show that the model improves the classifier's classification performance indicators F1, AUC, and G-means by 11.2%, 1.7%, and 12.8%, respectively. It achieves significant results in enhancing sample diversity, reducing class overlap, and improving the classifier's classification performance on imbalanced datasets.
Abstract:To solve the problems of low accuracy and easy false detection of small targets in current smoking behavior detection, an improved YOLOv5 recognition model YOLOv5s+is proposed. This model combines the backbone network of YOLOv5 with BoTNet to improve the feature extraction ability of the model, enabling it to detect smaller target objects; At the same time, the feature fusion part is improved by applying a weighted bidirectional feature pyramid BiFPN in the neck of the network model to efficiently fuse shallow position information and deep high-level semantic information, effectively improving detection accuracy. Integrate publicly available online datasets and self-made datasets into an office smoking experimental dataset, and compare the detection performance of the YOLOv5s+ model with the original YOLOv5 model on this dataset. The experimental results show that the average accuracy (mAP) of the improved model YOLOv5s+ is 81.8%, with an accuracy of 82.8% and a recall rate of 83.9%. Compared with the original model, it has improved by 5.4%, 4.1%, and 6.4%, respectively, and has achieved good detection of office smoking behavior.
Abstract:The graph convolutional neural network based on human skeleton data is not easily affected by background environmental noise and has strong robustness, which has become a research focus in the field of human action recognition at present. However, this network assigns the same weight to different neighborhoods in the same order, which limits its ability to capture spatial information correlations. To this end, a graph attention network weighted sum is introduced to sum the features of adjacent nodes, allowing each node to assign different weights based on its adjacent features to enhance feature extraction and learning effectiveness. At the same time, in order to solve the problem of representing the skeleton as an undirected graph, only the relationship between adjacent nodes or edges can be determined, which limits the ability to capture dependency relationships between nodes or edges. Introducing directed graph convolution, utilizing the feature information of first-order and second-order adjacent nodes for graph convolution, not only preserves the directional features of the directed graph, but also expands the perceptual domain of graph convolution, thereby extracting more features. The experiment shows that the proposed method can effectively improve the accuracy of action recognition.
Abstract:In the field of meteorology, the meteorological data detected by Doppler weather radar is stored in a three-dimensional polar coordinate system with the radar station as the origin, which has characteristics such as irregular shape and large data volume. The Marching Cubes (MC) algorithm is a classic algorithm in 3D reconstruction, but it has the disadvantages of low reconstruction efficiency and inability to directly process meteorological data when applied in the meteorological field. In order to achieve three-dimensional reconstruction of meteorological data, a radar data normalization and state labeling discrimination algorithm NBV-MC is proposed based on the MC algorithm. This algorithm normalizes and preprocesses radar based data files based on their characteristics, constructs trapezoidal hexahedral voxels using radar based data, marks the state of each hexahedral voxel, and dynamically determines whether it needs to be processed when traversing hexahedral voxels. The experimental results show that the NBV-MC algorithm not only solves the problem of being unable to be directly used in the MC algorithm due to the irregularity of meteorological data, but also effectively reduces the number of triangles required to draw contour surfaces while ensuring data authenticity and reconstruction efficiency, and improves reconstruction speed. Compared with the MC algorithm, the reconstruction efficiency of NBV-MC algorithm has increased by more than 77.70%, which is conducive to real-time scene interaction and facilitates meteorological researchers to directly analyze radar data.
Abstract:Due to the fact that extracting dynamic multi scene video keyframes directly from frame difference data often results in excessive redundant frames, the directional gradient histogram (HOG) feature has good stability for image brightness and scene changes. Therefore, a composite HOG feature clustering method for extracting keyframes from multi scene videos has been proposed to improve the efficiency of keyframe extraction. Firstly, by extracting the HOG features of video frames and introducing image information entropy, a composite feature vector is constructed to maintain the correlation of data features. Secondly, based on the composite feature vector, the difference data between video frames is calculated to determine the number of video segmentation shots and keyframe extractions; Again, considering both the intra shot frame set and the complete video frame set, select the video frames with high information entropy as the initial clustering centers without repetition to guide the clustering algorithm search direction, and extract video keyframes through K-means clustering. Compared with the traditional K-means clustering method, it was found that the proposed algorithm reduces redundancy by 0.003~0.015, improves precision by 0.14~0.21, reduces clustering time, and has better accuracy and efficiency.
Abstract:In order to achieve accurate Detection of Leaf blight of silage Maize, reduce the cost of manual diagnosis in field environment and reduce the impact of the disease, a modified YOLOv7-MLD(Maize Leaf-Blight Detection) model was proposed. Firstly, Diverse Branch Block (DBB) module was added to the backbone of YOLOv7 network to enhance its feature extraction capability. Then a Coordinate Attention module is added to the three output feature layers to enhance the ability of extracting disease features. Finally, the loss function is replaced by CIoU with SIoU to improve the convergence speed and regression accuracy of the bounding box. Experiments were carried out on a subset of maize leaf wilt disease data set, and the results showed that the AP value of YOLOV7-MLD model reached 84.2%, the F1 value increased by 5.9%, the accuracy rate and recall rate increased by 4.3% and 7.3%, respectively, compared with the original YOLOv7. The model can accurately locate and identify the leaf blight of silage maize in the complex field environment, and has very important practical significance for guiding the prevention and control of the early leaf blight of silage maize.
Abstract:A Transformer based end-to-end image level weakly supervised semantic segmentation network is proposed to address the complex scene and high annotation cost of remote sensing image semantic segmentation tasks. The network first improves the accuracy and granularity of the class activation map through a multi class label encoding module; Then, the affinity pseudo label generation module is used to further refine the representation of affinity relationships, generating high-precision affinity pseudo labels as segmentation supervision information, thereby improving the ability of weakly supervised networks; Simultaneously designing a mixed label data augmentation module to enhance the composition of remote sensing data; Finally, a mixed loss function with fusion affinity loss is provided to enhance the learning performance of the network. The experimental results on the ISAID dataset show that the model achieves an mIoU of 38.836% in segmentation results using image level labels, demonstrating better robustness and reliability compared to the control network. It has high application value in weakly supervised semantic segmentation of remote sensing images.
Abstract:In order to solve the problems of "hard integration" and "superficiality" in the ideological and political construction of data structure courses in information majors, in accordance with the requirements for the cultivation of engineering students in the "Guidelines for Ideological and Political Construction of Higher Education Curriculum", we will deeply explore the ideological and political elements related to data structures from the perspectives of engineering ethics education, national craftsmanship spirit, patriotism, and mission responsibility, and integrate them into professional course teaching. The practical results show that after the reform of ideological and political education in the curriculum, students have a clearer grasp of professional knowledge points, a stronger interest in learning, and achieved a silent educational effect. The ideological and political elements mentioned can also provide reference for ideological and political education in other courses of information majors
Keywords:curriculum ideological and political education;data structure;information major;teaching cases
Abstract:The new era has put forward new requirements for higher education teaching and talent cultivation, and the country and society's thirst for high-quality innovative talents is becoming increasingly urgent. In this context, new trends have emerged in the construction of university curricula, and the reform of curriculum and teaching is constantly deepening. To this end, the basic connotation of curriculum construction and teaching reform is first pointed out, which not only points out the core elements of 12 aspects, but also outlines the internal relationships of each element, and discusses the key elements among them; Then analyze the challenges and development trends faced by curriculum construction and teaching reform in the new era. Practice has shown that university teachers should always regard curriculum construction as the core work of talent cultivation, regard teaching reform as an important means to improve quality, continuously improve the comprehensive ability and level of curriculum teaching, create higher quality first-class courses, and lay a solid foundation for talent cultivation.