Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

114 results about "Learning by example" patented technology

Visual salience and semantic attribute based cross-modal image natural language description method

The invention belongs to the technical field of computer vision and natural language processing, and discloses a visual salience and semantic attribute based cross-modal image natural language description method. The method comprises the steps that multiscale deep visual features of all regions are extracted by adopting a convolutional neural network; by means of a pre-trained significance model,an image significance graph is returned, and an original image is weighted; a predefined dictionary is built to serve as a semantic attribute category, and semantic attribute detection is conducted ona visual significance image; semantic attributes are calculated through multi-instance learning; image features are weighted through the semantic attributes; visual-salience-based semantic attributefeatures are decoded through a long short-term memory network, and image description is generated. The method has the advantage of being high in accuracy and can be used for image retrieval under complex scenes, multi-objective image semantic understanding and the like.
Owner:XIDIAN UNIV

Network violent video identification method

The invention discloses a network violent video identification method based on multiple examples and multiple characteristics. The method for identifying the network violent videos comprises the steps of grasping violent videos, non-violent videos, comments on the violent videos, comments on the non-violent videos, brief introductions of the violent videos and brief inductions of the non-violent videos from a video sharing network, and structuring a video data training set; extracting textural characteristics from textural information of the training set, forming textural characteristic vectors to train a textural pre-classifier, and screening out candidate violent videos by using the pre-classifier; using a shot segmentation algorithm based on a self-adapting dual threshold for conducting segmentation on video segments of the candidate violent videos, extracting related visual characteristics and voice frequency characteristics of each scene to express the scene, taking each scene as an example of multi-example study, and taking video segments as a package; and using an MILES algorithm for converting the package into a single example, using a characteristic vector for training a classifier model, and using the classifier model for conducting classification on the candidate violent videos. By the utilization of the network violence video identification method, bad influences that the network violent videos are broadcasted without constrain are largely lightened.
Owner:人民中科(北京)智能技术有限公司

Artificial intelligence-based multi-label classification method and system of multi-level text

The invention relates to an artificial intelligence-based multi-label classification method and system of multi-level text. The method includes: 1) utilizing a neural network to construct a multi-label classification model of the multi-level text, and obtaining text class prediction results of training text according to the model; 2) carrying out learning on parameters of the multi-label classification model of the multi-level text according to existing text class labeling information in the training text and the text class prediction results, which are of the training text and are obtained inthe step 1), to obtain a multi-label classification model of the multi-level text with determined parameters; and 3) utilizing the multi-label classification model of the multi-level text with the determined parameters to classify to-be-classified text. The method infers labels of the formed text simply through the document-level labeling information, and can be well applied to scenes where labels of formed text are difficult to collect; compared with traditional multi-instance learning (MIL) methods, the method of the invention introduces minimal assumptions, and can better fit actual data;and the method of the invention has good scalability.
Owner:INST OF INFORMATION ENG CHINESE ACAD OF SCI

Semantic propagation and mixed multi-instance learning-based Web image retrieval method

The invention belongs to the technical field of image processing and particularly provides a semantic propagation and mixed multi-instance learning-based Web image retrieval method. Web image retrieval is performed by combining visual characteristics of images with text information. The method comprises the steps of representing the images as BoW models first, then clustering the images according to visual similarity and text similarity, and propagating semantic characteristics of the images into visual eigenvectors of the images through universal visual vocabularies in a text class; and in a related feedback stage, introducing a mixed multi-instance learning algorithm, thereby solving the small sample problem in an actual retrieval process. Compared with a conventional CBIR (Content Based Image Retrieval) frame, the retrieval method has the advantages that the semantic characteristics of the images are propagated to the visual characteristics by utilizing the text information of the internet images in a cross-modal mode, and semi-supervised learning is introduced in related feedback based on multi-instance learning to cope with the small sample problem, so that a semantic gap can be effectively reduced and the Web image retrieval performance can be improved.
Owner:XIDIAN UNIV

Plate and strip steel surface defect detection method based on saliency label information propagation model

The invention relates to the technical field of industrial surface defect detection, and provides a plate strip steel surface defect detection method based on a significance label information propagation model. The method comprises the following steps of firstly, acquiring a plate strip steel surface image I; then, extracting a bounding box from the image I, and executing a bounding box selectionstrategy; then, performing super-pixel segmentation on the image I, and extracting a feature vector from each super-pixel; then, constructing a significance label information propagation model, constructing a training set based on a multi-example learning framework to train a classification model based on a KISVM, classifying a test set by using the trained model to obtain a category label matrix,calculating a smooth constraint item and a high-level prior constraint item, and optimizing and solving a diffusion function; and finally, calculating a single-scale saliency map under multiple scales, and obtaining a final defect saliency map through multi-scale fusion. The surface defects of the strip steel can be efficiently, accurately and adaptively detected, a complete defect target can beuniformly highlighted, and a non-significant background area can be effectively inhibited.
Owner:NORTHEASTERN UNIV

Face comparing verification method based on multi-instance learning

The invention discloses a face comparing verification method based on multi-instance learning, applied to a people and certificate consistency verification occasion. The method performs face comparing verification based on the thought of multi-instance learning, and comprises the following steps of: S1, face image preprocessing; S2, face multi-instance learning training; and S3, face verification. The face image preprocessing step comprises face detection, feature point locating and DoG light treatment; the face multi-instance learning training step comprises face multi-instance definition, multi-instance feature extraction and multi-instance feature fusion; and the face verification step is to perform face consistency verification according to stock equity of each instance and similarity of matched instances in the step S2. The method solves a difficult problem of change of hair style, skin color, making up, micro plastic, and so on, in face comparing verification, provides an effective algorithm and a train of thought for face verification, and improves reliability of face verification. The method provided by the invention can be widely applied to the people and certificate consistency verification occasion for checking whether certificates, such as a 2nd-generation ID card, a passport, a driving license and a student ID card, are held by owners thereof or not.
Owner:GUANGDONG MICROPATTERN SOFTWARE CO LTD

Image multi-tag marking algorithm based on multi-example package feature learning

The invention discloses an image multi-tag marking algorithm based on multi-example package feature learning, and the algorithm comprises the steps: obtaining a set of image blocks of all training images; extracting the features of a color histogram and the features of a direction gradient histogram of each image block of the set of the training images; enabling one training image to serve as an image package, and obtaining an image package structure needed by a multi-example learning framework; enabling the examples in all image packages in the set to form a projection example set, enabling each image package to be projected towards the projection example set, and obtaining the projection features of the image packages; selecting the features with the high discrimination performance as the classification features of the image packages; importing the classification features of the image packages of the learned training image set into an SVM classifier for training, obtaining the parameters of a training model, and predicting a test image tag through employing a trained SVM classifier. The algorithm is simple in implementation, and a trainer is mature and reliable. The algorithm is quick in prediction, and achieves multiple image tags better.
Owner:SHANDONG INST OF BUSINESS & TECH

Object segmentation method based on multiple-instance learning and graph cuts optimization

The present invention discloses an object segmentation method based on multiple-instance learning and graph cuts optimization. The method comprises the first step of carrying out salient model construction by adopting a multiple-instance learning method on training images, and predicting packages and instances of a testing image by using a salient model, thus to obtain a saliency testing result of the testing image; a second step of introducing the saliency testing result of the testing image into a graph-cut frame, optimizing the graph-cut frame according to instance characteristic vectors and marks of the instance packages, acquiring a second-best solution of graph cuts optimization, and obtaining precise segmentation of an object. According to the method provided by the present invention, the saliency testing model is constructed by using the multiple-instance learning method and thus is suitable for images of specific types, the saliency testing result is used into an image segmentation method based on the graph theory so as to guide image segmentation, a graph cut model frame link is optimized, an agglomerative hierarchical clustering algorithm is adopted for solving, the segmentation result can thus well accords to semantic aware output, and an accurate object segmentation result can be obtained.
Owner:CHANGAN UNIV

Neural network-based movement recognition method

The invention discloses a neural network-based human body movement recognition method. The method includes the following steps of based on a video database, training N mutually independent 3D convolutional neural networks as a video feature extractor; according to the video feature extractor, training a multi-instance learning classifier; inputting a to-be-recognized video, extracting video features through the well trained network, and classifying movements by the classifier. According to the technical scheme of the invention, the influence of a large amount of noise features on a classification result is avoided. Meanwhile, the negative effect of a fixed sample length on a movement recognition result is also avoided.
Owner:TIANJIN UNIV

Video facial expression early detection method based on multi-instance learning

The invention discloses a video facial expression early detection method based on multi-instance learning. The method comprises the steps that (1) video data of training samples and to-be-tested samples is preprocessed, and a facial region in each frame image in a video is extracted; (2) an LBP descriptor is adopted to perform feature extraction on the facial region, obtained through preprocessing, in each frame image, and feature vectors of each frame image are obtained; (3) an expression early detection function is solved by the adoption of an extended structured output support vector machine based on multi-instance learning according to the feature vectors of the training samples; and (4) the early detection function obtained in the step (3) is used to perform facial expression early detection according to the features vector of the to-be-tested samples in the step (2), and an expression early detection result is obtained. Through the method, expressions can be monitored in real time, and the recognition rate is high.
Owner:SOUTHEAST UNIV

Weak supervision target detection method and system based on transfer learning

The invention discloses a weak supervision target detection method and system based on transfer learning; the method comprises the steps: extracting the features of an input strong supervision image and a weak supervision image through a deep convolutional neural network, extracting candidate boxes in the images through a region suggestion network, and obtaining the visual features of different candidate regions; performing feature extraction on category texts in the strong supervision data set and the weak supervision data set, establishing a semantic graph, and performing optimization by using a graph convolutional network to obtain semantic features of all category texts; employing dual-supervised average teacher network structure, which comprises a strong supervised classification and boundary regression student network, a weak supervised multi-instance learning student network and a classification and boundary regression teacher network; and aggregating bounding box information and classification information in the strong supervision data set and the weak supervision data set by using visual features and optimized semantic features, thereby performing bounding box regression and classification on candidate boxes. According to the invention, the weak supervision target detection effect is improved.
Owner:SHANGHAI JIAO TONG UNIV

Multi-instance weighted packet learning method for online uncertain image recognition

The invention provides a multi-instance weighted packet learning method for online uncertain image recognition, which reduces the influence of noise data on classification results in online image recognition through improving a multi-instance learning algorithm, and puts forward the ideal of endowing different weights to packet instances, updating a classifier continuously and adjusting the weights to increase classification accuracy. The multi-instance weighted packet learning method is different from the traditional multi-instance learning method for image recognition in that: during the traditional multi-instance learning, a training set is composed of a plurality of packets, each packet comprises a plurality of instances, and tags of the packet instances are unknown. The multi-instance weighted packet learning method for online uncertain image recognition reduces the interference of noise instances on classification prediction through weighing the instances, iterates the training classifier, trains instance weights, represents the packets and achieves higher classification accuracy.
Owner:SHENZHEN TISMART TECH CO LTD

Weighted extreme learning machine video target tracking method based on weighted multi-example learning

The invention discloses a weighted extreme learning machine video target tracking method based on weighted multi-example learning, solving the problem of bad tracking accuracy in the prior art. The method includes 1. initializing a Haar-like feature similar model pool and constructing a variety of feature model blocks, setting the weighted extreme learning machine network parameters; 2. extracting the training samples in the current frame and their feature blocks corresponding to the feature blocks of the different feature model blocks; 3. calculating the weighted multi-instance learning weight values; 4. constructing a plurality of networks corresponding to the different feature blocks and selecting the network with the largest similarity function value of the packet and the corresponding feature model block; 5. calculating the network global output weight values; 6. extracting the detection samples in the next frame and their corresponding feature blocks corresponding to the selected feature model blocks; 7. classifying the detection samples by means of the selected network and obtaining the target position of the next frame; and 8. repeating the above steps until the video is ended. According to the invention, the tracking accuracy is improved, and the target robustness tracking is realized.
Owner:XIDIAN UNIV

Electrocardiosignal classification method and device, electronic equipment and storage medium

The embodiment of the invention relates to an electrocardiosignal classification method and device, electronic equipment and a storage medium. The electrocardiosignal classification method disclosed by the embodiment of the invention comprises the following steps: acquiring electrocardiosignals; detecting a QRS complex wave from the electrocardiosignal; carrying out single heart beat cutting on the electrocardiosignals with the detected QRS composite waves and then packaging the electrocardiosignals into a plurality of heart beat packets; inputting the plurality of heart beat packets into a classifier model based on multi-example learning, obtaining a type identification result of each heart beat packet, wherein when the classifier model based on multi-example learning is used for classification, each heart beat packet is used as an example packet, each heart beat signal in the heart beat packet is used as an example in the example packet, and the type identification result comprises anormal rhythm type and an abnormal rhythm type. The electrocardiosignal classification method provided by the embodiment of the invention effectively reduces the time and labor cost spent in performing machine learning classification by manually labeling the heart beat.
Owner:GUANGZHOU SHIYUAN ELECTRONICS CO LTD

Three-branch convolutional network fabric defect detection method based on weak supervised learning

The invention provides a three-branch convolutional network fabric defect detection method based on weak supervised learning, and the method comprises the steps: firstly, building a multi-example learning detection network based on a mutual exclusion principle in a weak supervised network, so as to carry out the training through an image-level label; then, establishing a three-branch network framework, and adopting a long connection structure so as to extract and fuse the multi-level convolution feature map; utilizing the SE module and the cavity convolution to learn the correlation between channels and expand the convolution receptive field; and finally, calculating the positioning information of the target by using a class activation mapping method to obtain the attention mapping of thedefect image. According to the method, the problems of rich textural features and defect label missing contained in the fabric picture are comprehensively considered, and by adopting a weak supervision network mechanism and a mutual exclusion principle, the representation capability of the fabric picture is improved while the dependence on the label is reduced, so that the detection result has higher detection precision and adaptivity.
Owner:ZHONGYUAN ENGINEERING COLLEGE

Few-sample pedestrian re-identification method based on deep multi-example learning

The invention relates to a few-sample pedestrian re-identification method based on deep multi-example learning. The method comprises three stages: a network pre-training stage, a data set expansion stage and a network fine tuning stage. The method includes: after the pedestrian re-identification feature extraction sub-network is pre-trained, performing data expansion by utilizing a pedestrian keypoint feature region exchange algorithm; finely adjusting the pedestrian re-identification feature extraction sub-network and the feature aggregation sub-network by using the expanded data set; iteratively repeating data set expansion and network fine tuning until the feature extraction sub-network and the feature aggregation sub-network converge. Once the training is completed, the pedestrian re-identification model on the original domain is migrated and extended to the target domain by using few samples. On the premise of giving a small number of learning samples in the target domain, the pedestrian re-identification model can be effectively migrated and expanded to the target domain monitoring network, and the few-sample pedestrian re-identification method has the advantages of high accuracy, good robustness, good expansibility and mobility.
Owner:FUDAN UNIV
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products