Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

Real-time multi-modal language analysis system and method based on mobile edge intelligence

A language analysis and multi-modal technology, applied in the field of real-time multi-modal language analysis system, can solve problems such as difficult to implement, large energy consumption, difficult real-time multi-modal language analysis, etc.

Inactive Publication Date: 2020-04-03
HUNAN NORMAL UNIVERSITY
View PDF7 Cites 8 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] With the rapid development of artificial intelligence technology, speech recognition, image recognition, text recognition and other technologies have made breakthroughs in the field of computer human-computer interaction, but single-modal (single text, voice or image) language processing technology It is difficult to meet the complex scenes and environments in reality, but the analysis technology of multimodal language (combining multiple expressions such as text, voice and image) can allow the computer to hear and see clearly, and can better understand the meaning behind the language communication scene. Reducing misreading and misunderstanding in communication is a hot spot in future language communication analysis
[0003] However, multimodal language analysis technology is a cutting-edge multimodal artificial intelligence technology that requires complex artificial intelligence models as support. Real-time multimodal language analysis technology requires a large number of computing resources and extremely low communication delays. Difficult to achieve in the current computing environment
At the same time, if the mobile environment is further considered, the current mobile terminals often have limited energy consumption, and the calculation of the multimodal artificial intelligence model in the real-time multimodal language analysis system requires a lot of energy consumption. Modal form, difficult to achieve real-time multimodal language analysis
Therefore, the design of a real-time multimodal language analysis system and method based on a mobile environment is a huge technical challenge.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • Real-time multi-modal language analysis system and method based on mobile edge intelligence
  • Real-time multi-modal language analysis system and method based on mobile edge intelligence
  • Real-time multi-modal language analysis system and method based on mobile edge intelligence

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0085] The present invention will be further described below in conjunction with specific embodiment and accompanying drawing:

[0086] Such as figure 1 Shown, the system of multimodal language analysis system of the present invention comprises UAV, UGV and MGS three kinds of MEI servers, there are various single mode or multimodal tasks on every mobile terminal such as mobile phone and notebook, according to our method will each These tasks are offloaded to UAV, UGV and MGS for execution, the computing resources of the three are reduced in turn, and the flexibility of movement is increased in turn, realizing real-time and efficient multi-modal language analysis;

[0087] This solution divides the user's language data into three modes: text, voice and image, and assigns computing tasks to appropriate MEI servers for execution according to the difficulty of computing and analysis and the size of computing resources required.

[0088] The present invention also provides an onli...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a real-time multi-modal language analysis system and method based on mobile edge intelligence, and the system comprises three types of mobile edge intelligent servers: a mobilebase station (MGS), an unmanned vehicle (UGV) and an unmanned aerial vehicle (UAV), and the computing resources of the three types of mobile edge intelligent servers are sequentially reduced, and themoving flexibility is sequentially improved. According to the real-time multi-modal language analysis system, language data of a user is divided into three modalities, namely characters, voice and images, and a calculation task is allocated to a proper MEI server to be executed according to the calculation and analysis difficulty and the size of required calculation resources. According to the method, a real-time multi-modal language analysis and calculation problem in a dynamic environment is constructed, then a task unloading matrix and a resource allocation matrix are generated through a deep learning online optimization method, and meanwhile, movement path planning of an MEI server is automatically implemented according to channel conditions and interference during communication. Thetrained DNN can be suitable for a dynamic scene in which the number of multi-modal computing tasks changes, and has very high practicability.

Description

technical field [0001] The invention belongs to the technical field of mobile edge computing, and in particular relates to a real-time multimodal language analysis system and method based on mobile edge intelligence. Background technique [0002] With the rapid development of artificial intelligence technology, speech recognition, image recognition, text recognition and other technologies have made breakthroughs in the field of computer human-computer interaction, but single-modal (single text, voice or image) language processing technology It is difficult to meet the complex scenes and environments in reality, but the analysis technology of multimodal language (combining multiple expressions such as text, voice and image) can allow the computer to hear and see clearly, and can better understand the meaning behind the language communication scene. Reducing misreading and misunderstanding in communication is a hot spot in the future analysis of language communication. [000...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
IPC IPC(8): H04W24/02G06F9/50G06K9/62G06N3/04G06N3/08
CPCH04W24/02G06F9/5072G06N3/08G06N3/045G06F18/23
Inventor 江沸菠董莉王敏捷代建华王可之刘帅陈大卫
Owner HUNAN NORMAL UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products