Looking for breakthrough ideas for innovation challenges? Try Patsnap Eureka!

A real-time multimodal language analysis system and method based on mobile edge intelligence

A language analysis, multimodal technology, applied in the field of real-time multimodal language analysis system, can solve the problems of low communication delay, limited energy consumption, large amount of computing resources, etc., and achieve the effect of reducing complexity

Inactive Publication Date: 2021-06-04
HUNAN NORMAL UNIVERSITY
View PDF5 Cites 0 Cited by
  • Summary
  • Abstract
  • Description
  • Claims
  • Application Information

AI Technical Summary

Problems solved by technology

[0002] With the rapid development of artificial intelligence technology, speech recognition, image recognition, text recognition and other technologies have made breakthroughs in the field of computer human-computer interaction, but single-modal (single text, voice or image) language processing technology It is difficult to meet the complex scenes and environments in reality, but the analysis technology of multimodal language (combining multiple expressions such as text, voice and image) can allow the computer to hear and see clearly, and can better understand the meaning behind the language communication scene. Reducing misreading and misunderstanding in communication is a hot spot in future language communication analysis
[0003] However, multimodal language analysis technology is a cutting-edge multimodal artificial intelligence technology that requires complex artificial intelligence models as support. Real-time multimodal language analysis technology requires a large number of computing resources and extremely low communication delays. Difficult to achieve in the current computing environment
At the same time, if the mobile environment is further considered, the current mobile terminals often have limited energy consumption, and the calculation of the multimodal artificial intelligence model in the real-time multimodal language analysis system requires a lot of energy consumption. Modal form, difficult to achieve real-time multimodal language analysis
Therefore, the design of a real-time multimodal language analysis system and method based on a mobile environment is a huge technical challenge.

Method used

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
View more

Image

Smart Image Click on the blue labels to locate them in the text.
Viewing Examples
Smart Image
  • A real-time multimodal language analysis system and method based on mobile edge intelligence
  • A real-time multimodal language analysis system and method based on mobile edge intelligence
  • A real-time multimodal language analysis system and method based on mobile edge intelligence

Examples

Experimental program
Comparison scheme
Effect test

Embodiment Construction

[0085] The present invention will be further described below in conjunction with specific embodiment and accompanying drawing:

[0086] Such as figure 1 Shown, the system of multimodal language analysis system of the present invention comprises UAV, UGV and MGS three kinds of MEI servers, there are various single mode or multimodal tasks on every mobile terminal such as mobile phone and notebook, according to our method will each These tasks are offloaded to UAV, UGV and MGS for execution, the computing resources of the three are reduced in turn, and the flexibility of movement is increased in turn, realizing real-time and efficient multi-modal language analysis;

[0087] This solution divides the user's language data into three modes: text, voice and image, and assigns computing tasks to appropriate MEI servers for execution according to the difficulty of computing and analysis and the size of computing resources required.

[0088] The present invention also provides an onli...

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

PUM

No PUM Login to View More

Abstract

The invention discloses a real-time multimodal language analysis system and method based on mobile edge intelligence, including three types of mobile edge intelligent servers: mobile base station (MGS), unmanned vehicle (UGV) and unmanned aerial vehicle (UAV), three The computing resources of the user decrease in turn, and the flexibility of movement increases in turn; the real-time multi-modal language analysis system divides the user's language data into three modes: text, voice and image, according to the difficulty of calculation and analysis and the required It is necessary to calculate the size of the resource, and assign the calculation task to the appropriate MEI server for execution. The method of the present invention constructs a real-time multi-modal language analysis and calculation problem in a dynamic environment, and then generates a task offloading matrix and a resource allocation matrix through a deep learning online optimization method, and at the same time, automatically implements it according to the channel conditions and interference during communication The mobile path planning of the MEI server; the trained DNN can be applied to dynamic scenarios where the number of multi-modal computing tasks changes, and has strong practicability.

Description

technical field [0001] The invention belongs to the technical field of mobile edge computing, and in particular relates to a real-time multimodal language analysis system and method based on mobile edge intelligence. Background technique [0002] With the rapid development of artificial intelligence technology, speech recognition, image recognition, text recognition and other technologies have made breakthroughs in the field of computer human-computer interaction, but single-modal (single text, voice or image) language processing technology It is difficult to meet the complex scenes and environments in reality, but the analysis technology of multimodal language (combining multiple expressions such as text, voice and image) can allow the computer to hear and see clearly, and can better understand the meaning behind the language communication scene. Reducing misreading and misunderstanding in communication is a hot spot in the future analysis of language communication. [000...

Claims

the structure of the environmentally friendly knitted fabric provided by the present invention; figure 2 Flow chart of the yarn wrapping machine for environmentally friendly knitted fabrics and storage devices; image 3 Is the parameter map of the yarn covering machine
Login to View More

Application Information

Patent Timeline
no application Login to View More
Patent Type & Authority Patents(China)
IPC IPC(8): H04W24/02G06F9/50G06K9/62G06N3/04G06N3/08
CPCH04W24/02G06F9/5072G06N3/08G06N3/045G06F18/23
Inventor 江沸菠董莉王敏捷代建华王可之刘帅陈大卫
Owner HUNAN NORMAL UNIVERSITY
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Patsnap Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Patsnap Eureka Blog
Learn More
PatSnap group products