MPAI Publishes Working Draft of Use Cases and Functional Requirements of Multimodal Conversation (MPAI-MMC) Version 2

Geneva, Switzerland, April 23, 2022 --( The international, non-profit, unaffiliated Moving Picture, Audio and Data Coding by Artificial Intelligence (MPAI) standards developing organisation has concluded its 19th General Assembly. Among the outcomes is the publication of the working draft of the Use Cases and Functional Requirements of the planned Version 2 of the Multimodal Conversation (MPAI-MMC) standard. (

The MPAI process envisages that a standard be developed based on a Call for Technologies referring to two documents Functional Requirements and Framework Licence. While the MPAI-MMC V2 documents are still being finalised, MPAI offers an initial working draft ( of the Functional Requirements to alert the industry of its intention to initiate the development of the standard. This will happen when the Call for Technologies is published (planned to be the 13th of July 2022). Responses are expected to be submitted on the 10th of October 2022 and the standard to be published in the first few months of 2023.

Version 2 will substantially extend the capabilities of Version 1 of the MPAI-MMC standard ( by supporting three new use cases:
1. Conversation About a Scene: a human holds a conversation with a machine about objects in a scene of which the human is part. While conversing, the human points their fingers to indicare their interest in a particular object.
2. Human-Connected Autonomous Vehicle Interaction: a group of humans has a conversation on a domain-specific suject (travel by car) with a Connected Autonomous Vehicle. The machine understands the utterances, the emotion in the specch and in the faces, and the expression in their gestures. The machine manifests itself as the torso of an avatar whose face and head convey emotions congruent with the the speech it utters.
3. Avatar Videoconference. In this instance of Mixed-reality Collaborative Space (MCS), avatars represent humans participating in a videoconference. Avatars reproduce the movements of the torsoes of human participants with a high degree of accuracy.

MPAI develops data coding standards for applications that have AI as the core enabling technology. Any legal entity supporting the MPAI mission may join MPAI ( , if able to contribute to the development of standards for the efficient use of data.

So far, MPAI has developed 5 standards (see the list below), is currently engaged in extending two approved standards and is developing other 9.

Name of standard: AI Framework (approved)
- Acronym: MPAI-AIF
- link:
- Brief description: Specifies an infrastructure enabling the execution of implementations and access to the MPAI Store.

Name of standard: Context-based Audio Enhancement (developed)
-Acronym: MPAI-CAE
- Brief description: Improves the user experience of audio-related applications in a variety of contexts.

Name of standard: Compression and Understanding of Industrial Data (developed)
- Acronym: MPAI-CUI
- link:
- Brief description: Predicts the company performance from governance, financial, and risk data.

Name of standard: Governance of the MPAI Ecosystem (developed)
-Acronym: MPAI-GME
- Brief description: Establishes the rules governing the submission of and access to interoperable implementations.

Name of standard: Multimodal Conversation (approved)
-Acronym: MPAI-MMC
- link:
- Brief description: Enables human-machine conversation emulating human-human conversation.

Name of standard: Server-based Predictive Multiplayer Gaming (in development)
-Acronym: MPAI-SPG
- Brief description: Trains a network to com­pensate data losses and detects false data in online multiplayer gaming.

Name of standard: AI-Enhanced Video Coding (in development)
-Acronym: MPAI-EVC
- Brief description: Improves existing video coding with AI tools for short-to-medium term applications.

Name of standard: End-to-End Video Coding (in development)
-Acronym: MPAI-EEV
- Brief description: Explores the promising area of AI-based “end-to-end” video coding for longer-term applications.

Name of standard: Connected Autonomous Vehicles (in development)
-Acronym: MPAI-CAV
- Brief description: Specifies components for Environment Sensing, Autonomous Motion, and Motion Actuation.

Name of standard: Avatar Representation and Animation (in development)
-Acronym: MPAI-ARA
- Brief description: Specifies descriptors of avatars impersonating real humans.

Name of standard: Neural Network Watermarking (in development)
-Acronym: MPAI-NNW
- Brief description: Measures the impact of adding ownership and licensing information in models and inferences.

Name of standard: Integrative Genomic/Sensor Analysis (in development)
-Acronym: MPAI-GSA
- Brief description: Compresses high-throughput experiments data combining genomic/proteomic and other.

Name of standard: Mixed-reality Collaborative Spaces (in development)
-Acronym: MPAI-MCS
- Brief description: Supports collaboration of humans represented by avatars in virtual-reality spaces called Ambients

Name of standard: Visual Object and Scene Description (in development)
-Acronym: MPAI-OSD
- Brief description: Describes objects and their attributes in a scene and the semantic description of the objects.

Visit the MPAI web site (, contact the MPAI secretariat ( for specific information, subscribe to the MPAI Newsletter and follow MPAI on social media:
- LinkedIn (
- Twitter (
- Facebook ( ,
- Instagram (
- YouTube (

Most important: join MPAI (, share the fun, build the future.
Leonardo Chiariglione