[go: up one dir, main page]

WO2018177134A1 - Procédé permettant de traiter un contenu généré par l'utilisateur, support de stockage et terminal - Google Patents

Procédé permettant de traiter un contenu généré par l'utilisateur, support de stockage et terminal Download PDF

Info

Publication number
WO2018177134A1
WO2018177134A1 PCT/CN2018/079228 CN2018079228W WO2018177134A1 WO 2018177134 A1 WO2018177134 A1 WO 2018177134A1 CN 2018079228 W CN2018079228 W CN 2018079228W WO 2018177134 A1 WO2018177134 A1 WO 2018177134A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
image frame
terminal
feature
generated content
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2018/079228
Other languages
English (en)
Chinese (zh)
Inventor
杨田从雨
陈宇
张�浩
华有为
薛丰
肖鸿志
冯绪
吴昊
张振伟
欧义挺
董晓龙
戚广全
谢俊驰
谢斯豪
梁雪
段韧
张新磊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from CN201710199078.4A external-priority patent/CN107168619B/zh
Priority claimed from CN201710282661.1A external-priority patent/CN108334806B/zh
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Publication of WO2018177134A1 publication Critical patent/WO2018177134A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q50/00Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism

Definitions

  • the terminal can convert the selected image frame into a grayscale image, detect the edge in the grayscale image, thereby determining the grayscale change rate at the edge, and determining the sharpness according to the grayscale change rate.
  • access rights can be set when user-generated content is created. For example, if the content creator sets the access right that is only visible to the friend when creating the user-generated content, when the uploaded user account has a friend relationship with the creator's user account, the uploaded user account has the user-generated content. access permission. If the content creator sets access rights visible to everyone when creating user-generated content, any legitimate user account has access to the user-generated content.
  • the terminal uploads the selected image frame to the server, and the server queries the template image that matches the selected image frame.
  • the server queries the matching template image
  • the server feeds back the first notification to the terminal;
  • the server does not query the matching template image
  • the server registers the uploaded image frame as a template image, and feeds back the second notification to the terminal.
  • the terminal displays the content creation entry.
  • the corners of the mouth will rise. If the expression data extracted by the terminal including the facial feature data in the image frame is raised in the corner of the mouth, it can indicate that the emotional feature reflected by the face in the image frame is Happy. When people feel surprised, the mouth opens a large amount. If the terminal extracts the feature data extracted from the face feature data in the image frame, the mouth opening amplitude is large, and the face in the image frame can be represented. The emotional characteristics reflected are astonished.
  • the JPEG format refers to an image format compressed according to the international image compression standard.
  • the direction conforming to the emotional feature recognition condition may specifically be a direction when the angle between the central axis of the face image and the vertical direction in the image frame is not more than 45 degrees.
  • the voice emotional feature recognition result is “happy”.
  • the text obtained by the terminal to recognize the voice data is “I am very happy today”, including the emotional feature keyword “happy”, and the emotional feature mapped to “happy” is “happy”, then the voice emotional feature recognition result is “happy”.
  • the text obtained by the terminal recognizing the voice data is "I am very happy”, including the emotional feature keyword “happy”, and the emotional feature mapped to "happy” is "happy”, then the speech emotional feature recognition result is also "happy”.
  • the acoustic features include timbre and prosodic features.
  • the timbre refers to the sound of the sounding body. Different sounding bodies have different sounds due to different materials and structures.
  • the tone is characterized physics by spectral parameters.
  • the prosodic feature refers to the basic pitch and rhythm of the sound emitted by the sound body, and the rhythm feature is characterized by the fundamental frequency parameter, the duration distribution and the signal intensity.
  • the emotional feature type refers to the type of emotional features reflected by the face. Such as “happy”, “sad” or “anger”.
  • the confidence of the recognition result indicates that the facial emotion feature recognition result is the credibility of the real emotional feature of the face, and the higher the confidence of the recognition result, the higher the possibility that the face emotion feature recognition result is the real emotional feature of the face.
  • the emotional feature image library established in advance by the terminal may include a plurality of emotional feature image sets, and each of the emotional feature image sets reflects an emotional feature type.
  • the terminal may map an emotional feature image one by one corresponding to the intensity of the emotion.
  • the terminal searches for the emotional feature image set that is reflected in the emotional feature image database and the emotional feature type included in the speech emotional feature recognition result, and selects from the found emotional feature image set.
  • the speech emotion feature recognition result includes an emotional feature image corresponding to the emotional intensity.
  • step S1308 determining whether the facial emotion feature recognition result matches the voice emotion feature recognition result; if yes, the process goes to step S1309; if not, the process goes to step S1310.
  • S1316 Render user generated content in the played image frame according to the placement position.
  • the recognition result obtaining module 1703 is further configured to adjust the size of the image frame to a preset size; rotate the direction of the adjusted image frame to a direction that conforms to the emotional feature recognition condition; and send the rotated image frame to the server. Receiving a face emotion feature recognition result returned by the server for the transmitted image frame.
  • RAM is available in a variety of formats, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronization chain.
  • SRAM static RAM
  • DRAM dynamic RAM
  • SDRAM synchronous DRAM
  • DDRSDRAM double data rate SDRAM
  • ESDRAM enhanced SDRAM
  • Synchlink DRAM SLDRAM
  • Memory Bus Radbus
  • RDRAM Direct RAM
  • DRAM Direct Memory Bus Dynamic RAM
  • RDRAM Memory Bus Dynamic RAM

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

L'invention concerne un procédé permettant de traiter un contenu généré par l'utilisateur, ledit procédé consistant : à collecter des trames d'image à partir du monde réel ; à lire les trames d'image collectées trame par trame en fonction de la séquence temporelle de collecte ; à sélectionner une trame d'image parmi les trames d'image collectées ; à acquérir un contenu généré par l'utilisateur associé à une image de modèle correspondant à la trame d'image sélectionnée ; à acquérir une position de présentation du contenu généré par l'utilisateur dans l'image de modèle correspondante ; et, en fonction de la position de présentation, à rendre le contenu généré par l'utilisateur dans la trame d'image lue.
PCT/CN2018/079228 2017-03-29 2018-03-16 Procédé permettant de traiter un contenu généré par l'utilisateur, support de stockage et terminal Ceased WO2018177134A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN201710199078.4A CN107168619B (zh) 2017-03-29 2017-03-29 用户生成内容处理方法和装置
CN201710199078.4 2017-03-29
CN201710282661.1 2017-04-26
CN201710282661.1A CN108334806B (zh) 2017-04-26 2017-04-26 图像处理方法、装置和电子设备

Publications (1)

Publication Number Publication Date
WO2018177134A1 true WO2018177134A1 (fr) 2018-10-04

Family

ID=63674198

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/079228 Ceased WO2018177134A1 (fr) 2017-03-29 2018-03-16 Procédé permettant de traiter un contenu généré par l'utilisateur, support de stockage et terminal

Country Status (1)

Country Link
WO (1) WO2018177134A1 (fr)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109522799A (zh) * 2018-10-16 2019-03-26 深圳壹账通智能科技有限公司 信息提示方法、装置、计算机设备和存储介质
CN109670285A (zh) * 2018-11-13 2019-04-23 平安科技(深圳)有限公司 面部识别登陆方法、装置、计算机设备及存储介质
CN109840491A (zh) * 2019-01-25 2019-06-04 平安科技(深圳)有限公司 视频流播放方法、系统、计算机装置及可读存储介质
US11379683B2 (en) 2019-02-28 2022-07-05 Stats Llc System and method for generating trackable video frames from broadcast video
US12165645B1 (en) * 2020-05-28 2024-12-10 Amazon Technologies, Inc. Stylized text transcription for messaging

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110321082A1 (en) * 2010-06-29 2011-12-29 At&T Intellectual Property I, L.P. User-Defined Modification of Video Content
CN103426003A (zh) * 2012-05-22 2013-12-04 腾讯科技(深圳)有限公司 增强现实交互的实现方法和系统
CN104219559A (zh) * 2013-05-31 2014-12-17 奥多比公司 在视频内容中投放不明显叠加
CN107168619A (zh) * 2017-03-29 2017-09-15 腾讯科技(深圳)有限公司 用户生成内容处理方法和装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110321082A1 (en) * 2010-06-29 2011-12-29 At&T Intellectual Property I, L.P. User-Defined Modification of Video Content
CN103426003A (zh) * 2012-05-22 2013-12-04 腾讯科技(深圳)有限公司 增强现实交互的实现方法和系统
CN104219559A (zh) * 2013-05-31 2014-12-17 奥多比公司 在视频内容中投放不明显叠加
CN107168619A (zh) * 2017-03-29 2017-09-15 腾讯科技(深圳)有限公司 用户生成内容处理方法和装置

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109522799A (zh) * 2018-10-16 2019-03-26 深圳壹账通智能科技有限公司 信息提示方法、装置、计算机设备和存储介质
CN109670285A (zh) * 2018-11-13 2019-04-23 平安科技(深圳)有限公司 面部识别登陆方法、装置、计算机设备及存储介质
CN109840491A (zh) * 2019-01-25 2019-06-04 平安科技(深圳)有限公司 视频流播放方法、系统、计算机装置及可读存储介质
US11379683B2 (en) 2019-02-28 2022-07-05 Stats Llc System and method for generating trackable video frames from broadcast video
US11586840B2 (en) 2019-02-28 2023-02-21 Stats Llc System and method for player reidentification in broadcast video
US11830202B2 (en) 2019-02-28 2023-11-28 Stats Llc System and method for generating player tracking data from broadcast video
US11861848B2 (en) 2019-02-28 2024-01-02 Stats Llc System and method for generating trackable video frames from broadcast video
US11861850B2 (en) 2019-02-28 2024-01-02 Stats Llc System and method for player reidentification in broadcast video
US11935247B2 (en) 2019-02-28 2024-03-19 Stats Llc System and method for calibrating moving cameras capturing broadcast video
US12288342B2 (en) 2019-02-28 2025-04-29 Stats Llc System and method for player reidentification in broadcast video
US12165645B1 (en) * 2020-05-28 2024-12-10 Amazon Technologies, Inc. Stylized text transcription for messaging

Similar Documents

Publication Publication Date Title
US12271586B2 (en) Combining first user interface content into second user interface
US20240361881A1 (en) Updating avatar clothing for a user of a messaging system
US11094131B2 (en) Augmented reality apparatus and method
CN105808782B (zh) 一种图片标签的添加方法及装置
US12292299B2 (en) Augmented reality-based translations associated with travel
US12142278B2 (en) Augmented reality-based translation of speech in association with travel
WO2017157272A1 (fr) Procédé de traitement d'informations et terminal
WO2018177134A1 (fr) Procédé permettant de traiter un contenu généré par l'utilisateur, support de stockage et terminal
WO2021109678A1 (fr) Procédé et appareil de génération de vidéo, dispositif électronique et support de stockage
US12020383B2 (en) Facial synthesis in augmented reality content for third party applications
CN107168619B (zh) 用户生成内容处理方法和装置
US12148244B2 (en) Interactive augmented reality content including facial synthesis
US20250182524A1 (en) Facial synthesis in augmented reality content for online communities
US12148064B2 (en) Facial synthesis in augmented reality content for advertisements
CN113709545A (zh) 视频的处理方法、装置、计算机设备和存储介质
CN108334806B (zh) 图像处理方法、装置和电子设备
US20230326094A1 (en) Integrating overlaid content into displayed data via graphics processing circuitry and processing circuitry using a computing memory and an operating system memory

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18775573

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18775573

Country of ref document: EP

Kind code of ref document: A1