[go: up one dir, main page]

CN114840107B - Sketch data reuse and scene sketch auxiliary construction method and system - Google Patents

Sketch data reuse and scene sketch auxiliary construction method and system Download PDF

Info

Publication number
CN114840107B
CN114840107B CN202210237615.0A CN202210237615A CN114840107B CN 114840107 B CN114840107 B CN 114840107B CN 202210237615 A CN202210237615 A CN 202210237615A CN 114840107 B CN114840107 B CN 114840107B
Authority
CN
China
Prior art keywords
sketch
scene
video
panel
scene sketch
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210237615.0A
Other languages
Chinese (zh)
Other versions
CN114840107A (en
Inventor
马翠霞
刘舫
宋建成
邓小明
王宏安
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Software of CAS
Original Assignee
Institute of Software of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Software of CAS filed Critical Institute of Software of CAS
Publication of CN114840107A publication Critical patent/CN114840107A/en
Application granted granted Critical
Publication of CN114840107B publication Critical patent/CN114840107B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a method and a system for the reuse of sketch data and the auxiliary construction of a scene sketch, which belong to the field of computer vision, and extract and reuse the existing sketch materials based on sketch restoration, sketch-based image retrieval and other sketch intelligent processing technologies by utilizing the advantages of efficient information representation of sketch interaction; and assisting a user in constructing a scene sketch by using the extracted sketch materials and sketch data stored in a database, so that the method is used for subsequent applications such as sketch retrieval, video positioning and the like.

Description

一种草图数据重用与场景草图辅助构建方法及系统Method and system for reuse of sketch data and auxiliary construction of scene sketch

技术领域technical field

本发明属于计算机视觉领域,具体涉及一种草图数据重用与场景草图辅助构建方法及系统。The invention belongs to the field of computer vision, and in particular relates to a sketch data reuse and scene sketch auxiliary construction method and system.

背景技术Background technique

基于草图交互的可视媒体应用一直是人机交互、计算机视觉和多媒体领域的研究热点,如何优化草图数据的基础处理以及提高基于草图的可视媒体应用的效率是研究的重点问题。草图交互被广泛地应用于生活与工作中的各个方面,包括绘画、笔记速记、文档标注、互联网行业的网页用户界面(UI)和概念设计、电影动画行业的动画与影视制作等。近年来,草图交互相关研究和应用在工业界和学术界都引起了广泛关注,其中很重要的原因之一是触屏类硬件设备的蓬勃发展(例如微软公司Surface系列触控笔记本电脑、苹果公司的Apple pencil触控笔等)。人工智能时代下,一方面,用户对于草图数据的获取更加便捷;另一方面,基于深度学习技术的草图数据算法性能不断提高。基于草图交互的应用和任务也得到空前的普及。Visual media applications based on sketch interaction have always been a research hotspot in the fields of human-computer interaction, computer vision and multimedia. How to optimize the basic processing of sketch data and improve the efficiency of visual media applications based on sketches are the key issues of research. Sketch interaction is widely used in all aspects of life and work, including drawing, shorthand notes, document annotation, web user interface (UI) and concept design in the Internet industry, animation and film and television production in the film and animation industry, etc. In recent years, research and application related to sketch interaction have attracted widespread attention in both industry and academia. One of the important reasons is the vigorous development of touch-screen hardware devices (such as Microsoft Surface series touch laptops, Apple Inc. Apple pencil stylus, etc.). In the era of artificial intelligence, on the one hand, it is more convenient for users to obtain sketch data; on the other hand, the performance of sketch data algorithms based on deep learning technology continues to improve. Applications and tasks based on sketch interaction have also gained unprecedented popularity.

在草图交互任务与草图数据处理方面,利用基于深度学习的草图数据处理技术取得了巨大的进步,例如草图识别(sketch recognition)、基于草图的图像检索(sketch-based image retrieval,SBIR)、基于草图的图像生成(sketch-based image generation,SBIG)、草图解析(sketch parsing)、基于草图的视频摘要(sketch-based videosummarization)等。一些新的草图交互任务也被提出,例如草图生成(sketchgeneration)、基于草图交互的模型生成、草图抽象化、基于草图的图像编辑、草图分割等。SketchGAN(参考文献:Liu,Fang,Xiaoming Deng,Yu-Kun Lai,Yong-Jin Liu,Cuixia Ma,and Hongan Wang."Sketchgan:Joint sketch completion and recognition withgenerative adversarial network."In Proceedings of the IEEE/CVF Conference onComputer Vision and Pattern Recognition,pp.5830-5839.2019.)提出一种基于生成式对抗网络的草图补全与识别方法,对残缺草图数据进行有效的生成式补全与修复,同时利用补全后的草图数据进行草图识别,提高目前主流草图识别算法对残缺草图的识别准确率。In terms of sketch interaction tasks and sketch data processing, great progress has been made using sketch data processing technologies based on deep learning, such as sketch recognition, sketch-based image retrieval (sketch-based image retrieval, SBIR), sketch-based image generation (sketch-based image generation, SBIG), sketch parsing (sketch parsing), sketch-based video summarization (sketch-based videosummarization), etc. Some new sketch interaction tasks are also proposed, such as sketch generation, model generation based on sketch interaction, sketch abstraction, sketch-based image editing, sketch segmentation, etc. SketchGAN (References: Liu, Fang, Xiaoming Deng, Yu-Kun Lai, Yong-Jin Liu, Cuixia Ma, and Hongan Wang."Sketchgan: Joint sketch completion and recognition with generative adversarial network."In Proceedings of the IEEE/CVF Conference onComputer Vision and Pattern Recognition, pp.5830-5839.2019.) proposed a sketch completion and recognition method based on a generative adversarial network, which effectively generatively completes and repairs incomplete sketch data, and utilizes the completed sketch The data is used for sketch recognition to improve the recognition accuracy of current mainstream sketch recognition algorithms for incomplete sketches.

对于草图交互的研究,应对于实际应用的需求,向细粒度的方法发展。具体来说,相对于基于整体的任务(例如草图识别),近年来一些细粒度的草图交互任务得以被提出。目前大多数的基于草图的图像检索相关技术都建立在实例级、类别级检索的前提下,即:输入草图与待检索的图像对象均为单个物体;并且检索结果图像的物体与输入草图物体在类别上保持一致,即为正确的检索。Sketch Me that Shoe(参考文献:Yu,Qian,Feng Liu,Yi-Zhe Song,Tao Xiang,Timothy M.Hospedales,and Chen-Change Loy."Sketch me thatshoe."In Proceedings of the IEEE Conference on Computer Vision and PatternRecognition,pp.799-807.2016.)等研究实例级(instance-level)的基于草图的图像检索,不仅要求检索出的图像物体与输入的查询草图在类别上保持一致,而且要求它们外形、动作、方向等细节信息也相似。For the research on sketch interaction, we should develop towards fine-grained methods in response to the needs of practical applications. Specifically, relative to whole-based tasks such as sketch recognition, some fine-grained sketch interaction tasks have been proposed in recent years. At present, most of the sketch-based image retrieval related technologies are based on the premise of instance-level and category-level retrieval, that is, the input sketch and the image object to be retrieved are both single objects; Consistency in category is correct retrieval. Sketch Me that Shoe (References: Yu, Qian, Feng Liu, Yi-Zhe Song, Tao Xiang, Timothy M. Hospedales, and Chen-Change Loy."Sketch me that shoe."In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.799-807.2016.) and other researches on instance-level sketch-based image retrieval not only require the retrieved image objects to be consistent with the input query sketches in category, but also require their shape, action, Details like directions are similar.

草图特有的高度抽象性、直观性和简洁性,使得草图在人机交互、计算机视觉、多媒体、计算机图像学等各个领域被广泛应用。从1960年代至今,随着数据处理技术的不断提高,草图相关研究和应用不断得到优化。在智能人机交互领域中,尽管已经有许多学者将草图应用于可视媒体(图像、视频、动画等)的研究中,但在应用过程中仍有许多草图数据的处理和在可视媒体领域实际应用的关键技术尚未得到有效解决。在基于草图交互的应用中,一个至关重要的问题是草图数据的获取。虽然有几个重要的草图数据库陆续被提出,但草图数据仍然相对稀少且难以大批量获取。对于非专业画手,绘制出他们脑海中所想的物体的准确轮廓及恰当细节不是一件容易的事情,且往往耗费大量的时间,这也在一定程度上对草图交互在实际生活中的大量应用带来了阻碍。The unique high abstraction, intuition, and simplicity of sketches make sketches widely used in various fields such as human-computer interaction, computer vision, multimedia, and computer graphics. From the 1960s to the present, with the continuous improvement of data processing technology, the research and application of sketches have been continuously optimized. In the field of intelligent human-computer interaction, although many scholars have applied sketches to the research of visual media (images, videos, animations, etc.), there are still many sketch data processing and in the field of visual media in the application process. The key technology for practical application has not been effectively solved. In sketch-based interactive applications, a crucial issue is the acquisition of sketch data. Although several important sketch databases have been proposed one after another, sketch data is still relatively rare and difficult to obtain in large quantities. For non-professional painters, it is not easy to draw the exact outline and appropriate details of the objects in their minds, and it often takes a lot of time, which also affects the large number of sketch interactions in real life to a certain extent Apps get in the way.

发明内容Contents of the invention

为了克服现有基于草图交互的方法和应用对草图数据的重用率低、用户输入草图有难度等问题,本发明提出一种草图数据重用与场景草图辅助构建方法及系统,利用草图交互的高效信息表征的优势,基于草图修复以及基于草图的图像检索等草图智能处理技术,对已有的草图素材进行提取和重用;利用提取的草图素材以及数据库中存储的草图数据辅助用户进行场景草图构建,从而用于草图检索、视频定位等后续应用。In order to overcome the existing sketch interaction-based methods and applications that have a low reuse rate of sketch data and difficulty in inputting sketches by users, the present invention proposes a sketch data reuse and scene sketch auxiliary construction method and system, which utilizes the efficient information of sketch interaction Based on the advantages of representation, based on intelligent sketch processing technologies such as sketch repair and sketch-based image retrieval, the existing sketch materials are extracted and reused; the extracted sketch materials and the sketch data stored in the database are used to assist users in the construction of scene sketches, thereby It is used for subsequent applications such as sketch retrieval and video positioning.

本发明解决上述技术问题所采用的技术方案是:The technical solution adopted by the present invention to solve the problems of the technologies described above is:

一种草图数据重用与场景草图辅助构建方法,其步骤包括:A sketch data reuse and scene sketch auxiliary construction method, the steps of which include:

对用户在已有场景草图中选择的草图素材进行提取并保存到数据库中,并对其中残缺的草图素材进行自动修复;Extract the sketch material selected by the user in the existing scene sketch and save it in the database, and automatically repair the incomplete sketch material;

利用数据库中已有草图素材或者临时绘制草图,来构建场景草图或场景草图序列;Use existing sketch materials in the database or temporarily draw sketches to construct scene sketches or scene sketch sequences;

基于构建的场景草图进行图像检索,找到与场景草图最相似的图像;Perform image retrieval based on the constructed scene sketch to find the image most similar to the scene sketch;

基于构建的场景草图序列进行视频定位,找到与场景草图序列最相似的视频片段。Video localization is performed based on the constructed scene sketch sequence, and the video segment most similar to the scene sketch sequence is found.

进一步地,利用数据库中已有的草图素材通过拖拽、缩放和旋转中的一种或几种操作来构建场景草图。Further, the sketch material in the database is used to construct the sketch of the scene by one or several operations of dragging, zooming and rotating.

进一步地,图像检索方法为基于SceneSketcher方法并去掉该方法中类别敏感的交并比损失部分所得到的方法。Further, the image retrieval method is based on the SceneSketcher method and removes the class-sensitive intersection-over-union ratio loss part of the method.

进一步地,构建基于图像搜索方法的端对端的图像搜索模型,该图像搜索模型通过训练提高图像搜索精度。Furthermore, an end-to-end image search model based on the image search method is constructed, and the image search model improves image search accuracy through training.

进一步地,在基于构建的场景草图进行图像检索时,通过调整场景草图的姿态、大小、方向、位置中的一种或几种来找到与场景草图最相似的图像。Further, when performing image retrieval based on the constructed scene sketch, the most similar image to the scene sketch is found by adjusting one or more of the pose, size, direction, and position of the scene sketch.

进一步地,进行视频定位的步骤包括:基于卷积神经网络提取场景草图视觉特征;基于光流法提取视频动作特征;基于C3D网络提取视频特征;根据所述场景草图视觉特征、视频动作特征和视频特征利用回归算法进行视频定位。Further, the step of performing video positioning includes: extracting scene sketch visual features based on convolutional neural network; extracting video action features based on optical flow method; extracting video features based on C3D network; according to the scene sketch visual features, video action features and video Features utilize regression algorithms for video localization.

进一步地,在基于构建的场景草图序列进行视频定位时,通过调整场景草图的姿态、大小、方向、位置中的一种或几种来找到与场景草图序列最相似的视频片段。Further, when performing video positioning based on the constructed scene sketch sequence, the most similar video segment to the scene sketch sequence is found by adjusting one or more of the pose, size, direction, and position of the scene sketch.

一种草图数据重用与场景草图辅助构建系统,包括顶层的交互界面和底层的数据处理算法模块,该交互界面包括草图数据提取页面、场景草图构建与图像检索页面和基于场景草图的视频定位页面,以及用于切换页面的页面切换按钮;A sketch data reuse and scene sketch auxiliary construction system, including a top-level interactive interface and a bottom-level data processing algorithm module. The interactive interface includes a sketch data extraction page, a scene sketch construction and image retrieval page, and a scene sketch-based video positioning page. and a page switching button for switching pages;

草图数据提取页面包括:场景草图浏览面板和草图数据提取面板;The sketch data extraction page includes: scene sketch browsing panel and sketch data extraction panel;

场景草图构建与图像检索页面包括:场景草图构建画布、场景草图编辑工具面板、草图素材浏览和选择面板和图像检索结果显示面板;The scene sketch construction and image retrieval page includes: scene sketch construction canvas, scene sketch editing tool panel, sketch material browsing and selection panel, and image retrieval result display panel;

基于场景草图的视频定位页面包括:场景草图构建画布、场景草图编辑工具面板、草图素材浏览和选择面板、草图序列显示面板和视频定位结果显示面板;The video positioning page based on the scene sketch includes: the scene sketch construction canvas, the scene sketch editing tool panel, the sketch material browsing and selection panel, the sketch sequence display panel and the video positioning result display panel;

其中,场景草图浏览面板用于显示当前场景草图,框选草图对象元素和切合场景草图;草图数据提取面板用于提取框选的草图对象,并对残缺的草图素材进行自动修复;场景草图构建画布用于支持对构建场景草图的拖动、缩放和旋转操作;场景草图编辑工具面板用于提供对草图元素进行绘制和编辑的工具;草图素材浏览和选择面板用于展示系统推荐的草图素材列表;图像检索结果显示面板用于展示图像检索结果;草图序列显示面板是用于显示草图序列;视频定位结果显示面板用于显示定位的视频片段;Among them, the scene sketch browse panel is used to display the current scene sketch, frame-select sketch object elements and fit the scene sketch; the sketch data extraction panel is used to extract the frame-selected sketch objects, and automatically repair the incomplete sketch materials; the scene sketch construction canvas It is used to support dragging, zooming and rotating operations on the scene sketch; the scene sketch editing tool panel is used to provide tools for drawing and editing sketch elements; the sketch material browsing and selection panel is used to display the list of sketch materials recommended by the system; The image retrieval result display panel is used to display the image retrieval results; the sketch sequence display panel is used to display the sketch sequence; the video positioning result display panel is used to display the positioned video clips;

数据处理算法模块用于实现上述交互界面中的所有交互和数据处理。The data processing algorithm module is used to realize all interactions and data processing in the above-mentioned interactive interface.

进一步地,场景草图编辑工具面板提供画笔、橡皮擦、撤销、删除、清除和检索工具。Further, the scene sketch editing tool panel provides brush, eraser, undo, delete, clear and retrieve tools.

进一步地,数据处理算法模块含有基于图像搜索方法的端对端的图像搜索模型,该图像搜索模型通过训练提高图像搜索精度;该图像检索方法为基于SceneSketcher方法并去掉该方法中类别敏感的交并比损失部分所得到的方法。Further, the data processing algorithm module contains an end-to-end image search model based on the image search method, and the image search model improves image search accuracy through training; the image retrieval method is based on the SceneSketcher method and removes the category-sensitive cross-merge ratio in the method Lost part of the resulting method.

与现有技术相比,本发明的有益效果是:Compared with prior art, the beneficial effect of the present invention is:

1.本发明在首先对用户选择的草图物体进行修复,然后辅助用户绘制场景草图,并根据场景草图进行图像检索。1. The present invention first repairs the sketch object selected by the user, then assists the user to draw a scene sketch, and performs image retrieval according to the scene sketch.

2.本发明支持个性化草图交互:1)用户可以通过重新选择草图素材、调整草图物体的姿态和位置获得更准确、更理想的检索结果。并且,系统也提供Pen和Eraser等草图绘制工具来支持用户手动提取场景草图中的物体。2)用户可以利用个人收藏夹或者公共数据库中的草图素材进行场景草图构建,也可以使用创建的场景草图搜索图像。3)系统允许用户重新选择草图素材、调整场景草图中的各物体优化当前场景草图来检索到更符合预期的图像或定位到更精准的视频片段。2. The present invention supports personalized sketch interaction: 1) The user can obtain more accurate and ideal retrieval results by reselecting sketch materials and adjusting the posture and position of sketch objects. Moreover, the system also provides sketching tools such as Pen and Eraser to support users to manually extract objects in the scene sketch. 2) Users can use the sketch materials in personal favorites or public databases to construct scene sketches, and can also use the created scene sketches to search for images. 3) The system allows the user to reselect the sketch material, adjust each object in the scene sketch to optimize the current scene sketch to retrieve a more expected image or locate a more accurate video clip.

附图说明Description of drawings

图1为本发明实施例1中的一种草图数据重用与场景草图辅助构建系统交互界面图。FIG. 1 is an interactive interface diagram of a sketch data reuse and scene sketch auxiliary construction system in Embodiment 1 of the present invention.

图2为本发明实施例2中的一种草图数据重用与场景草图辅助构建流程图。Fig. 2 is a flow chart of sketch data reuse and scene sketch assistant construction in Embodiment 2 of the present invention.

图3为本发明实施例3中的基于草图序列的视频定位的交互界面图。FIG. 3 is an interactive interface diagram of video positioning based on a sketch sequence in Embodiment 3 of the present invention.

图4为基线系统交互界面图。Figure 4 is a diagram of the baseline system interaction interface.

图5为本发明系统和基线系统问卷调查中各指标的箱线图。Fig. 5 is a box plot of each indicator in the questionnaire survey of the system of the present invention and the baseline system.

图6为本发明的实验中被试用户构建的草图示例。Fig. 6 is an example of a sketch constructed by the tested user in the experiment of the present invention.

具体实施方式Detailed ways

为使本发明的上述特征和优点能更明显易懂,下文特举实施例,并配合所附图作详细说明如下。In order to make the above-mentioned features and advantages of the present invention more comprehensible, the following specific embodiments are described in detail in conjunction with the accompanying drawings.

实施例1Example 1

本实施例提出一种草图数据重用与场景草图辅助构建系统及方法,是针对草图数据重用与场景草图辅助构建问题所提出的常见草图素材的提取、草图数据的重用以及草图交互在图像检索和视频定位方面的技术方案。This embodiment proposes a system and method for sketch data reuse and scene sketch auxiliary construction, which is aimed at the problems of sketch data reuse and scene sketch auxiliary construction. Technical solutions for positioning.

本实施例提出的一种草图数据重用与场景草图辅助构建系统包括顶层的交互界面和底层的数据处理模块,交互界面如图1所示(图中界面显示的牛、人、草等画面属于本实施例中的一个示例),其包括:草图数据提取页面、场景草图构建与图像检索页面和基于场景草图的视频定位页面,用户可以通过页面切换按钮在三个页面之间切换操作。图1中上方图为草图数据提取页面,其包括②场景草图浏览面板和③草图数据提取面板;下方图为场景草图构建与图像检索页面,其包括④场景草图构建画布、⑤场景草图编辑工具面板、⑥草图素材浏览和选择面板和⑦图像检索结果显示面板。底层的数据处理模块用于处理顶层的界面的相关数据交互操作和数据计算。A kind of sketch data reuse and scene sketch auxiliary construction system proposed in this embodiment includes a top-level interactive interface and a bottom-level data processing module. An example in the embodiment), which includes: a draft data extraction page, a scene sketch construction and image retrieval page, and a scene sketch-based video positioning page, and the user can switch between the three pages through the page switching button. The upper picture in Figure 1 is the sketch data extraction page, which includes ②scene sketch browsing panel and ③sketch data extraction panel; the lower picture is the scene sketch construction and image retrieval page, which includes ④scene sketch construction canvas, ⑤scene sketch editing tool panel , ⑥ sketch material browsing and selection panel and ⑦ image retrieval result display panel. The underlying data processing module is used to process the related data interaction operation and data calculation of the top interface.

其中,①页面切换按钮:在页面直接进行切换操作;②场景草图浏览面板:显示当前场景草图,此时用户可以通过绘制矩形框的方式来框选感兴趣的草图对象元素,可以通过该面板上的左右箭头切换不同的场景草图(如图1中的中间图);③草图数据提取面板:当场景草图中的物体存在相互遮挡或重叠的现象时,系统支持自动修复草图物体,并保存到用户的个人收藏夹中以便后续的素材重用;④场景草图构建画布:支持用户构建场景草图的相关操作(包括拖动、缩放、旋转等);⑤场景草图编辑工具面板:从左到右包含Pen(画笔)、Eraser(橡皮擦)、Undo(撤销)、Delete(删除,即删除单个物体)、Clear(清除,即清除所有物体)和Retrieval(检索)等按钮,支持对草图元素的绘制和编辑;用户可以通过点击Retrieval按钮来根据当前构建的场景草图检索图像;⑥草图素材浏览和选择面板:展示系统推荐的草图素材列表;⑦图像检索结果显示面板:展示图像检索结果,其中在上方展示最匹配的图像大图,在下面列出与当前构建的场景草图相似的其它图像。本系统基于网站架构,可以适应手机、笔记本电脑、数字白板等多种设备,同时支持使用鼠标与在触屏设备上使用手指或笔进行输入标。Among them, ①page switching button: switch directly on the page; ②scene sketch browsing panel: displays the current scene sketch, at this time, the user can draw a rectangular frame to frame the interested sketch object elements, and can use this panel The left and right arrows of the left and right arrows switch between different scene sketches (as shown in the middle picture in Figure 1); ③Sketch data extraction panel: When the objects in the scene sketch are mutually occluded or overlapped, the system supports automatic repair of the sketch objects and saves them to the user ④Scene Sketch Construction Canvas: Supports related operations for users to build scene sketches (including dragging, zooming, rotating, etc.); ⑤Scene Sketch Editing Tool Panel: Contains Pen( Brush), Eraser (eraser), Undo (undo), Delete (delete, that is, delete a single object), Clear (clear, that is, clear all objects) and Retrieval (retrieval) and other buttons, support the drawing and editing of sketch elements; Users can click the Retrieval button to retrieve images based on the currently constructed scene sketch; ⑥Sketch material browsing and selection panel: display the list of sketch material recommended by the system; ⑦Image retrieval result display panel: display image retrieval results, among which the best match is displayed on the top Larger version of the image, with other images similar to the currently built scene sketch listed below. This system is based on the website architecture and can adapt to various devices such as mobile phones, laptops, and digital whiteboards. It also supports the use of mouse and fingers or pens on touch screen devices for input.

本实施例提出的一种草图数据重用与场景草图辅助构建方法基于上述系统,其包括以下步骤:A sketch data reuse and scene sketch auxiliary construction method proposed in this embodiment is based on the above-mentioned system, which includes the following steps:

S1:从现有场景草图中提取对象:本系统支持用户交互式的从场景草图中提取并保存其感兴趣的草图素材。用户从已有的场景草图中选择感兴趣的物体时,本系统可以对残缺的素材进行修复。本系统对草图交互方式进行以下优化:首先,当用户浏览场景草图(如草图画报或浏览故事板)时,场景草图浏览面板显示当前场景草图,用户可以框选出感兴趣的草图对象元素,由于场景草图中的元素普遍存在相互遮挡或重叠的现象,本系统可以对草图进行自动修复并存储到用户的个人收藏夹(Favorits);当用户对自动修复的草图效果不满意时,可以通过本系统提供的画笔(pen)、橡皮(eraser)等草图绘制工具来手动修改场景草图中的物体。其次,面对不具有专业绘画技能的普通用户,相比于要求用户徒手绘制脑海中完整的草图,本系统提供一些绘画提示或建议,或者利用用户个人收藏夹及系统公共数据库中已有的草图素材帮助用户快速构建新的草图。再次,除了通常的笔式交互界面提供的画笔和橡皮等草图输入工具以外,还可以提供更多的交互式编辑工具来辅助用户绘制草图。最后,本系统可以支持更灵活的、多个对象的草图绘制,同时便于用户绘制更加细节的信息(如物体的方向、背景等)。S1: Extract objects from existing scene sketches: This system supports users to interactively extract and save interested sketch materials from scene sketches. When the user selects the object of interest from the existing scene sketch, the system can repair the incomplete material. This system optimizes the sketch interaction mode as follows: First, when the user browses the scene sketch (such as a sketch pictorial or browses a storyboard), the scene sketch browsing panel displays the current scene sketch, and the user can select the interested sketch object elements. The elements in the scene sketch generally have the phenomenon of mutual occlusion or overlapping. This system can automatically repair the sketch and store it in the user's personal favorites (Favorits); when the user is not satisfied with the effect of the automatically repaired sketch, he can use this system Provides sketching tools such as pen and eraser to manually modify objects in the scene sketch. Secondly, in the face of ordinary users who do not have professional drawing skills, instead of requiring users to draw a complete sketch in their minds by hand, the system provides some drawing tips or suggestions, or utilizes the existing sketches in the user's personal favorites and the system's public database Materials help users quickly build new sketches. Again, in addition to the sketch input tools such as paintbrush and eraser provided by the usual pen-style interactive interface, more interactive editing tools can be provided to assist users in drawing sketches. Finally, the system can support more flexible sketching of multiple objects, and at the same time facilitate the user to draw more detailed information (such as the direction of the object, the background, etc.).

S2:场景草图构建:本系统支持用户重用数据库中已有的草图素材,通过对草图素材的拖拽、缩放和旋转等操作,快速、高效地进行场景草图构建。S2: Scene sketch construction: This system supports users to reuse the existing sketch materials in the database, and quickly and efficiently construct scene sketches by dragging, zooming and rotating the sketch materials.

S3:基于场景草图的图像检索:本系统提供基于用户构建的场景草图进行图像检索的功能,可以通过点击Retrieval按钮来根据当前构建的场景草图检索图像。用户可以在场景草图构建画布上通过拖动、缩放、旋转等操作,调整草图素材物体的姿态、大小、方向和位置,不断地调整当前构建的场景草图(如图1的中间图的各个场景草图),以及通过场景草图编辑工具面板的工具对草图元素进行绘制和编辑,以搜索到更符合预期的图像。本系统所用的图像检索方法是基于SceneSketcher(参考文献:Liu,Fang,et al."Scenesketcher:Fine-grained image retrieval with scene sketches."European Conference onComputer Vision.Springer,Cham,2020.)去掉了原方法中的类别敏感的交并比损失(category-wise IoU loss)部分,在保证检索模型精度不大幅度下降的情况下,构建端对端的基于草图的图像检索模型,加快模型的训练速度。S3: Image retrieval based on the scene sketch: This system provides the function of image retrieval based on the scene sketch constructed by the user, and the image can be retrieved according to the currently constructed scene sketch by clicking the Retrieval button. The user can adjust the posture, size, direction and position of the sketch material object by dragging, zooming, rotating and other operations on the scene sketch construction canvas, and constantly adjust the currently constructed scene sketch (the scene sketches in the middle picture of Figure 1 ), and use the tools in the scene sketch editing tool panel to draw and edit the sketch elements to search for images that are more in line with expectations. The image retrieval method used in this system is based on SceneSketcher (reference: Liu, Fang, et al. "Scenesketcher: Fine-grained image retrieval with scene sketches." European Conference on Computer Vision. Springer, Cham, 2020.) Removed the original method The category-wise IoU loss (category-wise IoU loss) part in the method builds an end-to-end sketch-based image retrieval model to speed up the training speed of the model while ensuring that the accuracy of the retrieval model does not drop significantly.

S4:基于场景草图序列的视频定位:本发明提供基于用户构建的场景草图序列进行视频定位的功能;用户可以不断地调整当前构建的场景草图,具体地可以在场景草图构建画布上通过拖动、缩放、旋转等操作,调整草图素材物体的姿态、大小、方向和位置,以定位到更符合预期的视频片段。所用视频定位方法包括基于卷积神经网络提取场景草图视觉特征,基于光流法提取动作特征,以及基于C3D网络(参考文献:Tran,Du,et al."Learningspatiotemporal features with 3d convolutional networks."Proceedings of theIEEE international conference on computer vision.2015.)提取视频特征,最后利用回归算法进行视频定位。S4: Video positioning based on the scene sketch sequence: the present invention provides the function of video positioning based on the scene sketch sequence constructed by the user; the user can continuously adjust the currently constructed scene sketch, specifically, on the scene sketch construction canvas by dragging, Scale, rotate and other operations, adjust the pose, size, direction and position of the sketch material object, so as to locate the video clip more in line with expectations. The video localization methods used include extracting visual features of scene sketches based on convolutional neural network, extracting action features based on optical flow method, and based on C3D network (reference: Tran, Du, et al."Learning spatial temporal features with 3d convolutional networks."Proceedings of theIEEE international conference on computer vision.2015.) Extract video features, and finally use regression algorithm for video positioning.

实施例2Example 2

本实施例给出提供一个草图数据重用与场景草图辅助构建系统及方法的具体应用实例,如图2所示,该方法包括以下步骤:This embodiment provides a specific application example of providing a sketch data reuse and scene sketch auxiliary construction system and method, as shown in Figure 2, the method includes the following steps:

S1:当用户浏览场景草图时,用户框选出感兴趣的草图物体,系统对其进行草图修复并存储到用户个人收藏夹(见图2的(1)),偶尔算法效果不佳时,系统提供草图绘制工具来支持用户手动提取场景草图中的物体。S1: When the user browses the scene sketch, the user selects the interested sketch object, and the system repairs the sketch and stores it in the user's personal favorites (see (1) in Figure 2). Occasionally, when the algorithm is not effective, the system Provides a sketch drawing tool to support users to manually extract objects in the scene sketch.

S2:用户利用个人收藏夹或者公共数据库中的草图素材进行场景草图构建(见图2的(2)),用户也可以使用创建的场景草图搜索图像(见图2的(3))。S2: The user uses the sketch material in personal favorites or public databases to construct the scene sketch (see (2) in Figure 2), and the user can also use the created scene sketch to search for images (see (3) in Figure 2).

S3:系统允许用户重新选择草图素材、调整场景草图中的各物体优化当前场景草图(见图2的(4))来检索到更符合预期的图像(见图2的(5))。S3: The system allows the user to reselect the sketch material, adjust each object in the scene sketch to optimize the current scene sketch (see (4) in Figure 2) to retrieve an image that is more in line with expectations (see (5) in Figure 2).

实施例3Example 3

本实施例提出一种草图数据重用与场景草图辅助构建系统及方法,是基于草图进行视频定位,要用到基于场景草图的视频定位页面,如图3所示,该页面包括:①场景草图构建画布;②场景草图编辑工具面板;③草图素材浏览和选择面板;④草图序列显示面板;⑤视频定位结果显示面板。其中①②③与实施例1中的相同,④草图序列显示面板是用于显示草图序列,⑤视频定位结果显示面板用于显示定位的视频片段。视频定位包括三个步骤:This embodiment proposes a sketch data reuse and scene sketch auxiliary construction system and method, which is based on the sketch for video positioning, and the video positioning page based on the scene sketch is used, as shown in Figure 3, the page includes: ① Scene sketch construction Canvas; ②Scene sketch editing tool panel; ③Sketch material browsing and selection panel; ④Sketch sequence display panel; ⑤Video positioning result display panel. Among them, ①②③ are the same as those in Embodiment 1, ④the sketch sequence display panel is used to display the sketch sequence, and ⑤the video positioning result display panel is used to display the positioned video clips. Video targeting consists of three steps:

S1:场景草图构建。用户根据视频定位的目标片段构建场景草图,本系统允许用户通过拖拽、移动、缩放和旋转等操作重用现有的草图素材加速构建过程。S1: Scene sketch construction. Users build scene sketches based on the target segments of video positioning. This system allows users to reuse existing sketch materials to speed up the construction process through operations such as dragging, moving, zooming, and rotating.

S2:草图序列生成。通过重复步骤S1合成草图序列,代表视频中需要定位的目标片段的关键帧,草图序列中元素的变化表示视频中相应对象的运动。S2: Sketch sequence generation. Synthesize a sketch sequence by repeating step S1, which represents the keyframe of the target segment to be located in the video, and the change of elements in the sketch sequence represents the motion of the corresponding object in the video.

S3:视频定位。系统根据用户创建的草图序列查找视频中最相似的片段。S3: Video positioning. The system finds the most similar segments in the video based on the sketch sequence created by the user.

以下对本发明提出的一种草图数据重用与场景草图辅助构建系统进行实验测试:The following is an experimental test of a sketch data reuse and scene sketch auxiliary construction system proposed by the present invention:

现设计一种基线系统作为对照系统来对本发明系统进行评估,该系统采用场景草图构建任务、问卷调查和访谈的方法进行研究。相比于本发明系统,基线系统没有场景草图素材提取和重用功能。如前文所述,现有的主流草图交互应用系统一般要求用户绘制完整的草图作为输入;类似地,在基线系统中,用户需要徒手绘制场景中的草图物体,构建场景草图来检索目标图像。图4为基线系统界面,包括:①场景草图构建画布;②场景草图编辑工具面板,从左到右同样包含Pen、Eraser、Undo、Delete、Clear和Retrieval;③图像检索结果显示面板。A baseline system is now designed as a control system to evaluate the system of the present invention. The system uses the scene sketch construction task, questionnaire survey and interview methods for research. Compared with the system of the present invention, the baseline system does not have scene sketch material extraction and reuse functions. As mentioned above, the existing mainstream sketch interactive application systems generally require users to draw complete sketches as input; similarly, in the baseline system, users need to draw sketch objects in the scene by hand to construct a scene sketch to retrieve the target image. Figure 4 shows the baseline system interface, including: ①Scene sketch construction canvas; ②Scene sketch editing tool panel, which also includes Pen, Eraser, Undo, Delete, Clear, and Retrieval from left to right; ③Image retrieval result display panel.

(1)参与者(1) Participants

招募16名参与者,其中8名男性和8名女性。被试用户的平均年龄为26.313岁,方差为3.260(男性:平均年龄=27.250,方差=3.845;女性:平均年龄=25.375,方差=2.446)。在测试之前,本实验向被试用户详细地解释了实验目的和过程,并让参与者先熟悉系统。每次实验历时大约半小时,所有参与者均不具备专业的草图绘画技能。16 participants were recruited, 8 males and 8 females. The average age of the tested users is 26.313 years old, and the variance is 3.260 (male: average age=27.250, variance=3.845; female: average age=25.375, variance=2.446). Before the test, this experiment explained the purpose and process of the experiment to the tested users in detail, and let the participants familiarize themselves with the system first. Each experiment lasted about half an hour, and none of the participants had professional sketching skills.

(2)实验任务和步骤(2) Experimental tasks and steps

被试用户需要分别使用本发明系统和基准系统完成实验任务。在每次实验中,向被试用户随机展示一张彩色图像作为目标(展示时长为2秒),被试用户需要在展示时间内记住图像的关键特征,然后根据脑海中的印象,利用本发明系统或基线系统构建场景草图,系统根据被试用户构建的场景草图从数据库中进行图像检索。基线系统仅提供给被试用户Pen、Eraser工具来绘制草图;本发明系统允许被试用户重用草图素材库中的草图物体来辅助构建过程。The tested users need to use the system of the present invention and the reference system to complete the experimental tasks respectively. In each experiment, a color image was randomly shown to the test users as a target (the display time was 2 seconds), and the test users were required to remember the key features of the image during the display time, and then according to the impression in their minds, use this The invention system or the baseline system constructs a scene sketch, and the system performs image retrieval from the database according to the scene sketch constructed by the tested user. The baseline system only provides the tested users with Pen and Eraser tools to draw sketches; the system of the present invention allows the tested users to reuse the sketch objects in the sketch material library to assist the construction process.

每位被试用户进行6次实验(其中4次在给定目标图像的情况下构建场景草图,2次自由绘制);实验共收集了由16位被试用户绘制的96张场景草图。Each test user conducted 6 experiments (4 of which were to construct scene sketches given the target image, and 2 free to draw); the experiment collected 96 scene sketches drawn by 16 test users.

(3)问卷调查(3) Questionnaire survey

在被试用户使用本发明系统和基准系统完成测试时,对被试用户进行问卷调查,定量地评估被试的用户体验。问卷调查的量标题涵盖了系统的可用性(用户体验的可用性度量标准,UMUX,参考文献:FINSTAD K.The usability metric for user experience[J].Interacting with Computers,2010,22(5):323-327.)、TAM技术接受模型(感知的实用性(PU)和感知的易用性(PEOU),参考文献:DAVIS F D,BAGOZZI R P,WARSHAW P R.Useracceptance of computer technology:a comparison of two theoretical models[J].Management science,1989,35(8):982-1003.)、以及AI界面问题。When the tested users use the system of the present invention and the benchmark system to complete the test, a questionnaire survey is conducted on the tested users to quantitatively evaluate the tested user experience. The volume title of the questionnaire covers the usability of the system (usability metric for user experience, UMUX, reference: FINSTAD K.The usability metric for user experience[J].Interacting with Computers,2010,22(5):323-327 .), TAM Technology Acceptance Model (Perceived Utility (PU) and Perceived Ease of Use (PEOU), References: DAVIS F D, BAGOZZI R P, WARSHAW P R. User acceptance of computer technology: a comparison of two theoretical models[ J].Management science,1989,35(8):982-1003.), and AI interface issues.

UMUX包含四个指标:UMUX consists of four metrics:

有效性(Effectiveness):(当前系统的)功能符合用户的要求;Effectiveness: the function (of the current system) meets the user's requirements;

满意度(Satisfaction):使用(当前系统)是令人沮丧的体验;Satisfaction: using (the current system) is a frustrating experience;

总体(Overall):(当前系统)易于使用;Overall: (the current system) is easy to use;

效率(Efficiency):用户不得不花太多时间在(当前系统)上进行更正。Efficiency: Users have to spend too much time making corrections (on the current system).

TAM技术接受模型主要评估用户对技术的接受程度。本实验分别在用户场景草图构建阶段和图像检索阶段测试TAM指标,其中包含两个方面:感知有用性(PU)和感知易用性(PEOU)。The TAM technology acceptance model mainly evaluates the user's acceptance of technology. This experiment tests the TAM metrics in the user scene sketch construction phase and the image retrieval phase respectively, which contains two aspects: Perceived Usefulness (PU) and Perceived Ease of Use (PEOU).

PEOU中有四个指标:There are four indicators in PEOU:

易于学习(Easy to learn):在(场景草图构建/图像检索)阶段,系统的操作易于理解;Easy to learn: In the stage of (scene sketch construction/image retrieval), the operation of the system is easy to understand;

易于使用(Easy to use):在(场景草图构建/图像检索)阶段,系统易于使用;Easy to use: the system is easy to use during the (scene sketch construction/image retrieval) stage;

舒适度(Comfortable):在(场景草图构建/图像检索)阶段,不需要大量的认知负担即可使用系统;Comfortable: During the (scene sketch construction/image retrieval) phase, the system can be used without significant cognitive load;

沟通性(Communicative):在(场景草图构建/图像检索)阶段,系统知道用户想要什么。Communicative: In the (scene sketch construction/image retrieval) stage, the system knows what the user wants.

PU中有三个指标:There are three indicators in PU:

有用性(Useful):在(场景草图构建/图像检索)阶段,系统可以提供帮助;Useful: In the (scene sketch construction/image retrieval) stage, the system can provide assistance;

省时(Time-saving):在(场景草图构建/图像检索)阶段,系统节省了时间;Time-saving: In the (scene sketch construction/image retrieval) stage, the system saves time;

完成度(Fulfilling):在(场景草图构建/图像检索)阶段,系统完成任务非常准确。Fulfilling: In the (scene sketch construction/image retrieval) phase, the system completes tasks very accurately.

另外,本实验还引入了AI用户界面常用指标中常用的三个指标(参考文献:OH C,SONG J,CHOI J,et al.I lead,you help but only with enough details:Understanding user experience of co-creation with artificial intelligence[C]//Proceedings of the 2018CHI Conference on Human Factors in ComputingSystems.2018:1-13.):In addition, this experiment also introduces three commonly used indicators in AI user interface (references: OH C, SONG J, CHOI J, et al. I lead, you help but only with enough details: Understanding user experience of co -creation with artificial intelligence[C]//Proceedings of the 2018CHI Conference on Human Factors in Computing Systems.2018:1-13.):

可控性(Controllability);Controllability;

创造力(Creativity);Creativity;

自由度(Degree of Freedom)。Degree of Freedom.

最后,本实验引入净推荐值(Net Promoter Score,NPS)指标,即调查用户是否愿意将系统推荐给其他人。Finally, this experiment introduces the Net Promoter Score (NPS) indicator, which is to investigate whether users are willing to recommend the system to others.

在每位被试用户完成所有实验任务时,需要填写0-7分的李克特量表(Likertscale)进行打分;对于净推荐值,被试用户在0-10之间打分。When each test user completes all the experimental tasks, they need to fill in the Likert scale of 0-7 points for scoring; for the net promoter score, the test users score between 0-10.

(4)访谈(4) Interview

为了更详细地了解本发明系统中的用户体验,本实验对被试用户进行了半结构化访谈。在访谈中,本实验询问被试用户对两种不同系统的想法,以及使用草图作为人机交互方式的体验。每次面试大约持续半小时。In order to understand the user experience in the system of the present invention in more detail, this experiment conducted semi-structured interviews with the tested users. In the interviews, the experiment asked the test users about their thoughts on the two different systems and their experiences using sketches as a means of human-computer interaction. Each interview lasts about half an hour.

(5)实验结果(5) Experimental results

本试验分别基于问卷调查中的数据和用户访谈对本发明系统与基线系统进行定量和定性的分析。This experiment carried out quantitative and qualitative analysis on the system of the present invention and the baseline system respectively based on the data in the questionnaire survey and user interviews.

(5.1)定量分析(5.1) Quantitative analysis

本实验应用配对t检验来检验本发明系统和基线系统的指标得分之间是否存在显着差异。本实验将显着性水平设置为0.05,即如果p<0.05则认为二者存在显著差异。图5为本发明系统和基线系统问卷调查中各指标的箱线图,可以看出,除Satisfaction、Easy toLearn,Controllability和Degree of Freedom四个指标外,本发明系统和基线系统的其他指标均存在显着差异。In this experiment, a paired t-test was used to test whether there was a significant difference between the index scores of the inventive system and the baseline system. In this experiment, the significance level is set at 0.05, that is, if p<0.05, it is considered that there is a significant difference between the two. Fig. 5 is the box diagram of each index in the questionnaire survey of the system of the present invention and the baseline system, as can be seen, except Satisfaction, Easy to Learn, Controllability and Degree of Freedom four indexes, other indexes of the system of the present invention and the baseline system all exist Significant differences.

用户体验的可用性度量:在effectiveness(t=3.60,p<.01)、efficiency(t=5.16,p<.01)、overall rating(t=-3.06,p<.01)三个指标上本发明系统和基线系统的得分均存在显著差异,且本发明系统的指标得分均优于基线系统。本发明系统和基线系统在satisfaction指标上的得分分别为(M=4.25,SD=1.48)和(M=3.50,SD=1.79),satisfaction指标无显著差异(t=1.44,p>.05)。Usability measurement of user experience: on the three indicators of effectiveness (t=3.60, p<.01), efficiency (t=5.16, p<.01), and overall rating (t=-3.06, p<.01), the present invention There are significant differences in the scores of the system and the baseline system, and the index scores of the system of the present invention are better than the baseline system. The scores of the system of the present invention and the baseline system on the satisfaction index are (M=4.25, SD=1.48) and (M=3.50, SD=1.79), respectively, and there is no significant difference in the satisfaction index (t=1.44, p>.05).

感知易用性(PEOU):本发明系统和基线系统在easy to learn(t=-0.22,p>.05)指标上不存在显著差异,即被试用户认为本发明系统(M=6.00,SD=1.10)与基线系统(M=6.06,SD=.68)都易于学习。本发明系统在ease of use指标得分(t=3.88,p<.01)上发现了显著性:被试用户在易用性指标上对本发明系统的打分(M=6.19,SD=.98)明显高于对基线系统的打分(M=5.00,SD=1.32)。在comfortability(t=3.67,p<.01)指标上也存在显著性,本发明系统的得分情况(M=5.75,SD=.77)优于基线系统(M=4.38,SD=1.20)。另外,本发明系统在communicative(t=4.58,p<.01)指标上的得分(M=5.81,SD=.98)也明显比基线系统的得分(M=4.25,SD=1.06)高。Perceived Ease of Use (PEOU): There is no significant difference between the system of the present invention and the baseline system on the easy to learn (t=-0.22, p>.05) index, that is, the tested users think that the system of the present invention (M=6.00, SD =1.10) and the baseline system (M=6.06, SD=.68) were both easy to learn. The system of the present invention found significance on the ease of use index score (t=3.88, p<.01): the tested users scored (M=6.19, SD=.98) on the ease of use index for the system of the present invention significantly Higher than the baseline system score (M=5.00, SD=1.32). Significance also exists in comfortability (t=3.67, p<.01), and the score of the system of the present invention (M=5.75, SD=.77) is better than that of the baseline system (M=4.38, SD=1.20). In addition, the score (M=5.81, SD=.98) of the inventive system on the communicative (t=4.58, p<.01) index was also significantly higher than that of the baseline system (M=4.25, SD=1.06).

感知有用性(PU):对于usefulness指标(t=5.23,p<.01),本发明系统的得分(t=5.23,p<.01)比基线系统的得分(M=4.00,SD=1.21)高。同样,本发明系统(M=6.13,SD=.89)比基线系统(M=3.25,SD=1.13)更节省时间(time-saving(t=7.67,p<.01))。并且,两个系统在fulfilling(t=5.57,p<.01)指标上的得分也存在显著差异,分别为本发明系统(M=5.81,SD=.66)、基线系统(M=3.75,SD=1.29)。Perceived usefulness (PU): For the usefulness indicator (t=5.23, p<.01), the score of the inventive system (t=5.23, p<.01) is higher than that of the baseline system (M=4.00, SD=1.21) high. Also, the inventive system (M=6.13, SD=.89) was more time-saving (t=7.67, p<.01) than the baseline system (M=3.25, SD=1.13). Moreover, there are also significant differences in the scores of the two systems on the fulfilling (t=5.57, p<.01) index, which are the system of the present invention (M=5.81, SD=.66), the baseline system (M=3.75, SD = 1.29).

AI用户界面常用指标:本发明系统与基线系统在creativity(t=-4.14,p<.01)指标上的得分存在显著差异,本发明系统的得分(M=5.13,SD=.89)比基线系统得分低(M=6.13,SD=.89),这也在一定程度上表明,素材重用在为场景草图构建提供便利的同时,限制了用户的创造力。另外,两个系统在controllability(t=1.86,p>.05)与degree offreedom(t=0.64,p>.05)两个指标上的得分相似,不存在显著性。AI user interface commonly used indicators: there is a significant difference between the score of the inventive system and the baseline system on the indicator of creativity (t=-4.14, p<.01), and the score of the inventive system (M=5.13, SD=.89) is higher than that of the baseline The system scored low (M=6.13, SD=.89), which also partly indicates that material reuse, while facilitating the construction of scene sketches, limits user creativity. In addition, the scores of the two systems on controllability (t=1.86, p>.05) and degree offreedom (t=0.64, p>.05) are similar, and there is no significance.

净推荐值(NPS):本发明系统的净推荐值得分为93.75\%,意味着15名被试用户打分在9分以上;基线系统的净推荐值得分为-37.5\%,意味着只有1名被试用户打分在9分以上,有8位用户打分为7分或者8分。Net Promoter Score (NPS): The NPS of the system of the present invention is divided into 93.75\%, which means that 15 tested users scored above 9 points; the Net Promoter Score of the baseline system is -37.5\%, which means that only 1 One of the tested users scored above 9 points, and 8 users scored 7 or 8 points.

(5.2)定性分析(5.2) Qualitative Analysis

在定性分析中,本实验旨在了解用户在使用本发明系统与基线系统时的主观想法、评论和建议。图6中展示了用户实验中被试用户构建的部分场景草图示例以及对应的图像检索目标,可以看出,用户基于本发明系统(SketchMaker)构建的草图质量更高。本试验对被试用户们在访谈中比较普遍和重要的观点进行了归纳:In qualitative analysis, this experiment aims to understand users' subjective thoughts, comments and suggestions when using the inventive system and the baseline system. Figure 6 shows an example of some scene sketches constructed by the tested users in the user experiment and the corresponding image retrieval targets. It can be seen that the sketches constructed by users based on the system (SketchMaker) of the present invention are of higher quality. This experiment summarizes the more common and important views of the test users in the interviews:

本发明系统用户界面易于使用;The system user interface of the present invention is easy to use;

本发明系统用户界面更加友好且使用高效灵感和创造力;The system user interface of the present invention is more friendly and uses efficient inspiration and creativity;

在访谈中本实验发现,用户对系统是否可以增强创造力的观点不一致,但用户更倾向于在检索复杂图像的时候使用本发明系统。In the interviews, this experiment found that users have different views on whether the system can enhance creativity, but users are more inclined to use the system of the present invention when retrieving complex images.

虽然本发明已以实施例公开如上,然其并非用以限定本发明,本领域的普通技术人员对本发明的技术方案进行的适当修改或者等同替换,均应涵盖于本发明的保护范围内,本发明的保护范围以权利要求所限定者为准。Although the present invention has been disclosed as above with the embodiments, it is not intended to limit the present invention. Appropriate modifications or equivalent replacements to the technical solutions of the present invention by those of ordinary skill in the art shall fall within the protection scope of the present invention. The scope of protection of the invention is defined by the claims.

Claims (9)

1. The method for reuse of the sketch data and auxiliary construction of the scene sketch is characterized by comprising the following steps:
extracting sketch materials selected by a user from the existing scene sketch, storing the sketch materials into a database, and automatically repairing incomplete sketch materials in the sketch materials;
constructing a scene sketch or a scene sketch sequence by utilizing the existing sketch materials in the database or temporarily drawing the sketch;
performing image retrieval based on the constructed scene sketch, and finding out an image most similar to the scene sketch;
video positioning is carried out based on the constructed scene sketch sequence, and video fragments most similar to the scene sketch sequence are found; the step of performing video localization includes: extracting visual features of a scene sketch based on a convolutional neural network; extracting video action features based on an optical flow method; extracting video features based on a C3D network; and carrying out video positioning by using a regression algorithm according to the visual features of the scene sketch, the video action features and the video features.
2. The method of claim 1, wherein the scene sketch is constructed by one or more of dragging, zooming, and rotating using sketch material already in the database.
3. The method of claim 1, wherein the image retrieval method is a method based on a sceneskeyer method and wherein the class-sensitive cross-ratio loss portion of the method is removed.
4. The method of claim 1, wherein an end-to-end image search model based on an image search method is constructed, the image search model improving image search accuracy and speed through training.
5. The method of claim 1, wherein the image most similar to the scene sketch is found by adjusting one or more of a pose, a size, a direction, and a position of the scene sketch when performing image retrieval based on the constructed scene sketch.
6. The method of claim 1, wherein in video localization based on the constructed sequence of scene sketches, video clips most similar to the sequence of scene sketches are found by adjusting one or more of the pose, size, direction, and position of the scene sketches.
7. A sketch data reuse and scene sketch auxiliary construction system for implementing the method according to any of claims 1-6, characterized by comprising a top layer interactive interface and a bottom layer data processing algorithm module, the interactive interface comprising a sketch data extraction page, a scene sketch construction and image retrieval page and a scene sketch based video positioning page, and a page switching button for switching pages;
the sketch data extraction page includes: a scene sketch browsing panel and a sketch data extraction panel;
the scene sketch construction and image retrieval page comprises: a scene sketch construction canvas, a scene sketch editing tool panel, a sketch material browsing and selecting panel and an image retrieval result display panel;
the video positioning page based on the scene sketch comprises: the scene sketch construction canvas, the scene sketch editing tool panel, the sketch material browsing and selecting panel, the sketch sequence display panel and the video positioning result display panel;
the scene sketch browsing panel is used for displaying a current scene sketch, selecting sketch object elements and cutting the scene sketch; the sketch data extraction panel is used for extracting sketch objects selected by a frame and automatically repairing incomplete sketch materials; the scene sketch construction canvas is used for supporting dragging, zooming and rotating operations for constructing the scene sketch; the scene sketch editing tool panel is used for providing tools for drawing and editing sketch elements; the sketch material browsing and selecting panel is used for displaying a sketch material list recommended by the system; the image retrieval result display panel is used for displaying image retrieval results; the sketch sequence display panel is used for displaying sketch sequences; the video positioning result display panel is used for displaying the positioned video clips;
the data processing algorithm module is used for realizing all interactions and data processing in the interaction interface.
8. The system of claim 7, wherein the scene sketch editing tool panel provides a brush, eraser, undo, delete, purge, and retrieve tool.
9. The system of claim 7, wherein the data processing algorithm module comprises an end-to-end image search model based on an image search method that improves image search accuracy and speed through training; the image retrieval method is a method based on a SceneSkeyer method and obtained by removing a category-sensitive cross ratio loss part in the method.
CN202210237615.0A 2021-04-28 2022-03-11 Sketch data reuse and scene sketch auxiliary construction method and system Active CN114840107B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110468472X 2021-04-28
CN202110468472 2021-04-28

Publications (2)

Publication Number Publication Date
CN114840107A CN114840107A (en) 2022-08-02
CN114840107B true CN114840107B (en) 2023-08-01

Family

ID=82562797

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210237615.0A Active CN114840107B (en) 2021-04-28 2022-03-11 Sketch data reuse and scene sketch auxiliary construction method and system

Country Status (1)

Country Link
CN (1) CN114840107B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2991037A1 (en) * 2013-04-23 2016-03-02 Tsinghua University Method of generating three-dimensional scene model
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101807198A (en) * 2010-01-08 2010-08-18 中国科学院软件研究所 Video abstraction generating method based on sketch
CN101763439B (en) * 2010-03-05 2012-09-19 中国科学院软件研究所 A sketch-based hypervideo construction method
US20130334300A1 (en) * 2011-01-03 2013-12-19 Curt Evans Text-synchronized media utilization and manipulation based on an embedded barcode
CN102663429B (en) * 2012-04-11 2014-03-26 上海交通大学 Method for motion pattern classification and action recognition of moving target
CN104794429B (en) * 2015-03-23 2018-10-23 中国科学院软件研究所 A kind of association visual analysis method towards monitor video
CN105468666B (en) * 2015-08-11 2019-09-17 中国科学院软件研究所 A kind of video content visual analysis method based on map metaphor
CN106599133B (en) * 2016-12-02 2020-04-17 中国科学院软件研究所 Monitoring video visual analysis method based on sketch interaction
US10304225B2 (en) * 2016-12-30 2019-05-28 Microsoft Technology Licensing, Llc Chart-type agnostic scene graph for defining a chart
CN109977882B (en) * 2019-03-29 2019-12-03 广东石油化工学院 A kind of half coupling dictionary is to the pedestrian of study again recognition methods and system
CN110147797B (en) * 2019-04-12 2021-06-01 中国科学院软件研究所 A method and device for sketch completion and recognition based on generative adversarial network
CN110933520B (en) * 2019-12-10 2020-10-16 中国科学院软件研究所 A surveillance video display method and storage medium based on spiral abstract
CN111291212B (en) * 2020-01-24 2022-10-11 复旦大学 Method and system for zero-shot sketch image retrieval based on graph convolutional neural network
CN112069336B (en) * 2020-08-04 2022-10-14 中国科学院软件研究所 Fine-grained image retrieval method and system based on scene sketch

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2991037A1 (en) * 2013-04-23 2016-03-02 Tsinghua University Method of generating three-dimensional scene model
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study

Also Published As

Publication number Publication date
CN114840107A (en) 2022-08-02

Similar Documents

Publication Publication Date Title
US8874525B2 (en) Hierarchical display and navigation of document revision histories
US8701002B2 (en) Hierarchical display and navigation of document revision histories
Zimmer et al. 3d bat: A semi-automatic, web-based 3d annotation toolbox for full-surround, multi-modal data streams
US8533595B2 (en) Hierarchical display and navigation of document revision histories
JP6098120B2 (en) Assembly sequence generation program, assembly sequence generation apparatus, and manufacturing method
US20210110601A1 (en) Collaborative 3d modeling system
US20120272173A1 (en) Hierarchical display and navigation of document revision histories
US20120272151A1 (en) Hierarchical display and navigation of document revision histories
EP2180700A1 (en) Interface system for editing video data
Woodbury et al. Interactive design galleries: A general approach to interacting with design alternatives
CN102346633A (en) System and method for interactive filter
CN109635102B (en) Theme model lifting method based on user interaction
Chen et al. History assisted view authoring for 3d models
CN105930773A (en) Motion identification method and device
CN106126085A (en) A kind of multi-point touch gesture interaction three-dimensional modeling method of facing mobile apparatus
US11132821B1 (en) Providing graphical user interface tools for efficiently selecting handles in vector artwork on touch-based devices
CN114840107B (en) Sketch data reuse and scene sketch auxiliary construction method and system
JP2013025745A (en) Method for creating teacher data, program for creating teacher data, and teacher data creation device
Appert et al. Reciprocal drag-and-drop
Drossis et al. 3D visualization and multimodal interaction with temporal information using timelines
CN1131774A (en) handwritten character input device
Wang et al. Interactive multi-scale structures for summarizing video content
Weng et al. Dualslide: global-to-local sketching interface for slide content and layout design
Prakljačić et al. Tool for image annotation in context of modern object detection
Ward et al. Merging technology and users: applying image browsing to the fashion industry for design inspiration

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant