WO1999045483A1 - Method and system for generating semantic visual templates for image and video retrieval - Google Patents

Method and system for generating semantic visual templates for image and video retrieval Download PDF

Info

Publication number
WO1999045483A1
WO1999045483A1 PCT/US1999/004776 US9904776W WO9945483A1 WO 1999045483 A1 WO1999045483 A1 WO 1999045483A1 US 9904776 W US9904776 W US 9904776W WO 9945483 A1 WO9945483 A1 WO 9945483A1
Authority
WO
WIPO (PCT)
Prior art keywords
query
visual
concept
subset
user
Prior art date
Application number
PCT/US1999/004776
Other languages
French (fr)
Other versions
WO1999045483A9 (en
Inventor
Shih-Fu Chang
William Chen
Hari Sundaram
Original Assignee
The Trustees Of Columbia University In The City Of New York
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by The Trustees Of Columbia University In The City Of New York filed Critical The Trustees Of Columbia University In The City Of New York
Priority to KR1020007009804A priority Critical patent/KR20010041607A/en
Priority to CA002322448A priority patent/CA2322448A1/en
Priority to EP99911110A priority patent/EP1066572A1/en
Priority to JP2000534957A priority patent/JP2002506255A/en
Publication of WO1999045483A1 publication Critical patent/WO1999045483A1/en
Publication of WO1999045483A9 publication Critical patent/WO1999045483A9/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5854Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using shape and object relationship
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5862Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/71Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7834Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using audio features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7847Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
    • G06F16/786Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using motion, e.g. object motion or camera motion

Definitions

  • the invention relates to database still image, video and audio retrieval and, more particularly, to techniques which facilitate access to database items.
  • the database can be indexed using a collection of visual templates.
  • the visual templates represent semantic concepts or categories, e.g. skiing, sunset and the like.
  • SVT semantic visual templates
  • Semantic visual templates can be established by an interactive process between a user and a system.
  • the user can provide the system with an initial sketch or example image, as a seed to the system to automatically generate other representations of the same concept.
  • the user then can pick those views for inclusion that are plausible for representing the concept.
  • the database can be searched with it, for the user to provide relevancy feedback on the returned results.
  • SVT's the user can interact with the system at concept level. In forming new concepts, pre-existing SVT's can be used..
  • Fig. 1 is a schematic of an interactive technique for generating a library or collection of semantic visual templates in accordance with a preferred embodiment of the invention.
  • Fig. 2 is a diagram which illustrates a concept having necessary and sufficient conditions.
  • Fig. 3 is a diagram which illustrates query generation.
  • Fig. 4 is a schematic of an interactive system in accordance with a preferred further embodiment of the invention, including audio processing.
  • Fig. 5 shows a set of icons exemplifying the concept "high jump”.
  • Fig. 6 shows a set of icons exemplifying the concept "sunset”.
  • Fig. 7 shows a semantic visual template for the concept "slalom"
  • VideoQ VideoQ
  • each object can be characterized by salient attributes such as color, texture, size, shape and motion, for example.
  • video object database consisting of all the objects extracted from the scene and their attributes.
  • Visual Templates A visual template represents an idea, in the form of a sketch or an animated sketch. As a single visual template may be a poor representative of a class of interest, a library of visual templates can be assembled, containing representative templates for different semantic classes. For example, when searching for video clips of the class Sunset, one could select one or more visual templates corresponding to the class and use similarity-based querying to find video clips of sunsets.
  • An important advantage of using a visual template library lies in linkage of a low-level visual feature representation to high-level semantic concepts. For example, if a user enters a query in a constrained natural language form as described in the above-referenced patent applications, visual templates can be used to transform the natural language query into automated queries specified by visual attributes and constraints. When visual content in the repository or database is not indexed textually, customary textual search methods cannot be applied directly.
  • a semantic visual template is the set of visual templates associated with a particular semantic. This notion of an SVT has certain key properties as follows: Semantic visual templates are general in nature. For a given concept, there should be a set of visual templates that cover that concept well. Examples of successful SVT's are Sunset, High Jump, Down-hill Skiing.
  • a semantic visual template for a concept should be small but cover a large percentage of relevant images and videos in the collection, for high precision-recall performance.
  • a semantic visual template can be understood further as a set of icons or example scenes/objects that represent the semantic with which the template is associated. From a semantic visual template, feature vectors can be extracted for querying. The icons are animated sketches.
  • the features associated with each object and their spatial and temporal relationships are important. Histograms, texture and structural information are examples of global features that can be part of such a template. The choice between an icon-based realization versus a feature vector set formed out of global characteristics depends upon the semantic to be represented.
  • each template contains multiple icons, example scenes/objects to represent a concept.
  • the elements of the set can overlap in their coverage. Desirably, coverage is maximized with a minimal template set.
  • Each icon for a concept e.g. down-hill ski, sunset, beach crowd, is a visual representation consisting of graphic objects resembling the actual objects in a scene.
  • Each object is associated with a set of visual attributes, e.g. color, shape, texture, motion. The relevancy of each attribute and each object to the concept is also specified.
  • sun object may be optional, as there may be sunset videos in which the sun is not visible.
  • high jump the motion attribute of the foreground object is mandatory, the texture attribute of the background is non-mandatory, and both are more relevant than other attributes.
  • Fig. 5 shows several potential icons for "high jump", and Fig. 6 for "sunset".
  • the optimal set of icons should be chosen based on relevancy feedback and maximal coverage in terms of recall as described below in further detail.
  • the positive coverage sets for different visual templates may overlap. Therefore, it is an objective to find a small set of visual templates with large, minimally overlapping positive coverage.
  • Users may provide initial conditions for effective visual templates. For example, a user may use a yellow circle (foreground) and a light-red rectangle (background) as an initial template for retrieving sunset scenes. Also, users may indicate weights and relevancy of different objects, attributes, and necessary conditions pertaining to the context by answering an interactive questionnaire. The questionnaire is sensitive to the current query that the user has sketched out on a sketchpad, for example.
  • the search system Given the initial visual template and relevancy of all visual attributes in the template, the search system will return a set of most similar images/video to the user. Given the returned results, the user can provide subjective evaluation of the returned results. The, precision of the results and positive coverage, i.e. recall can be computed.
  • the system can determine an optimal strategy for altering the initial visual query and generate modified queries based on: 1.
  • Such features are embodied in a technique as conceptually exemplified by Fig. 1, with specific illustration of a query for the concept "high jump".
  • the query includes three objects, namely two stationary rectangular background fields and an object which moves to the lower right.
  • four qualities are specified with associated weights, e.g. color, texture, shape and size, represented in Fig. 1 by vertical bars.
  • a new query can be formed by stepping at least one of the qualities, at which point user interaction can be invoked for deciding as to plausibility for inclusion as an icon in the template.
  • this template can be used for a database search. The results of the search can be evaluated for recall and precision. If acceptable, the template can be stored as a semantic visual template for "high jump".
  • the fundamental video data unit may be termed a video shot, comprising multiple segmented video objects.
  • the lifetime of any particular video object may be equal to or less than the duration of the video shot.
  • a similarity measure D between a member of the SVT set and a video shot can be defined as
  • O are the objects specified in the template
  • O' are the matched objects for Ofact
  • d f is the feature distance between its arguments
  • d s is the similarity between the spatial-temporal structure in the template and that among matched objects in the video shot
  • ⁇ f and ⁇ s are the normalized weights for the feature distance and the structure dissimilarity.
  • the query procedure is to generate a candidate list for each object in the query.
  • the distance D is the minimum over all possible sets of matched objects that satisfy the spatial-temporal restrictions. For example, if the semantic template has three objects and two candidate objects are kept for each single object query, there will be at most eight potential candidate sets of objects considered in computing the minimal distance in Equation 1. Given N objects in the query, this appears to require searching over all sets of
  • Each video object, O is used to query the entire object database, resulting in a list of matched objects which can be kept short by using a threshold. Only objects included in this list are then considered as candidate objects matching O,
  • the candidate objects on the list are then joined, resulting in the final set of matched objects on which the spatial-temporal structure relationships will be verified.
  • Template generation Two-way interaction is used between a user and the system for generating the templates. Given the initial scenario and using relevancy feedback, the technique converges on a small set of icons that gives maximum recall.
  • a user furnishes an initial query as a sketch of the concept for which a template is to be generated, consisting of objects with spatial and temporal constraints. The user can also specify whether the object is mandatory. Each object has features to which the user assigns relevancy weights.
  • the initial query can be regarded as a point in a high-dimensional feature space into which all videos in the database can be mapped.
  • a step size can be determined with the help of the weight that the user has specified along with the initial query, which weight can be regarded as a measure for the degree of relevancy attributed by the user to the feature of the object. Accordingly, a low weight results in coarse quantization and vice versa, e.g. when
  • is the jump distance corresponding to a feature
  • is the weight associated with the feature
  • the feature pace is quantized into hyper-rectangles. For example, for color the cuboids can ge generated using the metric for the LUV space along with ⁇ ( ⁇ ).
  • an additional join can be included in step 2 with respect to the candidate lists for each object.
  • a “sunset” may be described at the object level as well as the global level.
  • a global-level description may take the form of a color or texture histogram.
  • An object-level description may be a collection of two objects such as the sky and the sun. These objects may be further quantified using feature level descriptors.
  • a concept e.g. Sunset
  • N necessary
  • S sufficient
  • Additional templates may be generated manually i.e. as the user inputs additional queries. The task is undertaken for each concept. Necessary conditions can be imposed on a concept, thereby automatically generating additional templates, given an initial query template.
  • the user interacts with the system through a "concept questionnaire", to specify necessary conditions for the semantic searched for. These conditions may also be global, e.g. the global color distribution, the relative spatial and temporal interrelationships etc.
  • the system moves in the feature space to generate additional templates, with the user's original one as a starting point.
  • This generation is also modified by the relevancy feedback given to the system by the user.
  • new rules can be determined pertaining to the necessary conditions. These can be used further to modify the template generation procedure.
  • the rules are generated by looking at the correlation between the conditions deemed necessary for a concept with the videos that have been marked as relevant by the user.
  • a query, for a "crowd of people", in VideoQ is in the form of a sketch.
  • the user has specified a visual query with an object, giving weights for color and size, but is unable to specify a more detailed description in the form of either texture (of the crowd) or the relative spatial and temporal movements that characterize the concept of a crowd of people.
  • texture of the crowd
  • relative spatial and temporal movements that characterize the concept of a crowd of people.
  • the system identifies the video clips relevant to the concept that the user is interested in. Now, since the system knows that texture and the spatial and temporal arrangements are necessary to the concept, it seeks to determine consistent patterns amongst the features deemed necessary, amongst the relevant videos. These patterns are then returned to the user, who is asked if they are consistent with the concept that he is searching for. If the user accepts these patterns as consistent with the concept, then they will be used to generate new query templates, as illustrated by Fig. 3. Including this new rule has two-fold impact on query template generation, namely it improves the speed of the search and increases the precision of the returned results.
  • the query defines a feature space where the search is executed.
  • the feature space is defined by the attributes and relevancy weights of the visual template.
  • the attributes define the axes of the feature space, and the relevancy weights stretch/compress the associated axes.
  • each video shot can be represented as a point in this space.
  • the visual template covers a portion of this space. Since the visual template can differ in feature and in character (global against object level), the spaces that are defined by the templates differ and are non-overlapping.
  • Selection of a few features may be insufficient to determine a concept, but it may be adequately represented by a suitable selection differing as to weight, for example. Thus, a concept can be mapped into a feature space.
  • a concept is not limited to a single feature space nor to a single cluster.
  • sunsets cannot be totally characterized by a single color or a single shape.
  • it is important to determine not only the global static features and weights relating to a concept, but also those features and weights that can vary.
  • the search for concepts starts by specifying certain global constants. Through a context questionnaire, the number of objects in the search is determined, and the global features that are necessary to each object. These represent constraints in the search process that do not vary.
  • a user gives an initial query specifying features and setting weights.
  • a set intersection is taken with the set of necessary conditions defined by the user. The necessary conditions are left unchanged. Changes are made to the template based on changes to those features deemed sufficient. If the sets do not intersect, rules are derived that characterize the concept based on the necessary conditions and relevancy feedback.
  • the threshold determines the number of non-overlapping coverings possible. The number of coverings determines the size and number of jumps possible along that particular feature.
  • the algorithm performs a breadth first search and is guided by three criteria:
  • the greedy algorithm going in the direction of increasing recall Compute all possible initial jumps. Convert each jump into the corresponding visual template. Execute the query and collate all the results. Show the results to the user for relevancy feedback and chose those results that maximize incremental recall as possible points of subsequent query.
  • Keywords accompanying the data can either be generated manually or are obtained by association, i.e. keywords are extracted from the accompanying text (in the case of an image) or the captions that accompany videos.
  • VideoQ provides a "language" for inputting the query, in terms of a sketch. There is a simple correspondence between what exists in VideoQ and its natural language counterpart, as illustrated by Table 1.
  • a constrained language set can be used, with a set of allowable words.
  • a sentence is parsed into classes such as nouns, verbs, adjectives, and adverbs to generate a motion model of the video sequence.
  • nouns the objects
  • An noun (i.e. scenario/object) database may initially include a hundred scenes or so, and be extensible by user interaction. Each object may have a shape description that is modified by the various modifiers such as adjectives (color, texture), verbs (walked), adverbs (slowly). This can then be inserted into the VideoQ palette, where it may be subject to further refinement.
  • the parser When the parser encounters a word that is absent from its modifier database(i.e. the databases corresponding respectively to verbs, adverbs, prepositions, adjectives), it then looks up a thesaurus to determine if synonyms of that word are present in its database, and uses them instead. If that fails, it returns a message to indicate an invalid string.
  • modifier database i.e. the databases corresponding respectively to verbs, adverbs, prepositions, adjectives
  • the parser When the parser encounters a word that it cannot classify, the user must either modify the text or, if the word is a noun (like "Bill"), then he can indicate to the system the class (in this case a noun), and additionally indicate that the word refers to a human being. If the user indicates a noun that is absent from the system databases, then the user is prompted to draw that object in the sketch pad so that the system can learn about the object. In the database, attributes such as motion, color, texture and shapes can be generated at the object level, so that one level of matching can be at that level.
  • the audio stream can be used that accompanies a video, as illustrated by Fig. 4. Indeed, if the audio is closely correlated to the video, it may be the single most important source of the semantic content of the video.
  • a set of keywords can be generated, 10-20 per video sequence, for example. Then the search at the keyword level can be joined to the search that at the model level. Those videos can then be ranked the highest which match at the keyword (semantic) level as well as the motion-model level.
  • Semantic visual templates for retrieving video shots of slalom skiers Semantic visual templates for retrieving video shots of slalom skiers.
  • the system asks and the user answers questions regarding context.
  • the semantic visual template is labeled "slalom”.
  • the query is specified as object-based, including two objects.
  • the large blank background represents the ski slope and the smaller foreground object the skier with its characteristic zigzag motion trail.
  • the system generates a set or test icons from which the user selects plausible feature variations in the skier's color and motion trajectory.
  • the four selected colors and the three selected motion trails are joined to form 12 possible skiers.
  • the list of skiers is joined with the single background, resulting in the 12 icons of Fig. 7 where groups of three adjacent icons are understood as having the same.color.
  • the user chooses a candidate set to query the system.
  • the system retrieves the 20 closest video shots.
  • the user provides relevancy feedback to guide the system to a small set of exemplar for slalom skiers.
  • the database contains nine high jumpers in 2589 video shots.

Abstract

For database image/video retrieval, a semantic visual template (SVT) is a set of icons of example scenes/objects that characterize a concept, e.g. skiing, sunset and the like. SVTs provide for two-way interaction between a user and a system. The user can provide the system with an initial sketch or example image, as a seed to the system to automatically generate other representations of the same concept. The user then can pick those views for inclusion that are plausible for representing the concept. When an SVT has been established, the database can be searched with it, for the user to provide relevancy feedback on the returned results. With established SVTs, the user can interact with the system at concept level. In forming new concepts, pre-existing SVTs can be used. A limited vocabulary can be parsed, in conjunction with semantic visual templates for querying the system.

Description

METHOD AND SYSTEM FOR GENERATING SEMANTIC VISUAL TEMPLATES FOR IMAGE AND VIDEO RETRIEVAL
Technical Field
The invention relates to database still image, video and audio retrieval and, more particularly, to techniques which facilitate access to database items.
Background of the Invention Increasingly, as images and videos are produced, disseminated and stored in digital form, tools and systems for searching and retrieving visual information are becoming important. However, while efficient "search engines" have become widely available for text data, corresponding tools for searching for visual image and video data have remained elusive. Typically, in available image and video databases, keyword techniques are used to index and retrieve images. Known retrieval systems of this type suffer from a number of drawbacks, as images may not be associated with textual information, as manual inclusion of captions is time consuming and subjective, and as textual annotations are usually extrinsic, failing to represent the intrinsic visual characteristics of a scene. For example, a textual description such as "A man stands against a brick wall" conveys little visual information about either the man or the brick wall. Such visual information is often vital in retrieving a particular video.
Recently, researchers have begun to explore new forms of image and video repository retrieval. Such exploration is based on similarity of visual attributes, e.g. color, texture, shape and spatial and temporal relationships amongst the objects that make up a video. In this paradigm, queries are usually specified by giving an example image or a visual sketch. Then, retrieval systems return the image or video that have the highest similarity to the given example or sketch.
Summary of the Invention For ease of retrieval of images and videos from a database, the database can be indexed using a collection of visual templates. Preferably, in accordance with an aspect of the invention, the visual templates represent semantic concepts or categories, e.g. skiing, sunset and the like. There results an architecture using semantic visual templates (SVT), with each SVT standing for a concept and consisting of queries that describe the concept well.
Semantic visual templates can be established by an interactive process between a user and a system. The user can provide the system with an initial sketch or example image, as a seed to the system to automatically generate other representations of the same concept. The user then can pick those views for inclusion that are plausible for representing the concept. When an SVT has been established, the database can be searched with it, for the user to provide relevancy feedback on the returned results. With established SVT's , the user can interact with the system at concept level. In forming new concepts, pre-existing SVT's can be used.. Provided further is a technique for parsing a limited vocabulary of words in conjunction with semantic visual templates for querying the system.
Brief Description of the Drawing
Fig. 1 is a schematic of an interactive technique for generating a library or collection of semantic visual templates in accordance with a preferred embodiment of the invention.
Fig. 2 is a diagram which illustrates a concept having necessary and sufficient conditions.
Fig. 3 is a diagram which illustrates query generation.
Fig. 4 is a schematic of an interactive system in accordance with a preferred further embodiment of the invention, including audio processing.
Fig. 5 shows a set of icons exemplifying the concept "high jump".
Fig. 6 shows a set of icons exemplifying the concept "sunset".
Fig. 7 shows a semantic visual template for the concept "slalom"
Detailed Description
Incorporated herein by reference are U.S. provisional patent application No. 60/045,637, filed May 5, 1997 and PCT International Application No. PCT US98/09124 filed May 5, 1998 wherein Canada, Japan, the Republic of Korea and the United States of America are designated, describing techniques for object-based spatial and temporal visual searching using visual templates to search for images and video in different categories. Such search techniques, referred to as VideoQ, can be used in conjunction with the present invention.
In a video stream, the beginning and end of a scene can be determined using VideoQ, for example, which technique can be used further to compensate for camera motion for extracting objects in the scene. Still using VideoQ, each object can be characterized by salient attributes such as color, texture, size, shape and motion, for example. There results a video object database consisting of all the objects extracted from the scene and their attributes.
Visual Templates A visual template represents an idea, in the form of a sketch or an animated sketch. As a single visual template may be a poor representative of a class of interest, a library of visual templates can be assembled, containing representative templates for different semantic classes. For example, when searching for video clips of the class Sunset, one could select one or more visual templates corresponding to the class and use similarity-based querying to find video clips of sunsets. An important advantage of using a visual template library lies in linkage of a low-level visual feature representation to high-level semantic concepts. For example, if a user enters a query in a constrained natural language form as described in the above-referenced patent applications, visual templates can be used to transform the natural language query into automated queries specified by visual attributes and constraints. When visual content in the repository or database is not indexed textually, customary textual search methods cannot be applied directly.
Semantic Visual Templates (SVT
A semantic visual template is the set of visual templates associated with a particular semantic. This notion of an SVT has certain key properties as follows: Semantic visual templates are general in nature. For a given concept, there should be a set of visual templates that cover that concept well. Examples of successful SVT's are Sunset, High Jump, Down-hill Skiing.
A semantic visual template for a concept should be small but cover a large percentage of relevant images and videos in the collection, for high precision-recall performance.
Our procedures for finding semantic visual templates for different concepts are systematic, efficient, and robust. Efficiency refers to the convergence to a small visual template set. Robustness is demonstrated by applying the new library of templates to new image and video collections. With reference to VideoQ, a semantic visual template can be understood further as a set of icons or example scenes/objects that represent the semantic with which the template is associated. From a semantic visual template, feature vectors can be extracted for querying. The icons are animated sketches. In VideoQ, the features associated with each object and their spatial and temporal relationships are important. Histograms, texture and structural information are examples of global features that can be part of such a template. The choice between an icon-based realization versus a feature vector set formed out of global characteristics depends upon the semantic to be represented. For example, a sunset scene may be adequately represented by a couple of objects, while a waterfall or a crowd is better represented using a global feature set. Hence, each template contains multiple icons, example scenes/objects to represent a concept. The elements of the set can overlap in their coverage. Desirably, coverage is maximized with a minimal template set. Each icon for a concept, e.g. down-hill ski, sunset, beach crowd, is a visual representation consisting of graphic objects resembling the actual objects in a scene. Each object is associated with a set of visual attributes, e.g. color, shape, texture, motion. The relevancy of each attribute and each object to the concept is also specified. For example, for "sunset", color and spatial structures of the objects such as sun and sky are more relevant. For a sunset scene, the sun object may be optional, as there may be sunset videos in which the sun is not visible. For the concept "high jump", the motion attribute of the foreground object is mandatory, the texture attribute of the background is non-mandatory, and both are more relevant than other attributes. Some concepts may need just one object to represent the global attributes of the scene.
Fig. 5 shows several potential icons for "high jump", and Fig. 6 for "sunset". The optimal set of icons should be chosen based on relevancy feedback and maximal coverage in terms of recall as described below in further detail. We have devised efficient techniques for generating semantic visual templates for various concepts. Each semantic concept may have a few representative visual templates, which can be used to retrieve a significant portion of images and video, for positive coverage or high recall from the repository. The positive coverage sets for different visual templates may overlap. Therefore, it is an objective to find a small set of visual templates with large, minimally overlapping positive coverage.
Users may provide initial conditions for effective visual templates. For example, a user may use a yellow circle (foreground) and a light-red rectangle (background) as an initial template for retrieving sunset scenes. Also, users may indicate weights and relevancy of different objects, attributes, and necessary conditions pertaining to the context by answering an interactive questionnaire. The questionnaire is sensitive to the current query that the user has sketched out on a sketchpad, for example.
Given the initial visual template and relevancy of all visual attributes in the template, the search system will return a set of most similar images/video to the user. Given the returned results, the user can provide subjective evaluation of the returned results. The, precision of the results and positive coverage, i.e. recall can be computed.
The system can determine an optimal strategy for altering the initial visual query and generate modified queries based on: 1. The relevancy factor of each visual attribute obtained by the user questionnaire,
2. Precision-recall performance of the previous query, and 3. Information about feature level distribution of images and video in the repository.
Such features are embodied in a technique as conceptually exemplified by Fig. 1, with specific illustration of a query for the concept "high jump". The query includes three objects, namely two stationary rectangular background fields and an object which moves to the lower right. For each object in the query, four qualities are specified with associated weights, e.g. color, texture, shape and size, represented in Fig. 1 by vertical bars. A new query can be formed by stepping at least one of the qualities, at which point user interaction can be invoked for deciding as to plausibility for inclusion as an icon in the template. Once a suitable number of icons have been assembled into a tentative template, this template can be used for a database search. The results of the search can be evaluated for recall and precision. If acceptable, the template can be stored as a semantic visual template for "high jump".
Template Metric The fundamental video data unit may be termed a video shot, comprising multiple segmented video objects. The lifetime of any particular video object may be equal to or less than the duration of the video shot. A similarity measure D between a member of the SVT set and a video shot can be defined as
D = min { ωf {l] d^O,, O',) + ωs ds } (1)
where the O, are the objects specified in the template, O', are the matched objects for O„ df is the feature distance between its arguments, ds is the similarity between the spatial-temporal structure in the template and that among matched objects in the video shot, ωf and ωs are the normalized weights for the feature distance and the structure dissimilarity. The query procedure is to generate a candidate list for each object in the query. Then, the distance D is the minimum over all possible sets of matched objects that satisfy the spatial-temporal restrictions. For example, if the semantic template has three objects and two candidate objects are kept for each single object query, there will be at most eight potential candidate sets of objects considered in computing the minimal distance in Equation 1. Given N objects in the query, this appears to require searching over all sets of
N objects that appear together in a video shot. However, for computational economy, the following more economical procedure can be adopted:
1. Each video object, O, , say, is used to query the entire object database, resulting in a list of matched objects which can be kept short by using a threshold. Only objects included in this list are then considered as candidate objects matching O,
2. The candidate objects on the list are then joined, resulting in the final set of matched objects on which the spatial-temporal structure relationships will be verified.
Template generation Two-way interaction is used between a user and the system for generating the templates. Given the initial scenario and using relevancy feedback, the technique converges on a small set of icons that gives maximum recall. A user furnishes an initial query as a sketch of the concept for which a template is to be generated, consisting of objects with spatial and temporal constraints. The user can also specify whether the object is mandatory. Each object has features to which the user assigns relevancy weights.
The initial query can be regarded as a point in a high-dimensional feature space into which all videos in the database can be mapped. For automatic generation of the set of test icons it will be necessary to make jumps in each of the features of each of the objects, after quantizing the space. For quantizing, a step size can be determined with the help of the weight that the user has specified along with the initial query, which weight can be regarded as a measure for the degree of relevancy attributed by the user to the feature of the object.. Accordingly, a low weight results in coarse quantization and vice versa, e.g. when
Δ(ω) = l/(a-ω + b) (2)
where Δ is the jump distance corresponding to a feature, ω is the weight associated with the feature, and a and b are parameters which are chosen such that Δ(0) = 1 and Δ(l) = d0 which is a system parameter related to thresholding, set at 0.2 in a prototype system. Using the jump distance, the feature pace is quantized into hyper-rectangles. For example, for color the cuboids can ge generated using the metric for the LUV space along with Δ(ω).
To prevent the total number of possible icons from increasing rapidly, joint variation of the features is prevented, e.g. as follows:
1. For each feature in the object, the user picks a plausible set for that feature. 2. The system then performs a join on the set of features associated with the object.
3. The user then picks the joins that most likely represent variations of the object, resulting in a candidate icon list.
In a multiple object case, an additional join can be included in step 2 with respect to the candidate lists for each object. Once a list of plausible scenarios has been generated, the system is queried using the icons which the user has picked. Using relevancy feedback on the returned results, with the user labeling the returned results as positive or negative, those icons are determined which result in maximum recall.
Concept Covers
Sufficiently many "coverings" are desired of a concept, for a user to search a database. Each covering can reside in a different feature space, e.g. a "sunset" may be described at the object level as well as the global level. A global-level description may take the form of a color or texture histogram. An object-level description may be a collection of two objects such as the sky and the sun. These objects may be further quantified using feature level descriptors.
As illustrated by Fig. 2, a concept (e.g. Sunset) has two different kinds of conditions, necessary (N) and sufficient (S). A semantic visual template is a sufficient condition on the concept and not a necessary one, so that a particular SVT need not cover the concept to its full extent. Additional templates may be generated manually i.e. as the user inputs additional queries. The task is undertaken for each concept. Necessary conditions can be imposed on a concept, thereby automatically generating additional templates, given an initial query template.
The user interacts with the system through a "concept questionnaire", to specify necessary conditions for the semantic searched for. These conditions may also be global, e.g. the global color distribution, the relative spatial and temporal interrelationships etc. Once the necessary and sufficient conditions for the concept are established, the system moves in the feature space to generate additional templates, with the user's original one as a starting point. This generation is also modified by the relevancy feedback given to the system by the user. By analyzing the relevancy feedback, new rules can be determined pertaining to the necessary conditions. These can be used further to modify the template generation procedure. The rules are generated by looking at the correlation between the conditions deemed necessary for a concept with the videos that have been marked as relevant by the user. This principle of determining rules (or implications) is akin to the techniques found in "data mining" as described in the above-identified patent applications and in papers by S. Brin et al., "Dynamic Itemset Counting and Implication Rules for Market Basket Data", ACM SIGMOD Conference on Management of Data, 1997, pp. 255- 246 and S. Brin et al., "Beyond Market Baskets: Generalizing Association Rules to Correlations", ACM SIGMOD Conference on Management of Data, 1997, pp. 265- Rule Generation Example
A query, for a "crowd of people", in VideoQ is in the form of a sketch. The user has specified a visual query with an object, giving weights for color and size, but is unable to specify a more detailed description in the form of either texture (of the crowd) or the relative spatial and temporal movements that characterize the concept of a crowd of people. However, since he feels that the idea of a "crowd" is strongly characterized by a texture and by relative spatial and temporal arrangements of people, he lists them down as necessary conditions.
Through the process of feedback, the system identifies the video clips relevant to the concept that the user is interested in. Now, since the system knows that texture and the spatial and temporal arrangements are necessary to the concept, it seeks to determine consistent patterns amongst the features deemed necessary, amongst the relevant videos. These patterns are then returned to the user, who is asked if they are consistent with the concept that he is searching for. If the user accepts these patterns as consistent with the concept, then they will be used to generate new query templates, as illustrated by Fig. 3. Including this new rule has two-fold impact on query template generation, namely it improves the speed of the search and increases the precision of the returned results.
Generating Concept Covers The query defines a feature space where the search is executed. The feature space is defined by the attributes and relevancy weights of the visual template. In particular, the attributes define the axes of the feature space, and the relevancy weights stretch/compress the associated axes. Within this composite feature space, each video shot can be represented as a point in this space. The visual template covers a portion of this space. Since the visual template can differ in feature and in character (global against object level), the spaces that are defined by the templates differ and are non-overlapping.
Selection of a few features may be insufficient to determine a concept, but it may be adequately represented by a suitable selection differing as to weight, for example. Thus, a concept can be mapped into a feature space.
A concept is not limited to a single feature space nor to a single cluster. For example, with respect to the class of sunset video sequences, sunsets cannot be totally characterized by a single color or a single shape. Thus, it is important to determine not only the global static features and weights relating to a concept, but also those features and weights that can vary.
The search for concepts starts by specifying certain global constants. Through a context questionnaire, the number of objects in the search is determined, and the global features that are necessary to each object. These represent constraints in the search process that do not vary.
A user gives an initial query specifying features and setting weights. A set intersection is taken with the set of necessary conditions defined by the user. The necessary conditions are left unchanged. Changes are made to the template based on changes to those features deemed sufficient. If the sets do not intersect, rules are derived that characterize the concept based on the necessary conditions and relevancy feedback.
The relevancy weight of each feature indicates the tolerance that the user desires along each feature. This tolerance is then mapped to a distance threshold along each feature, e.g. d(ω) = l/(a-ω + c), defining a hyper-ellipsoid in the feature space searched in. The threshold determines the number of non-overlapping coverings possible. The number of coverings determines the size and number of jumps possible along that particular feature. The algorithm performs a breadth first search and is guided by three criteria:
First, the greedy algorithm going in the direction of increasing recall: Compute all possible initial jumps. Convert each jump into the corresponding visual template. Execute the query and collate all the results. Show the results to the user for relevancy feedback and chose those results that maximize incremental recall as possible points of subsequent query.
Second, a logarithmic search, as subsequent queries are searched by taking a smaller jump in a local region. The rationale is that the current query point produced good results and should be search carefully for further templates. Searching stops when enough visual templates have been generated to cover above 70 percent of the concept (i.e recall is above 70%).
Third, as the breadth first search often yields too many possibilities to examine at once, feature level distributions are used to guide the search. The distribution along each feature has been pre-calculated. This information is used to select jumps to regions that have a high concentration of video shots and avoid sparse ones.
Language Integration with SVT
Known text based queries on images and videos rely on matching on keywords accompanying the image or the video. Keywords accompanying the data can either be generated manually or are obtained by association, i.e. keywords are extracted from the accompanying text (in the case of an image) or the captions that accompany videos.
Such an approach precludes the possibility of any practical system containing a very large database of videos or images for several reasons such as:
It is not feasible to generate annotations manually for the existing database of videos.
Most videos do not contain captions. There may not be an immediate correlation between the accompanying caption and the video. For example, during a baseball game the commentators may be talking about the exploits of Babe Ruth, who is not present in the game that is being played, and a text-based keyword on locating videos containing "Babe Ruth" will incorrectly display this video. Generating semantic content about a video by analysis of the video stream alone amounts to the computer vision problem which is known to be difficult. A more practicable approach is to simultaneously make use of the visual content such as the motion of objects, attributes like color and texture, with the descriptive power of natural languages. The user types in a string, which the system parses into a video model.
VideoQ provides a "language" for inputting the query, in terms of a sketch. There is a simple correspondence between what exists in VideoQ and its natural language counterpart, as illustrated by Table 1.
Attribute NL Type Motion Verb
Color, Texture Adjective
Shape Noun
Spatial/temporal Preposition/Conjunction
Table 1
A constrained language set can be used, with a set of allowable words. A sentence is parsed into classes such as nouns, verbs, adjectives, and adverbs to generate a motion model of the video sequence.
For example for the phrase "Bill walked slowly towards the sunset", the system can parse as shown in Table 2.
Word NL-Tvpe
Bill Noun
Walked Verb
Slowly Adverb
Towards Preposition
Sunset Noun
Table 2 For verbs, adverbs, adjectives and prepositions, a small but fixed database can be used, as these are modifiers (or descriptors) on nouns (the objects). An noun (i.e. scenario/object) database may initially include a hundred scenes or so, and be extensible by user interaction. Each object may have a shape description that is modified by the various modifiers such as adjectives (color, texture), verbs (walked), adverbs (slowly). This can then be inserted into the VideoQ palette, where it may be subject to further refinement.
When the parser encounters a word that is absent from its modifier database(i.e. the databases corresponding respectively to verbs, adverbs, prepositions, adjectives), it then looks up a thesaurus to determine if synonyms of that word are present in its database, and uses them instead. If that fails, it returns a message to indicate an invalid string.
When the parser encounters a word that it cannot classify, the user must either modify the text or, if the word is a noun (like "Bill"), then he can indicate to the system the class (in this case a noun), and additionally indicate that the word refers to a human being. If the user indicates a noun that is absent from the system databases, then the user is prompted to draw that object in the sketch pad so that the system can learn about the object. In the database, attributes such as motion, color, texture and shapes can be generated at the object level, so that one level of matching can be at that level.
As a further source of information, the audio stream can be used that accompanies a video, as illustrated by Fig. 4. Indeed, if the audio is closely correlated to the video, it may be the single most important source of the semantic content of the video. From the audio stream, a set of keywords can be generated, 10-20 per video sequence, for example. Then the search at the keyword level can be joined to the search that at the model level. Those videos can then be ranked the highest which match at the keyword (semantic) level as well as the motion-model level.
EXAMPLES
Semantic visual templates for retrieving video shots of slalom skiers.
1. The system asks and the user answers questions regarding context. The semantic visual template is labeled "slalom". The query is specified as object-based, including two objects.
2. The user sketches the initial query. The large blank background represents the ski slope and the smaller foreground object the skier with its characteristic zigzag motion trail.
3. Maximum relevancy weights are assigned to all the features associated with the background and the skier. The background feature is specified to remain static while those of the skier can vary during template generation.
4. The system generates a set or test icons from which the user selects plausible feature variations in the skier's color and motion trajectory.
5. The four selected colors and the three selected motion trails are joined to form 12 possible skiers. The list of skiers is joined with the single background, resulting in the 12 icons of Fig. 7 where groups of three adjacent icons are understood as having the same.color.
6. The user chooses a candidate set to query the system. The system retrieves the 20 closest video shots. The user provides relevancy feedback to guide the system to a small set of exemplar for slalom skiers.
Sunsets. A database was used which includes 72 sunsets in more than 1952 video shots. Using just an initial sketch, without semantic visual templates, a recall of 10 % and a precision of 35% were realized. Using semantic visual templates, 8 icons were generated, with which 36 sunset were found, for a recall of 50% and a precision of 24%.
High Jumpers. The database contains nine high jumpers in 2589 video shots.
Without semantic visual templates, recall was 44% and precision 20%. With semantic visual templates recall was improved to 56% and precision to 25%. The system converged to a single icon different from the initial sketch provided by the user.

Claims

1. A computerized method for generating a visual template for a concept, comprising the steps of: a. obtaining at least one initial query for the concept; b. generating at least one additional query related to the initial query; c. producing the additional query for inspection for appropriateness with respect to the concept; and d. in case of appropriateness, including the additional query in the visual template for the concept.
2. The method of claim 1, wherein each query is represented by an icon/example image.
3. The method of claim 1, wherein the initial query is obtained via a sketchpad.
4. The method of claim 1, wherein generating the additional query comprises stepping a query feature with a step size which is inversely related to a weight associated with the query feature.
5. The method of claim 1, wherein generating the additional icon comprises forming a join of plausible feature values.
6. The method of claim 1, wherein appropriateness is ascertained by two-way user interaction.
7. A computerized method for querying a video database for a concept using a subset of a natural language in conjunction with semantic visual templates, comprising the steps of: a. obtaining a textual query; b. parsing the query to generate visual attributes; c. using the visual attributes for forming a visual query; d. using the visual query to retrieve information; and e. displaying the information.
8. The method of claim 7, wherein the textual query is obtained from a keyboard.
9. The method of claim 7, wherein the subset of the natural language comprises a small set of nouns, verbs, prepositions, adjectives and adverbs.
10. The method of claim 7, further comprising a step of expanding the subset in an interactive fashion.
11. The method of claim 7, wherein parsing comprises the steps of:
(i) establishing a correspondence between the query and the natural language subset;
(ii) labeling different parts of the query as a noun, verb, adjective or preposition; and (iii) obtaining clarification if a word in the query is absent from the natural language subset and labeling the word accordingly.
12. The method of claim 7, wherein forming the visual query comprises the step of establishing a correspondence between the natural language subset and a set of semantic visual templates generated by the method of claim 1, with the semantic visual template being a visual embodiment of a noun in the query, an adjective serving to modify the visual embodiment of the nouns, a verb serving to embody an action, and a preposition serving to establish a spatial and temporal order needed in forming a visual query.
13. A computerized system for generating a visual template for a concept, comprising: a. means for obtaining at least one initial query for the concept; b. means for generating at least one additional query related to the initial query; c. means for producing the additional query for inspection for appropriateness with respect to the concept; and d. means in case of appropriateness for including the additional query in the visual template for the concept.
14. The system of claim 1, wherein each query is represented by an icon/example image.
15. The system of claim 1, wherein the means for obtaining the initial query comprises a sketchpad.
16. The system of claim 1, wherein generating the additional query comprises stepping a query feature with a step size which is inversely related to a weight associated with the query feature.
17. The system of claim 1, wherein generating the additional icon comprises forming a join of plausible feature values.
18. The method of claim 1, wherein appropriateness is ascertained by two- way interaction between the system and a user.
19. A computerized system for querying a video database for a concept using a subset of a natural language in conjunction with semantic visual templates, comprising: a. means for obtaining a textual query; b. means for parsing the query to generate visual attributes; c. means for using the visual attributes for forming a visual query; d. means for using the visual query to retrieve information; and e. means for displaying the information.
20. The system of claim 19, wherein the means for obtaining the textual query comprises a keyboard.
21. The system of claim 19, wherein the subset of the natural language comprises a small set of nouns, verbs, prepositions, adjectives and adverbs.
22. The system of claim 19, further comprising means for expanding the subset in an interactive fashion.
23. The method of claim 19, wherein the means for parsing comprises:
(i) means for establishing a correspondence between the query and the natural language subset;
(ii) mans for labeling different parts of the query as a noun, verb, adjective or preposition; and
(iii) means for obtaining clarification if a word in the query is absent from the natural language subset and labeling the word accordingly.
24. The system of claim 19, wherein the means for forming the visual query comprises means for establishing a correspondence between the natural language subset and a set of semantic visual templates generated by the system of claim 13, with the semantic visual template being a visual embodiment of a noun in the query, an adjective serving to modify the visual embodiment of the nouns, a verb serving to embody an action, and a preposition serving to establish a spatial and temporal order needed in forming a visual query.
PCT/US1999/004776 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video retrieval WO1999045483A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
KR1020007009804A KR20010041607A (en) 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video retrieval
CA002322448A CA2322448A1 (en) 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video retrieval
EP99911110A EP1066572A1 (en) 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video retrieval
JP2000534957A JP2002506255A (en) 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video verification

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US7678198P 1998-03-04 1998-03-04
US60/076,781 1998-03-04

Publications (2)

Publication Number Publication Date
WO1999045483A1 true WO1999045483A1 (en) 1999-09-10
WO1999045483A9 WO1999045483A9 (en) 2000-10-12

Family

ID=22134152

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1999/004776 WO1999045483A1 (en) 1998-03-04 1999-03-04 Method and system for generating semantic visual templates for image and video retrieval

Country Status (5)

Country Link
EP (1) EP1066572A1 (en)
JP (1) JP2002506255A (en)
KR (1) KR20010041607A (en)
CA (1) CA2322448A1 (en)
WO (1) WO1999045483A1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002023891A3 (en) * 2000-09-13 2002-05-30 Koninkl Philips Electronics Nv Method for highlighting important information in a video program using visual cues
US6563959B1 (en) 1999-07-30 2003-05-13 Pixlogic Llc Perceptual similarity image retrieval method
US7031555B2 (en) 1999-07-30 2006-04-18 Pixlogic Llc Perceptual similarity image retrieval
EP2100260A1 (en) * 2006-12-01 2009-09-16 Google, Inc. Identifying images using face recognition
GB2466245A (en) * 2008-12-15 2010-06-23 Univ Sheffield Crime Scene Mark Identification System
US8190604B2 (en) 2008-04-03 2012-05-29 Microsoft Corporation User intention modeling for interactive image retrieval
US8463045B2 (en) 2010-11-10 2013-06-11 Microsoft Corporation Hierarchical sparse representation for image retrieval
US9147125B2 (en) 2013-05-03 2015-09-29 Microsoft Technology Licensing, Llc Hand-drawn sketch recognition
US9317533B2 (en) 2010-11-02 2016-04-19 Microsoft Technology Licensing, Inc. Adaptive image retrieval database
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
US11347786B2 (en) 2013-11-27 2022-05-31 Hanwha Techwin Co., Ltd. Image search system and method using descriptions and attributes of sketch queries

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116992294B (en) * 2023-09-26 2023-12-19 成都国恒空间技术工程股份有限公司 Satellite measurement and control training evaluation method, device, equipment and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5493677A (en) * 1994-06-08 1996-02-20 Systems Research & Applications Corporation Generation, archiving, and retrieval of digital images with evoked suggestion-set captions and natural language interface
US5708805A (en) * 1992-10-09 1998-01-13 Matsushita Electric Industrial Co., Ltd. Image retrieving apparatus using natural language
US5724484A (en) * 1991-03-20 1998-03-03 Hitachi, Ltd. Data processing methods and apparatus for supporting analysis/judgement

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5724484A (en) * 1991-03-20 1998-03-03 Hitachi, Ltd. Data processing methods and apparatus for supporting analysis/judgement
US5708805A (en) * 1992-10-09 1998-01-13 Matsushita Electric Industrial Co., Ltd. Image retrieving apparatus using natural language
US5493677A (en) * 1994-06-08 1996-02-20 Systems Research & Applications Corporation Generation, archiving, and retrieval of digital images with evoked suggestion-set captions and natural language interface

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6563959B1 (en) 1999-07-30 2003-05-13 Pixlogic Llc Perceptual similarity image retrieval method
US7031555B2 (en) 1999-07-30 2006-04-18 Pixlogic Llc Perceptual similarity image retrieval
WO2002023891A3 (en) * 2000-09-13 2002-05-30 Koninkl Philips Electronics Nv Method for highlighting important information in a video program using visual cues
US9053357B2 (en) 2006-12-01 2015-06-09 Google Inc. Identifying images using face recognition
EP2100260A1 (en) * 2006-12-01 2009-09-16 Google, Inc. Identifying images using face recognition
US8085995B2 (en) 2006-12-01 2011-12-27 Google Inc. Identifying images using face recognition
EP2100260B1 (en) * 2006-12-01 2012-05-23 Google, Inc. Identifying images using face recognition
AU2007325117B2 (en) * 2006-12-01 2012-10-11 Google Llc Identifying images using face recognition
US9552511B2 (en) 2006-12-01 2017-01-24 Google Inc. Identifying images using face recognition
US8190604B2 (en) 2008-04-03 2012-05-29 Microsoft Corporation User intention modeling for interactive image retrieval
GB2466245A (en) * 2008-12-15 2010-06-23 Univ Sheffield Crime Scene Mark Identification System
US9317533B2 (en) 2010-11-02 2016-04-19 Microsoft Technology Licensing, Inc. Adaptive image retrieval database
US8463045B2 (en) 2010-11-10 2013-06-11 Microsoft Corporation Hierarchical sparse representation for image retrieval
US9147125B2 (en) 2013-05-03 2015-09-29 Microsoft Technology Licensing, Llc Hand-drawn sketch recognition
US9870516B2 (en) 2013-05-03 2018-01-16 Microsoft Technology Licensing, Llc Hand-drawn sketch recognition
US11347786B2 (en) 2013-11-27 2022-05-31 Hanwha Techwin Co., Ltd. Image search system and method using descriptions and attributes of sketch queries
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
CN106126581B (en) * 2016-06-20 2019-07-05 复旦大学 Cartographical sketching image search method based on deep learning

Also Published As

Publication number Publication date
JP2002506255A (en) 2002-02-26
KR20010041607A (en) 2001-05-25
WO1999045483A9 (en) 2000-10-12
CA2322448A1 (en) 1999-09-10
EP1066572A1 (en) 2001-01-10

Similar Documents

Publication Publication Date Title
Srihari et al. Intelligent indexing and semantic retrieval of multimodal documents
Cheng et al. Semantic visual templates: linking visual features to semantics
Meghini et al. A model of multimedia information retrieval
Chen et al. A novel video summarization based on mining the story-structure and semantic relations among concept entities
US7502780B2 (en) Information storage and retrieval
EP1565846B1 (en) Information storage and retrieval
US7610306B2 (en) Multi-modal fusion in content-based retrieval
Hsu et al. Reranking methods for visual search
US8140550B2 (en) System and method for bounded analysis of multimedia using multiple correlations
US20040107221A1 (en) Information storage and retrieval
WO1999045483A1 (en) Method and system for generating semantic visual templates for image and video retrieval
Ciocca et al. Quicklook2: An integrated multimedia system
Paz-Trillo et al. An information retrieval application using ontologies
Pradhan et al. A query model to synthesize answer intervals from indexed video units
Baan et al. Lazy users and automatic video retrieval tools in (the) lowlands
Srihari et al. A model for multimodal information retrieval
Doulaverakis et al. Ontology-based access to multimedia cultural heritage collections-The REACH project
Mallik et al. Multimedia ontology learning for automatic annotation and video browsing
Kutics et al. Use of adaptive still image descriptors for annotation of video frames
Tanaka et al. Organization and retrieval of video data
Chen et al. Generating semantic visual templates for video databases
Faudemay et al. Intelligent delivery of personalised video programmes from a video database
Srihari et al. Finding pictures in context
Affendey et al. Video data modelling to support hybrid query
Gauch et al. The VISION digital video library project

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): CA JP KR US

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
ENP Entry into the national phase

Ref document number: 2322448

Country of ref document: CA

Ref country code: CA

Ref document number: 2322448

Kind code of ref document: A

Format of ref document f/p: F

ENP Entry into the national phase

Ref country code: JP

Ref document number: 2000 534957

Kind code of ref document: A

Format of ref document f/p: F

WWE Wipo information: entry into national phase

Ref document number: 1020007009804

Country of ref document: KR

WWE Wipo information: entry into national phase

Ref document number: 1999911110

Country of ref document: EP

AK Designated states

Kind code of ref document: C2

Designated state(s): CA JP KR US

AL Designated countries for regional patents

Kind code of ref document: C2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

COP Corrected version of pamphlet

Free format text: PAGES 1/3-3/3, DRAWINGS, REPLACED BY NEW PAGES 1/4-4/4; DUE TO LATE TRANSMITTAL BY THE RECEIVING OFFICE

WWP Wipo information: published in national office

Ref document number: 1999911110

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 09623277

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 1020007009804

Country of ref document: KR

WWW Wipo information: withdrawn in national office

Ref document number: 1999911110

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1020007009804

Country of ref document: KR