WO2008150544A1 - Improving audio/video fingerprint search accuracy using multiple search combining - Google Patents

Improving audio/video fingerprint search accuracy using multiple search combining Download PDF

Info

Publication number
WO2008150544A1
WO2008150544A1 PCT/US2008/007054 US2008007054W WO2008150544A1 WO 2008150544 A1 WO2008150544 A1 WO 2008150544A1 US 2008007054 W US2008007054 W US 2008007054W WO 2008150544 A1 WO2008150544 A1 WO 2008150544A1
Authority
WO
WIPO (PCT)
Prior art keywords
fingerprint
query
measure
search
difference
Prior art date
Application number
PCT/US2008/007054
Other languages
French (fr)
Inventor
Wenju Jiang
Claus Bauer
Original Assignee
Dolby Laboratories Licensing Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dolby Laboratories Licensing Corporation filed Critical Dolby Laboratories Licensing Corporation
Priority to US12/663,057 priority Critical patent/US8266142B2/en
Priority to EP08794341A priority patent/EP2168061A1/en
Priority to CN2008800188850A priority patent/CN101681381B/en
Priority to JP2010511189A priority patent/JP5090523B2/en
Publication of WO2008150544A1 publication Critical patent/WO2008150544A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/63Querying
    • G06F16/632Query formulation
    • G06F16/634Query by example, e.g. query by humming
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Definitions

  • the invention relates to audio and video fingerprinting. More particularly the invention relates to improving the accuracy of fingerprint search accuracy.
  • One prospective application is identifying an excerpt of audio or video within a repository of known content. This can be useful in monitoring illegal download/transfer of content on the Internet. It can also be useful in providing intelligent access to content that users are listening to or watching and are interested in, but for which they do not know the title or author information.
  • the user can, for example, place a cell phone against the loudspeaker currently playing a song, and let the cell phone operator's software find out the song's title, its artist and album information, etc.
  • Such identification capability is generally implemented by first generating feature information (called fingerprints) designed to identify uniquely audio and video signals, and then performing some form of pattern matching search between the fingerprints from the repository/database and those from the excerpt in the search query.
  • fingerprints feature information
  • Such a database is often implemented in the form of a search tree, although other data structures are also possible.
  • each fingerprint corresponds to a certain segment of audio or video. So a two second audio fingerprint would correspond to a two second segment of audio.
  • a fingerprint is " usually implemented as a concatenation of small blocks of feature information that are often called signatures.
  • a two second fingerprint may, for example, consist of two hundred 10 millisecond (ms) long signatures, wherein each signature is computed from 10 ms of new audio or video information.
  • the pattern matching of fingerprints is therefore a process of comparing corresponding signatures. This is illustrated in FIG. 1 that shows the creation of a fingerprint from signatures.
  • a well- designed fingerprint generation method should select whichever signature is closer in terms of timing to better match the corresponding signature from the query.
  • the goal of a fingerprint search system is to find a query fingerprint's corresponding counterpart fingerprint, and the counterpart should minimize its time misalignment with the query fingerprint, if such misalignment exists at all.
  • the query excerpt may have undergone some editing or processing steps, such as recapturing sound played from a loudspeaker using a cell phone, there may be some distortions in the captured audio/video signals. As a result, the resulting fingerprints may also change slightly with respect to their counterparts in the database, assuming there is a count erpart.
  • the possibility of distortions in the excerpt means the best match in such a search is often not an identical match, but a form of closest match.
  • To define the closest match requires the definition of a measure of difference between two fingerprints.
  • a commonly used measure of difference is the Hamming distance, that is, the number of differing bits between the fingerprint from the query excerpt and that from the database.
  • the corresponding criterion for the closest match is thus the fingerprint from the database that has the minimum Hamming distance from the fingerprint from the query excerpt.
  • the Hamming distance between two fingerprints divided by the number of bits in a fingerprint is often referred to as the bit error rate (BER).
  • the BER is an example of measure of relative difference.
  • the minimum Hamming distance criterion works well when the BER between the fingerprint from the excerpt and its counterpart is small. However, as BER increases, the search result producing the minimum Hamming distance increasingly does not find the actual counterpart. Fortunately, in most fingerprint search applications, it is only necessary to identify the correct audio/video piece, but not necessarily the corresponding segment. But when BER increases further, the search result may even find the wrong audio/video piece, let alone the correct segment within that piece. The BER depends on both the level of distortion in the query exce ⁇ t, and the robustness of the fingerprint extraction method with respect to such distortions.
  • an exce ⁇ t does not belong to any piece in the database.
  • the excerpt may be the recording of a new composition of music. Because no search algorithm can know beforehand (without being told) whether an exce ⁇ t belongs to the database or not, the best it can do is still apply the same criterion of minimum Hamming distance, but expecting that the minimum Hamming distance found in such cases will be much more different (preferably higher) than that of an exce ⁇ t originally from the database, and use some threshold to determine whether the exce ⁇ t is from the database.
  • the exce ⁇ t belongs to the database, and the search returns the correct audio/video piece (finding the correct piece is enough, here it is not necessary to find the correct counte ⁇ art segment).
  • the exce ⁇ t belongs to the database, and the search returns the wrong audio/video piece.
  • FIG. 2 shows an example of BER distribution for three different possible outcomes of a single search. Each of these outcomes would generate a corresponding probability-density-function (PDF) distribution of the BER.
  • PDF probability-density-function
  • the BER of the first outcome should generally be smaller than the BER of the second and third outcomes, as illustrated in FIG. 2.
  • FIG. 1 is a schematic example of the manner in which a particular type of fingerprint may be generated.
  • FIG. 2 shows an idealized example of a probability-density- function (PDF) as a function of the bit error rate (BER) that may result for each of three possible outcomes of a single fingerprint matching search.
  • PDF probability-density- function
  • BER bit error rate
  • FIG. 3 is a schematic conceptualization of an example of fingerprint-matching search results combining using majority voting with respect to three searches.
  • FIG. 4a is a schematic conceptualization of an example of fingerprint-matching search results with respect to three searches in which time locations are consistent for two of three potential matches.
  • FIG. 4b is a schematic conceptualization of an example of fingerprint-matching search results with respect to three searches in which time locations are not consistent for two of three potential matches.
  • FIG. 5 is an example of the manner in which a confidence level may be mapped as a function of the bit error rate (BER).
  • BER bit error rate
  • FIG. 6 is an idealized example useful in understanding one way, using linear regression in this example, that relative time locations between the time of the fingerprint query and the time of the fingerprint found in a database may be determined.
  • FIG. 7 is an example of the manner in which a confidence level may be mapped as a function of time difference in relative time locations between the time of the fingerprint query and the time of the fingerprint found in a database.
  • a method for use in identifying a segment of audio and/or video information comprises obtaining a query fingerprint at each of a plurality of spaced-apart time locations in said segment, searching fingerprints in a database for a potential match for each such query fingerprint, obtaining a confidence level of a potential match to a found fingerprint in the database for each such query fingerprint, and combining the results of searching for potential matches, wherein each potential match result is weighted by a respective confidence level.
  • a confidence level may be a function of at least one or both of (1) a measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time location of a found fingerprint.
  • the measure of difference may be based on Hamming distance, or its relative difference variant such as the bit error rate, or it may be based on an Lp norm in which a p-th root of the sum of the p-th power of absolute values of difference over each dimension of a vector that constitutes a fingerprint is obtained.
  • the Lp norm measure may be an L2 norm difference measure in which a square root of the sum of the square of absolute values of difference is obtained.
  • the measure of difference may be obtained by auto-correlation in which the value of correlation indicates the closeness of the match.
  • the confidence level may be a combination of respective confidence levels of each of a function of at least one or both of (1) the measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time locations of a found fingerprint.
  • the combination of the respective confidence levels may be, for example, a summation of respective confidence levels or a multiplicative product of respective confidence levels.
  • a confidence level may be a function of at least the relative timing relationship between the time location of the query fingerprint and the time location of the potentially matching found fingerprint in the database, in which case there are at least three spaced- apart time locations.
  • the spaced-apart time locations in the segment of audio and/or video information may be spaced apart so as to minimize correlation between or among query fingerprints. There may be multiple sets of spaced-apart time locations with respect to each of which sets the results of searching for potential matches is combined. The results of searching may be combined within each set and the results of such combinations then may be further combined. Alternatively, the results of searching may be combined across all sets. Fingerprints in the database may have related metadata about said audio and/or video information. Such metadata may include at least a piece title of the audio and/or video information.
  • a query fingerprint may be obtained at each of said plurality of spaced-apart time locations in substantially the same way.
  • a simple but useful method of combining results from multiple searches is majority voting. For example, as illustrated in FIG. 3, if one conducts three searches, two of them return piece A, one returns piece B, then we conclude that the final search result should be piece A. If we get three different answers, piece A, B and C, then there is no majority winner, and the answer is inconclusive. In this case, it is best to conclude that the excerpt does not belong to the database, because if it did belong, it is relatively unlikely three searches will give three different results.
  • a voting scheme should be designed to maximize the number of correct answers and minimize the number of wrong answers, using probabilistic analysis. For example, for pieces originally from the database, if a single search has 98% accuracy, then combining three searches with majority voting will draw the wrong conclusion only if at least two out of three searches are wrong, which is 3xO.98xO.O2 2 + 0.02 3 « 1.2xlO ⁇ 3 , so the combined accuracy should be almost 99.9%. This probabilistic analysis assumes each search to be independent of other searches, i.e., no correlation between searches.
  • Exemplary pseudocode for search combining using majority voting may be expressed as follows:
  • the time intervals d[l ], d[2], ..., in FIG. 3 are used to conduct searches at different locations, and, in general, they should be chosen to minimize the overlaps among their corresponding search segments in order to reduce any undesirable correlation between search results.
  • the combining process can be further cascaded by additional iteration(s) of multiple search(es) in the event of inconclusive votes, and either consider the votes in the new iteration only, or combine some or all of the previous votes with a voting scheme chosen by the implementer. It would seem that a simple scheme like majority voting can give very good results.
  • Accuracy may be increased, even in the presence of such burst errors, by employing aspects of the present invention. This may be accomplished, for example, by combining not only the piece's title information from each search result, but also accessory information such as the time location of the found segment within the found piece, and the BER of the search result. For example in FIG. 4(a), if two of three searches return piece A, and the time locations of these two search segments differ by five seconds, then the time locations of the found segments in these two search results should also differ by five seconds (or within a certain error margin). If this is not satisfied, as in FIG. 4(b), it is almost certain that this is a false majority winner.
  • Exemplary pseudocode for search combining using t 0 ff Set accessory information may be expressed as follows:
  • search[i] means the result from the i-th search and "search[i]. piece” is the found piece's identification information such as its title.
  • Min t offset dist For a large N number of searches, clearly the value of Min t offset dist would be smaller if more of the searches return the same piece, provided that the result is correct. For instance, all else being equal, the minimum value among five samples is certainly smaller than the minimum value among three samples. So, we can define more thresholds, e.g., d_B, d_C, in addition to d_A, depending on how many searches return the same piece. This way, one can use the accessory information (W tset in this example) intelligently based on the majority voting information (how many searches return the same piece), and achieve higher accuracy.
  • the spacing between the query fingerprints versus, the spacing between the fingerprints from the search results will no longer be the same. Instead they will drift apart slowly with respect to the spacing between the query fingerprints. To deal with this, one can either make sure that the spacing between the query fingerprints will never cause the deviation of t o i ⁇ set to exceed a threshold such as d_A.
  • the exact amount of adjustment in such dynamic thresholding can be determined by the implementer, using probabilistic analysis to ensure that the number of false matches resulting from such dynamic thresholding is negligible. If the result is inconclusive after combining three searches, instead of giving up and claiming that the excerpt does not belong to the database, one may perform more searches and combine all the results. For example, in the case of FIG. 4(b), one may perform two more searches. If the excerpt really belongs to piece A, but the second search somehow returns the wrong time location, then with two more searches it is very likely one of these two searches will also return piece A with the right time location. However, when combining five searches, the definition of a voting scheme becomes more complicated. In an example of a simplified implementation, one may define that the two searches (out of five) returning the same piece and that are most consistent in terms of time locations as the winner, in the manner that the "min” and "argmin” operators are used in the pseudocode above.
  • fingerprint search accuracy may be improved by combining results from multiple searches.
  • results are intelligently combined.
  • a type of voting based on available information in the search results is employed to improve fingerprint search accuracy.
  • the main available information is the (supposed) title of the searched piece (the query excerpt).
  • the accessory available information which can substantially improve the accuracy in combining the results, includes parameters such as Wset (as described above) or t d i sta nc e and a measure of distance, such as BER.
  • the parameter trance represents the linearity of time locations between the query piece and the found piece(s): the closer the time locations are to a straight line, obtainable by linear regression, the more confident one should be about the match in the results.
  • the t d i stance parameter is described further below in connection with FIG. 6.
  • the BER derived from Hamming distance, indicates how different the found piece is from the query piece. The higher the BER is, the more different they are from each other, and the lower the BER is, the less different and more resembling they are with respect to each other. All else being equal, should be more confident of a match with a lower BER than a match with a higher BER.
  • the values of accessory information exhibit a degree of confidence in their corresponding search results.
  • One approach to intelligently combine them, outlined here, is to use confidence levels of individual values of accessory information as weights during the combination of results.
  • FIG. 5 shows an example of confidence level as a function of BER. One may refer to such function as ⁇ (BER).
  • a fingerprint may be viewed as a vector of numbers.
  • the measure of difference may be an Lp norm measure in which a p-th root of the sum of the p-th power of absolute values of distance over each dimension of a vector that constitutes a fingerprint is obtained.
  • the BER is based on Hamming distance, which is a Ll norm difference measure (sum of absolute values of difference at each dimension, where each dimension is 1 bit). There could also be L2 no ⁇ n difference measure (where a square root of sum of square of absolute values of difference is obtained), although L2 norm is similar to Ll norm when each dimension is 1 bit. There could be also be a difference measure by auto-correlation, where a higher value denotes closer match.
  • time que ry is the time location of the query fingerprint
  • time found is the time location of a potentially matching fingerprint in the found piece
  • the pair of these two values is shown as a diamond shape in FIG. 6.
  • Linear regression is then performed on these diamond points.
  • the deviation of each diamond point from its resulting linear regression prediction (the straight line) is referred to as the time location distance or trance, shown as vertical lines in FIG. 6.
  • time location distance or trance shown as vertical lines in FIG. 6.
  • FIG. 7 we may refer to such function as/(t d i Stance )- Any suitable type of linear regression may be employed. Any of various techniques for removing outliers may be employed prior to performing linear regression or as part of the linear regression process.
  • the t d j stance parameter is a more generalized form of the t o ff S et parameter.
  • An example using the t o n set parameter is shown in FIGS. 4a and 4b. It measures the offset between a time location in the query excerpt and the corresponding time location in a match. If consecutive fingerprint matches are correct it should roughly be a constant, but not necessarily close to zero (because of the uncertain absolute time of the query material) value when there is no speed-up or speed-down.
  • the parameter td, stance is more general, as shown in FIG. 6.
  • Exemplary pseudocode for combining confidence levels and determining the outcome of search results may be expressed as:
  • the combining is simply a summation operation, although one may use other operations such as multiplication (the confidence levels would then have to be non-negative), as long as the combining operation is a monotonic non-decreasing function with respect to each of the confidence levels. If/s() always returns 1 , and/() always returns 0, and the variable "threshold" has value N/2 where N is the number of searches conducted, then the procedure expressed in the pseudocode is reduced to a simple majority voting scheme. In practice, one should choose the such that they maximize the probability that the final output is correct.
  • the invention may be implemented in hardware or software, or a combination of both (e.g., programmable logic arrays). Unless otherwise specified, the methods or algorithms included as part of the invention are not inherently related to any particular computer or other apparatus. In particular, various general -purpose machines may be used with programs written in accordance with the teachings herein, or it may be more convenient to construct more specialized apparatus (e.g., integrated circuits) to perform the required method steps. Thus, the invention may be implemented in one or more computer programs executing on one or more programmable computer systems each comprising at least one processor, at least one data storage system (including volatile and non-volatile memory and/or storage elements), at least one input device or port, and at least one output device or port.
  • programmable computer systems each comprising at least one processor, at least one data storage system (including volatile and non-volatile memory and/or storage elements), at least one input device or port, and at least one output device or port.
  • Program code is applied to input data to perform the functions described herein and generate output info ⁇ nation.
  • the output information is applied to one or more output devices, in known fashion.
  • Each such program may be implemented in any desired computer language (including machine, assembly, or high level procedural, logical, or object oriented programming languages) to communicate with a computer system. In any case, the language may be a compiled or interpreted language.
  • Each such computer program is preferably stored on or downloaded to a storage media or device (e.g., solid state memory or media, or magnetic or optical media) readable by a general or special purpose programmable computer, for configuring and operating the computer when the storage media or device is read by the computer system to perform the procedures described herein.
  • the inventive system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer system to operate in a specific and predefined manner to perform the functions described herein.

Abstract

A method for use in identifying a segment of audio and/or video information comprises obtaining a query fingerprint at each of a plurality of spaced-apart time locations in said segment, searching fingerprints in a database for a potential match for each such query fingerprint, obtaining a confidence level of a potential match to a found fingerprint in the database for each such query fingerprint, and combining the results of searching for potential matches, wherein each potential match result is weighted by a respective confidence level. A confidence level may be a function of at least one or both of (1) a measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time location of a found fingerprint.

Description

Description
Improving Audio/Video Fingerprint Search Accuracy Using Multiple Search Combining
Technical Field
The invention relates to audio and video fingerprinting. More particularly the invention relates to improving the accuracy of fingerprint search accuracy.
Background A rt
With the rapid growth of digital content, there is an increasing demand for properly managing and locating the content. One prospective application is identifying an excerpt of audio or video within a repository of known content. This can be useful in monitoring illegal download/transfer of content on the Internet. It can also be useful in providing intelligent access to content that users are listening to or watching and are interested in, but for which they do not know the title or author information. The user can, for example, place a cell phone against the loudspeaker currently playing a song, and let the cell phone operator's software find out the song's title, its artist and album information, etc.
Such identification capability is generally implemented by first generating feature information (called fingerprints) designed to identify uniquely audio and video signals, and then performing some form of pattern matching search between the fingerprints from the repository/database and those from the excerpt in the search query. Such a database is often implemented in the form of a search tree, although other data structures are also possible. Generally, each fingerprint corresponds to a certain segment of audio or video. So a two second audio fingerprint would correspond to a two second segment of audio. A fingerprint is" usually implemented as a concatenation of small blocks of feature information that are often called signatures. A two second fingerprint may, for example, consist of two hundred 10 millisecond (ms) long signatures, wherein each signature is computed from 10 ms of new audio or video information. The pattern matching of fingerprints is therefore a process of comparing corresponding signatures. This is illustrated in FIG. 1 that shows the creation of a fingerprint from signatures.
In order to perform proper pattern matching between a fingerprint from a query excerpt and those from the database, it is crucial to have proper time alignment between the two during a comparison. To ensure that, usually all fingerprints starting at every possible time offset are added to the database to guarantee that at least one of them will have time alignment that is close enough to the query fingerprint. If a signature is 10 ms long, then a two second fingerprint is shifted every 10 ms over a two second sliding window and then added to the database. This is also illustrated in FIG. 1 , and in this case it creates a 99.5% overlap between successive fingerprints, but such redundancy is usually required to ensure good search performance. For any remaining time misalignment less than 10 ms (or in general the duration of each signature), a well- designed fingerprint generation method should select whichever signature is closer in terms of timing to better match the corresponding signature from the query. In short, the goal of a fingerprint search system is to find a query fingerprint's corresponding counterpart fingerprint, and the counterpart should minimize its time misalignment with the query fingerprint, if such misalignment exists at all.
Because the query excerpt may have undergone some editing or processing steps, such as recapturing sound played from a loudspeaker using a cell phone, there may be some distortions in the captured audio/video signals. As a result, the resulting fingerprints may also change slightly with respect to their counterparts in the database, assuming there is a count erpart.
The possibility of distortions in the excerpt means the best match in such a search is often not an identical match, but a form of closest match. To define the closest match requires the definition of a measure of difference between two fingerprints. For example, a commonly used measure of difference is the Hamming distance, that is, the number of differing bits between the fingerprint from the query excerpt and that from the database. With this definition of measure of difference, the corresponding criterion for the closest match is thus the fingerprint from the database that has the minimum Hamming distance from the fingerprint from the query excerpt. The Hamming distance between two fingerprints divided by the number of bits in a fingerprint is often referred to as the bit error rate (BER). The BER is an example of measure of relative difference. The minimum Hamming distance criterion works well when the BER between the fingerprint from the excerpt and its counterpart is small. However, as BER increases, the search result producing the minimum Hamming distance increasingly does not find the actual counterpart. Fortunately, in most fingerprint search applications, it is only necessary to identify the correct audio/video piece, but not necessarily the corresponding segment. But when BER increases further, the search result may even find the wrong audio/video piece, let alone the correct segment within that piece. The BER depends on both the level of distortion in the query exceφt, and the robustness of the fingerprint extraction method with respect to such distortions.
Furthermore, it is possible that an exceφt does not belong to any piece in the database. For example, the excerpt may be the recording of a new composition of music. Because no search algorithm can know beforehand (without being told) whether an exceφt belongs to the database or not, the best it can do is still apply the same criterion of minimum Hamming distance, but expecting that the minimum Hamming distance found in such cases will be much more different (preferably higher) than that of an exceφt originally from the database, and use some threshold to determine whether the exceφt is from the database.
Therefore, there are three possible outcomes after a single search operation (wherein only one query fingeφrint is used to search the database), before applying any thresholds (for example in terms of BER):
1. The exceφt belongs to the database, and the search returns the correct audio/video piece (finding the correct piece is enough, here it is not necessary to find the correct counteφart segment).
2. The exceφt belongs to the database, and the search returns the wrong audio/video piece.
3. The exceφt does not belong to the database, and because the search always returns some audio/video piece, the answer will always be wrong.
FIG. 2 shows an example of BER distribution for three different possible outcomes of a single search. Each of these outcomes would generate a corresponding probability-density-function (PDF) distribution of the BER. For a well-designed fingeφrint extraction algorithm, the BER of the first outcome should generally be smaller than the BER of the second and third outcomes, as illustrated in FIG. 2.
However, if the BER of the second and third outcomes have very similar PDF distributions, it would be difficult to distinguish between an exceφt belonging to the database but having a wrong search result, and an exceφt that doesn't belong there. Furthermore, for pieces originally from the database, after applying common audio/video distortions such as codec compression, the search result in a typical implementation is correct (in terms of identifying the correct piece) usually ranges from 90 to 99%, depending on the fingeφrint duration and the type of distortions, before applying any BER threshold. This is good but a higher level of accuracy is certainly desirable, and after applying BER thresholding (say at BER = 0.2 in FIG. 2), the ratio of correct search results only decreases slightly because the tail of the BER distribution of outcome one is discarded to avoid falsely picking up too much of the head distribution of outcome two. This means tweaking the BER threshold alone cannot lead to very high accuracy (say 99.9%) in a single search.
Description of the Drawings
FIG. 1 is a schematic example of the manner in which a particular type of fingerprint may be generated.
FIG. 2 shows an idealized example of a probability-density- function (PDF) as a function of the bit error rate (BER) that may result for each of three possible outcomes of a single fingerprint matching search.
FIG. 3 is a schematic conceptualization of an example of fingerprint-matching search results combining using majority voting with respect to three searches.
FIG. 4a is a schematic conceptualization of an example of fingerprint-matching search results with respect to three searches in which time locations are consistent for two of three potential matches.
FIG. 4b is a schematic conceptualization of an example of fingerprint-matching search results with respect to three searches in which time locations are not consistent for two of three potential matches.
FIG. 5 is an example of the manner in which a confidence level may be mapped as a function of the bit error rate (BER).
FIG. 6 is an idealized example useful in understanding one way, using linear regression in this example, that relative time locations between the time of the fingerprint query and the time of the fingerprint found in a database may be determined.
FIG. 7 is an example of the manner in which a confidence level may be mapped as a function of time difference in relative time locations between the time of the fingerprint query and the time of the fingerprint found in a database.
Disclosure of the Invention
In accordance with aspects of the present invention, a method for use in identifying a segment of audio and/or video information comprises obtaining a query fingerprint at each of a plurality of spaced-apart time locations in said segment, searching fingerprints in a database for a potential match for each such query fingerprint, obtaining a confidence level of a potential match to a found fingerprint in the database for each such query fingerprint, and combining the results of searching for potential matches, wherein each potential match result is weighted by a respective confidence level. A confidence level may be a function of at least one or both of (1) a measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time location of a found fingerprint. The measure of difference may be based on Hamming distance, or its relative difference variant such as the bit error rate, or it may be based on an Lp norm in which a p-th root of the sum of the p-th power of absolute values of difference over each dimension of a vector that constitutes a fingerprint is obtained. The Lp norm measure may be an L2 norm difference measure in which a square root of the sum of the square of absolute values of difference is obtained. As a further alternative, the measure of difference may be obtained by auto-correlation in which the value of correlation indicates the closeness of the match.
The confidence level may be a combination of respective confidence levels of each of a function of at least one or both of (1) the measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time locations of a found fingerprint. The combination of the respective confidence levels may be, for example, a summation of respective confidence levels or a multiplicative product of respective confidence levels. A confidence level may be a function of at least the relative timing relationship between the time location of the query fingerprint and the time location of the potentially matching found fingerprint in the database, in which case there are at least three spaced- apart time locations.
The spaced-apart time locations in the segment of audio and/or video information may be spaced apart so as to minimize correlation between or among query fingerprints. There may be multiple sets of spaced-apart time locations with respect to each of which sets the results of searching for potential matches is combined. The results of searching may be combined within each set and the results of such combinations then may be further combined. Alternatively, the results of searching may be combined across all sets. Fingerprints in the database may have related metadata about said audio and/or video information. Such metadata may include at least a piece title of the audio and/or video information.
A query fingerprint may be obtained at each of said plurality of spaced-apart time locations in substantially the same way.
A simple but useful method of combining results from multiple searches is majority voting. For example, as illustrated in FIG. 3, if one conducts three searches, two of them return piece A, one returns piece B, then we conclude that the final search result should be piece A. If we get three different answers, piece A, B and C, then there is no majority winner, and the answer is inconclusive. In this case, it is best to conclude that the excerpt does not belong to the database, because if it did belong, it is relatively unlikely three searches will give three different results.
This can be generalized to any number of searches, using majority voting (wherein the piece getting more than half of the votes wins) as illustrated by the pseudocode shown below, or any voting scheme desired by the implementer (for example the piece with the most votes and without a tie wins). A voting scheme should be designed to maximize the number of correct answers and minimize the number of wrong answers, using probabilistic analysis. For example, for pieces originally from the database, if a single search has 98% accuracy, then combining three searches with majority voting will draw the wrong conclusion only if at least two out of three searches are wrong, which is 3xO.98xO.O22 + 0.023 « 1.2xlO~3, so the combined accuracy should be almost 99.9%. This probabilistic analysis assumes each search to be independent of other searches, i.e., no correlation between searches.
Exemplary pseudocode for search combining using majority voting may be expressed as follows:
N = 3;
Do N fingerprint searches, at time t, t+d[l], ..., t+d[N-l] of the excerpt, respectively;
If (> N/2 searches return the same piece A)
Return piece A; Else Return "exceφt does not belong to the database";
The time intervals d[l ], d[2], ..., in FIG. 3 are used to conduct searches at different locations, and, in general, they should be chosen to minimize the overlaps among their corresponding search segments in order to reduce any undesirable correlation between search results. The combining process can be further cascaded by additional iteration(s) of multiple search(es) in the event of inconclusive votes, and either consider the votes in the new iteration only, or combine some or all of the previous votes with a voting scheme chosen by the implementer. It would seem that a simple scheme like majority voting can give very good results. The example above suggests an accuracy of almost 99.9%, but the actual accuracy may be only marginally better than the original 98% when there is strong burstiness or correlation of errors in the search: if one search gives the wrong result, a search at a nearby time location will likely give the wrong result as well, and often the same wrong result. This can happen for example due to certain regions of audio/video (such as high-motion segments in video or quiet audio signals) being more sensitive to distortion (low bit-rate compression or background noise, respectively). This would mislead a majority decision algorithm into thinking it found the right answer, when it has not. In addition, if a piece does not belong to the database, one must also deal with the case where more than one search may return the same piece, since it would result in a wrong majority winner.
Accuracy may be increased, even in the presence of such burst errors, by employing aspects of the present invention. This may be accomplished, for example, by combining not only the piece's title information from each search result, but also accessory information such as the time location of the found segment within the found piece, and the BER of the search result. For example in FIG. 4(a), if two of three searches return piece A, and the time locations of these two search segments differ by five seconds, then the time locations of the found segments in these two search results should also differ by five seconds (or within a certain error margin). If this is not satisfied, as in FIG. 4(b), it is almost certain that this is a false majority winner. This is a fair assumption because if a search returns the wrong piece, it is almost certainly due to accident that it happens to be the result producing the minimum Hamming distance. Therefore, the time location of the found segment should be relatively random, rather than deterministic. And consequently, the probability that two wrong search results returning the same wrong piece and a pair of consistently close time locations is fairly low. The exact probability can be estimated by analyzing the PDF distribution of the time locations of both correct and wrong search results.
To generalize this comparison with N searches, it means the differences between the time locations of those searches should also be the same or close across the queries and the search results. However, to make this comparison easier to implement in software, one can calculate the offset (i.e., difference) between the time locations of the original piece and the excerpt in a search result, and this value is referred to as tofiset in FIGS. 4(a) and 4(b). The offset for the first search is tOιτset,ι, for the second search is tOset,2, respectively. For each pair of search results that return the same piece, one may calculate the difference between their toπSet- Then, for the pair that achieves the minimum difference, if such minimum difference is less than a certain threshold (say 0.2sec), its piece title may then be considered the final answer. This is illustrated in the following pseudocode for search combining using accessory information, where Min_t_offset_dist is the minimum difference, and Candidate indicates for which (pair of) search this minimum is achieved. A threshold d_A with an example value of 0.2 seconds is used to determine whether the result is trustable enough.
Exemplary pseudocode for search combining using t0ffSet accessory information may be expressed as follows:
N - 3;
Do N fingerprint searches, at time t, t+d[l], ..., t+d[N-l] of the excerpt, respectively;
Min_t_offset_dist = m\n(tφel>l - tφel J , where search[i].piece == search[j].piece, i < j)
Candidate = argmin(^e/ , - tφel y , where search[i].piece -= searcfι[j].piece,i < j)
d_A = 0.2; If (Min_t_offset_dist < d_A && Candidate has > N/2 votes)
Return search[Candidate].piece; Else
Return "excerpt does not belong to the database"; In the above pseudocode, "search[i]" means the result from the i-th search and "search[i]. piece" is the found piece's identification information such as its title. A more generalized way to use accessory information timing information, described below, employs a time measure tdiStance rather than tOfl-set.
For a large N number of searches, clearly the value of Min t offset dist would be smaller if more of the searches return the same piece, provided that the result is correct. For instance, all else being equal, the minimum value among five samples is certainly smaller than the minimum value among three samples. So, we can define more thresholds, e.g., d_B, d_C, in addition to d_A, depending on how many searches return the same piece. This way, one can use the accessory information (Wtset in this example) intelligently based on the majority voting information (how many searches return the same piece), and achieve higher accuracy.
If the query material has experienced some speed-up or speed-down, which is sometimes the case in radio programs, the spacing between the query fingerprints versus, the spacing between the fingerprints from the search results will no longer be the same. Instead they will drift apart slowly with respect to the spacing between the query fingerprints. To deal with this, one can either make sure that the spacing between the query fingerprints will never cause the deviation of toset to exceed a threshold such as d_A. Or, if one has a sufficient number of votes, one can observe the timing correlation between the query fingerprints and the fingerprints from the search results, which should produce a linear trend with a slope slight higher or lower than one depending on whether it experienced speed-up or speed-down, assuming no other distortions would contribute to errors in the timing of fingerprints from the search results. Then, we can use a measure of confidence of the linear trend (such as based on linear regression) to determine the confidence of the combined result. An example of such a linear regression is shown in FIG. 6 discussed below.
Similar to the concept of Wfset, if the BER in a search result is very low, which indicates a very close match, one can be highly confident in the accuracy of the result. However, if the BER is high, then the result is less trustable. Furthermore, when more than one piece of accessory information is used in a logical AND manner, one is not constrained to use only fixed thresholds for them. For example, if the threshold d_A is initially 0.2 seconds, and the BER threshold is 12%, but the BER of a current search is extremely low, such as 0.1%, then one can relax the threshold d_A from 0.2sec to say 0.3sec with a strong confidence. This may be referred to as dynamic thresholding. The exact amount of adjustment in such dynamic thresholding can be determined by the implementer, using probabilistic analysis to ensure that the number of false matches resulting from such dynamic thresholding is negligible. If the result is inconclusive after combining three searches, instead of giving up and claiming that the excerpt does not belong to the database, one may perform more searches and combine all the results. For example, in the case of FIG. 4(b), one may perform two more searches. If the excerpt really belongs to piece A, but the second search somehow returns the wrong time location, then with two more searches it is very likely one of these two searches will also return piece A with the right time location. However, when combining five searches, the definition of a voting scheme becomes more complicated. In an example of a simplified implementation, one may define that the two searches (out of five) returning the same piece and that are most consistent in terms of time locations as the winner, in the manner that the "min" and "argmin" operators are used in the pseudocode above.
Thus, in accordance with aspects of the present invention, fingerprint search accuracy may be improved by combining results from multiple searches. Preferably, in accordance with other aspects of the present invention, such results are intelligently combined. In the aspects of the present invention described above, a type of voting based on available information in the search results is employed to improve fingerprint search accuracy. The main available information is the (supposed) title of the searched piece (the query excerpt). The accessory available information, which can substantially improve the accuracy in combining the results, includes parameters such as Wset (as described above) or tdistance and a measure of distance, such as BER. The parameter trance represents the linearity of time locations between the query piece and the found piece(s): the closer the time locations are to a straight line, obtainable by linear regression, the more confident one should be about the match in the results. The tdistance parameter is described further below in connection with FIG. 6. Similarly, the BER, derived from Hamming distance, indicates how different the found piece is from the query piece. The higher the BER is, the more different they are from each other, and the lower the BER is, the less different and more resembling they are with respect to each other. All else being equal, should be more confident of a match with a lower BER than a match with a higher BER. Therefore, the values of accessory information exhibit a degree of confidence in their corresponding search results. One approach to intelligently combine them, outlined here, is to use confidence levels of individual values of accessory information as weights during the combination of results. FIG. 5 shows an example of confidence level as a function of BER. One may refer to such function as ^(BER).
Although in the previous example BER is mapped to a confidence level, such a measure of difference is not critical to the invention and other forms of difference measure may also be used. A fingerprint may be viewed as a vector of numbers. In the case of Hamming distance and BER, it is viewed as a vector of binary numbers, i.e., a binary vector. Each dimension of such a vector is just one bit, but one may envision a fingerprint vector of numbers other than binary, and the size of each dimension may even be different across different dimensions, although it should be predictable. Thus, more generally, the measure of difference may be an Lp norm measure in which a p-th root of the sum of the p-th power of absolute values of distance over each dimension of a vector that constitutes a fingerprint is obtained. The BER is based on Hamming distance, which is a Ll norm difference measure (sum of absolute values of difference at each dimension, where each dimension is 1 bit). There could also be L2 noπn difference measure (where a square root of sum of square of absolute values of difference is obtained), although L2 norm is similar to Ll norm when each dimension is 1 bit. There could be also be a difference measure by auto-correlation, where a higher value denotes closer match.
Similarly, one can measure how far the time location of a potential match to a query fingerprint in a database is away from its supposed location as predicted, for example, by linear regression. This is shown in FIG. 6, where timequery is the time location of the query fingerprint, and timefound is the time location of a potentially matching fingerprint in the found piece, and the pair of these two values is shown as a diamond shape in FIG. 6. Linear regression is then performed on these diamond points. The deviation of each diamond point from its resulting linear regression prediction (the straight line) is referred to as the time location distance or trance, shown as vertical lines in FIG. 6. Then one can map this distance to a confidence level, as shown in FIG. 7. We may refer to such function as/(tdiStance)- Any suitable type of linear regression may be employed. Any of various techniques for removing outliers may be employed prior to performing linear regression or as part of the linear regression process.
As mentioned above, the tdjstance parameter is a more generalized form of the toffSet parameter. An example using the tonset parameter is shown in FIGS. 4a and 4b. It measures the offset between a time location in the query excerpt and the corresponding time location in a match. If consecutive fingerprint matches are correct it should roughly be a constant, but not necessarily close to zero (because of the uncertain absolute time of the query material) value when there is no speed-up or speed-down. The parameter td,stance is more general, as shown in FIG. 6. It measures the vertical distance between a sample point (whose X and Y coordinates are the time locations in the query excerpt and the match, respectively) and the straight line obtained by any of various suitable forms of linear regression. The td,stance parameter is supposed to be closed to zero if it corresponds to a good match. The notion of trance works even when there is a speed-up or speed- down, because it uses linear regression and it does not assume the slope of linear regression to be one, as does the tOιrset parameter. It will also be noted that because the toffset parameter assumes a slope of one, it is possible to check the consistency of t0flset for only two time locations. However, when using trance, because one does not assume the slope to be one, one needs at least three time locations to obtain a non-trivial linear regression (although one may, if desired, use at least two time locations to obtain a linear regression, and then discard any slope that is too far from slope of one).
Note that in FIG. 7, the confidence level drops more rapidly than in FIG. 5. This is because we expect that in typical implementations, the tolerance for BER may be relatively high, whereas any non-trivial amount of td,stance is an indication that the result is much less reliable. In addition, since two points always produce a line that passes through them perfectly, the use of confidence level based on td,stance should only be considered where there are three or more points.
An example of a procedure, in accordance with aspects of the present invention, for combining confidence levels and determining the outcome of search results intelligently is shown in the following pseudocode.
Exemplary pseudocode for combining confidence levels and determining the outcome of search results may be expressed as:
For ( X in set of unique pieces in the results) { For (i in where search [i]. piece == X) {
Confidencel =y#(search[i].BER); If (the set where results[i].piece == X has >=3 points)
Confidence2 =/(search[i].t_distance); Else Confidence2 = 0; Total_score[X] += Confidence 1 + Confidence!;
}
Pick X for which Total_score[X] is the highest; If (Total_score[X] > threshold)
Return "the query excerpt is piece X"; Else
Return "unknown query excerpt";
As indicated in the pseudocode, first map the BER and trance to their corresponding confidence levels, and then combine them. In this example, the combining is simply a summation operation, although one may use other operations such as multiplication (the confidence levels would then have to be non-negative), as long as the combining operation is a monotonic non-decreasing function with respect to each of the confidence levels. If/s() always returns 1 , and/() always returns 0, and the variable "threshold" has value N/2 where N is the number of searches conducted, then the procedure expressed in the pseudocode is reduced to a simple majority voting scheme. In practice, one should choose the
Figure imgf000014_0001
such that they maximize the probability that the final output is correct.
Implementation The invention may be implemented in hardware or software, or a combination of both (e.g., programmable logic arrays). Unless otherwise specified, the methods or algorithms included as part of the invention are not inherently related to any particular computer or other apparatus. In particular, various general -purpose machines may be used with programs written in accordance with the teachings herein, or it may be more convenient to construct more specialized apparatus (e.g., integrated circuits) to perform the required method steps. Thus, the invention may be implemented in one or more computer programs executing on one or more programmable computer systems each comprising at least one processor, at least one data storage system (including volatile and non-volatile memory and/or storage elements), at least one input device or port, and at least one output device or port. Program code is applied to input data to perform the functions described herein and generate output infoπnation. The output information is applied to one or more output devices, in known fashion. Each such program may be implemented in any desired computer language (including machine, assembly, or high level procedural, logical, or object oriented programming languages) to communicate with a computer system. In any case, the language may be a compiled or interpreted language. Each such computer program is preferably stored on or downloaded to a storage media or device (e.g., solid state memory or media, or magnetic or optical media) readable by a general or special purpose programmable computer, for configuring and operating the computer when the storage media or device is read by the computer system to perform the procedures described herein. The inventive system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer system to operate in a specific and predefined manner to perform the functions described herein.
A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, some of the steps described above may be order independent, and thus can be performed in an order different from that described. Accordingly, other embodiments are within the scope of the following claims.

Claims

Claims
1. A method for use in identifying a segment of audio and/or video information, comprising obtaining a query fingerprint at each of a plurality of spaced-apart time locations in said segment, searching fingerprints in a database for a potential match for each such query fingerprint, obtaining a confidence level of a potential match to a found fingerprint in the database for each such query fingerprint, and combining the results of searching for potential matches, wherein each potential match result is weighted by a respective confidence level.
2. A method according to claim 1 wherein a confidence level is a function of at least one or both of (1) a measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time location of a found fingerprint.
3. A method according to claim 2 wherein the measure of difference is based on Hamming distance.
4. A method according to claim 3 wherein the measure of difference is the bit error rate.
5. A method according to claim 2 wherein the measure of difference is an Lp norm in which a p-th root of the sum of the p-th power of absolute values of distance over each dimension of a vector that constitutes a fingerprint is obtained.
6. A method according to claim 5 wherein the measure of difference is an L2 norm distance measure in which a square root of the sum of the square of absolute values of distance is obtained.
7. A method according to claim 2 wherein the measure of difference is obtained by auto-correlation in which the value of correlation indicates the closeness of the match.
8. A method according to any one of claims 2 through 7 wherein said confidence level is a combination of respective confidence levels of each of a function of at least one or both of (1) the measure of difference between a query fingerprint and a found fingerprint and (2) the relative timing relationship between the time location of a query fingerprint and the time locations of a found fingerprint.
9. A method according to claim 8 wherein said combination of the respective confidence levels is a summation of respective confidence levels.
10. A method according to claim 8 wherein said combination of the respective confidence levels is a multiplicative product of respective confidence levels.
11. A method according to any one of claims 1-10 wherein said time locations are spaced apart so as to minimize correlation of the search results between or among query fingerprints.
12. A method according to any one of claims 1-1 1 wherein there are multiple sets of spaced-apart time locations with respect to each of which sets the results of searching for potential matches is combined.
13. A method according to claim 12 wherein the results of searching are combined within each set and the results of such combinations are further combined.
14. A method according to claim 12 wherein the results of searching are combined across all sets.
15. A method according to any one of claim 1-14 wherein fingerprints in the database have related metadata about said audio and/or video information.
16. A method according to claim 15 wherein said metadata includes at least a piece title of the audio and/or video information.
17. A method according to any one of claims 1 through 16 wherein a query fingerprint is obtained at each of said plurality of spaced-apart time locations in substantially the same way.
18. A method according to any one of claim 2-17 wherein a confidence level is a function of at least the relative timing relationship between the time location of the query fingerprint and the time location of the potentially matching found fingerprint in the database, and there are at least three spaced-apart time locations.
19. Apparatus adapted to perform the methods of any one of claims 1 through 18.
20. A computer program, stored on a computer-readable medium for causing a computer to perform the methods of any one of claims 1 through 18.
PCT/US2008/007054 2007-06-06 2008-06-04 Improving audio/video fingerprint search accuracy using multiple search combining WO2008150544A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US12/663,057 US8266142B2 (en) 2007-06-06 2008-06-04 Audio/Video fingerprint search accuracy using multiple search combining
EP08794341A EP2168061A1 (en) 2007-06-06 2008-06-04 Improving audio/video fingerprint search accuracy using multiple search combining
CN2008800188850A CN101681381B (en) 2007-06-06 2008-06-04 Improving audio/video fingerprint search accuracy using multiple search combining
JP2010511189A JP5090523B2 (en) 2007-06-06 2008-06-04 Method and apparatus for improving audio / video fingerprint search accuracy using a combination of multiple searches

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US93361407P 2007-06-06 2007-06-06
US60/933,614 2007-06-06

Publications (1)

Publication Number Publication Date
WO2008150544A1 true WO2008150544A1 (en) 2008-12-11

Family

ID=39764983

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2008/007054 WO2008150544A1 (en) 2007-06-06 2008-06-04 Improving audio/video fingerprint search accuracy using multiple search combining

Country Status (6)

Country Link
US (1) US8266142B2 (en)
EP (1) EP2168061A1 (en)
JP (1) JP5090523B2 (en)
CN (1) CN101681381B (en)
TW (1) TWI447601B (en)
WO (1) WO2008150544A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011059504A (en) * 2009-09-11 2011-03-24 Victor Co Of Japan Ltd Content identification method, content identification system, content retrieval device, and content use device
WO2012066577A1 (en) * 2010-11-18 2012-05-24 Sling Media Pvt Ltd Automated searching for solutions to support self-diagnostic operations of web-enabled devices

Families Citing this family (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100215210A1 (en) * 2008-05-21 2010-08-26 Ji Zhang Method for Facilitating the Archiving of Video Content
US20100215211A1 (en) * 2008-05-21 2010-08-26 Ji Zhang System for Facilitating the Archiving of Video Content
US8611701B2 (en) * 2008-05-21 2013-12-17 Yuvad Technologies Co., Ltd. System for facilitating the search of video content
US8548192B2 (en) * 2008-05-22 2013-10-01 Yuvad Technologies Co., Ltd. Method for extracting a fingerprint data from video/audio signals
WO2009140824A1 (en) * 2008-05-22 2009-11-26 Yuvad Technologies Co., Ltd. A system for identifying motion video/audio content
US20100169911A1 (en) * 2008-05-26 2010-07-01 Ji Zhang System for Automatically Monitoring Viewing Activities of Television Signals
US8195689B2 (en) 2009-06-10 2012-06-05 Zeitera, Llc Media fingerprinting and identification system
US8335786B2 (en) * 2009-05-28 2012-12-18 Zeitera, Llc Multi-media content identification using multi-level content signature correlation and fast similarity search
CN102216952B (en) * 2008-11-17 2013-06-05 杜比实验室特许公司 Media fingerprints that reliably correspond to media content with projection of moment invariants
US20100138411A1 (en) * 2008-11-30 2010-06-03 Nexidia Inc. Segmented Query Word Spotting
CN102460470B (en) 2009-06-11 2014-12-03 杜比实验室特许公司 Trend analysis in content identification based on fingerprinting
US8644622B2 (en) * 2009-07-30 2014-02-04 Xerox Corporation Compact signature for unordered vector sets with application to image retrieval
US8677400B2 (en) 2009-09-30 2014-03-18 United Video Properties, Inc. Systems and methods for identifying audio content using an interactive media guidance application
US8161071B2 (en) 2009-09-30 2012-04-17 United Video Properties, Inc. Systems and methods for audio asset storage and management
US8428955B2 (en) * 2009-10-13 2013-04-23 Rovi Technologies Corporation Adjusting recorder timing
US20110085781A1 (en) * 2009-10-13 2011-04-14 Rovi Technologies Corporation Content recorder timing alignment
US8682145B2 (en) 2009-12-04 2014-03-25 Tivo Inc. Recording system based on multimedia content fingerprints
US8606293B2 (en) 2010-10-05 2013-12-10 Qualcomm Incorporated Mobile device location estimation using environmental information
US8483725B2 (en) 2010-12-03 2013-07-09 Qualcomm Incorporated Method and apparatus for determining location of mobile device
EP2659481B1 (en) 2010-12-30 2016-07-27 Dolby Laboratories Licensing Corporation Scene change detection around a set of seed points in media data
CN102622353B (en) * 2011-01-27 2013-10-16 天脉聚源(北京)传媒科技有限公司 Fixed audio retrieval method
US9143571B2 (en) 2011-03-04 2015-09-22 Qualcomm Incorporated Method and apparatus for identifying mobile devices in similar sound environment
US8706499B2 (en) * 2011-08-16 2014-04-22 Facebook, Inc. Periodic ambient waveform analysis for enhanced social functions
US9113202B1 (en) * 2011-09-21 2015-08-18 Google Inc. Inverted client-side fingerprinting and matching
US8433577B2 (en) 2011-09-27 2013-04-30 Google Inc. Detection of creative works on broadcast media
US8892572B2 (en) * 2011-12-30 2014-11-18 Cellco Partnership Video search system and method of use
US8681950B2 (en) 2012-03-28 2014-03-25 Interactive Intelligence, Inc. System and method for fingerprinting datasets
US9202255B2 (en) * 2012-04-18 2015-12-01 Dolby Laboratories Licensing Corporation Identifying multimedia objects based on multimedia fingerprint
US9553756B2 (en) * 2012-06-01 2017-01-24 Koninklijke Kpn N.V. Fingerprint-based inter-destination media synchronization
US8959022B2 (en) * 2012-07-03 2015-02-17 Motorola Solutions, Inc. System for media correlation based on latent evidences of audio
CN103575323B (en) * 2012-07-30 2016-03-30 日电(中国)有限公司 Take detection method and device
CN103021440B (en) * 2012-11-22 2015-04-22 腾讯科技(深圳)有限公司 Method and system for tracking audio streaming media
US9529907B2 (en) * 2012-12-31 2016-12-27 Google Inc. Hold back and real time ranking of results in a streaming matching system
US20140280304A1 (en) * 2013-03-15 2014-09-18 Steven D. Scherf Matching versions of a known song to an unknown song
US9659014B1 (en) * 2013-05-01 2017-05-23 Google Inc. Audio and video matching using a hybrid of fingerprinting and content based classification
CN103440330A (en) * 2013-09-03 2013-12-11 网易(杭州)网络有限公司 Music program information acquisition method and equipment
TWI527025B (en) 2013-11-11 2016-03-21 財團法人資訊工業策進會 Computer system, audio matching method, and computer-readable recording medium thereof
GB2523311B (en) * 2014-02-17 2021-07-14 Grass Valley Ltd Method and apparatus for managing audio visual, audio or visual content
CN104991946B (en) * 2015-07-13 2021-04-13 联想(北京)有限公司 Information processing method, server and user equipment
US9836535B2 (en) * 2015-08-25 2017-12-05 TCL Research America Inc. Method and system for content retrieval based on rate-coverage optimization
CN105138886B (en) * 2015-08-26 2017-03-22 江苏久祥汽车电器集团有限公司 Robot biometric identification system
US10606879B1 (en) * 2016-02-29 2020-03-31 Gracenote, Inc. Indexing fingerprints
CN106910494B (en) 2016-06-28 2020-11-13 创新先进技术有限公司 Audio identification method and device
TWI612516B (en) * 2016-08-25 2018-01-21 財團法人資訊工業策進會 Audio fingerprint recognition apparatus, audio fingerprint recognition method and computer program product thereof
CN107515931B (en) * 2017-08-28 2023-04-25 华中科技大学 Repeated data detection method based on clustering
US11687801B2 (en) * 2018-04-23 2023-06-27 Qliktech International Ab Knowledge graph data structures and uses thereof
US10832692B1 (en) * 2018-07-30 2020-11-10 Amazon Technologies, Inc. Machine learning system for matching groups of related media files
CN111008301B (en) * 2019-12-19 2023-08-15 新华智云科技有限公司 Method for searching video by using graph
CN111489757B (en) * 2020-03-26 2023-08-18 北京达佳互联信息技术有限公司 Audio processing method, device, electronic equipment and readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003009277A2 (en) * 2001-07-20 2003-01-30 Gracenote, Inc. Automatic identification of sound recordings
WO2005050620A1 (en) * 2003-11-18 2005-06-02 Koninklijke Philips Electronics N.V. Matching data objects by matching derived fingerprints
FR2863080A1 (en) * 2003-11-27 2005-06-03 Advestigo METHOD FOR INDEXING AND IDENTIFYING MULTIMEDIA DOCUMENTS
US20060229878A1 (en) * 2003-05-27 2006-10-12 Eric Scheirer Waveform recognition method and apparatus

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7289643B2 (en) * 2000-12-21 2007-10-30 Digimarc Corporation Method, apparatus and programs for generating and utilizing content signatures
US5841888A (en) * 1996-01-23 1998-11-24 Harris Corporation Method for fingerprint indexing and searching
US5715518A (en) * 1996-03-06 1998-02-03 Cellular Technical Services Company, Inc. Adaptive waveform matching for use in transmitter identification
US7013301B2 (en) * 2003-09-23 2006-03-14 Predixis Corporation Audio fingerprinting system and method
US6990453B2 (en) 2000-07-31 2006-01-24 Landmark Digital Services Llc System and methods for recognizing sound and music signals in high noise and distortion
US7277766B1 (en) * 2000-10-24 2007-10-02 Moodlogic, Inc. Method and system for analyzing digital audio files
JP4723171B2 (en) * 2001-02-12 2011-07-13 グレースノート インク Generating and matching multimedia content hashes
US7127106B1 (en) * 2001-10-29 2006-10-24 George Mason Intellectual Properties, Inc. Fingerprinting and recognition of data
US6859804B2 (en) * 2002-06-11 2005-02-22 The Regents Of The University Of California Using histograms to introduce randomization in the generation of ensembles of decision trees
US8055503B2 (en) * 2002-10-18 2011-11-08 Siemens Enterprise Communications, Inc. Methods and apparatus for audio data analysis and data mining using speech recognition
US20040153444A1 (en) * 2003-01-30 2004-08-05 Senders Steven L. Technique for effectively providing search results by an information assistance service
US7970644B2 (en) * 2003-02-21 2011-06-28 Accenture Global Services Limited Electronic toll management and vehicle identification
US20040258280A1 (en) * 2003-06-23 2004-12-23 Canadian Bank Note Company, Ltd. Optical document authentication system
ATE509207T1 (en) 2003-07-10 2011-05-15 Skf Ab BEARING UNIT WITH SOLDER OR HARD SOLDER CONNECTION
CN100521781C (en) * 2003-07-25 2009-07-29 皇家飞利浦电子股份有限公司 Method and device for generating and detecting fingerprints for synchronizing audio and video
WO2005041109A2 (en) * 2003-10-17 2005-05-06 Nielsen Media Research, Inc. Methods and apparatus for identifiying audio/video content using temporal signal characteristics
US7519954B1 (en) * 2004-04-08 2009-04-14 Mcafee, Inc. System and method of operating system identification
US8688248B2 (en) * 2004-04-19 2014-04-01 Shazam Investments Limited Method and system for content sampling and identification
DE102004023436B4 (en) * 2004-05-10 2006-06-14 M2Any Gmbh Apparatus and method for analyzing an information signal
EP1779659B1 (en) * 2004-08-12 2020-10-07 Gracenote Inc. Selection of content from a stream of video or audio data
US20060265436A1 (en) * 2005-05-20 2006-11-23 Andrew Edmond Grid network for distribution of files
CA2631270A1 (en) * 2005-11-29 2007-06-07 Google Inc. Detecting repeating content in broadcast media
US20090006337A1 (en) * 2005-12-30 2009-01-01 Mediaguide, Inc. Method and apparatus for automatic detection and identification of unidentified video signals
US7899625B2 (en) * 2006-07-27 2011-03-01 International Business Machines Corporation Method and system for robust classification strategy for cancer detection from mass spectrometry data
US7979464B2 (en) * 2007-02-27 2011-07-12 Motion Picture Laboratories, Inc. Associating rights to multimedia content
US7983915B2 (en) * 2007-04-30 2011-07-19 Sonic Foundry, Inc. Audio content search engine

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003009277A2 (en) * 2001-07-20 2003-01-30 Gracenote, Inc. Automatic identification of sound recordings
US20060229878A1 (en) * 2003-05-27 2006-10-12 Eric Scheirer Waveform recognition method and apparatus
WO2005050620A1 (en) * 2003-11-18 2005-06-02 Koninklijke Philips Electronics N.V. Matching data objects by matching derived fingerprints
FR2863080A1 (en) * 2003-11-27 2005-06-03 Advestigo METHOD FOR INDEXING AND IDENTIFYING MULTIMEDIA DOCUMENTS

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CANO P ET AL: "A review of algorithms for audio fingerprinting", PROCEEDINGS OF THE 2003 IEEE RADAR CONFERENCE. HUNTSVILLE, AL, MAY 5 - 8, 2003; [IEEE RADAR CONFERENCE], NEW YORK, NY : IEEE, US, 9 December 2002 (2002-12-09), pages 169 - 173, XP010642539, ISBN: 978-0-7803-7920-6 *
See also references of EP2168061A1 *
WELSH M ET AL: "QUERYING LARGE COLLECTIONS OF MUSIC FOR SIMILARITY", UC BERKELEY TECHNICAL REPORT, XX, XX, no. -1096, 1 November 1999 (1999-11-01), XP008027813 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011059504A (en) * 2009-09-11 2011-03-24 Victor Co Of Japan Ltd Content identification method, content identification system, content retrieval device, and content use device
WO2012066577A1 (en) * 2010-11-18 2012-05-24 Sling Media Pvt Ltd Automated searching for solutions to support self-diagnostic operations of web-enabled devices

Also Published As

Publication number Publication date
CN101681381A (en) 2010-03-24
TW200912681A (en) 2009-03-16
JP2010530100A (en) 2010-09-02
US20100205174A1 (en) 2010-08-12
EP2168061A1 (en) 2010-03-31
TWI447601B (en) 2014-08-01
JP5090523B2 (en) 2012-12-05
CN101681381B (en) 2012-11-07
US8266142B2 (en) 2012-09-11

Similar Documents

Publication Publication Date Title
US8266142B2 (en) Audio/Video fingerprint search accuracy using multiple search combining
Miller et al. Audio fingerprinting: nearest neighbor search in high dimensional binary spaces
US6931413B2 (en) System and method providing automated margin tree analysis and processing of sampled data
Cano et al. A review of algorithms for audio fingerprinting
EP2685450B1 (en) Device and method for recognizing content using audio signals
CN102959624B (en) System and method for audio media recognition
Haitsma et al. A highly robust audio fingerprinting system with an efficient search strategy
US8238669B2 (en) Detection and classification of matches between time-based media
EP1550297B1 (en) Fingerprint extraction
Wei et al. Frame fusion for video copy detection
US20070071330A1 (en) Matching data objects by matching derived fingerprints
CN101465122A (en) Method and system for detecting phonetic frequency spectrum wave crest and phonetic identification
BRPI0112901B1 (en) methods to recognize an audio sample, and, computer system to perform the same
JP2008504741A (en) Method for characterizing the overlap of two media segments
JP2008530597A (en) Automatic identification of repeated material in audio signals
CN111651527B (en) Identity association method, device, equipment and storage medium based on track similarity
Haitsma et al. An efficient database search strategy for audio fingerprinting
CN112732972B (en) Audio fingerprint generation system and method
Herley Accurate repeat finding and object skipping using fingerprints
CN110061976B (en) Unknown protocol frame sequence extraction method and system based on data mining
WO2010038187A1 (en) Method for data clusters indexing, recognition and retrieval in presence of noise
Cha An Efficient Search Algorithm for Fingerprint Databases.
Assent et al. Speeding up complex video copy detection queries
Carvalhido Real-time audio fingerprinting for advertising detection in streaming broadcast content
CN115691553A (en) Copyright identification method for video background music

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200880018885.0

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08794341

Country of ref document: EP

Kind code of ref document: A1

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2010511189

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2008794341

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 12663057

Country of ref document: US