US 12,105,753 B2
Automated cover song identification
Markus K. Cremer, Orinda, CA (US); Zafar Rafii, Berkeley, CA (US); Robert Coover, Orinda, CA (US); and Prem Seetharaman, Chicago, IL (US)
Assigned to Gracenote, Inc., New York, NY (US)
Filed by Gracenote, Inc., Emeryville, CA (US)
Filed on Sep. 16, 2022, as Appl. No. 17/946,915.
Application 17/946,915 is a continuation of application No. 17/065,479, filed on Oct. 7, 2020, granted, now 11,461,390.
Application 17/065,479 is a continuation of application No. 15/698,557, filed on Sep. 7, 2017, granted, now 10,803,119, issued on Oct. 13, 2020.
Claims priority of provisional application 62/441,469, filed on Jan. 2, 2017.
Prior Publication US 2023/0008776 A1, Jan. 12, 2023
This patent is subject to a terminal disclaimer.
Int. Cl. G06F 16/683 (2019.01); G06Q 50/18 (2012.01)
CPC G06F 16/683 (2019.01) [G06Q 50/184 (2013.01)] 21 Claims
OG exemplary drawing
 
15. A method comprising:
executing a constant Q transform on time slices of query audio data to output constant Q transformed time slices;
binarizing the constant Q transformed time slices to output binarized and constant transformed time slices;
executing a two-dimensional Fourier transform on time windows within the binarized and constant Q transformed time slices to output two-dimensional Fourier transforms of the time windows;
generating a reference data structure based on a sequential order of the two-dimensional Fourier transforms;
storing the reference data structure in a database; and
identifying a query data structure associated with query audio data as a cover rendition of the audio data based on a comparison of the query data structure and the reference data structure using a similarity matrix, wherein the similarity matrix indicates at least one degree to which reference portions of the reference data structure are associated with query portions of the query data structure, and wherein the at least one degree satisfies a corresponding threshold.