Sound feature priority alignment
First Claim
Patent Images
1. A method implemented by one or more computing devices, the method comprising:
- identifying features of sound data from a plurality of recordings;
calculating values for individual frames of the sound data from the plurality of recordings, each of the values based on a similarity value scaled by a priority value, wherein;
the similarity value corresponds to a similarity determined using a similarity matrix built by comparing a first set of individual frames of a first of the plurality of recordings to a second set of individual frames of a second of the plurality of recordings, wherein the similarity matrix represents a comparison between the first set of individual frames and the second set of individual frames computed using a normalized inner product of compared frames; and
the priority value assigned to individual frames of the sound data based on the identified features of the sound data as the identified features occur in the respective individual frames of the sound data, the priority value based on how speech characteristics of a particular frame compare to speech characteristics of other individual frames of the sound data; and
aligning the sound data from the plurality of recordings based at least in part on the calculated values that are based on the similarity value scaled by the priority value, wherein individual frames of the sound data from the plurality of recordings are more likely to be aligned between different recordings when the priority value assigned to the individual frames of the sound data based on the identified features is higher than priority values assigned to other individual frames; and
outputting the aligned sound data.
2 Assignments
0 Petitions
Accused Products
Abstract
Sound feature priority alignment techniques are described. In one or more implementations, features of sound data are identified from a plurality of recordings. Values are calculated for frames of the sound data from the plurality of recordings. The values are based on similarity of the frames of the sound data from the plurality of recordings to each other, the similarity based on the identified features and a priority that is assigned based on the identified features of respective frames. The sound data from the plurality of recordings is then aligned based at least in part on the calculated values.
-
Citations
20 Claims
-
1. A method implemented by one or more computing devices, the method comprising:
-
identifying features of sound data from a plurality of recordings; calculating values for individual frames of the sound data from the plurality of recordings, each of the values based on a similarity value scaled by a priority value, wherein; the similarity value corresponds to a similarity determined using a similarity matrix built by comparing a first set of individual frames of a first of the plurality of recordings to a second set of individual frames of a second of the plurality of recordings, wherein the similarity matrix represents a comparison between the first set of individual frames and the second set of individual frames computed using a normalized inner product of compared frames; and the priority value assigned to individual frames of the sound data based on the identified features of the sound data as the identified features occur in the respective individual frames of the sound data, the priority value based on how speech characteristics of a particular frame compare to speech characteristics of other individual frames of the sound data; and aligning the sound data from the plurality of recordings based at least in part on the calculated values that are based on the similarity value scaled by the priority value, wherein individual frames of the sound data from the plurality of recordings are more likely to be aligned between different recordings when the priority value assigned to the individual frames of the sound data based on the identified features is higher than priority values assigned to other individual frames; and outputting the aligned sound data. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 19)
-
-
11. A system comprising:
-
at least one module implemented at least partially in hardware and configured to identify features of sound data; and one or more modules implemented at least partially in hardware and configured to automatically generate sound feature rules, the sound feature rules generated based on the identified features of the sound data, the one or more modules further configured to use the sound feature rules to assign a priority value to individual frames of the sound data based on features that are identified in the frames of the sound data, the one or more modules further configured to use the priority value to scale a similarity value of frames of different recordings to determine an alignment of sound data between the different recordings, the similarity value determined using a similarity matrix built by comparing a first set of individual frames of a first of the plurality of recordings to a second set of individual frames of a second of the plurality of recordings, wherein the similarity matrix represents a comparison between the first set of individual frames and the second set of individual frames computed using a normalized inner product of compared frames, the sound feature rules specifying that frames of the sound data having a higher energy are assigned a higher priority value for alignment than frames of the sound data having a lower energy. - View Dependent Claims (12, 13, 18)
-
-
14. A computing device comprising:
-
one or more processors; and one or more computer-readable storage media having instructions stored thereon that, responsive to execution by the one or more processors of the computing device, causes the computing device to perform operations comprising; aligning sound data from different recordings of a plurality of recordings to generate aligned sound data, the aligning performed based at least in part using sound feature rules to assign a priority value to individual frames in the sound data based on features that are identified in the individual frames of the sound data and scaling a similarity value by the priority value, the similarity value corresponding to a similarity determined using a similarity matrix built by comparing a first set of individual frames of a first of the plurality of recordings to a second set of individual frames of a second of the plurality of recordings, wherein the similarity matrix represents a comparison between the first set of individual frames and the second set of individual frames computed using a normalized inner product of compared frames, the sound feature rules configured such that priority values are increased to place higher importance on frames of the sound data with phrase onsets; and outputting the generated aligned sound data. - View Dependent Claims (15, 16, 17, 20)
-
Specification