AU2002301619B2 - Audio Segmentation with the Bayesian Information Criterion - Google Patents
Audio Segmentation with the Bayesian Information Criterion Download PDFInfo
- Publication number
- AU2002301619B2 AU2002301619B2 AU2002301619A AU2002301619A AU2002301619B2 AU 2002301619 B2 AU2002301619 B2 AU 2002301619B2 AU 2002301619 A AU2002301619 A AU 2002301619A AU 2002301619 A AU2002301619 A AU 2002301619A AU 2002301619 B2 AU2002301619 B2 AU 2002301619B2
- Authority
- AU
- Australia
- Prior art keywords
- sequence
- frame
- audio samples
- audio
- frames
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
- 230000011218 segmentation Effects 0.000 title description 22
- 238000000034 method Methods 0.000 claims description 60
- 238000009826 distribution Methods 0.000 claims description 43
- 230000007704 transition Effects 0.000 claims description 23
- 230000006870 function Effects 0.000 claims description 8
- 230000015654 memory Effects 0.000 claims description 8
- 102100035353 Cyclin-dependent kinase 2-associated protein 1 Human genes 0.000 claims 1
- 230000008859 change Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 7
- 238000012552 review Methods 0.000 description 7
- 230000005236 sound signal Effects 0.000 description 6
- 238000003860 storage Methods 0.000 description 6
- 239000013598 vector Substances 0.000 description 6
- 230000001360 synchronised effect Effects 0.000 description 5
- 238000007476 Maximum Likelihood Methods 0.000 description 4
- 238000013179 statistical model Methods 0.000 description 4
- 238000013459 approach Methods 0.000 description 3
- 238000000605 extraction Methods 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000005070 sampling Methods 0.000 description 3
- 238000001228 spectrum Methods 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Landscapes
- Signal Processing For Digital Recording And Reproducing (AREA)
Description
-i- S&F Ref: 611119
AUSTRALIA
PATENTS ACT 1990 COMPLETE SPECIFICATION FOR A STANDARD PATENT
ORIGINAL
Name and Address of Applicant Actual Inventor(s): Address for Service: Invention Title: Canon Kabushiki Kaisha 30-2, Shimomaruko 3-chome, Ohta-ku Tokyo 146 Japan Timothy John Wark Spruson Ferguson St Martins Tower,Level 31 Market Street Sydney NSW 2000 (CCN 3710000177) Audio Segmentation with the Bayesian Information Criterion ASSOCIATED PROVISIONAL APPLICATION DETAILS [33] Country [31] Applic. No(s) AU PR8470 AU PR8471 [32] Application Date 25 Oct 2001 25 Oct 2001 The following statement is a full description of this invention, including the best method of performing it known to me/us:- 5815c
A
AUDIO SEGMENTATION WITH THE BAYESIAN INFORMATION
CRITERION
Technical Field of the Invention The present invention relates generally to the segmentation of audio streams and, in particular, to the use of the Bayesian Information Criterion as a method of segmentation.
Background Art There is an increasing demand for automated computer systems that extract meaningful information from large amounts of data. One such application is the extraction of information from continuous streams of audio. Such continuous audio streams may include speech from, for example, a news broadcast or a telephone conversation, or non-speech, such as music or background noise.
In order for a system to be able to extract information from the continuous audio stream, the system is typically first required to segment the continuous audio stream into homogeneous segments, each segment including audio from only one speaker or other constant acoustic condition. Once the segment boundaries have been located, each segment may be processed individually to, for example, classify the information contained within each of the segments.
Whilst a number of techniques have been proposed in a somewhat ad-hoc manner for segmenting audio in specific applications, one of the most successful approaches that has been used is an approach based on the Bayesian Information Criterion (BIC). The BIC is a model selection criterion known in statistical literature and is used to 611119AU.doc
I
-2determine the positions of segment boundaries by determining the most likely positions where the signal characteristics change. When applied to audio segmentation, the BIC is used to determine whether a section of audio is better described by one statistical model or two different statistical models, hence allowing a segmentation decision to be made. It also gives a criterion to determine whether the change at this point is significant, or not.
Previous systems performing audio segmentation with the BIC have made the assumption that the statistical model characterising each audio segment is a Gaussian process. However, the Gaussian model tends not to hold very well when only a small amount of data is available for the audio stream between segment changes. Thus, segmentation performs very poorly with the Gaussian BIC under these conditions.
Another major setback for BIC-based segmentation systems is the computation time required to segment large audio streams. This is due to the fact that previous BIC systems have used multi-dimensional features for describing important characteristics within the audio stream, such multi-dimensional features being those of the mel-cepstral vectors or linear predictive coefficients.
Summary of the Invention It is an object of the present invention to substantially overcome, or at least ameliorate, one or more disadvantages of existing arrangements.
According to an aspect of the invention, there is provided a method of segmenting a sequence of audio samples into a plurality of homogeneous segments, said method comprising the steps of: forming a sequence of frames along said sequence of audio samples, each said frame comprising a number of said audio samples; 611119AU.doc -3extracting, for each said frame, a single-dimensional data feature, said data features forming a sequence of said data features each corresponding to one of said frames; and detecting one or more transition points in said sequence of data features by applying the Bayesian Information Criterion to said sequence of data features, said transition points defining said homogeneous segments.
Other aspects of the invention are also disclosed.
Brief Description of the Drawings One or more embodiments of the present invention will now be described with reference to the drawings, in which: Fig. 1 shows a schematic block diagram of a system upon which audio segmentation can be practiced; Fig. 2 shows a flow diagram of a method for segmenting a sequence of sampled audio from unknown origin into homogeneous segments; Fig. 3A shows a flow diagram of a method for detecting a single transition-point within a sequence of frame features; Fig. 3B shows a flow diagram of a method for detecting multiple transition-point within a sequence of frame features; Figs. 4A and 4B show a sequence of frames and the sequence of frames being divided at into two segments; Fig. 5A illustrates a distribution of example frame features and the distribution of a Gaussian event model that best fits the set of frame features obtained from a segment of speech; 611119AU.doc -4- Fig. 5B illustrates a distribution of the example frame features of Fig. 5A and the distribution of a Laplacian event model that best fits the set of frame features; Fig. 6A illustrates a distribution of example frame features and the distribution of a Gaussian event model that best fits the set of frame features obtained from a segment of music; Fig. 6B illustrates a distribution of the example frame features of Fig. 6A and the distribution of a Laplacian event model that best fits the set of frame features; Fig. 7 illustrates the formation of frames from the sequence of audio samples, the extraction of the sequence frame features, and the detection of segments within the 1o sequence of frame features; and Fig. 8 shows a media editor within which the method for segmenting a sequence of sampled audio into homogeneous segments may be practiced.
Detailed Description including Best Mode Some portions of the description which follow are explicitly or implicitly presented in terms of algorithms and symbolic representations of operations on data within a computer memory. These algorithmic descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, conceived to be a self-consistent sequence of steps leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated.
611119AU.doc It should be borne in mind, however, that the above and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels.
Unless specifically stated otherwise, and as apparent from the following, it will be appreciated that throughout the present specification, discussions refer to the action and processes of a computer system, or similar electronic device, that manipulates and transforms data represented as physical (electronic) quantities within the registers and memories of the computer system into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission or display devices.
Where reference is made in any one or more of the accompanying drawings to steps and/or features, which have the same reference numerals, those steps and/or features have for the purposes of this description the same function(s) or operation(s), unless the contrary intention appears.
Fig. 1 shows a schematic block diagram of a system 100 upon which audio segmentation can be practiced. The system 100 comprises a computer module 101, such as a conventional general-purpose computer module, input devices including a keyboard 102, pointing device 103 and a microphone 115, and output devices including a display device 114 and one or more loudspeakers 116.
The computer module 101 typically includes at least one processor unit 105, a memory unit 106, for example formed from semiconductor random access memory (RAM) and read only memory (ROM), input/output interfaces including a video interface 107 for the video display 114, an 1/O interface 113 for the keyboard 102, the pointing device 103 and interfacing the computer module 101 with a network 118, such as the Internet, and an audio interface 108 for the microphone 115 and the loudspeakers 116.
611119AU.doc -6- A storage device 109 is provided and typically includes a hard disk drive and a floppy disk drive. A CD-ROM or DVD drive 112 is typically provided as a non-volatile source of data. The components 105 to 113 of the computer module 101, typically communicate via an interconnected bus 104 and in a manner which results in a conventional mode of operation of the computer module 101 known to those in the relevant art.
Audio data for processing by the system 100, and in particular the processor 105, may be derived from a compact disk or video disk inserted into the CD-ROM or DVD drive 112 and may be received by the processor 105 as a data stream encoded in a particular format. Audio data may alternatively be derived from downloading audio data from the network 118. Yet another source of audio data may be recording audio using the microphone 115. In such a case, the audio interface 108 samples an analog signal received from the microphone 115 and provides the audio data to the processor 105 in a particular format for processing and/or storage on the storage device 109.
The audio data may also be provided to the audio interface 108 for conversion into an analog signal suitable for output to the loudspeakers 116.
Fig. 2 shows a flow diagram of a method 200 of segmenting an audio stream in the form of a sequence x(n) of sampled audio from unknown origin into homogeneous segments. The method 200 is preferably implemented in the system 100 by a software program executed by the processor 105. A homogeneous segment is a segment only containing samples from a source having constant acoustic characteristic, such as from a particular human speaker, a type of background noise, or a type of music. It is assumed that the audio stream is appropriately digitised at a sampling rate F. Those skilled in the art would understand the steps required for converting an analog audio stream into the sequence x(n) of sampled audio. In an example arrangement, the audio stream is sampled 611119AU.doc -7at a sampling rate F of 16 kHz and the sequence x(n) of sampled audio is stored on the storage device 109 in a form such as a.wav file or a raw file. The method 200 starts in step 202 where the sequence x(n) of sampled audio are read from the storage device 109 and placed in memory 106.
Fig. 7 illustrates such a sequence x(n) of sampled audio. In order for the Bayesian Information Criterion (BIC) to be applied to the sequence x(n) of sampled audio, one or more features must be extracted for each small, incremental interval of K samples along the sequence An underlying assumption is that the properties of the audio signal change relative slowly in time, and that each extracted feature provides a succinct description of important characteristics of the audio signal in the associated interval. Ideally, such features extract enough information from the underlying audio signal so that the subsequent segmentation algorithm can perform well, and yet be compact enough that segmentation can be performed very quickly.
Referring again to Fig. 2, in step 204 the processor 105 forms interval windows or frames, each containing K audio samples. In the example, a frame of 20 ms is used, which corresponds to K=320 samples at the sampling rate F of 16 kHz. Further, the frames are overlapping, with the start position of the next frame positioned only 10 ms later in time, or 160 samples later, providing a shift-time of 10 mins. The forming of frames 701 to 704 and extraction of features 711 to 714 are also illustrated in Fig. 7.
Referring again to Fig. 2, in step 206 a Hamming window function of the same length as that of the frames, i.e. K samples long, is applied by the processor 105 to the sequence samples x(n) in each frame to give a modified set of windowed audio samples s(i,k) for frame i, with ke The purpose of applying the Hamming window is to 611119AU.doc -8reduce the side-lobes created when applying the Fast Fourier Transform (FFT) in subsequent operations.
In step 208 the bandwidth BW(i) of the modified set of windowed audio samples s(i,k) of the i'th frame is calculated by the processor 105 as follows: 2 S,()dco BW f( (1) SS,(cw)do where is the power spectrum of the modified windowed audio samples s(i,k) of the i'th frame, co is a signal frequency variable for the purposes of calculation, and FC is the frequency centroid, defined as: d FC(i) jS(o dw (2) The Simpson's integration is used to evaluate the integrals. The Fast Fourier Transform is used to calculate the power spectrum Si(6)) whereby the samples s(i,k), having length K, are zero padded until the next highest power of 2 is reached. Thus, in the example where the length of the samples s(k) is 320, the FFT would be applied to a vector of length 512, formed from 320 modified windowed audio samples s(i,k) and 192 zero components.
In step 210 the energy E(i) of the modified set of windowed audio samples s(i,k) of the i'th frame is calculated by the processor 105 as follows: l (3) A frame feature f(i) for each frame i is calculated by the processor 105 in step 212 by weighting the frame bandwidth BW(i) by the frame energy This forces a bias 611119AU.doc -9in the measurement of bandwidth BW(i) in those frames i that exhibit a higher energy and are thus more likely to come from an event of interest, rather than just background noise. The frame feature f(i) is thus calculated as being: f E(i)B W(i) (4) Steps 206 to 212 jointly extract the frame feature f(i) from the sequence x(n) of audio samples and the frame i. The frame feature f(i) shown in Equation is a single dimensional feature providing a great reduction in the computation time when it is applied to the Bayesian Information Criterion over systems that use a multi-dimensional feature vector such as mel-cepstral vectors or linear predictive coefficients. Mel-cepstral features seek to extract information from a signal by "binning" the magnitudes of the power spectrum in bins centred at various frequencies. A Discrete Cosine Transform (DCT) is then applied in order to produce a vector of coefficients, typically in the order of 12 to 16. In a similar way linear-predictive coefficients (LPC) are derived by modelling the signal as an auto-regressive (AR) time-series, where the coefficients of the time-series become the features f(i) again having a dimension of 12 to 16.
The BIC is used in step 220 by the processor 105 to segment the sequence of frame features f(i) into hiomogeneous segments, such as the segments illustrated in Fig.
7. The output of step 220 is one or more frame numbers of the frames where changes in acoustic characteristic were detected. In order to provide the output in a user-friendly manner, the processor 105 converts each frame number received from step 220 into time in seconds, the time being from the start point of the audio signal. This conversion is done by the processor 105 in step 225 by multiplying each output frame number by the window-shift. In the example where the window-shift of 10 ms is used, the output frame numbers are multiplied by 10 ms to get the segment boundaries in seconds.
611119AU.doc In an alternative arrangement where the audio data is associated with a video sequence, the output may be stored as metadata of the video sequence. The metadata may be used to assist in segmentation of the video, for example.
The BIC used in step 220 will now be described in more detail. The value of the BIC is a statistical measure for how well a model represents a set of features and is calculated as: BIC log(L) D log(N) 2 where L is the maximum-likelihood probability for a chosen model to represent the set of features D is the dimension of the model which is 1 when the frame feature f(i) of Equation is used, and N is the number of features f(i) being tested against the model.
The maximum-likelihood L is calculated by finding the parameters 0 of the model that maximise the probability of the featuresj(i) being from that model. Thus, for a set of parameters 0, the maximum-likelihood L is: L max P(f (6) Segmentation using the BIC operates by testing whether the sequence of features f(i) are better described by a single-distribution event model, or a twin-distribution event model, where the first m number of frames, those being frames are from a first source and the remainder of the N frames, those being frames are from a second source. The frame m is accordingly termed the change-point. To allow a comparison, a criterion difference ABIC is calculated between the BIC using the twindistribution event model with that using the single-distribution event-model. As the change-point m approaches a transition in acoustic characteristics, the criterion difference ABIC typically increases, reaching a maximum at the transition, and reducing again 611119AU.doc 1I towards the end of the N frames under consideration. If the maximum criterion difference ABIC is above a predefined threshold, then the two-distribution event model is deemed a more suitable choice, indicating a significant transition in acoustic characteristics at the change-point m where the criterion difference ABIC reached a maximum.
Current BIC segmentation systems assume that the features f(i) are best represented by a Gaussian event model having a probability density function of the form: g(fl() exp{ (fi O t- (7) )2 2 (2r)T j xE where t is the mean vector of the features and E is the covariance matrix.
Fig. 5A illustrates a distribution 500 of frame features Jfi), where the frame featuresf(i) were obtained from an audio stream of duration I second containing voice.
Also illustrated is the distribution of a Gaussian event model 502 that best fits the set of frame features f(i).
It is proposed that frame features fji) representing the characteristics of audio signals such as a particular speaker or block of music, is better represented by a leptokurtic distribution, particularly where the number N of features being tested against the model is small. A leptokurtic distribution is a distribution that is more peaky than a Gaussian distribution, such as a Laplacian distribution. Fig. 5B illustrates the distribution 500 of the same frame featuresf(i) as those of Fig. 5A, together with the distribution of a Laplacian event model 505 that best fits the set of frame featuresf(i). It can be seen that the Laplacian event model gives a much better characterisation of the feature distribution 500 than the Gaussian event model.
This proposition is further illustrated in Figs. 6A and 6B wherein a distribution 600 of frame features f(i) obtained from an audio stream of duration 1 second containing 611119AU.doc 12music is shown. The distribution of a Gaussian event model 602 that best fits the set of frame features f(i) is shown in Fig. 6A, and the distribution of a Laplacian event model 605 is illustrated in Fig. 6B.
A quantitative measure to substantiate that the Laplacian distribution provides a better description of the distribution characteristics of the features f(i) for short events rather than the Gaussian model is the Kurtosis statistical measure which provides a measure of the "peakiness" of a distribution and may be calculated for a sample set X as: xE(X 3 (8) (var(X))2 For a true Gaussian distribution, the Kurtosis measure will be 0, whilst for a true Laplacian distribution the Kurtosis measure will be 3. In the case of the distributions 500 and 600 shown in Figs. 5A and 6A, the Kurtosis measures c were 2.33 and 2.29 respectively, hence the distributions 500 and 600 are more Laplacian in nature rather than Gaussian.
The Laplacian probability density function in one dimension is: 1 f i (9) g(f p, a) exp (9) where p is the mean of the frame features f(i) and a is their standard deviation. In a higher order feature space with frame features each having dimension D, the feature distribution is represented as: 2 2 S(2 2 i t ifi f ti f t where v D) 2 and is the modified Bessel function of the third kind.
611119AU.doc 13 Whilst the method 200 can be used with multi-dimensional features the rest of the analysis is contained to the one-dimensional space due to the use of the onedimensional featurefji) shown in Equation Given N frame featuresJ(i) as illustrated in Fig. 4A, the maximum likelihood L for the set of frame featuresj(i) falling under a single Laplacian distribution is: L (2)2 -exp- 2 If(i)-K (11) where a is the standard deviation of the frame features}(i) and/I is the mean of the frame features Equation (11) may be simplified providing: N2~ IVJN L (2r2 Y exp Z If)-P (12) Ia 1=1* The maximum log-likelihood log(L), assuming natural logs, for all N frame featuresj(i) to fall under a single Laplacian event model is thus: log(L)= N(2, 2 N'2 'V (13) 2 (o ,=1 Fig. 4B shows the N frames being divided at frame m into two segments 550 and 555, with the first mi number of frames forming segment 550 and the remainder of the N frames forming segment 555. A log-likelihood ratio R(m) of a twina Laplacian distribution event model to a single Laplacian distribution event model, with the division at frame mi and assuming segment 550 is from a first source and segment 555 is from a second source, is: R(m) log(L, log(L 2 log(L) (14) where: 611119AU.doc 0 a -14log(L, 2 c 2 2 and log()= 2 i Ef()- 2 (16) log(L 2 2 ~mIIf U2 2 (o-2 wherein {k,7, 1 and 2 ,c1r 2 are the means and standard deviations of the frame featuresj(i) before and after the change point m.
The criterion difference ABIC for the Laplacian case having a change point m is calculated as: ABIC(m)=
D
-§og m(N (17) In a simplest of cases where only a single transition is to be detected in a section of audio represented by a sequence of N frame features the most likely transition point i is given by: m arg max ABIC(m)) (18) Fig. 3A shows a flow diagram of a method 300 for detecting a single transitionpoint 6t within a sequence of N frame featuresJ(i) that may be substituted as step 220 in method 200 shown in Fig. 2. When more than one transition-point rh(j) is to be detected, the method 400 shown in Fig. 3B is substituted as step 220 in method 200 (Fig.
Method 400 uses method 300 as is described below.
Method 300, performed by the processor 105, receives a sequence of N' frame featuresj(i) as input. When method 300 is substituted as step 220 in method 200, then the number of frames N' equals the number of features N. In step 305 the change-point m is 611119AU.doc set by the processor 105 to 1. The change-point m sets the point dividing the sequence of VN frame featuresfi) into two separate sequences namely m] and [rm+l; AI].
Step 310 follows where the processor 105 calculates the log-likelihood ratio R(m) by first calculating the means and standard deviations and {p 2 ,'r 2 of the frame featuresfli) before and after the change-point m. Equations (15) and (16) are then calculated by the processor 105, and the results are substituted into Equation (14).
The criterion difference ABIC for the Laplacian case having the change-point m is then calculated by the processor 105 using Equation (17) in step 315.
In step 320 the processor 105 determines whether the change point m has reached the end of the sequence of length If the change-point m has not reached the end of the sequence, then the change-point m is incremented by the processor 105 in step 325 and steps 310 to 320 are repeated for the next change-point m. When the processor 105 determines in step 320 that the change-point in has reached the end of the sequence, then the method 300 proceeds to step 330 where the processor 105 determines whether a significant change in the sequence of N' frame features J(i) occurred by determining whether the maximum criterion difference max[ ,BIC(m)] has a value that is greater than a predetermined threshold. In the example, the predetermined threshold is set to 0. If the change was determined by the processor 105 in step 330 to be significant, then the method proceeds to step 335 where the most likely transition-point m is determined using Equation and the result is provided to step 225 (Fig. 2) for processing and output to the user. Alternatively, in step 340 the null string is provided as output to step 225 (Fig. 2) which in turn informs the user that no significant transition could be detected in the audio signal.
611119AU.doc 16- Fig. 313 shows a flow diagram of the method 400 for detecting multiple transition-points m(j) within the sequence of N frame features fli) that may be used as step 220 in method 200 shown in Fig. 2. Method 400 thus receives the sequence of N frame featuresJ{i) from step 212 (Fig. 2) and provides the result to step 225 (Fig. 2) for processing and output to the user. Given an audio stream that is assumed to contain an unknown number of transition points th(j), the method 400 operates principally by analysing short sequences of frame features with each sequence consisting of N,,in frame featuresJ(i), and detecting a single transition-point 4(j) within each sequence, if it occurs, using method 300 (Fig. 3A). Once all the transition-points M(j) are detected, i0 the method 400 performs a second pass wherein each of the transition-points ih(j) detected are verified as being significant by analysing the sequence of frame features included in the segments either side of the transition-point 6i(j) under consideration, and eliminating any transition-points fi(j) verified not to be significant. The verified significant transition-points ih'(j) are then provided to step 225 (Fig. 2) for processing and output to the user.
Method 400 starts in step 405 where the sequence of frame features are defined by the processor 105 as being the sequence The first sequence includes N,,i features and method 400 is therefore initiated with a=1 and The number of features is variable and is determined for each application. By varying Ni,, the user can control whether short or spurious events should be detected or ignored, where the requirement being different with each scenario. In example, a minimum segment length of 1 second is assumed, thus given that the frame features f(i) are 611119AU.doc j 17extracted every 10 ms, being the window shift time, the number of features is set to 100.
Step 410 follows where the processor 105 detects a single transition-point th(j) within the sequence if it occurs, using method 300 (Fig. 3A) with In step 415 the processor 105 determines whether the output received from step 410, i.e.
method 300, is a transition-point rh(j) or a null string indicating that no transition-point occurred in the sequence If a transition-point ih(j) was detected in the sequence then the method 400 proceeds to step 420 where that transition-point h(j) is stored in the memory 106. Step 425 follows wherein a next sequence Wa)fb)] is defined by the processor 105 by setting a=rh(j)+6\ and where 81 is a predetermined small number of frames.
If the processor 105 determines in step 415 that no significant transition-point th(j) was detected in the sequence f(a)fjb)], then the sequence is lengthened by the processor 105 in step 430 by appending a small number 52 of frame featuresj(i) to the sequence by defining b=b+82. From either step 425 or 430 the method 400 proceeds to step 435 where the processor 105 determines whether all NV frame featuresj(i) have been considered. If all N frame featuresJ(i) have not been considered, then control is passed by the processor 105 to step 410 from where steps 410 to 435 are repeated until all the frame featuresj(i) have been considered.
The method 400 then proceeds to step 440, which is the start of the second pass.
In the second pass the method 400 verifies each of the N, transition-points 6h(j) detected in steps 405 to 435. The transition-point are verified by analysing the sequence of 611119AU.doc -18 frame features included in the segments either side of a transition-point mz(j) under consideration thus, when considering the transition-point it(j), the sequence -1)+1)t6 (j 1+ is analysed, with the verified transition-point r'(j) being set to 0. Accordingly, step 440 starts by setting a counterj to 1 and n to 0. Step 445 follows where the processor 105 detects a single transition-point th within the sequence 1 if it occurs, using again method 300 (Fig.
3A). In step 450 the processor 105 determines whether the output received from step 445, i.e. method 300, is a transition-point tr or a null string indicating that no significant transition-point m occurred in the sequence +1 If a transition-point A was detected in the sequence (j 1 then the method 400 proceeds to step 455 where that transition-point rh is stored in memory 106 and in a sequence of verified transition-points Step 460 follows wherein the counter j is incremented and n is reset to 0 by the processor 105. Alternatively if the processor 105 in step 450 determined that no significant transition-point m was detected by step 445, then the sequence [f(rh'(j +1 is merged by the processor 105 in step 465. The counter n is also incremented thereby extending the sequence of feature framesf(i) under consideration to the next transition-point h(j).
From either step 460 or 465 the method 400 proceeds to step 470 where it is determined by the processor 105 whether all the transition-points m(j) have been considered for verification. If any transition-points m(j) remain, control is returned to step 445 from where steps 445 to 470 are repeated until all the transition-points rh(j) 611119AU.doc '-9 19have been considered. The method 400 then passes the sequence of verified transitionpoints dh'(G) to step 225 (Fig. 2) for processing and output to the user.
Fig. 8 shows a media editor 800 within which the method 200 (Fig. 2) of segmenting a sequence of sampled audio into homogeneous segments may be practiced.
In particular, the media editor 800 is a graphical user interface, formed on display 114 of system 100 (Fig. of a media editor application, which is executed on the processor 105. The media editor 800 is operable by a user who wishes to review recorded media clips, which may include audio data and/or audio data synchronised with a video sequence, and wishes to construct a home production from the recorded media clips.
The media editor 800 includes a browser screen 810 which allows the user to search and/or browse a database or directory structure for media clips and into which files containing media clips may be loaded. The media clips may be stored as ".wav", ".mpg" files or files in other formats, and typically is loaded from a CD-ROMiDVD inserted into the CD-ROM/DVD drive 112 (Fig. 1).
Each file containing a media clip may be represented by an icon 804 once loaded into the browser screen 810. The icon 804 may be a keyframe when the file contains video data. When an icon 804 is selected by the user, its associated media content is transferred to the review/edit screen 812. More than one icon 804 may be selected, in which case the selected media content will be placed in the review/edit screen one after the other.
After selecting the aforementioned icons 804, a play button 814 on the review/edit screen 812 may be pressed. The media clip(s) associated with the aforementioned selected icon(s) 804 are played from a selected position and in the desired sequence, in a contiguous fashion as a single media presentation, and continues until the 611119AU.doc I.L i., 20 end of the presentation at which point playback stops.. In the case where the media clip(s) contains video and audio data, then the video is displayed within the display area 840 of the review/edit screen 812, while the synchronised audio content is played over the loadspeakers 116 (Fig. Alternatively, when the media clip only contains an audio sequence, then the audio is played over the loadspeakers 116. Optionally, some waveformn representation of the audio sequence may be displayed in the display area 840.
A playlist summary bar 820 is also provided on the review/edit screen 812, presenting to the user an overall timeline representation of the entire production being considered. The playlist summary bar 820 has a playlist scrubber 825, which moves along the playlist summary bar 820 and indicates the relative position within the presentation presently being played. The user may browse the production by moving the playlist scrubber 825 along the playlist summary bar 820 to a desired position to commerce play at that desired position. The review/edit screen 812 typically also includes other viewing controls including a pause button, a fast forward button, a rewind button, a frame step forward button, a frame step reverse button, a clip-index forward button, and a clip-index reverse button. The viewer play controls, referred to collectively as 850, may be activated by the user to initiate various kinds of playback within the presentation.
The user may also initiate a segmentation function for segmenting the audio sequence associated with the selected media clip(s). Method 200 (Fig. 2) will read in the audio sequence and return transition-points 6t'(q) as semantic event boundary locations.
In one implementation, the transition-points fh'(g) determined by method 200 (Fig. 2) are indicated as transition lines 822 on the playlist summary bar 820. The transition lines 822 illustrate borders of segments, such as segment 830. The length of the playlist 611119AU.doc -21 summary bar between the respective transition lines 822 represents the proportionate duration of an individual segment compared to the overall presentation duration.
In the case where the media clip(s) includes synchronised video and audio sequences, the transition lines 822 resulting from the audio segmentation also provides segmentation of the synchronised video sequence, based on the homogeneity of the audio sequence. Accordingly, the transition lines 822 also provide segmentation of the associated video.
The segments are selectable and manipulable by common editing commands such as "drag and drop", "copy", "paste", "delete" and so on. Automatic "snapping" is also provided whereby, in a drag and drop operation, a dragged segment is automatically inserted at a point between two other segments, thereby retaining the unity of the segments.
The user may thus edit the presentation, with the knowledge that the segment contained between consecutive transition lines 822 represents media content where the audio sequence is homogeneous. Such a segment could represent an event where only silence exists or one person is talking or one type of music is playing in the background.
For example, the user may delete segments containing silence by selecting such segments and deleting them. If the segment contained a video sequence with synchronised audio, then the associated video would also be deleted. Similar conditions apply to the other commands.
In another example the segments provide to the user an advantageous means for compiling a presentation of audio sequences wherein a particular speaker is talking. The user only needs to listen to a small part of each segment to identify whether the segment contains that speaker. There is no need for an exhaustive search for transition points, 611119AU.doc Mi-, I 22 which typically includes many pausing, rewinding and play operations to find such transition points.
Yet another application of the segmentation method 200 described herein is in an automatic audio classification system. In such a system, a media sequence which includes an audio sequence is first segmented using method 200 to determine the transition-points Known techniques may then be used to extract clip-level features from the audio samples within each segment. The extracted clip-level features are next classified against models of events of interest using statistical models known in the art. A label is then attached to each segment.
The models of events of interest are typically obtained through a training stage wherein the user obtains clip-level features from manually labelled segments of interest.
Such may be provided as described above in relation to Fig. 8.
The foregoing describes only some embodiments of the present invention, and modifications and/or changes can be made thereto without departing from the scope and spirit of the invention, the embodiment(s) being illustrative and not restrictive.
In the context of this specification, the word "comprising" means "including principally but not necessarily solely" or "having" or "including" and not "consisting only of". Variations of the word comprising, such as "comprise" and "comprises" have corresponding meanings.
611119AU.doc
Claims (5)
1. A method of segmenting a sequence of audio samples into a plurality of homogeneous segments, said method comprising the steps of: forming a sequence of frames along said sequence of audio samples, each said frame comprising a number of said audio samples; extracting, for each said frame, a single-dimensional data feature, said data features forming a sequence of said data features each corresponding to one of said frames; and detecting one or more transition points in said sequence of data features by applying the Bayesian Information Criterion to said sequence of data features, said transition points defining said homogeneous segments.
2. The method as claimed in claim 1, wherein a leptokurtic distribution is used as an event model in said Bayesian Information Criterion.
3. The method according to claim 2 wherein said leptokurtic distribution is a Laplacian distribution.
4. The method as claimed in any one of claims 1 to 3, wherein step for each said frame comprises the sub-steps of: (bl) extracting a bandwidth from the audio samples of said frame; (b2) extracting an energy value from the audio samples of said frame; and (b3) forming said data feature by biasing said bandwidth of said frame with said energy value of said frame. 611119AU.doc
24- The method as claimed in claim 4, wherein said data feature is formed by multi- plying said bandwidth of said frame with said energy value of said frame. 6. The method as claimed in any one of claims 1 to 5 wherein said frames are overlapping. 7. The method as claimed in any one of claims 1 to 6 comprising the further step following step of: (al) applying a Hamming window function to said audio samples in each of said frames. 8. An apparatus for segmenting a sequence of audio samples into a plurality of homogeneous segments, said apparatus comprising: means for forming a sequence of frames along said sequence of audio samples, each said frame comprising a number of said audio samples; means for extracting, for each said frame, a single-dimensional data feature, said data features forming a sequence of said data features each corresponding to one of said frames; and means for detecting one or more transition points in said sequence of data features by applying the Bayesian Information Criterion to said sequence of data features, said transition points defining said homogeneous segments. 9. The apparatus as claimed in claim 8, wherein a leptokurtic distribution is used as an event model in said Bayesian Information Criterion. The apparatus according to claim 9 wherein said leptokurtic distribution is a Laplacian distribution. 611119AU.doc 1 0 25 11. The apparatus as claimed in any one of claims 8 to 10, wherein said means for extracting said data feature for each frame comprises: means for extracting a bandwidth from the audio samples of said frame; means for extracting an energy value from the audio samples of said frame; and means for forming said data feature by biasing said bandwidth of said frame with said energy value of said frame. 12. The apparatus as claimed in claim 11, wherein said data feature is formed by multi-plying said bandwidth of said frame with said energy value of said frame. 13. The apparatus as claimed in any one of claims 8 to 12 wherein said frames are overlapping. 14. The apparatus as claimed in any one of claims 8 to 13 further comprising means for applying a Hamming window function to said audio samples in each of said frames before said data feature is extracted. A program stored on a memory medium for segmenting a sequence of audio samples into a plurality of homogeneous segments, said program comprising: code for forming a sequence of frames along said sequence of audio samples, each said frame comprising a number of said audio samples; code for extracting, for each said frame, a single-dimensional data feature, said data features forming a sequence of said data features each corresponding to one of said frames; and code for detecting one or more transition points in said sequence of data features by applying the Bayesian Information Criterion to said sequence of data features, said transition points defining said homogeneous segments. 611119AU.doc -a -26- 16. The program as claimed in claim 15, wherein a leptokurtic distribution is used as an event model in said Bayesian Information Criterion. 17. A method of segmenting a sequence of audio samples into a plurality of homogeneous segments, said apparatus being substantially as described with reference to Figs. 2 and 3A or 3B. 18. An apparatus for segmenting a sequence of audio samples into a plurality of homogeneous segments, said apparatus being substantially as described with reference to the drawings. Dated this Twenty-second day of OCTOBER 2002 CANON KABUSHIKI KAISHA Patent Attorneys for the Applicant SPRUSON&FERGUSON 611119AU.doc
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| AU2002301619A AU2002301619B2 (en) | 2001-10-25 | 2002-10-24 | Audio Segmentation with the Bayesian Information Criterion |
Applications Claiming Priority (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| AUPR8470 | 2001-10-25 | ||
| AUPR8471 | 2001-10-25 | ||
| AUPR8471A AUPR847101A0 (en) | 2001-10-25 | 2001-10-25 | A modified approach to audio segmentation with the bayesian information criterion using the laplacian distribution |
| AUPR8470A AUPR847001A0 (en) | 2001-10-25 | 2001-10-25 | A single-dimensional feature for fast audio segmentation using the bayesian information criterion |
| AU2002301619A AU2002301619B2 (en) | 2001-10-25 | 2002-10-24 | Audio Segmentation with the Bayesian Information Criterion |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| AU2002301619A1 AU2002301619A1 (en) | 2003-06-12 |
| AU2002301619B2 true AU2002301619B2 (en) | 2004-08-26 |
Family
ID=39266588
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| AU2002301619A Ceased AU2002301619B2 (en) | 2001-10-25 | 2002-10-24 | Audio Segmentation with the Bayesian Information Criterion |
Country Status (1)
| Country | Link |
|---|---|
| AU (1) | AU2002301619B2 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111444967A (en) * | 2020-03-30 | 2020-07-24 | 腾讯科技(深圳)有限公司 | Training method, generation method, device, equipment and medium for generating confrontation network |
| CN119167224A (en) * | 2024-07-29 | 2024-12-20 | 哈尔滨理工大学 | A method for identifying human activity changes based on Gaussian segmentation model |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6424946B1 (en) * | 1999-04-09 | 2002-07-23 | International Business Machines Corporation | Methods and apparatus for unknown speaker labeling using concurrent speech recognition, segmentation, classification and clustering |
| US6567775B1 (en) * | 2000-04-26 | 2003-05-20 | International Business Machines Corporation | Fusion of audio and video based speaker identification for multimedia information access |
-
2002
- 2002-10-24 AU AU2002301619A patent/AU2002301619B2/en not_active Ceased
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6424946B1 (en) * | 1999-04-09 | 2002-07-23 | International Business Machines Corporation | Methods and apparatus for unknown speaker labeling using concurrent speech recognition, segmentation, classification and clustering |
| US6567775B1 (en) * | 2000-04-26 | 2003-05-20 | International Business Machines Corporation | Fusion of audio and video based speaker identification for multimedia information access |
Non-Patent Citations (1)
| Title |
|---|
| Ferriros et al., Acoustic Change Detection and Clustering on Broadcast News, ICSI March 2000 * |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111444967A (en) * | 2020-03-30 | 2020-07-24 | 腾讯科技(深圳)有限公司 | Training method, generation method, device, equipment and medium for generating confrontation network |
| CN111444967B (en) * | 2020-03-30 | 2023-10-31 | 腾讯科技(深圳)有限公司 | Training method, generating method, device, equipment and medium for generating countermeasure network |
| CN119167224A (en) * | 2024-07-29 | 2024-12-20 | 哈尔滨理工大学 | A method for identifying human activity changes based on Gaussian segmentation model |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| EP1547060B1 (en) | System and method for generating an audio thumbnail of an audio track | |
| US7243062B2 (en) | Audio segmentation with energy-weighted bandwidth bias | |
| US6697564B1 (en) | Method and system for video browsing and editing by employing audio | |
| Tzanetakis et al. | Marsyas: A framework for audio analysis | |
| US6490553B2 (en) | Apparatus and method for controlling rate of playback of audio data | |
| US7027124B2 (en) | Method for automatically producing music videos | |
| US10134440B2 (en) | Video summarization using audio and visual cues | |
| EP1081960B1 (en) | Signal processing method and video/voice processing device | |
| JP4345321B2 (en) | Method for automatically creating an optimal summary of linear media and product with information storage media for storing information | |
| US7263485B2 (en) | Robust detection and classification of objects in audio using limited training data | |
| US7266287B2 (en) | Using background audio change detection for segmenting video | |
| US20030112265A1 (en) | Indexing video by detecting speech and music in audio | |
| US20050241465A1 (en) | Musical composition reproduction method and device, and method for detecting a representative motif section in musical composition data | |
| EP1944770A2 (en) | System and method for video production | |
| JP2005322401A (en) | Method, apparatus and program for generating media segment library, custom stream generation method and custom media stream transmission system | |
| JP2007519987A (en) | Integrated analysis system and method for internal and external audiovisual data | |
| JP2007065659A (en) | Extraction and matching of characteristic fingerprint from audio signal | |
| JP2006524359A (en) | Analysis of time characteristics displayed in parameters | |
| JP2008185626A (en) | Highlight scene detection device | |
| JP2006319980A (en) | Video summarizing apparatus, method and program using event | |
| US20070061728A1 (en) | Time approximation for text location in video editing method and apparatus | |
| JP3437617B2 (en) | Time-series data recording / reproducing device | |
| JP3475317B2 (en) | Video classification method and apparatus | |
| JP3757719B2 (en) | Acoustic data analysis method and apparatus | |
| WO2012164818A1 (en) | Region of interest identification device, region of interest identification method, region of interest identification program, and region of interest identification integrated circuit |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FGA | Letters patent sealed or granted (standard patent) | ||
| MK14 | Patent ceased section 143(a) (annual fees not paid) or expired |