[go: up one dir, main page]

CN105898169A - Video processing method and device - Google Patents

Video processing method and device Download PDF

Info

Publication number
CN105898169A
CN105898169A CN201510511387.1A CN201510511387A CN105898169A CN 105898169 A CN105898169 A CN 105898169A CN 201510511387 A CN201510511387 A CN 201510511387A CN 105898169 A CN105898169 A CN 105898169A
Authority
CN
China
Prior art keywords
voice messaging
video
recording
recorded
sub
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510511387.1A
Other languages
Chinese (zh)
Inventor
李瑞科
姜乐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LeTV Information Technology Beijing Co Ltd
Original Assignee
LeTV Information Technology Beijing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by LeTV Information Technology Beijing Co Ltd filed Critical LeTV Information Technology Beijing Co Ltd
Priority to CN201510511387.1A priority Critical patent/CN105898169A/en
Publication of CN105898169A publication Critical patent/CN105898169A/en
Pending legal-status Critical Current

Links

Landscapes

  • Television Signal Processing For Recording (AREA)

Abstract

An embodiment of the invention provides a video processing method and device that are used for adding captions in a recorded video and correcting a defect of technologies of the prior art via which video shooting is homogenized. The video processing method comprises the following steps: when a start-recording-voice trigger event is detected, the recorded video is played and voice information is recorded; when a conclude-recording-voice trigger event is detected, the video is stopped and voice information recording operation is ceased, caption information is obtained after recorded voice information is identified, and the caption information is inserted into the video according to time information of the recorded voice information.

Description

A kind of method for processing video frequency and device
Technical field
The present embodiments relate to video capture technical field, particularly relate to a kind of method for processing video frequency and dress Put.
Background technology
Along with the development of intelligent mobile terminal technology, using mobile terminal to carry out short video capture becomes and works as One of lower popular application.At present, mobile terminal is utilized to carry out the style of shooting of video capture single, i.e. User utilizes the aids such as template, the filter that video capture application program provides to carry out video capture. After video capture completes, user can publish to website and for other users viewing or download, also Can be locally stored at mobile terminal, the video shot cannot be entered by user according to the wish of oneself Row processes, and uses the video of user's shooting to tend to homogeneity.
In actual application, user there may be following demand: adds captions in the video shot, And existing video capture method cannot realize this function, therefore, how to increase in the video of shooting Captions become and utilize mobile terminal to carry out one of video capture technical field technical problem urgently to be resolved hurrily.
Summary of the invention
The embodiment of the present invention provides a kind of method for processing video frequency and device, in order to increase in the video of shooting Captions, solve the defect of video capture homogeneity in prior art.
The embodiment of the present invention provides a kind of method for processing video frequency, including:
Commence play out, when beginning recorded speech trigger event being detected, the video shot to go forward side by side lang Message breath is recorded;
When end recorded speech trigger event being detected, stop playing described video and recorded speech information;
It is identified obtaining caption information to the voice messaging recorded;
Described caption information is inserted in described video by the temporal information according to the voice messaging recorded.
The embodiment of the present invention provides a kind of video process apparatus, including:
Control unit, has shot for commencing play out when beginning recorded speech trigger event being detected Video and carry out voice messaging recording;And when end recorded speech trigger event being detected, stop Play described video and recorded speech information;
Voice recognition unit, for being identified obtaining caption information to the voice messaging recorded;
Caption information inserts unit, is used for the temporal information according to the voice messaging recorded by described captions Information is inserted in described video.
The embodiment of the present invention provides a kind of video processing equipment, and including processor and memorizer, processor can For the program read in memorizer, perform following process: beginning recorded speech trigger event detected Time commence play out the video shot and carry out voice messaging recording;End recorded speech detected During trigger event, stop playing described video and recorded speech information;The voice messaging recorded is known Do not obtain caption information;Described caption information is inserted into institute by the temporal information according to the voice messaging recorded State in video.
The method for processing video frequency of embodiment of the present invention offer and device, after video capture completes, by inciting somebody to action The speech recognition that user records is captions and is inserted in the video shot, and user can be shooting Video adds captions so that the video of shooting more individual character, improves Consumer's Experience.
Accompanying drawing explanation
In order to be illustrated more clearly that the embodiment of the present invention or technical scheme of the prior art, below will be to reality Execute the required accompanying drawing used in example or description of the prior art to be briefly described, it should be apparent that under, Accompanying drawing during face describes is some embodiments of the present invention, for those of ordinary skill in the art, On the premise of not paying creative work, it is also possible to obtain other accompanying drawing according to these accompanying drawings.
Fig. 1 a is that user needs processing completing video after in the embodiment of the present invention, video capture completes Time display interface schematic diagram;
Fig. 1 b is the interface schematic diagram of the voice recording page in the embodiment of the present invention;
Fig. 1 c in the embodiment of the present invention for provide captions on/off switch interface schematic diagram;
Fig. 1 d is the video display effect schematic diagram that with the addition of captions in the embodiment of the present invention;
Fig. 2 is method for processing video frequency flow chart in the embodiment of the present invention;
Fig. 3 is video capture apparatus structure schematic diagram in the embodiment of the present invention.
Detailed description of the invention
For making the purpose of the embodiment of the present invention, technical scheme and advantage clearer, below in conjunction with this Accompanying drawing in bright embodiment, is clearly and completely described the technical scheme in the embodiment of the present invention, Obviously, described embodiment is a part of embodiment of the present invention rather than whole embodiments.Based on Embodiment in the present invention, those of ordinary skill in the art are obtained under not making creative work premise The every other embodiment obtained, broadly falls into the scope of protection of the invention.
Embodiment one
In order to improve Consumer's Experience, in the embodiment of the present invention, user is after shooting completes video, permissible Video for shooting adds captions.As shown in Figure 1a, in the embodiment of the present invention, after video capture completes User needs to complete display interface schematic diagram when video processes, and user needs to complete for shooting Video add captions time, can click on interpolation aside button trigger video processing applications program start.Depending on After frequency processes application program launching, enter the voice recording page, as shown in Figure 1 b, for the voice recording page Interface schematic diagram.User can start to record button by click and start recorded speech, triggers Speech Record Flow process processed, user click on start to record button while, commence play out completed video, user according to The video record voice messaging play.When user needs to terminate voice recording flow process, click on and terminate to record Button, stops playing video simultaneously.
Based on this, as in figure 2 it is shown, the method for processing video frequency embodiment flow process provided for the embodiment of the present invention Figure, may comprise steps of:
S21, detect beginning recorded speech trigger event time commence play out the video shot and carry out Voice messaging is recorded.
Beginning recorded speech trigger event is detected detecting that user determines when clicking on and start to record button, Now commence play out and completed video, and enter voice recording flow process.
S22, when end recorded speech trigger event being detected, stop playing described video and recorded speech Information.
End recorded speech trigger event is detected detecting that user determines when clicking on end recording button, Now stop playing having completed video, terminate voice recording flow process.
S23, to record voice messaging be identified obtaining caption information.
S24, according to the temporal information of voice messaging recorded, the caption information obtained is inserted in video.
During it should be noted that be embodied as, it is also possible to the language recorded while recorded speech information Message breath is identified, and is inserted in video and shows, i.e. step S23, step S24 can be with Step S21 performs simultaneously.
In step S23, can there is a following two embodiment:
The first embodiment, it is identified at network side
Concrete, when end recorded speech trigger event being detected, send voice to network side server Identify request, wherein carry the voice messaging of recording, network side server utilize speech recognition technology Return after the voice messaging received is identified as caption information.
The second embodiment, it is identified in terminal unit this locality
Under this embodiment, need exploit person to be previously written in video processing applications program in advance and turn over Translate storehouse, store when installing video processing applications program to terminal unit local, to identify what user recorded Voice.Being limited by the memory space of terminal unit, the language comprised in the translation storehouse being previously written may be also Imperfect, for example, it is possible to be only only written the language that major part user commonly uses, such as Chinese, English, right Can be not written in some uncommon language, to save the memory space of terminal unit.
It addition, compared with network side recognition method, identify that in terminal unit this locality response speed is relatively fast, User is immediately available for identifying the caption information of correspondence after recording terminates.
In the embodiment that network side is identified, due to the available translated resources of network side server More, therefore, it can be identified as the voice messaging that user records the caption information of different language, and its Recognition result is more accurate.But, it is relatively big by network environment influence, if network environment is preferable, eventually The delay that end equipment obtains caption information is less, if network environment is poor, terminal unit obtains captions letter The delay of breath is bigger.
So that the caption information syncretizing effect that video playback and voice messaging identification obtain is more preferable, specifically During enforcement, video playback and voice recording can be controlled during carrying out subtitle recognition and Tong Bu carry out.Under Face is illustrated by embodiment two.
Embodiment two
Commence play out video when beginning recorded speech trigger event being detected, and obtain video start broadcast Put time point;Stop playing video when end recorded speech trigger event being detected, and obtain video Terminate play time.Based on this when inserting caption information in video, caption information can be inserted In the broadcast start time point obtained and the video terminated between play time.
Such as, commence play out video when beginning recorded speech trigger event being detected, and get video Broadcast start time point be 5s, detect end recorded speech trigger event time stop play video, And the end play time getting video is 15s, then the caption information that will identify that is inserted into video 5s-15s between.
For the recognition result delay issue of network side identification, in the embodiment of the present invention, end detected After recorded speech trigger event, postponing more than predetermined threshold value if obtaining caption information, such as, captions are believed Breath postpones 2s and just obtains, then, when inserting caption information in video, need to increase the time postponed.Example As, user records the voice messaging of 5s, obtains the caption information identified, this situation after 2s Under, it is assumed that the broadcast start time point of acquisition is 5s, and end play time is 10s, to regarding When inserting caption information in Pin, need to increase the delay of 2s, 5s-12s will be inserted into by caption information Between video in.If obtaining caption information to postpone to be not more than predetermined threshold value, then it is negligible, I.e. carry out inserting still according to the broadcast start time point obtained and end play time, Ji Jiangshi The caption information not gone out is inserted in the video between 5s-10s.
Based on implementing the embodiment that in two, video playback is Tong Bu with voice recording, in embodiment three, user is permissible Complete the recording of voice messaging several times, and be identified respectively obtaining captions, according to the voice recording time It is inserted in the video content of correspondence.Based on this, embodiments provide embodiment three.
Embodiment three
The video completed for one section of shooting, user can be inserted into this after recorded speech information several times and regard In Pin.Such as, user have taken the video of one section of 20s, can recorded speech information in four times.Each Section voice messaging is corresponding with one section of video therein.When being embodied as, record one section of voice messaging also After obtaining the caption information of its correspondence, it is inserted into the video of corresponding play time according to recording time point In.For example, it is assumed that the recording time point recording first paragraph voice messaging is 0s-5s, broadcasting of its correspondence Putting time point is 0s-5s, then, after obtaining the captions identified, insert it into 0s-5s's In video.Assume that the recording time point recording second segment voice messaging is 5s-12s, broadcasting of its correspondence Putting time point is 5s-12s, then, after obtaining the captions identified, insert it into 5s-12s Video in, by that analogy.
If in recording process, network environment is poor, for example, it is assumed that record the record of first paragraph voice messaging Time point processed is 0s-5s, and the play time of its correspondence is 0s-5s, 2s after recording terminates Obtain the caption information identified, then the caption information of acquisition be inserted in 0s-7s video, And the beginning recording time of second segment voice messaging is from the beginning of 7s, it is assumed that record second segment voice messaging Recording time point is 7s-15s, and the play time of its correspondence is 7s-15s, then obtain knowledge After the captions not gone out, insert it in the video of 7s-15s.In processing procedure, need all the time Keep recording time point consistent with play time.
User is recording whole voice messagings or during recorded speech information, if used The voice messaging of one section of video has been recorded at family several times, the voice messaging i.e. recorded by least one according to The sub-voice messaging composition of recording time arrangement.User can delete the unsatisfied voice messaging of recording and lay equal stress on New recording, when being embodied as, can implement, according to embodiment four, the mistake that sub-video is deleted and again recorded Journey.
Embodiment four
In the present embodiment, when the arbitrary sub-voice messaging trigger event of deletion being detected, delete corresponding son Voice messaging.Accordingly, user is when deleting the sub-voice messaging recorded, if again detected To when starting recorded speech trigger event, can prompt the user whether again to record deleted sub-voice letter Breath, determines whether again to record deleted sub-voice messaging according to the selection of user.Again record determining During the sub-voice messaging that system is deleted, commence play out from the beginning recording time point being deleted sub-voice messaging Video also re-starts voice messaging recording;And be identified obtaining captions to the voice messaging again recorded The caption information that identification is obtained by information the temporal information according to recorded speech information again inserts this video In.
It is also preferred that the left when again recording deleted sub-voice messaging, the embodiment of the present invention provides following two Kind again record the embodiment of deleted sub-voice messaging:
The first embodiment, records deleted sub-voice messaging the most again.
Under this embodiment, detect again record deleted sub-voice messaging start record language During sound trigger event, commence play out video again from the beginning recording time point being deleted sub-voice messaging Carry out voice messaging recording, until it reaches this stops when being deleted the end recording time point of sub-voice messaging Recorded speech information also stops playing video.
Such as, user is 4 sub-voice messagings of video record of one section of 20s, it is assumed that the son that user deletes The recording time point of voice messaging is 5s-12s, then the video playback time point of its correspondence is 5s- 12s, user, after deleting this sub-voice messaging, commences play out video from play time 5s, Carry out voice recording simultaneously, when play time arrives 12s, stop terminating voice recording and stopping Play video.
The second embodiment, to recording time point user select delete sub-voice messaging after whole Sub-voice messaging is recorded again.
Under this embodiment, the beginning recorded speech again recording deleted sub-voice messaging detected During trigger event, commence play out described video from the beginning recording time point being deleted sub-voice messaging and lay equal stress on Newly carry out voice messaging recording, until it reaches the end recording time of recording time voice messaging the latest Stop recorded speech information during point and stop playing described video.
Such as, user is 4 sub-voice messagings of video record of one section of 20s, it is assumed that the son that user deletes The recording time point of voice messaging is 5s-12s, then the video playback time point of its correspondence is 5s- 12s, user, after deleting this sub-voice messaging, commences play out video from play time 5s, Carry out voice recording simultaneously, when play time arrives 20s, stop terminating voice recording and stopping Play video.Certainly, in recording process, it is corresponding that user can also record 5s-20s several times Voice messaging.Such as, user records 5s-8s for the first time, and second time records 8s-15s, the Record 15s-20s three times.Every time during recorded speech, when video playback time point and voice recording Between point keep consistent.
When being embodied as, in order to improve the accuracy that speech recognition is captions further, user can also divide Section recorded video, is finally each section of video record voice messaging.Such as, 4 shooting 20s of user's part Video, it is assumed that shooting the video segment of 5s, 6s, 5s, 4s, afterwards, user is respectively directed to again every time Each section of video utilizes said method to be that it adds captions.
When being embodied as, it is also possible to captions on/off switch is provided, touch according to the unlatching/closedown detected Captions are opened/closed to the event of sending out.As illustrated in figure 1 c, for providing the interface signal of captions on/off switch Figure, in Fig. 1 c, the lower right corner is the switch that captions are opened/closed, concrete, when switch is shape shown in Fig. 1 c During state, captions can be opening, when switch is triggered and slides to the left, can be closed by captions.As Shown in Fig. 1 d, for the addition of the video display effect schematic diagram of captions.
In the method for processing video frequency that the embodiment of the present invention provides, user can be that the video of oneself shooting adds Captions so that the video shot more has individual character, improves Consumer's Experience.
One is provided based on same inventive concept, embodiment five with the embodiment of the present application one~embodiment four Plant video capture processing means, owing to this device solves principle and the process of above-mentioned video capture of problem Method is similar, and therefore the enforcement of said apparatus may refer to the enforcement of method, repeats no more in place of repetition.
Embodiment five
As it is shown on figure 3, be video capture processing means example structure schematic diagram in the present invention, permissible Including:
Control unit 31, has shot for commencing play out when beginning recorded speech trigger event being detected Become video and carry out voice messaging recording;And when end recorded speech trigger event being detected, stop Only play described video and recorded speech information;
Voice recognition unit 32, for being identified obtaining caption information to the voice messaging recorded;
Caption information inserts unit 33, is used for the temporal information according to the voice messaging recorded by described captions Information is inserted in described video.
When being embodied as, voice recognition unit 32, may be used for sending speech recognition request to network side, Wherein said speech recognition request carries the voice messaging of recording;Receive described network side according to institute's predicate Message ceases the caption information identified;Or the voice messaging of recording is identified as caption information in this locality.
The video process apparatus that the embodiment of the present invention provides, it is also possible to include acquiring unit, wherein:
Acquiring unit, for when beginning recorded speech trigger event being detected, obtaining opening of described video Beginning play time;And when end recorded speech trigger event being detected, obtain the knot of described video Bundle play time;Caption information inserts unit 33, for described caption information is inserted into described beginning In video between play time and end play time.
When being embodied as, the voice messaging of recording is believed according to the sub-voice of recording time arrangement by least one Breath composition.
Based on this, the video process apparatus that the embodiment of the present invention provides can also include:
Delete unit, for when the arbitrary sub-voice messaging trigger event of deletion being detected, deleting correspondence Sub-voice messaging.
Control unit 31, it is also possible to for after deleting the sub-voice messaging that element deletion is corresponding, in inspection When measuring the beginning recorded speech trigger event again recording deleted sub-voice messaging, from deleted son The beginning recording time point of voice messaging commences play out described video and re-starts voice messaging recording;
Voice recognition unit 32, it is also possible to for being identified obtaining captions to the voice messaging again recorded Information;
Caption information inserts unit 33, it is also possible to for the temporal information according to the voice messaging again recorded Caption information identification obtained inserts in described video.
Wherein, control unit 31, specifically for again recording deleted sub-voice messaging detecting When starting recorded speech trigger event, commence play out from the beginning recording time point being deleted sub-voice messaging Described video also re-starts voice messaging recording, until it reaches the described end being deleted sub-voice messaging Stop recorded speech information during recording time point and stop playing described video;Or again record detecting When making the beginning recorded speech trigger event of the sub-voice messaging being deleted, from being deleted sub-voice messaging Start recording time point commence play out described video and re-start voice messaging recording, until it reaches record Stop recorded speech information during the end recording time point of time sub-voice messaging the latest and stop playing institute State video.
For convenience of description, above each several part is divided by function and retouches respectively for each module (or unit) State.Certainly, can be the function of each module (or unit) same or multiple when implementing the present invention Software or hardware realize.When being embodied as, the said equipment identification device can be arranged in terminal unit.
The embodiment of the present invention can be passed through hardware processor (hardware processor) and realize Fig. 3 Shown correlation function.When being embodied as, processor may be used for reading the program in memorizer, Perform following process to realize the correlation function shown in Fig. 3: detect that beginning recorded speech triggers thing Commence play out the video shot during part and carry out voice messaging recording;Detecting that language is recorded in end During sound trigger event, stop playing described video and recorded speech information;The voice messaging recorded is carried out Identification obtains caption information;Described caption information is inserted into by the temporal information according to the voice messaging recorded In described video.
Device embodiment described above is only schematically, wherein said illustrates as separating component Unit can be or may not be physically separate, the parts shown as unit can be or Person may not be physical location, i.e. may be located at a place, or can also be distributed to multiple network On unit.Some or all of module therein can be selected according to the actual needs to realize the present embodiment The purpose of scheme.Those of ordinary skill in the art are not in the case of paying performing creative labour, the most permissible Understand and implement.
Through the above description of the embodiments, those skilled in the art is it can be understood that arrive each reality The mode of executing can add the mode of required general hardware platform by software and realize, naturally it is also possible to by firmly Part.Based on such understanding, the portion that prior art is contributed by technique scheme the most in other words Dividing and can embody with the form of software product, this computer software product can be stored in computer can Read in storage medium, such as ROM/RAM, magnetic disc, CD etc., including some instructions with so that one Computer equipment (can be personal computer, server, or the network equipment etc.) performs each to be implemented The method described in some part of example or embodiment.
Last it is noted that above example is only in order to illustrate technical scheme, rather than to it Limit;Although the present invention being described in detail with reference to previous embodiment, the ordinary skill of this area Personnel it is understood that the technical scheme described in foregoing embodiments still can be modified by it, or Person carries out equivalent to wherein portion of techniques feature;And these amendments or replacement, do not make corresponding skill The essence of art scheme departs from the spirit and scope of various embodiments of the present invention technical scheme.

Claims (14)

1. a method for processing video frequency, it is characterised in that including:
Commence play out, when beginning recorded speech trigger event being detected, the video shot to go forward side by side lang Message breath is recorded;
When end recorded speech trigger event being detected, stop playing described video and recorded speech information;
It is identified obtaining caption information to the voice messaging recorded;
Described caption information is inserted in described video by the temporal information according to the voice messaging recorded.
Method the most according to claim 1, it is characterised in that the voice messaging recorded is known The caption information not obtained, specifically includes:
Sending speech recognition request to network side, wherein said speech recognition request carries the voice of recording Information;Receive the caption information that described network side identifies according to described voice messaging;Or
The voice messaging of recording is identified as caption information in this locality.
Method the most according to claim 1, it is characterised in that also include:
When beginning recorded speech trigger event being detected, obtain the broadcast start time point of described video;
When end recorded speech trigger event being detected, obtain the end play time of described video; And
Described caption information is inserted in described video by the temporal information according to the voice messaging recorded, tool Body includes:
Described caption information is inserted into described broadcast start time point and terminates regarding between play time In Pin.
Method the most according to claim 1, it is characterised in that the voice messaging of recording is by least one The individual sub-voice messaging composition according to recording time arrangement.
Method the most according to claim 4, it is characterised in that described method also includes: in detection To when deleting arbitrary sub-voice messaging trigger event, delete corresponding sub-voice messaging.
Method the most according to claim 5, it is characterised in that the sub-voice letter that described deletion is corresponding Include after breath:
When the beginning recorded speech trigger event again recording deleted sub-voice messaging being detected, from The beginning recording time point being deleted sub-voice messaging commences play out described video and re-starts voice messaging Record;
It is identified the voice messaging again recorded obtaining caption information and according to recorded speech information again Temporal information caption information that identification is obtained insert in described video.
Method the most according to claim 6, it is characterised in that from being deleted opening of sub-voice messaging Beginning recording time point commences play out described video and re-starts voice messaging recording, specifically includes:
When the beginning recorded speech trigger event again recording deleted sub-voice messaging being detected, from The beginning recording time point being deleted sub-voice messaging commences play out described video and re-starts voice messaging Record, until it reaches described be deleted sub-voice messaging end recording time point time stop recorded speech letter Cease and stop playing described video;Or
When the beginning recorded speech trigger event again recording deleted sub-voice messaging being detected, from The beginning recording time point being deleted sub-voice messaging commences play out described video and re-starts voice messaging Record, until it reaches stop during the end recording time point of recording time voice messaging the latest recording language Message ceases and stops playing described video.
8. a video process apparatus, it is characterised in that including:
Control unit, has shot for commencing play out when beginning recorded speech trigger event being detected Video and carry out voice messaging recording;And when end recorded speech trigger event being detected, stop Play described video and recorded speech information;
Voice recognition unit, for being identified obtaining caption information to the voice messaging recorded;
Caption information inserts unit, for being believed by described captions according to the temporal information of the voice messaging recorded Breath is inserted in described video.
Device the most according to claim 8, it is characterised in that
Described voice recognition unit, specifically for sending speech recognition request, wherein said language to network side Sound identification request carries the voice messaging of recording;Receive described network side according to described voice messaging identification The caption information gone out;Or the voice messaging of recording is identified as caption information in this locality.
Device the most according to claim 8, it is characterised in that also include acquiring unit, wherein:
Described acquiring unit, for when beginning recorded speech trigger event being detected, obtaining described video Broadcast start time point;And when end recorded speech trigger event being detected, obtain described video End play time;
Described caption information inserts unit, for described caption information is inserted into described broadcast start time In video between point and end play time.
11. devices according to claim 8, it is characterised in that the voice messaging of recording is by least One sub-voice messaging composition according to recording time arrangement.
12. devices according to claim 11, it is characterised in that also include:
Delete unit, for when the arbitrary sub-voice messaging trigger event of deletion being detected, deleting correspondence Sub-voice messaging.
13. devices according to claim 12, it is characterised in that
Described control unit, is additionally operable to after the sub-voice messaging that described deletion element deletion is corresponding, When the beginning recorded speech trigger event again recording deleted sub-voice messaging being detected, from being deleted The beginning recording time point of sub-voice messaging commences play out described video and re-starts voice messaging recording;
Described voice recognition unit, is additionally operable to be identified the voice messaging again recorded obtaining captions letter Breath;
Described caption information inserts unit, and being additionally operable to the temporal information according to the voice messaging again recorded will Identify that the caption information obtained inserts in described video.
14. devices according to claim 13, it is characterised in that
Described control unit, specifically for detecting the beginning again recording deleted sub-voice messaging During recorded speech trigger event, commence play out described from the beginning recording time point being deleted sub-voice messaging Video also re-starts voice messaging recording, until it reaches the described end being deleted sub-voice messaging is recorded Stop recorded speech information during time point and stop playing described video;Or again record quilt detecting During the beginning recorded speech trigger event of sub-voice messaging deleted, from the beginning of being deleted sub-voice messaging Recording time point commences play out described video and re-starts voice messaging recording, until it reaches recording time Stop during the end recording time point of sub-voice messaging the latest recorded speech information stopping play described in regard Frequently.
CN201510511387.1A 2015-08-19 2015-08-19 Video processing method and device Pending CN105898169A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510511387.1A CN105898169A (en) 2015-08-19 2015-08-19 Video processing method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510511387.1A CN105898169A (en) 2015-08-19 2015-08-19 Video processing method and device

Publications (1)

Publication Number Publication Date
CN105898169A true CN105898169A (en) 2016-08-24

Family

ID=57002778

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510511387.1A Pending CN105898169A (en) 2015-08-19 2015-08-19 Video processing method and device

Country Status (1)

Country Link
CN (1) CN105898169A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106506335A (en) * 2016-11-10 2017-03-15 北京小米移动软件有限公司 Method and device for sharing video files
CN106851401A (en) * 2017-03-20 2017-06-13 惠州Tcl移动通信有限公司 A kind of method and system of automatic addition captions
CN109495792A (en) * 2018-11-30 2019-03-19 北京字节跳动网络技术有限公司 A kind of subtitle adding method, device, electronic equipment and the readable medium of video
CN110166816A (en) * 2019-05-29 2019-08-23 上海乂学教育科技有限公司 The video editing method and system based on speech recognition for artificial intelligence education

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102148939A (en) * 2010-02-09 2011-08-10 佛山市南海海信多媒体技术有限公司 Method, device and television for real-time displaying subtitles of television program
CN103414949A (en) * 2013-08-09 2013-11-27 冠捷显示科技(厦门)有限公司 Multimedia editing system and method based on smart television
CN103458321A (en) * 2012-06-04 2013-12-18 联想(北京)有限公司 Method and device for loading subtitles
WO2014162310A4 (en) * 2013-04-02 2014-11-20 Nir Igal Automatic generation of a database for speech recognition from video captions
CN104202542A (en) * 2014-08-28 2014-12-10 深圳市大疆创新科技有限公司 Automatic subtitle generating method and device for video camera

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102148939A (en) * 2010-02-09 2011-08-10 佛山市南海海信多媒体技术有限公司 Method, device and television for real-time displaying subtitles of television program
CN103458321A (en) * 2012-06-04 2013-12-18 联想(北京)有限公司 Method and device for loading subtitles
WO2014162310A4 (en) * 2013-04-02 2014-11-20 Nir Igal Automatic generation of a database for speech recognition from video captions
CN103414949A (en) * 2013-08-09 2013-11-27 冠捷显示科技(厦门)有限公司 Multimedia editing system and method based on smart television
CN104202542A (en) * 2014-08-28 2014-12-10 深圳市大疆创新科技有限公司 Automatic subtitle generating method and device for video camera

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106506335A (en) * 2016-11-10 2017-03-15 北京小米移动软件有限公司 Method and device for sharing video files
CN106506335B (en) * 2016-11-10 2019-08-30 北京小米移动软件有限公司 Method and device for sharing video files
CN106851401A (en) * 2017-03-20 2017-06-13 惠州Tcl移动通信有限公司 A kind of method and system of automatic addition captions
CN109495792A (en) * 2018-11-30 2019-03-19 北京字节跳动网络技术有限公司 A kind of subtitle adding method, device, electronic equipment and the readable medium of video
CN110166816A (en) * 2019-05-29 2019-08-23 上海乂学教育科技有限公司 The video editing method and system based on speech recognition for artificial intelligence education

Similar Documents

Publication Publication Date Title
CN108924586B (en) Video frame detection method and device and electronic equipment
WO2019109643A1 (en) Video recommendation method and apparatus, and computer device and storage medium
US9881215B2 (en) Apparatus and method for identifying a still image contained in moving image contents
US20140064693A1 (en) Method and System for Video Event Detection for Contextual Annotation and Synchronization
CN111091811B (en) Method and device for processing voice training data and storage medium
US10645468B1 (en) Systems and methods for providing video segments
CN108008898B (en) Page information acquisition method and device, computer equipment and storage medium
WO2021002965A1 (en) Highlights video player
CN103440330A (en) Music program information acquisition method and equipment
CN108966016B (en) Video clip rebroadcasting method and device and terminal equipment
CN104995639A (en) Terminal and method for managing video file
CN109618224A (en) Video data processing method, device, computer readable storage medium and equipment
CN105898169A (en) Video processing method and device
CN112887480A (en) Audio signal processing method and device, electronic equipment and readable storage medium
CN112380922A (en) Method and device for determining compound video frame, computer equipment and storage medium
CN115080792A (en) Video association method and device, electronic equipment and storage medium
CN117768430A (en) Method, device and equipment for sending and receiving live pictures in session
JP2025515613A (en) Method, apparatus, device and storage medium for audio editing - Patents.com
CN105808231A (en) System and method for recording script and system and method for playing script
CN111147778B (en) Track recording method and device based on digital video recorder and storage medium
CN112843696A (en) Shooting method, shooting device, electronic equipment and storage medium
CN103309993B (en) The extracting method of a kind of key word and device
EP4354883A1 (en) Video processing method and apparatus, and electronic device
CN114363720B (en) Video slicing method, system, equipment and medium based on computer vision
KR102066750B1 (en) Terminal apparatus and method for controlling record file

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160824

WD01 Invention patent application deemed withdrawn after publication