TWM664095U - A system for music teaching - Google Patents
A system for music teaching Download PDFInfo
- Publication number
- TWM664095U TWM664095U TW113203432U TW113203432U TWM664095U TW M664095 U TWM664095 U TW M664095U TW 113203432 U TW113203432 U TW 113203432U TW 113203432 U TW113203432 U TW 113203432U TW M664095 U TWM664095 U TW M664095U
- Authority
- TW
- Taiwan
- Prior art keywords
- practitioner
- musical instrument
- posture
- practice
- module
- Prior art date
Links
- 238000012937 correction Methods 0.000 claims abstract description 19
- 238000012549 training Methods 0.000 claims description 51
- 230000004044 response Effects 0.000 claims description 11
- 238000013473 artificial intelligence Methods 0.000 claims description 5
- 230000000007 visual effect Effects 0.000 claims description 3
- 238000000034 method Methods 0.000 abstract description 43
- 230000008569 process Effects 0.000 abstract description 21
- 238000005516 engineering process Methods 0.000 abstract description 11
- 230000036544 posture Effects 0.000 description 92
- 210000003811 finger Anatomy 0.000 description 8
- 238000003384 imaging method Methods 0.000 description 8
- 238000001514 detection method Methods 0.000 description 7
- 230000033001 locomotion Effects 0.000 description 7
- 238000013135 deep learning Methods 0.000 description 6
- 210000000707 wrist Anatomy 0.000 description 6
- 239000000203 mixture Substances 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 210000001513 elbow Anatomy 0.000 description 4
- 210000001145 finger joint Anatomy 0.000 description 4
- 238000009434 installation Methods 0.000 description 4
- 101000827703 Homo sapiens Polyphosphoinositide phosphatase Proteins 0.000 description 3
- 102100023591 Polyphosphoinositide phosphatase Human genes 0.000 description 3
- 210000003423 ankle Anatomy 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 210000003127 knee Anatomy 0.000 description 3
- 238000009527 percussion Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 210000002832 shoulder Anatomy 0.000 description 3
- 230000009471 action Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 238000007405 data analysis Methods 0.000 description 2
- 238000013434 data augmentation Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 210000004247 hand Anatomy 0.000 description 2
- 210000001624 hip Anatomy 0.000 description 2
- 230000000750 progressive effect Effects 0.000 description 2
- 101001121408 Homo sapiens L-amino-acid oxidase Proteins 0.000 description 1
- 102100026388 L-amino-acid oxidase Human genes 0.000 description 1
- 101100012902 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) FIG2 gene Proteins 0.000 description 1
- 101100233916 Saccharomyces cerevisiae (strain ATCC 204508 / S288c) KAR5 gene Proteins 0.000 description 1
- 230000000903 blocking effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 210000002478 hand joint Anatomy 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 208000037805 labour Diseases 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000003739 neck Anatomy 0.000 description 1
- 238000004886 process control Methods 0.000 description 1
- 238000012827 research and development Methods 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
- 210000001364 upper extremity Anatomy 0.000 description 1
Landscapes
- Auxiliary Devices For Music (AREA)
Abstract
本創作公開了一種音樂教學的系統,包括影像模塊,樂器識別模塊,人體姿態估計模塊;所述影像模塊用於獲取練習者的練習影像數據;所述樂器識別模塊用於從所述練習影像數據中獲取樂器數據信息以及與所述練習者相關的人體數據信息;所述人體姿態估計模塊用於根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型;根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。本創作採用人體姿態信息和樂器信息相結合的方式,解決了現有技術中不能在練習過程中呈現練習者的人體姿態情況,進而實現對練習過程中人體姿態的校正的問題。This invention discloses a music teaching system, including an image module, a musical instrument recognition module, and a human posture estimation module; the image module is used to obtain the practice image data of the practitioner; the musical instrument recognition module is used to obtain musical instrument data information and human body data information related to the practitioner from the practice image data; the human body posture estimation module is used to determine the practice type of the practitioner based on the human body data information and the musical instrument data information; the human body data information is marked according to the practice type to obtain the practice posture of the practitioner. This creation adopts a method of combining human body posture information and musical instrument information to solve the problem that the existing technology cannot present the practitioner's human body posture during the practice process, and thus realizes the correction of human body posture during the practice process.
Description
本創作屬於音樂教學領域,具體涉及一種用於音樂教學的系統。 This work belongs to the field of music teaching, specifically involving a system used for music teaching.
目前的音樂教學設備例如鍵盤類樂器、管弦樂、打擊樂、彈撥樂器等教學系統是基於傳感器對琴弦、琴鍵進行彈撥、對鼓面進行按壓或敲擊的位置,使用力度等因素,把練習者的曲目演奏情況和曲譜進行對比,以判斷練習者的彈奏好壞。這種方法實際上就是根據練習者使用身體(如手指)或借助其他器具與樂器接觸,把力量施加在樂器後,導致樂器發出聲音,是把音色、音階與樂譜上的音符在比較,這種比較方法無法全面反映練習者在練習過程中人體姿態相對於樂器的位置是否正確,因而無法對練習者為何出現演奏錯誤或動作不到位進行根本性原因的發現,彈奏時學生的彈奏姿勢常常被忽略,錯誤的彈奏姿勢會對學生的身體造成很大損傷。對於練習者在彈奏過程中出現的問題,需要老師或者設備給出有效的校正指示,進而改善演奏。 Current music teaching equipment, such as keyboard instruments, orchestral instruments, percussion instruments, and plucked instruments, uses sensors to detect the position and force of the strings and keys plucked, the drums pressed or struck, and the like, and compares the student's performance with the music score to determine whether the student is playing well or not. This method is actually based on the fact that the practitioner uses the body (such as fingers) or other instruments to contact the instrument, and then applies force to the instrument, causing the instrument to make a sound. It is to compare the timbre, scale and notes on the score. This comparison method cannot fully reflect whether the position of the practitioner's body posture relative to the instrument during the practice process is correct, so it is impossible to find the fundamental reasons why the practitioner makes playing errors or inadequate movements. The playing posture of students is often ignored when playing, and the wrong playing posture will cause great damage to the student's body. For the problems that the practitioner encounters during the playing process, the teacher or equipment needs to give effective correction instructions to improve the performance.
為了解決上述問題,有鑒於此,亟需提供一種用於音樂教學的系統,以便實現在練習過程中結合樂器類型呈現練習者對應於樂器類型的人體姿態情況。 In order to solve the above problems, it is urgent to provide a system for music teaching so as to realize the combination of musical instrument types and present the body posture of the practitioner corresponding to the musical instrument type during the practice process.
有鑑於此,本創作人特地針對可調整練習者的練習姿態之音樂教學系統加以研究及改良,期以一較佳設計改善上述問題,並在經過長期研發及不斷測試後,始有本創作之問世。 In view of this, the author of this invention has specially researched and improved the music teaching system that can adjust the practitioner's practice posture, hoping to improve the above problems with a better design. After long-term research and development and continuous testing, this creation was finally released.
為了至少解決上述問題,本創作提出了一種用於音樂教學系統,包括:影像模塊,樂器識別模塊,人體姿態估計模塊;所述影像模塊用於獲取練習者的練習影像數據;所述樂器識別模塊用於獲取所述練習影像數據中的一種或多種樂器數據信息;所述人體姿態估計模塊用於從所述練習影像數據中獲取與所述練習者相關的人體數據信息;根據所述人體數據信息和所述樂器數據信息,判斷練習者的練習類型;根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。 In order to at least solve the above problems, this invention proposes a music teaching system, including: an image module, a musical instrument recognition module, and a human posture estimation module; the image module is used to obtain the practice image data of the practitioner; the musical instrument recognition module is used to obtain one or more musical instrument data information in the practice image data; the human posture estimation module is used to obtain human data information related to the practitioner from the practice image data; the practice type of the practitioner is determined according to the human data information and the musical instrument data information; the human data information is marked according to the practice type to obtain the practice posture of the practitioner.
在第一方面中,本創作提供一種用於音樂教學系統,包括影像模塊,樂器識別模塊,人體姿態估計模塊;所述影像模塊用於獲取練習者的練習影像數據;所述樂器識別模塊用於獲取所述練習影像數據中的一種或多種樂器數據信息;所述人體姿態估計模塊用於從所述練習影像數據中獲取與所述練習者相關的人體數據信息;根據所述人體數據信息和所述樂器數據信息,判斷練習者的練習類型;根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。 In the first aspect, the invention provides a music teaching system, including an image module, a musical instrument identification module, and a human posture estimation module; the image module is used to obtain the practice image data of the practitioner; the musical instrument identification module is used to obtain one or more musical instrument data information in the practice image data; the human posture estimation module is used to obtain human data information related to the practitioner from the practice image data; the practice type of the practitioner is determined according to the human data information and the musical instrument data information; the human data information is marked according to the practice type to obtain the practice posture of the practitioner.
可選地,將所述練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示。 Optionally, the exercise posture of the practitioner is compared with the image reference data or the exercise template image data to determine whether the exercise posture of the practitioner is correct; and in response to determining that the exercise posture of the practitioner is incorrect, a correction instruction for correcting the exercise posture of the practitioner is output.
可選地,還包括:影像參照模塊和判斷模塊;所述影像參照模塊用於獲取影像參照數據或練習者的練習模板影像數據;所述判斷模塊用於將所述練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示。 Optionally, it further includes: an image reference module and a judgment module; the image reference module is used to obtain image reference data or training template image data of the trainee; the judgment module is used to compare the training posture of the trainee with the image reference data or the training template image data to judge whether the training posture of the trainee is correct; and in response to judging that the training posture of the trainee is incorrect, output a correction instruction for correcting the training posture of the trainee.
可選地,所述練習影像數據至少包括可見光影像數據、不可見光影像數據或聲納影像數據中的一種或多種。 Optionally, the training image data includes at least one or more of visible light image data, invisible light image data or sonar image data.
可選地,根據所述樂器數據信息對樂器類型進行分類以形成樂器分類庫;以及根據所述樂器分類信息將所述樂器數據信息存儲到相應的樂器分類庫。 Optionally, the types of musical instruments are classified according to the musical instrument data information to form a musical instrument classification library; and the musical instrument data information is stored in the corresponding musical instrument classification library according to the musical instrument classification information.
可選地,其中根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型還包括:根據所述樂器數據信息和所述人體數據信息進行判定,定位練習者與練習姿態相關的關節點位置;基於關節點位置判斷練習者的類型。 Optionally, judging the practice type of the practitioner according to the human body data information and the musical instrument data information further includes: determining according to the musical instrument data information and the human body data information, locating the joint positions of the practitioner related to the practice posture; and judging the type of the practitioner based on the joint position.
可選地,所述練習影像數據還包括階段性的練習模板影像數據,其中所述階段性與時間和/或練習難度相關。 Optionally, the training image data also includes phased training template image data, wherein the phases are related to time and/or training difficulty.
可選地,通過執行人工智能連續影像對比來獲得所述練習影像數據與影像參照數據的對比結果;基於所述對比結果判定練習者的練習優劣度;以及基於所述練習優劣度輸出針對練習者的姿態校正建議。 Optionally, the comparison result of the training image data and the image reference data is obtained by executing artificial intelligence continuous image comparison; the training quality of the trainee is determined based on the comparison result; and posture correction suggestions for the trainee are output based on the training quality.
可選地,還包括以視覺和/或聽覺的方式輸出所述校正指示。 Optionally, it also includes outputting the correction instruction in a visual and/or auditory manner.
在第二方面中,本創作提供一種用於音樂教學方法,包括獲取練習者的練習影像數據;從所述練習影像數據中獲取樂器數據信息以及與所述練 習者相關的人體數據信息;根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型;根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。 In the second aspect, the invention provides a method for music teaching, including obtaining a practitioner's practice video data; obtaining musical instrument data information and human body data information related to the practitioner from the practice video data; judging the practice type of the practitioner based on the human body data information and the musical instrument data information; marking the human body data information according to the practice type to obtain the practice posture of the practitioner.
可選地,將所述練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示。 Optionally, the exercise posture of the practitioner is compared with the image reference data or the exercise template image data to determine whether the exercise posture of the practitioner is correct; and in response to determining that the exercise posture of the practitioner is incorrect, a correction instruction for correcting the exercise posture of the practitioner is output.
可選地,所述練習影像數據至少包括可見光影像數據、不可見光影像數據或聲納影像數據中的一種或多種。 Optionally, the training image data includes at least one or more of visible light image data, invisible light image data or sonar image data.
可選地,根據所述樂器數據信息對樂器類型進行分類以形成樂器分類庫;以及根據所述樂器分類信息將所述樂器數據信息存儲到相應的樂器分類庫。 Optionally, the types of musical instruments are classified according to the musical instrument data information to form a musical instrument classification library; and the musical instrument data information is stored in the corresponding musical instrument classification library according to the musical instrument classification information.
可選地,其中根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型還包括:根據所述樂器數據信息和所述人體數據信息進行判定,定位練習者與練習姿態相關的關節點位置;基於關節點位置判斷練習者的類型。 Optionally, judging the practice type of the practitioner according to the human body data information and the musical instrument data information further includes: determining according to the musical instrument data information and the human body data information, locating the joint positions of the practitioner related to the practice posture; and judging the type of the practitioner based on the joint position.
可選地,所述練習影像數據還包括階段性的練習模板影像數據,其中所述階段性與時間和/或練習難度相關。 Optionally, the training image data also includes phased training template image data, wherein the phases are related to time and/or training difficulty.
可選地,通過執行人工智能連續影像對比來獲得所述練習影像數據與影像參照數據的對比結果;基於所述對比結果判定練習者的練習優劣度;以及基於所述練習優劣度輸出針對練習者的姿態校正建議。 Optionally, the comparison result of the training image data and the image reference data is obtained by executing artificial intelligence continuous image comparison; the training quality of the trainee is determined based on the comparison result; and posture correction suggestions for the trainee are output based on the training quality.
可選地,還包括以視覺和/或聽覺的方式輸出所述校正指示。 Optionally, it also includes outputting the correction instruction in a visual and/or auditory manner.
與現有技術相比,本創作具有以下優點: 本創作提供一種用於音樂教學的系統,本創作通過在練習者練習過程中獲取練習影像數據,從練習影像數據中提取樂器數據信息和練習者的人體數據信息;結合人體數據信息和樂器數據信息對練習者的練習類型進行判斷,依據練習類型對人體數據信息進行標記,以獲得練習者對應於練習者正在使用的練習樂器的練習姿態,從而便於將練習姿態情況的影像進行存儲、比對、或通過數據傳輸呈現給遠程的老師。進一步地,通過將練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示,進而實現可以智能、直觀地給出對應於練習者所使用的器材的練習姿勢的判斷和校正建議。 Compared with the existing technology, this invention has the following advantages: This invention provides a system for music teaching. This invention obtains the practice image data of the practitioner during the practice process, extracts the instrument data information and the body data information of the practitioner from the practice image data; judges the practice type of the practitioner by combining the body data information and the instrument data information, and marks the body data information according to the practice type to obtain the practice posture of the practitioner corresponding to the practice instrument that the practitioner is using, so as to facilitate the storage, comparison, or presentation of the image of the practice posture to a remote teacher through data transmission. Furthermore, by comparing the exercise posture of the practitioner with the image reference data or the exercise template image data, it is determined whether the exercise posture of the practitioner is correct; and in response to determining that the exercise posture of the practitioner is incorrect, a correction instruction for correcting the exercise posture of the practitioner is output, thereby realizing the intelligent and intuitive provision of the judgment and correction suggestions corresponding to the exercise posture of the equipment used by the practitioner.
〔本創作〕 [Original work]
100、200:方法 100, 200: Method
300:系統 300:System
301:影像模塊 301: Image module
302:樂器識別模塊 302: Musical instrument identification module
303:人體姿態估計模塊 303: Human body posture estimation module
304:影像參照模塊 304: Image reference module
305:判斷模塊 305: Judgment module
500:設備組成 500: Equipment composition
501:主機 501:Host
502:第一攝像頭 502: First Camera
503:第二攝像頭 503: Second camera
504:琴面 504: piano surface
505:固定杆 505:Fixed rod
S101、S102、S103、S104、S201、S202、S203、S204、S205:步驟 S101, S102, S103, S104, S201, S202, S203, S204, S205: Steps
通過參考附圖閱讀下文的詳細描述,本創作示例性實施方式的上述以及其他目的、特徵和優點將變得易於理解。在附圖中,以示例性而非限制性的方式示出了本創作的若干實施方式,並且相同或對應的標號表示相同或對應的部分,其中:[圖1]示出了本創作第一實施例的用於音樂教學的方法的示例性流程圖。 By reading the detailed description below with reference to the attached figures, the above and other purposes, features and advantages of the exemplary implementation of the present invention will become easy to understand. In the attached figures, several implementations of the present invention are shown in an exemplary and non-restrictive manner, and the same or corresponding numbers represent the same or corresponding parts, among which: [Figure 1] shows an exemplary flow chart of the method for music teaching of the first embodiment of the present invention.
[圖2]示出了本創作第二實施例的用於音樂教學的方法的示例性流程圖。 [Figure 2] shows an exemplary flow chart of a method for music teaching in the second embodiment of the present invention.
[圖3]示出了本創作第三實施例的用於音樂教學的系統的方塊示意圖。 [Figure 3] shows a block diagram of a system for music teaching in the third embodiment of the present invention.
[圖4]示出了本創作實施例的音樂練習者人體姿態關節點位識別圖像特徵示範。 [Figure 4] shows a demonstration of the image features of the joint position recognition of a music practitioner's human body posture in the present invention embodiment.
[圖4a]示出了本創作實施例的根據圖4提取的關鍵關節點位圖示範性示例。 [Figure 4a] shows an exemplary example of a key node bitmap extracted according to Figure 4 of the present invention.
[圖5a]示出了本創作一種實施例的影像模塊的設備組成的示範性示例。 [Figure 5a] shows an exemplary example of the device composition of the imaging module of one embodiment of the present invention.
[圖5b]示出了本創作另外一種實施例的影像模塊的設備組成的示範性示例。 [Figure 5b] shows an exemplary example of the device composition of the imaging module of another embodiment of the present invention.
[圖5c]示出了本創作另外一種實施例的一種用於音樂教學的電子設備的示範性示例。 [Figure 5c] shows an exemplary example of an electronic device for music teaching in another embodiment of the present invention.
本創作係一種用於音樂教學的系統,其實施手段、特點及其功效,茲舉數種較佳可行實施例並配合圖式於下文進行詳細說明,俾供 鈞上深入瞭解並認同本創作。 This creation is a system for music teaching. Its implementation methods, features and effects are described in detail below with several preferred feasible embodiments and diagrams, so that you can have a deeper understanding and agree with this creation.
下面將結合本創作實施例中的附圖,對本創作實施例中的技術方案進行清楚、完整地描述,顯然,所描述的實施例是本創作一部分實施例,而不是全部的實施例。基於本創作中的實施例,本創作所屬技術領域中具有通常知識者在沒有做出進步性勞動前提下所獲得的所有其他實施例,都屬於本創作保護的範圍。 The following will combine the attached figures in the embodiments of this invention to clearly and completely describe the technical solutions in the embodiments of this invention. Obviously, the described embodiments are part of the embodiments of this invention, not all of them. Based on the embodiments in this invention, all other embodiments obtained by a person with ordinary knowledge in the technical field to which this invention belongs without making progressive labor are within the scope of protection of this invention.
應當理解,本創作的說明書和申請專利範圍中使用的術語“包括”和“包含”指示所描述特徵、整體、步驟、操作、元素和/或組件的存在,但並不排除一個或多個其它特徵、整體、步驟、操作、元素、組件和/或其集合的存在或添加。 It should be understood that the terms "include" and "comprising" used in the description and patent application of this invention indicate the existence of the described features, wholes, steps, operations, elements and/or components, but do not exclude the existence or addition of one or more other features, wholes, steps, operations, elements, components and/or their collections.
還應當理解,在此本創作說明書中所使用的術語僅僅是出於描述特定實施例的目的,而並不意在限定本創作。如在本創作說明書和申請專利範圍中所使用的那樣,除非上下文清楚地指明其它情況,否則單數形式的“一”、“一個”及“該”意在包括複數形式。還應當進一步理解,在本創作說明書和 申請專利範圍中使用的術語“和/或”是指相關聯列出的項中的一個或多個的任何組合以及所有可能組合,並且包括這些組合。 It should also be understood that the terms used in this invention specification are only for the purpose of describing specific embodiments and are not intended to limit the invention. As used in this invention specification and the scope of the patent application, the singular forms "a", "an" and "the" are intended to include plural forms unless the context clearly indicates otherwise. It should also be further understood that the term "and/or" used in this invention specification and the scope of the patent application refers to any combination and all possible combinations of one or more of the associated listed items, and includes these combinations.
如在本說明書和申請專利範圍中所使用的那樣,術語“如果”可以依據上下文被解釋為“當...時”或“一旦”或“響應於確定”或“響應於檢測到”。類似地,短語“如果確定”或“如果檢測到[所描述條件或事件]”可以依據上下文被解釋為意指“一旦確定”或“響應於確定”或“一旦檢測到[所描述條件或事件]”或“響應於檢測到[所描述條件或事件]”。 As used in this specification and claims, the term "if" may be interpreted as "when" or "upon" or "in response to determining" or "in response to detecting", depending on the context. Similarly, the phrase "if it is determined" or "if [described condition or event] is detected" may be interpreted as meaning "upon determination" or "in response to determining" or "upon detection of [described condition or event]" or "in response to detecting [described condition or event]", depending on the context.
以下結合附圖對本創作的優選實施例進行說明,應當理解,此處所描述的優選實施例僅用於說明和解釋本公開,並不用於限定本公開。 The following is a description of the preferred embodiments of this invention in conjunction with the attached figures. It should be understood that the preferred embodiments described here are only used to illustrate and explain this disclosure, and are not used to limit this disclosure.
針對現有技術中不能直觀呈現練習者人體姿態的問題,本創作的第一實施例提供一種用於音樂教學的方法100,下面參照附圖1詳細描述。 In view of the problem that the existing technology cannot intuitively present the body posture of the practitioner, the first embodiment of this creation provides a method 100 for music teaching, which is described in detail below with reference to Figure 1.
如圖1所示,一種用於音樂教學的方法100包括:在步驟S101中:獲取練習者的練習影像數據。 As shown in FIG1 , a method 100 for music teaching includes: in step S101: obtaining the practice image data of the practitioner.
優選地,所述練習影像數至少包括可見光影像數據、不可見光影像數據或聲納影像數據中的一種或多種。其中可見光影像可以是普通照相機或者攝像機等攝影設備拍攝的影像,不可見光影像數據是指紅外相機、紅外攝像機等拍攝的影像。 Preferably, the training image data includes at least one or more of visible light image data, invisible light image data or sonar image data. The visible light image may be an image taken by a photographic device such as an ordinary camera or video camera, and the invisible light image data refers to an image taken by an infrared camera, infrared video camera, etc.
在步驟S102中:從所述練習影像數據中獲取樂器數據信息以及與所述練習者相關的人體數據信息。 In step S102: musical instrument data information and human body data information related to the practitioner are obtained from the practice image data.
優選地,根據所述樂器數據信息對樂器類型進行分類以形成樂器分類庫;以及根據所述樂器分類信息將所述樂器數據信息存儲到相應的樂器分類庫。 Preferably, the types of musical instruments are classified according to the musical instrument data information to form a musical instrument classification library; and the musical instrument data information is stored in the corresponding musical instrument classification library according to the musical instrument classification information.
其中獲取樂器數據信息可以包括一種或多種樂器,根據樂器的外形或者音色信息對練習影像數據中的樂器進行分類形成樂器分類庫,例如樂器分類庫可以定義為管弦樂器、鍵盤樂器、打擊樂器等等。 The acquired instrument data information may include one or more instruments. The instruments in the practice image data are classified according to the appearance or timbre information of the instruments to form an instrument classification library. For example, the instrument classification library may be defined as orchestral instruments, keyboard instruments, percussion instruments, etc.
進一步地,還包括根據練習者需求添加練習者自定義的樂器分類庫。例如,練習者可以根據需求對樂器進行改進,形成新的樂器類型,進而形成新的樂器分類庫。練習者也可以根據需求對已有的樂器分類庫進行分類的重新組合。 Furthermore, it also includes adding a practitioner-defined instrument classification library according to the practitioner's needs. For example, the practitioner can improve the instrument according to the needs, form a new instrument type, and then form a new instrument classification library. The practitioner can also re-classify the existing instrument classification library according to the needs.
在步驟S103中:根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型。 In step S103: judging the practitioner's practice type based on the human body data information and the musical instrument data information.
優選地,根據所述樂器數據信息和所述人體數據信息進行判定,定位練習者與練習姿態相關的關節點位置;基於關節點位置判斷練習者的類型。 Preferably, the position of the joints of the practitioner related to the practice posture is located based on the musical instrument data information and the human body data information; and the type of the practitioner is determined based on the joint position.
其中,對關節點位置定位後形成關節點位的識別圖像,關節點位的識別圖像特徵示範請參考圖4。所述關節點位的識別圖像基於人體姿態估計得出,人體姿態估計是指通過計算機算法在圖像或視頻中定位人體關鍵點(如肩、肘、腕、髖膝、膝、踝,手指等)。 Among them, after locating the joint positions, an identification image of the joint points is formed. Please refer to Figure 4 for the demonstration of the characteristics of the identification image of the joint points. The identification image of the joint points is obtained based on human posture estimation. Human posture estimation refers to locating key points of the human body (such as shoulders, elbows, wrists, hips, knees, ankles, fingers, etc.) in images or videos through computer algorithms.
在步驟S104中:根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態;下面結合圖4中進行烏克麗麗練習的小女孩的圖像以及圖4a示出的上述小女孩的關節點識別標識,舉例說明上述方法100的具體實現過程:根據步驟S101,通過影像設備獲取烏克麗麗練習者或者訓練過程的圖像或視頻,圖像的獲取可以採用接入網路的攝像機或者手機拍攝,拍攝的 影像通過網路上傳到雲端服務器進行後續的數據分析,或者採用在本地設備上進行後續的數據分析,例如通過手機上的APP實現。 In step S104: the human body data information is marked according to the exercise type to obtain the exercise posture of the practitioner; the specific implementation process of the above method 100 is illustrated by combining the image of the little girl practicing ukulele in FIG4 and the joint node identification of the little girl shown in FIG4a: according to step S101, the image or video of the ukulele practitioner or the training process is obtained through an imaging device. The image can be obtained by shooting with a camera or mobile phone connected to the network, and the shot image is uploaded to the cloud server through the network for subsequent data analysis, or the subsequent data analysis is performed on the local device, for example, through an APP on a mobile phone.
根據步驟S102所述的從所述練習影像數據中獲取樂器數據信息以及與所述練習者相關的人體數據信息。具體地,在如圖4這樣的實際場景中,首先需要在上述小女孩進行烏克麗麗練習過程的影像信息(圖像或視頻幀)中定位人體,例如,在小女孩演奏烏克麗麗過程中,通過定位小女孩人體部位關鍵關節點的位置,即,在人體區域內,首先識別關鍵的關節點的位置,例如主體軀幹的大關節點,例如頭部、頸部、肩部、肘部、腕部、手部、髖部、膝蓋和腳踝等,並對可以明顯體現烏克麗麗演奏過程的身體軀幹姿態的關鍵關節點的位置進行標識,例如手部關節點進行詳細標識。然後,從練習者練習過程的影像信息中提取樂器部分的圖像,將提取的樂器部分的圖像與現有的樂器數據庫比對可以直接識別出練習者所使用的器材,或者根據前述身體關鍵關節點的信息識別出練習者所使用的器材,進而得到相關的樂器數據信息,比如通過手部的姿態即大致區分出是鍵盤類樂器還是琴弦類樂器,比如彈鋼琴和彈吉烏克麗麗手部姿態有比較明顯的區別。進一步地,還可以根據樂器數據信息和練習者的前述的人體姿態識別出練習者的演奏場景。其中演奏場景包括樂器器材的數據信息(具體到烏克麗麗還是吉他,或者鋼琴還是電子琴等類似的某一類樂器中的一種)、多個拍攝視角等信息,其中樂器數據信息也可以採用聲音識別的方式識別得出。 According to step S102, musical instrument data information and human body data information related to the practitioner are obtained from the practice image data. Specifically, in an actual scene such as FIG4 , it is first necessary to locate the human body in the image information (image or video frame) of the above-mentioned little girl practicing the ukulele. For example, in the process of the little girl playing the ukulele, by locating the positions of the key nodes of the little girl's body parts, that is, in the human body area, first identify the positions of the key nodes, such as the major nodes of the main body trunk, such as the head, neck, shoulders, elbows, wrists, hands, hips, knees and ankles, etc., and mark the positions of the key nodes of the body trunk posture that can clearly reflect the ukulele playing process, such as the hand nodes are identified in detail. Then, images of musical instruments are extracted from the video information of the practitioner's practice process, and the extracted images of musical instruments are compared with the existing musical instrument database to directly identify the instrument used by the practitioner, or to identify the instrument used by the practitioner based on the aforementioned information of the key nodes of the body, and then obtain relevant musical instrument data information, such as roughly distinguishing whether it is a keyboard instrument or a string instrument through the hand posture, for example, there is a relatively obvious difference between the hand postures of playing the piano and playing the ukulele. Furthermore, the performance scene of the practitioner can also be identified based on the musical instrument data information and the aforementioned human body posture of the practitioner. The performance scene includes the data information of the musical instrument (specifically, whether it is a ukulele or a guitar, or a piano or an electronic keyboard, etc.), multiple shooting angles, etc. The musical instrument data information can also be identified by sound recognition.
根據步驟S103,根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型;根據圖4a中身體軀幹姿態和手部關節的標識信息以及樂器的圖像,可以判斷出小女孩彈奏的樂器是烏克麗麗,進一步根據手部的姿態, 可以得出小女孩的練習類型,即小女孩是初學者或者專業演奏者等不同的演奏練習階段,其中對彈奏過程中動作質量影響最大的是手部各個關鍵關節點的位置,所以,對手部關鍵關節點的標識和呈現,能夠清晰呈現彈奏動作中存在的問題。類似地,結合實際情況,對於練習者進行其他類型樂器的演奏,採用如上的方法可以得出練習者的練習類型或者練習階段。其中,人體區域關節點的具體的識別技術可以採用計算機視覺和深度學習技術,如卷積神經網路等。 According to step S103, the practice type of the practitioner is determined according to the human body data information and the musical instrument data information; according to the identification information of the body trunk posture and hand joints and the image of the musical instrument in FIG4a, it can be determined that the musical instrument played by the little girl is ukulele, and further according to the hand posture, the practice type of the little girl can be obtained, that is, the little girl is a beginner or a professional performer, etc., and the position of each key node of the hand has the greatest impact on the quality of the action during the playing process. Therefore, the identification and presentation of the key nodes of the hand can clearly present the problems existing in the playing action. Similarly, combined with actual situations, for practitioners playing other types of musical instruments, the above method can be used to derive the practitioner's practice type or practice stage. Among them, the specific recognition technology of human regional nodes can adopt computer vision and deep learning technology, such as convolutional neural network, etc.
識別關鍵關節點的位置可以採用關鍵點自動檢測算法,通常需要利用計算機視覺和深度學習技術在圖像或視頻幀中自動識別關鍵關節點的位置。自動檢測算法需要大量已標注的數據進行訓練,以便在新的圖像上實現準確的關鍵點識別。大量已標注的數據輸入現有的預訓練姿態識別模型(如OpenPose、PoseNet或AlphaPose等基礎模型)進行訓練,使用預訓練模型提取特徵,進而得到關鍵關節位置的識別並進行標記。這些預訓練姿態識別模型在通用人體姿態識別方面表現良好,並可以處理多種場景和姿勢。可以理解的是上面的描述僅僅是示例性的而非限制性的,本創作所屬技術領域中具有通常知識者根據本創作的指導可以對上述方法進行改變而不脫離本創作的精神和實質。 To identify the location of key nodes, an automatic key point detection algorithm can be used, which usually requires the use of computer vision and deep learning technology to automatically identify the location of key nodes in images or video frames. The automatic detection algorithm requires a large amount of annotated data for training in order to achieve accurate key point recognition on new images. A large amount of annotated data is input into the existing pre-trained pose recognition model (such as OpenPose, PoseNet or AlphaPose and other basic models) for training, and the pre-trained model is used to extract features, thereby obtaining the identification and marking of the key joint locations. These pre-trained pose recognition models perform well in general human pose recognition and can handle a variety of scenes and poses. It is understood that the above description is merely exemplary and not restrictive, and a person with ordinary knowledge in the technical field to which this creation belongs can modify the above method according to the guidance of this creation without departing from the spirit and essence of this creation.
根據步驟S104,當採集到音樂演奏練習者的圖像或視頻時,對攝像或者照相設備捕捉到的圖像或視頻數據進行分析,根據前述的練習類型提取當前練習者的關鍵點坐標,標記關鍵關節點坐標位置,其中坐標位置可以是相對值。進而綜合人體數據信息的標記獲得練習者的練習姿態。進一步地,所述練習姿態的結果用於分析練習者的技巧、評估動作質量等任務。 According to step S104, when the image or video of the music performance practitioner is collected, the image or video data captured by the camera or photographic equipment is analyzed, and the key point coordinates of the current practitioner are extracted according to the aforementioned practice type, and the key node coordinate positions are marked, where the coordinate positions can be relative values. Then, the marking of the human body data information is integrated to obtain the practitioner's practice posture. Furthermore, the result of the practice posture is used to analyze the practitioner's skills, evaluate the quality of the movement, and other tasks.
根據本創作的一個優選實施例,所述步驟S101中的練習影像數據還包括階段性的練習模板影像數據,其中所述階段性與時間和/或練習難度相 關。所述練習影像數據可以用於存儲練習者以往練習過程中每個練習階段的練習數據,以備練習者自己或者老師對練習過程中各個階段的練習數據進行對比,用於判斷練習者的演奏水平是否有所提升或者有什麼缺陷。 According to a preferred embodiment of the present invention, the practice image data in step S101 also includes staged practice template image data, wherein the stage is related to time and/or practice difficulty. The practice image data can be used to store the practice data of each practice stage in the practice process of the practitioner in the past, so that the practitioner himself or the teacher can compare the practice data of each stage in the practice process to determine whether the practicer's performance level has improved or has any defects.
針對現有技術中不能以圖像或者音頻的方式將練習者人體姿態進行分析並給出校正建議的問題,本創作的第二實施例提供一種用於音樂教學的方法。 In order to solve the problem that the existing technology cannot analyze the body posture of the practitioner in the form of images or audio and provide correction suggestions, the second embodiment of this invention provides a method for music teaching.
如圖2所示,一種用於音樂教學的方法200包括:在步驟S201中:獲取練習者的練習影像數據。 As shown in FIG2 , a method 200 for music teaching includes: in step S201: obtaining the practice image data of the practitioner.
在步驟S202中:從所述練習影像數據中獲取樂器數據信息以及與所述練習者相關的人體數據信息。 In step S202: musical instrument data information and human body data information related to the practitioner are obtained from the practice image data.
在步驟S203中:根據所述人體數據信息和所述樂器數據信息判斷練習者的練習類型。 In step S203: judging the practitioner's practice type based on the human body data information and the musical instrument data information.
在步驟S204中:根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。 In step S204: mark the human body data information according to the exercise type to obtain the exercise posture of the practitioner.
在步驟S205中:將所述練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示。 In step S205: the exercise posture of the practitioner is compared with the image reference data or the exercise template image data to determine whether the exercise posture of the practitioner is correct; and in response to determining that the exercise posture of the practitioner is incorrect, a correction instruction for correcting the exercise posture of the practitioner is output.
其中,影像參照數據是從有名的音樂家或某年齡段的音樂佼佼者彈奏時的影像生成的影像參照數據或者由老師根據練習者的進度,製作的針對於練習者一個練習階段的練習模板影像數據,所述練習模板影像數據可以是老師示範或者練習者在老師指導下做出的標準動作。 The image reference data is generated from the image of a famous musician or a music leader of a certain age group playing, or is a training template image data for a training stage of a trainee made by a teacher according to the trainee's progress. The training template image data can be a teacher's demonstration or a standard movement made by the trainee under the guidance of the teacher.
進一步地,通過執行人工智能連續影像對比來獲得所述練習影像數據與影像參照數據的對比結果;基於所述對比結果判定練習者的練習優劣度;以及基於所述練習優劣度輸出針對練習者的姿態校正建議。 Furthermore, the comparison result of the training image data and the image reference data is obtained by executing artificial intelligence continuous image comparison; the training quality of the trainee is determined based on the comparison result; and posture correction suggestions for the trainee are output based on the training quality.
在實際應用中,練習者通過借助人工智能連續影像對比功能獲得自己一個階段的練習過程的評價,基於前述利用計算機視覺和深度學習技術練習者一個階段的連續圖像或視頻幀中自動識別關鍵關節點的位置並進行標注,得到練習者多個練習階段的影像標注和對比結果,通過對比結果,進而得出練習過程的評價。比如得出的是有所改善還是退步的評價以及練習過程中一直存在的比較嚴重的問題。 In actual application, the practitioner obtains an evaluation of his/her own practice process at a certain stage by using the continuous image comparison function of artificial intelligence. Based on the aforementioned use of computer vision and deep learning technology, the positions of key nodes in the continuous images or video frames of the practitioner at a certain stage are automatically identified and annotated, and the image annotation and comparison results of the practitioner at multiple practice stages are obtained. Through the comparison results, the evaluation of the practice process is obtained. For example, whether there is improvement or regression, and the more serious problems that have always existed in the practice process.
如圖3所示,本創作的第三實施例還提供一種用於音樂教學的系統300,包括:影像模塊301,樂器識別模塊302,人體姿態估計模塊303,影像參照模塊304和判斷模塊305。 As shown in FIG3 , the third embodiment of the present invention also provides a system 300 for music teaching, including: an image module 301, a musical instrument recognition module 302, a human posture estimation module 303, an image reference module 304 and a judgment module 305.
所述影像模塊301用於獲取和存儲練習者的練習影像數據;所述樂器識別模塊302用於獲取所述練習影像數據中的一種或多種樂器數據信息;所述人體姿態估計模塊303從所述練習影像數據中獲取與所述練習者相關的人體數據信息;根據所述人體數據信息和所述樂器數據信息,判斷練習者的練習類型;根據所述練習類型對人體數據信息進行標記,以獲得練習者的練習姿態。 The image module 301 is used to obtain and store the practice image data of the practitioner; the musical instrument identification module 302 is used to obtain one or more musical instrument data information in the practice image data; the human body posture estimation module 303 obtains human body data information related to the practitioner from the practice image data; determines the practice type of the practitioner according to the human body data information and the musical instrument data information; and marks the human body data information according to the practice type to obtain the practice posture of the practitioner.
其中,人體姿態估計包括判斷身體動作,包括區別於不同樂器練習中的人體重要大關節的位置和指關節,尤其是所有指關節的動作(其他領域的人體姿態估計只有17點或33點,很少會使用所有指關節作為參數。)。 Among them, human posture estimation includes judging body movements, including the positions of important joints and finger joints in different instrument practices, especially the movements of all finger joints (human posture estimation in other fields only has 17 or 33 points, and rarely uses all finger joints as parameters.)
比如鋼琴練習時,需要用到抬臂,甩臂、立指、滑指、固定腕、肘關節。 For example, when practicing the piano, you need to raise your arms, swing your arms, stand your fingers, slide your fingers, and fix your wrists and elbows.
比如小提琴的:握弓、執弓、連弓、頓弓、虎口保持鬆弛等。 For example, for violin: holding the bow, holding the bow, stringing the bow, staccato bow, keeping the base of the thumb relaxed, etc.
比如打擊樂的手掌彎曲舉至胸前等。 For example, for percussion instruments, the palms are bent and raised to the chest.
為了實現人體姿態估計,可以會採用人體姿態識別技術。人體姿態識別技術可以包括以下幾個步驟:人體檢測:在圖像或視頻幀中定位人體,為後續的姿態識別提供空間範圍。 In order to achieve human posture estimation, human posture recognition technology may be used. Human posture recognition technology may include the following steps: Human body detection: locate the human body in the image or video frame to provide a spatial range for subsequent posture recognition.
關鍵關節點檢測:在人體區域內,識別關鍵關節點的位置,例如頭部、頸部、肩部、肘部、腕部、髖部、膝蓋和腳踝等。人體姿態識別算法包括OpenPose或DeepLabCut等。這些算法可以在給定示範的圖像上輸出關鍵點的二維坐標或三維坐標,這些坐標可以進一步用於分析人體姿態。 Key node detection: Identify the locations of key nodes in the human body, such as the head, neck, shoulders, elbows, wrists, hips, knees, and ankles. Human posture recognition algorithms include OpenPose or DeepLabCut. These algorithms can output the two-dimensional or three-dimensional coordinates of key points on a given demonstration image, which can be further used to analyze human posture.
姿態構建:將檢測到的關鍵關節點連接起來,構建人體的骨架模型,從而描述人體的姿態。 Posture construction: Connect the detected key nodes to construct a human skeleton model to describe the human posture.
對攝像頭或深度相機捕捉到的圖像或視頻數據進行分析,提取練習者的關鍵點坐標,標注關鍵關節點。 Analyze the images or video data captured by the camera or depth camera, extract the key coordinates of the practitioner, and mark the key nodes.
隨著深度學習和卷積神經網路(CNN)的快速發展,人體姿態識別技術取得了顯著的進步,通常採用深度學習的方法。一些典型的深度學習方法,如HourglassNetwork、、PoseNet和AlphaPose等,已經在人體姿態識別任務上取得了很高的準確率和實時性能。 With the rapid development of deep learning and convolutional neural networks (CNN), human posture recognition technology has made significant progress, usually using deep learning methods. Some typical deep learning methods, such as HourglassNetwork, PoseNet and AlphaPose, have achieved high accuracy and real-time performance in human posture recognition tasks.
通過上面的人體姿態識別進而得到對人體姿態的描述,即,實現人體姿態估計。 Through the above human posture recognition, we can get a description of the human posture, that is, realize human posture estimation.
具體地,例如,在圖4中,例如對鋼琴練習者或者鍵盤使用者的人體姿態識別通常聚焦於手部、手指、手腕和上肢等關鍵區域。以下是關於上述兩種練習者人體姿態估計的示範性方法。 Specifically, for example, in Figure 4, human posture recognition for piano players or keyboard users usually focuses on key areas such as hands, fingers, wrists, and upper limbs. The following is an exemplary method for human posture estimation for the above two types of players.
1)明確關鍵關節集:針對上述練習者的姿態識別任務,明確關鍵關節集。對於鋼琴練習者,這通常包括手腕、掌部、各指關節等。確保在圖像中標注的關鍵關節能夠涵蓋鋼琴演奏過程中的主要動作。 1) Clarify the key joint set: For the posture recognition task of the above-mentioned practitioners, clarify the key joint set. For piano practitioners, this usually includes the wrist, palm, and finger joints. Make sure that the key joints marked in the image can cover the main movements in the piano playing process.
2)選擇合適的圖像和視頻:為了獲取高質量的標注數據,應選擇清晰、高分辨率的圖像和視頻,避免運動模糊和遮擋。此外,為了提高識別的準確度,儘量選擇多角度、多場景和多種練習者類型的數據。 2) Select appropriate images and videos: In order to obtain high-quality annotation data, clear, high-resolution images and videos should be selected to avoid motion blur and occlusion. In addition, in order to improve the accuracy of recognition, try to select data from multiple angles, multiple scenes, and multiple types of practitioners.
3)使用專門的手部姿態識別模型:可以使用專門針對手部姿態識別的預訓練模型,如Hand3D、DeepHPS或OpenPose手部模塊模型等。這些模型在手部姿態識別任務上表現良好,可以作為基礎模型進行針對上述練習者的影像進行學習。 3) Use a specialized hand posture recognition model: You can use a pre-trained model specifically for hand posture recognition, such as Hand3D, DeepHPS, or OpenPose hand module models. These models perform well in hand posture recognition tasks and can be used as a basic model to learn from the images of the above practitioners.
4)將預訓練模型應用於上述練習者的姿態識別任務。首先,使用預訓練模型提取特徵,然後使用專門收集的上述練習者數據集進行訓練和微調,以便更好地適應上述練習場景。 4) Apply the pre-trained model to the pose recognition task of the above-mentioned practitioner. First, extract features using the pre-trained model, and then use the specially collected above-mentioned practitioner dataset for training and fine-tuning to better adapt to the above-mentioned practice scenarios.
5)數據增強:為了提高模型的泛化能力,可以使用數據增強技術,如圖像旋轉、翻轉、裁剪、縮放、亮度和對比度調整等,來增加訓練數據的多樣性。這有助於讓模型在不同場景和視角下獲得更好的表現。 5) Data augmentation: In order to improve the generalization ability of the model, data augmentation techniques can be used, such as image rotation, flipping, cropping, scaling, brightness and contrast adjustment, etc., to increase the diversity of training data. This helps the model to perform better in different scenes and perspectives.
6)通過使用上述訓練後的模型進行新圖像的關節點識別和標注。將音樂器材特徵和人體姿態數據進行融合,形成一個綜合數據集。還可以使用時間戳將音樂特徵和姿態數據同步,以便在分析中考慮它們之間的關係。 6) Use the trained model to identify and annotate new images. Fuse the musical instrument features and human posture data to form a comprehensive dataset. You can also use timestamps to synchronize the music features and posture data so that their relationship can be considered in the analysis.
所述影像參照模塊304用於獲取影像參照數據或練習者的練習模板影像數據;所述判斷模塊305用於將所述練習者的練習姿態與影像參照數據或練習模板影像數據進行比對,以判斷所述練習者的練習姿態是否正確;以及響應於判斷所述練習者的練習姿態不正確,輸出用於校正所述練習者的練習姿態的校正指示。 The image reference module 304 is used to obtain image reference data or the image data of the trainee's training template; the judgment module 305 is used to compare the trainee's training posture with the image reference data or the training template image data to judge whether the trainee's training posture is correct; and in response to judging that the trainee's training posture is incorrect, output a correction instruction for correcting the trainee's training posture.
其中,上述三個實施例中的各個步驟的全部或部份功能可以利用本地設備的API調用雲端的設備或其他分布式設備之算力協助完成。 Among them, all or part of the functions of each step in the above three embodiments can be completed by using the API of the local device to call the computing power of the cloud device or other distributed devices.
圖4示出了本創作實施例的用於音樂教學的方法中的關節點位識別的圖像特徵示範示例。 FIG4 shows an example of an image feature demonstration of a joint point position recognition method used in music teaching according to an embodiment of the present invention.
圖5a示出了本創作一種實施例的影像模塊的設備組成500,本創作的影像模塊可以包括從下到上依次安裝於固定杆505上的感應器501、第一攝像頭502和第二攝像頭503。進一步地,所述感應器501可以是產生紅外線的雷射器或者產生其他顏色光(例如綠光)的雷射器。優選地,所述雷射器例如可以是紅外一字雷射器。需要說明的是,本創作的影像模塊選擇紅外線作為信號檢測的光源,是因為紅外雷射具有低功耗、集成度效果好和發光頻率單一等優點,並且由於樂器的演奏界面(例如鍵盤)通常是一個平面,因此不能採用傳統的單束雷射作為光源。基於上述原理,本創作採用紅外一字雷射器作為檢測光源。特別地,紅外一字雷射器可配置成可調發射角度和方向,如果演奏的樂器是鋼琴時,所發射的紅外線可以與琴面(504)平行,並且其安裝位置靠近琴面,以使得其發射的雷射平面緊貼琴面。 FIG5a shows a device composition 500 of an imaging module of an embodiment of the present invention. The imaging module of the present invention may include a sensor 501, a first camera 502, and a second camera 503, which are sequentially mounted on a fixed rod 505 from bottom to top. Furthermore, the sensor 501 may be a laser that generates infrared rays or a laser that generates other color lights (e.g., green light). Preferably, the laser may be, for example, an infrared single-beam laser. It should be noted that the imaging module of the present invention selects infrared rays as the light source for signal detection because infrared lasers have the advantages of low power consumption, good integration effect, and a single luminous frequency, and because the playing interface of a musical instrument (e.g., a keyboard) is usually a plane, a traditional single-beam laser cannot be used as a light source. Based on the above principle, this invention uses an infrared straight laser as the detection light source. In particular, the infrared straight laser can be configured to have an adjustable emission angle and direction. If the instrument being played is a piano, the infrared rays emitted can be parallel to the piano surface (504), and its installation position is close to the piano surface so that the laser plane it emits is close to the piano surface.
所述第一攝像頭502和第二攝像頭503可以是廣角攝像頭,其配置用於分別採集所述樂器數據信息和人體數據信息。具體地,當演奏者進行演奏時,其手指或敲擊器具會阻擋並反射按壓或敲擊鍵位置的紅外線,此時該按壓或敲擊鍵即被觸發,於是攝像頭實時拍攝到手指或敲擊器具對紅外線進行阻擋時的圖像,並將該圖像傳送給人體姿態估計模塊和/或判斷模塊做進一步分析和處理。另外,所述第一攝像頭502和第二攝像頭503的安裝高度可以根據其拍攝的範圍進行調整,其拍攝的範圍應能分別覆蓋樂器整體情況和人體關鍵關節部位。 The first camera 502 and the second camera 503 may be wide-angle cameras, which are configured to collect the musical instrument data information and human body data information respectively. Specifically, when the performer is playing, his fingers or striking instruments will block and reflect the infrared rays at the position of the pressed or struck keys, and the pressed or struck keys will be triggered at this time, so the camera will take real-time images of the fingers or striking instruments blocking the infrared rays, and transmit the images to the human body posture estimation module and/or judgment module for further analysis and processing. In addition, the installation heights of the first camera 502 and the second camera 503 can be adjusted according to their shooting ranges, and their shooting ranges should be able to cover the overall situation of the musical instrument and the key joints of the human body respectively.
如圖5b所示,本創作另外一種實施例的影像模塊的設備組成,本創作的影像模塊可以包括從下到上依次安裝於固定杆505上的第一攝像頭502和第二攝像頭503。 As shown in FIG. 5b, the device composition of the imaging module of another embodiment of the present invention, the imaging module of the present invention may include a first camera 502 and a second camera 503 sequentially mounted on a fixing rod 505 from bottom to top.
所述第一攝像頭502和第二攝像頭503可以是具有廣角攝像頭,其配置用於分別採集所述樂器數據信息和人體數據信息。具體地,當演奏者進行演奏時,攝像頭實時拍攝到手指或敲擊器具時的圖像,並將該圖像傳送給人體姿態估計模塊和/或判斷模塊做進一步分析和處理。另外,所述第一攝像頭502和第二攝像頭503的安裝高度可以根據其拍攝的範圍進行調整,其拍攝的範圍應能分別覆蓋樂器整體情況和人體關鍵關節部位。 The first camera 502 and the second camera 503 may be wide-angle cameras, which are configured to collect the musical instrument data information and human body data information respectively. Specifically, when the performer is playing, the camera captures the image of the fingers or the striking instruments in real time, and transmits the image to the human body posture estimation module and/or judgment module for further analysis and processing. In addition, the installation height of the first camera 502 and the second camera 503 can be adjusted according to the range of their shooting, and the range of their shooting should be able to cover the overall situation of the musical instrument and the key joints of the human body respectively.
如圖5c所示,本創作另外一種實施例的提供一種用於音樂教學的電子設備,包括:主機501,第一攝像頭502和第二攝像頭503;主機501用於接收第一攝像頭502和第二攝像頭503的拍攝的圖像;在本實施例中僅示意性地給出第一攝像頭502和第二攝像頭503安裝於固定杆505上,並不對其安裝方式和位置進行限定;主機501還包括處理器;以及存儲器,其用於存儲用於音樂教學的 程序指令,當所述處理器運行該程序指令時,使得所述電子設備執行根據第一實施例任一項所述方法。 As shown in FIG. 5c, another embodiment of the present invention provides an electronic device for music teaching, including: a host 501, a first camera 502 and a second camera 503; the host 501 is used to receive images taken by the first camera 502 and the second camera 503; in this embodiment, the first camera 502 and the second camera 503 are only schematically shown to be installed on a fixed rod 505, and the installation method and position thereof are not limited; the host 501 also includes a processor; and a memory, which is used to store program instructions for music teaching, and when the processor runs the program instructions, the electronic device executes any method according to the first embodiment.
上述兩個實施例中所述的第一攝像頭,第二攝像頭,感應器三者是”或”的關係,可以配合使用,也可以獨立設置,可設置於不同的模塊或設備上(比如不同的電腦主機或者手機上),在具體實施過程中,第一攝像頭和第二攝像頭的功能並不做限制。 The first camera, the second camera, and the sensor described in the above two embodiments are in an "or" relationship and can be used together or independently, and can be set on different modules or devices (such as different computer hosts or mobile phones). In the specific implementation process, the functions of the first camera and the second camera are not limited.
本創作的技術方案尤其適用於在AI晶片中對練習者的影像數據進行處理。 The technical solution of this creation is particularly suitable for processing the image data of practitioners in AI chips.
綜上,本創作通過獲取練習過程中的影像數據並對練習過程進行過程性控制,將練習姿態情況的影像進行存儲、標記處理,直接呈現於教學設備中或者通過數據傳輸呈現給遠程的老師,進而,通過對練習者影像數據的分析對比,實現可以智能、直觀地給出對應於練習者所使用的器材的練習姿勢的判斷和校正建議。 In summary, this creation obtains image data during the practice process and performs process control on the practice process, stores and marks images of the practice posture, and directly presents them in the teaching equipment or presents them to remote teachers through data transmission. Furthermore, through the analysis and comparison of the image data of the practitioners, it can intelligently and intuitively give judgments and correction suggestions on the practice posture corresponding to the equipment used by the practitioners.
綜觀上述,本創作所揭露之技術手段不僅為前所未見,且確可達致預期之目的與功效,故兼具新穎性與進步性,誠屬專利法所稱之新型無誤,以其整體結構而言,確已符合專利法之法定要件,爰依法提出新型專利申請。 In summary, the technical means disclosed in this creation are not only unprecedented, but can also achieve the expected purpose and effect. Therefore, it is both novel and progressive, and is truly a new type as defined by the Patent Law. In terms of its overall structure, it has indeed met the statutory requirements of the Patent Law, and therefore a new type patent application is filed in accordance with the law.
惟以上所述者,僅為本創作之較佳實施例,當不能以此作為限定本創作之實施範圍,即大凡依本創作申請專利範圍及說明書內容所作之等效變化與修飾,皆應仍屬於本創作專利涵蓋之範圍內。 However, the above is only the best implementation example of this creation, and it cannot be used to limit the scope of implementation of this creation. In other words, all equivalent changes and modifications made according to the scope of the patent application of this creation and the content of the specification should still fall within the scope of the patent of this creation.
300:系統 300:System
301:影像模塊 301: Image module
302:樂器識別模塊 302: Musical instrument identification module
303:人體姿態估計模塊 303: Human body posture estimation module
304:影像參照模塊 304: Image reference module
305:判斷模塊 305: Judgment module
Claims (9)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202320815179 | 2023-04-12 | ||
| CN2023208151790 | 2023-04-12 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| TWM664095U true TWM664095U (en) | 2024-12-11 |
Family
ID=94735807
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| TW113203432U TWM664095U (en) | 2023-04-12 | 2024-04-09 | A system for music teaching |
Country Status (1)
| Country | Link |
|---|---|
| TW (1) | TWM664095U (en) |
-
2024
- 2024-04-09 TW TW113203432U patent/TWM664095U/en unknown
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20180315329A1 (en) | Augmented reality learning system and method using motion captured virtual hands | |
| TWI878086B (en) | Method, device and computer readable storage medium for music teaching | |
| CN113762133A (en) | Self-weight fitness auxiliary coaching system, method and terminal based on human body posture recognition | |
| CN113657184B (en) | Piano playing fingering evaluation method and device | |
| WO2022052941A1 (en) | Intelligent identification method and system for giving assistance with piano teaching, and intelligent piano training method and system | |
| CN113409651B (en) | Live fitness method, system, electronic device, and storage medium | |
| Ng et al. | 3D augmented mirror: a multimodal interface for string instrument learning and teaching with gesture support | |
| CN112488047B (en) | A method for intelligent recognition of piano fingering | |
| CN113657185A (en) | Intelligent auxiliary method, device and medium for piano practice | |
| WO2024212940A1 (en) | Method and device for music teaching, and computer-readable storage medium | |
| CN114170868A (en) | Intelligent piano training method and system | |
| KR20130067856A (en) | Apparatus and method for performing virtual musical instrument on the basis of finger-motion | |
| CN114092971B (en) | A human motion assessment method based on visual images | |
| TWM664095U (en) | A system for music teaching | |
| CN119399242B (en) | Ping-pong ball auxiliary training method, storage medium, device and terminal equipment | |
| KR102171319B1 (en) | Appratus for writing motion-script, appratus for self-learning montion and method for using the same | |
| WO2023105601A1 (en) | Information processing device, information processing method, and program | |
| CN114360052A (en) | An intelligent somatosensory coaching system based on AlphaPose and joint point angle matching algorithm | |
| Kerdvibulvech et al. | Real-time guitar chord estimation by stereo cameras for supporting guitarists | |
| CN117077084B (en) | Dance scoring method based on space-time heterogeneous double-flow convolutional network | |
| CN113255470A (en) | Multi-mode piano partner training system and method based on hand posture estimation | |
| Kerdvibulvech et al. | Guitarist fingertip tracking by integrating a Bayesian classifier into particle filters | |
| Enkhbat et al. | Using Hybrid Models for Action Correction in Instrument Learning Based on AI | |
| Zhao et al. | A study on attention-based fine-grained image recognition: Towards musical instrument performing hand shape assessment | |
| TW202419138A (en) | Golf teaching method and golf teaching system |