Pachet et al., 2008 - Google Patents
Hit Song Science Is Not Yet a Science.Pachet et al., 2008
View PDF- Document ID
- 6185201801226902007
- Author
- Pachet F
- Roy P
- Publication year
- Publication venue
- ISMIR
External Links
Snippet
We describe a largeYscale experiment aiming at validating the hypothesis that the popularity of music titles can be predicted from global acoustic or human features. We use a 32.000 title database with 632 manuallyYentered labels per title including 3 related to the …
- 238000000034 method 0 abstract description 11
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/3074—Audio data retrieval
- G06F17/30755—Query formulation specially adapted for audio data retrieval
- G06F17/30758—Query by example, e.g. query by humming
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/3074—Audio data retrieval
- G06F17/30743—Audio data retrieval using features automatically derived from the audio content, e.g. descriptors, fingerprints, signatures, MEP-cepstral coefficients, musical score, tempo
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/3074—Audio data retrieval
- G06F17/30749—Audio data retrieval using information manually generated or using information not derived from the audio data, e.g. title and artist information, time and location information, usage information, user ratings
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS
- G10H2210/00—Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
- G10H2210/031—Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/30—Information retrieval; Database structures therefor; File system structures therefor
- G06F17/30017—Multimedia data retrieval; Retrieval of more than one type of audiovisual media
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/121—Musical libraries, i.e. musical databases indexed by musical parameters, wavetables, indexing schemes using musical parameters, musical rule bases or knowledge bases, e.g. for automatic composing methods
- G10H2240/131—Library retrieval, i.e. searching a database or selecting a specific musical piece, segment, pattern, rule or parameter set
- G10H2240/141—Library retrieval matching, i.e. any of the steps of matching an inputted segment or phrase with musical database contents, e.g. query by humming, singing or playing; the steps may include, e.g. musical analysis of the input, musical feature extraction, query formulation, or details of the retrieval process
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS
- G10H2240/00—Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
- G10H2240/075—Musical metadata derived from musical analysis or for use in electrophonic musical instruments
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS
- G10H2250/00—Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
- G10H2250/131—Mathematical functions for musical analysis, processing, synthesis or composition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00-G10L21/00
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Pachet et al. | Hit Song Science Is Not Yet a Science. | |
| Li et al. | Toward intelligent music information retrieval | |
| Li et al. | A comparative study on content-based music genre classification | |
| Salamon et al. | Tonal representations for music retrieval: from version identification to query-by-humming | |
| Bergstra et al. | Aggregate features and AdaBoost for music classification | |
| Casey et al. | Content-based music information retrieval: Current directions and future challenges | |
| Typke | Music retrieval based on melodic similarity | |
| Tingle et al. | Exploring automatic music annotation with" acoustically-objective" tags | |
| Kacprzyk | Studies in Computational Intelligence, Volume 274 | |
| Sordo et al. | Annotating Music Collections: How Content-Based Similarity Helps to Propagate Labels. | |
| Javed et al. | A new rhythm in a1: Convolutional neural networks for music genre classification | |
| Prockup et al. | Modeling Genre with the Music Genome Project: Comparing Human-Labeled Attributes and Audio Features. | |
| Zhang et al. | Influence of musical elements on the perception of ‘Chinese style’in music | |
| Kostek et al. | Creating a reliable music discovery and recommendation system | |
| Vatolkin et al. | Multi-objective investigation of six feature source types for multi-modal music classification | |
| Pachet et al. | Improving multilabel analysis of music titles: A large-scale validation of the correction approach | |
| Dwivedi et al. | Generative Adversarial Networks Based Framework for Music Genre Classification | |
| Herrera et al. | Simac: Semantic interaction with music audio contents | |
| Yeh et al. | Popular music representation: chorus detection & emotion recognition | |
| Chordia et al. | Extending Content-Based Recommendation: The Case of Indian Classical Music. | |
| Poonia et al. | Music genre classification using machine learning: A comparative study | |
| Geetha Ramani et al. | Improvised emotion and genre detection for songs through signal processing and genetic algorithm | |
| Yang et al. | Improving Musical Concept Detection by Ordinal Regression and Context Fusion. | |
| Jun et al. | Music retrieval and recommendation scheme based on varying mood sequences | |
| Won | Representation learning for music classification and retrieval: bridging the gap between natural language and music semantics |