Generator Musik Instrumental Otomatis Pada Akapela Berbasis Recurrent Neural Network Dengan Memanfaatkan Masukan Latent Autoencoder Dan Discrete Cosine Transform

Julian, Saputra (2023) Generator Musik Instrumental Otomatis Pada Akapela Berbasis Recurrent Neural Network Dengan Memanfaatkan Masukan Latent Autoencoder Dan Discrete Cosine Transform. Undergraduate Thesis thesis, Institut Teknologi Telkom Purwokerto.

[img] Text
1_Cover.pdf

Download (4MB)
[img] Text
3_Abstract (Inggris).pdf

Download (55kB)
[img] Text
2_Abstrak (Indonesia).pdf

Download (123kB)
[img] Text
4_Bab I.pdf

Download (92kB)
[img] Text
5_Bab II.pdf

Download (429kB)
[img] Text
6_Bab III.pdf

Download (2MB)
[img] Text
8_Bab V.pdf

Download (78kB)
[img] Text
7_Bab IV.pdf
Restricted to Registered users only

Download (5MB) | Request a copy
[img] Text
9_Daftar Pustaka.pdf

Download (108kB)

Abstract

Not everyone can create an instrument or background music for a song. Besides having knowledge of music, they should also possess experience in composing melodies. Based on this issue, this research designs a model for an automatic music instrument generator where the input is in the form of human vocal sounds with lyrics and tones. This input is then processed using a selected model and architecture to produce harmonious musical instrument notes as the output. In previous studies, researchers applied Recurrent Neural Network (RNN) as a method for automatic music composition, although the results were not yet perfect. In this study, a combination of Autoencoder utilizing Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN) architecture is used, with inputs consisting of human vocal sounds and instrumental sounds. This research presents three schemes: Autoencoder-RNN, Autoencoder-Comb-RNN, and Discrete Cosine Transform (DCT)-RNN schemes. These three schemes have different types of sound inputs, namely short sound, comb-filtered sound, and DCT-based frequency sound. By combining autoencoder and RNN models, different models with varying specifications are obtained and their performance is evaluated in terms of both loss and creativity. The best model obtained in this research is the DCT-RNN scheme, specifically Model DCT-17, with an evaluation of performance resulting in an average loss of 0.02993620155, an average bias of 0.01588229974, and a standard deviation of 0.02059837636. Listening tests of the generated sound output within the frequency range of 0-1600 Hertz revealed clear, tonal, and harmonious sounds. This indicates that within this frequency range, there are numerous sound signals with diverse patterns compared to frequencies above it, as the generated sounds tend to have higher frequencies. Keywords : music instrument, autoencoder, convolutional neural network, recurrent neural network, discrete cosine transform

Item Type: Thesis (Undergraduate Thesis)
Subjects: T Technology > TA Engineering (General). Civil engineering (General)
Divisions: Faculty of Informatics > Informatics Engineering
Depositing User: pustakawan ittp
Date Deposited: 04 Aug 2023 07:04
Last Modified: 04 Aug 2023 07:04
URI: http://repository.ittelkom-pwt.ac.id/id/eprint/9841

Actions (login required)

View Item View Item