0% found this document useful (0 votes)
16 views

Automatic Music Generation

Uploaded by

vedantbarguje1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views

Automatic Music Generation

Uploaded by

vedantbarguje1
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 16

Automatic Music

Generation
Presented By:
GUNJAL PRATIK PRAKASH(24070149008)
BARGUJE VEDANT KRUSHNA(24070149031)
BIRWADKAR PRAJWAL SUNIL(24070149003)
Content

• Introduction
• Literature Study
• Methodology
• Implementation
• Conclusion & Future Work
Introduction
• Algorithmic music generation is a difficult subject that has been
extensively researched in recent decades. MarkoV models or graph-
based energy reduction algorithms, which both produce carefully
planned melodic characteristics, are two common techniques for
making algorithmic music.
• Despite the fact that these techniques can yield unique compositions,
the music they produce frequently features repetitive sequences and
lacks the theme patterns that are common in most musical works .
Large-scale corpuses may now be able to produce novel music thanks
to recent advancements in recurrent network topologies and the
expansion of computing power.
• The most well-known recurrent network for simulating long-term
dependence is the Long Short-Term Memory (LSTM) network, which
Hochreiter and Schmidhuber (5) created in 1997. Cho et al.’s Gated
Recurrent Units (GRU) have been utilized to successfully replicate long-
term dependencies in a number of generic sequence modeling
applications.
• We believe that by using LSTM and GRU networks for algorithmic
music production, we may produce works that sound distinctive and
are musically cohesive while also more correctly representing the long-
term theme structure of musical compositions.
Literature Study:
Sr. Paper Title Publisher Year Take-away points
numbe
r
1 FLUX that Plays Music Zhengcong Fei, 1 September FluxMusic, is Transformers for generating
Mingyuan Fan, 2024 music from text. It works by transforming
Changqian Yu, and text descriptions into mel-spectrogram
Junshi Huang representations using attention
mechanisms.
2 Automatic Music Generator Using Zayed University, 9 December The music generated was listenable and
Recurrent Neural Network. Abu Dhabi, United 2019 interesting which the highest score is
Arab Emirates double stacked layer GRU model with a
score of 6.85 out of 10.
3 Generating Music by Fine-Tuning Tayba Asgher Dept. 2016 the ability to train models to generate
Recurrent Neural Networks with of Computer Science pleasantsounding melodies, we believe our
Reinforcement Learning Riphah International approach of using RL to fine-tune RNN
University Lahore, Pa models could be promising for a number of
kistan. applications..
Sr.no Paper title publisher year Take away points
4 Music Generation by Deep Christopher Lueg 16 October 2018. The use of deep
Learning – Challenges and learning architectures
Directions∗
University of
Technology, and techniques for the
generation of music (as
Sydney, Australia well as other artistic
content) is a growing
area of research.
However, there remain
open challenges such
as control, structure,
creativity

5 Music Generation Douglas Dec-2017 The performance of this


model can be more
Using Three-layered AberdeenGoogleZur impressive if the dataset is
LSTM ich,Switzerland altered to include more
tunes, in variety, and that
of multiple instruments.
The training, if done more
rigorously, can yield better
results as well
s.no Paper title Publisher year Take away points
6 Generating Publish with 2017 Qualitatively, the
Polyphonic Music Hindawi output of our
Using Tied Parallel model has
Networks. measure-level
structure, and in
some cases
successfully
reproduces
complex rhythms,
melodies, and
counterpoint.
Project flow and Methodology:
Data Collection:
• Gather a diverse dataset of music in a format suitable for your model. This can include MIDI files, sheet
music, or audio recordings. The larger and more diverse the dataset, the better your model can learn .
Data Preprocessing:
• Convert the music data into a format that can be used for training. For example, you might: Represent notes
as numerical values or one-hot encoded vectors. Create sequences of notes or chords from the raw data.
Normalize the data, ensuring it’s in a consistent key and tempo . Data should be divided into training and
validation sets.
Model Architecture:
• Design your RNN model. Due to its capacity to recognise long-term dependencies in music, the LSTM (Long
Short-Term Memory) neural network is a common option. Define the input and output layers, hidden layers,
and additional hyperparameters. Consider the architecture’s depth and width, depending on your
computational resources and the complexity of the music you want to generate.
Training:
Feed the preprocessed data into your RNN model. Train the model using backpropagation and
gradient descent. The objective is to reduce the discrepancy between the musical compositions
and the training data. Choose an appropriate loss function, such as cross-entropy loss.
Experiment with hyperparameters like learning rate and batch size to optimize training.
Fine-Tuning:
Adjust various parameters to control the quality and creativity of the generated music. Consider
using temperature to control randomness. A higher temperature produces more random
outputs, while a lower temperature makes the output more deterministic. Experiment with the
length of the generated sequences to influence the structure of the music.
Evaluation and Iteration:
• Evaluate the generated music’s quality by listening to it. Check to see if it matches the
intended style and structure against the training data. If the results are unsatisfactory, iterate
on your model, data or hyperparameters. You might need to train the model for more
epochs, refine the preprocessing steps, or collect more diverse data. Continue the cycle of
training, generating, evaluating, and iterating until you achieve the desired music generation
quality .
• Model Training:
• Train the model using the training dataset. You can use mini-batches for efficiency and
adjust the number of epochs accordingly.Monitor the training process and evaluate the
model's performance on the validation set to avoid overfitting.Save the trained model
weights for future use.
• Music Generation:
• Load the trained model weights.Provide an initial seed sequence to start the music
generation process.Use the trained RNN model to predict the next set of musical notes
based on the seed sequence.Sample from the predicted probabilities to generate a diverse
range of musical outputs.Repeat the prediction step with the updated sequence, iteratively
generating longer music sequences.
Conclusion

• his project has demonstrated that algorithmic music generation


with waveforms as input is possible with the use of recurrent
neural networks, particularly the LSTM network.
• Interesting future directions include investigating the effect of
adding layers of recurrent units and discovering the impact that
additional layers have on performance. Similarly, it would be
worthwhile to perform network training across genres with a
substantially larger corpus.
• Unfortunately, due to the time and memory constraints of the
Stanford Rye clusters, we were unable to pursue these directions
fully. We are eager to experiment with more complex architectures
and larger corpus sizes to see how well our preliminary results
generalize and evaluate the effect of network depth on music
generation performance
Future Work

• there are several feature enhancements you can consider. First, you can focus on
refining the model architecture by exploring different RNN variants such as
stacked or bidirectional LSTM/GRU layers. This exploration aims to improve the
quality of the generated music. Additionally, fine-tuning the model through
hyperparameter adjustments, like tuning the learning rate, batch size, or
sequence length, can further enhance the model's performance.
• Another avenue to explore is conditional generation, where you can train the
RNN to generate music in specific styles or genres by incorporating genre labels
or genre-specific features. Integrating additional data sources, such as lyrics or
artist-specific patterns, can also diversify and enhance the generated music.
Thank you

You might also like