intelligent signal processing coursework for midterm introduction the
Search for question
Question
Intelligent Signal Processing
Coursework for midterm
Introduction
The midterm coursework for Intelligent Signal Processing consists of a series of four individual
exercises. These exercises cover the first five topics of the course:
Digitising, representing, and storing audio signals
-
Editing and processing digital audio
-
-
Frequency domain representations
Extracting features from audio signals
Speech recognition.
The exercises are strongly based on the subjects covered during the course, but also invite the
student for further investigation.
It is recommended that the students carefully read all the sections of this document, both to ensure
a good understanding of the coursework exercises, in addition to knowing what to submit.
Exercise 1
Description
The goal of this exercise is to create a web-based audio application using p5.js and its library
p5.sound that processes a pre-recorded sound file, sending the processed audio signal to the
computer's speakers or audio output. Optionally, the user could also record the processed audio
signal as a digital audio file on the computer's drive.
The application should include the following effects: low-pass filter, waveshaper distortion, dynamic
compressor, reverb and master volume. skip
skip
pause
play
stop
to
start
to
end
loop
record
low-pass filter
dynamic compressor
master
volume
cutoff
frequency resonance
attack
knee
release
dry/wet
output
level
ratio
threshold
reverb
reverb
duration
decay
rate
dry/wet
output
level
waveshaper distortion
distortion
amount oversample
spectrum in
reverse
о
dry/wet
output
level
output
dry/wet
level
口
spectrum out
Figure 1. Schema of the GUI of the application.
Sound File
Low Pass
Filter
Waveshaper
Distortion
Dynamic
Compressor
Reverb
Master
Volume
Figure 2. Internal signal flow of the application.
Regarding the pre-recorded sound file, you should record in Audacity these two lines from the poem
If by Rudyard Kipling:
If you can dream – and not make dreams your master;
-
If you can think - and not make thoughts your aim;
The audio must be recorded at an optimal recording level without clipping.
The recording must also be edited in Audacity, in order to remove possible silences at the beginning
and end of the file. Finally, the recording must be normalised and saved as a WAV format, at 48 kHz
and 24-bits. The functionality of the application should meet the following requirements:
1. The application should include the playback controls and effects controls shown in Image 1.
2. Internally, the effects must be connected in a chain, as shown in Image 2.
3. The application should include a Record button that allows the user to start/stop recording
the processed signal as a WAV file.
4. The application must display both the spectrum of the original sound and the spectrum of
the processed sound.
Ideas for further development:
1. Enhance the filter effect by adding a type selector that allows the user to select between a
low-pass, high-pass or band-pass filter.
2. Allow the user to select between the live microphone input and the pre-recorded audio file
as the audio source for the application.
3. Configure a delay audio effect and add this to the audio chain before the dynamic
compressor.
List of deliverables
For Exercise 1, you should submit in a ZIP file:
•
The source code of the application exercise 1.
•
•
•
•
A link to the application running in a web page using the Coursera static web page function.
A screencast recording demonstrating that the application meets all the requirements and
shows implementation of the further developments (maximum length of two minutes).
A link to the application running in a web page using the Coursera static web page function.
A written report in PDF format, approximately 500 words. This report must include:
A brief description of the processes of audio recording, editing, processing and
saving in Audacity. This section must include at least two screenshots of Audacity
showing both the original recorded voice, and the recorded voice after editing and
normalising it.
о
о
Marking criteria
A brief description of the main characteristics of each effect and how they have
been programmed.
A brief analysis of the application discussing how the low-pass filter and the master
volume effects affect the sound's spectrum. This should also be illustrated through
screenshots.
A brief description of the further development implemented.
The screencast recording has a maximum length of two minutes, and it
demonstrates that the application meets all the requirements and shows the
further developments implemented.
The sound file has been satisfactorily recorded, edited, processed and saved.
Done?
Marks
1
1 The application includes the requested playback controls, and these have been
satisfactorily implemented. In particular, the Record button allows the user to
record the processed audio signal in WAV format.
1
The effects have been connected in a chain. The chain is functioning properly, and
the user can listen to the processed audio signal.
1
The filters have been correctly configured, and include the requested controls.
1
The written report includes a brief description of the processes of audio recording,
editing, processing and saving in Audacity.
1
The written report includes a brief description of the main characteristics of each
effect and how they have been programmed.
1
The written report includes a brief analysis of the application discussing how the
low-pass filter and the master volume effects affect the sound's spectrum.
The application includes further development.
Total
1
2
10
Exercise 2
Description
A famous DJ has contacted you to develop an interactive web-based application for visualising his
music during its concerts. The application must be based on p5.js, p5.speech and the JavaScript
audio feature extraction library Meyda.
Task 1
First, to evaluate your skills, you are asked to perform the following task. The DJ sends you three
sounds (Ex2_sound1.wav, Ex2_sound2.wav and Ex2_sound3.wav) and you have to select Meyda
audio features that could help represent these sounds visually in an appropriate manner. For
example, if the 'brightness' of one of the sounds radically changes over time, to select an audio
feature that measures the brightness of this sound could be a good choice from a perspective of
producing visual impact.
To perform Task 1, you have to fill in the following table. You have to select three Meyda audio
features for each sound and justify your selections.
Sound 1
Sound 2
Meyda audio features
Justification Sound 3
Task 2
The second task consists of creating the aforementioned web-based application for audio
visualisation. The application (exercise 2) will use the song Kalte_Ohren_(_Remix_).mp3 (*) as an
audio source.
Figure 3. Idea for the audio visualisation application.
You could use the image of Figure 3 as an inspiration. The visual variables could include:
1. Number of rectangles.
2. Rectangle size.
3. Rectangle fill colour.
4. Rectangle border size.
5. Rectangle border colour.
6. Rectangle fill colour opacity.
7. Rectangle border opacity.
8. Rectangle rotation.
9. Background colour.
You have the full freedom to choose which audio features to use and how to map them to the visual
variables.