Development/Summer of Code/2019/AcousticBrainz
Proposed mentors: ruaok or alastairp
Languages/skills: Python, Postgres, Flask
Forum for discussion
(see also: GSoC - Getting started)
If you want to work on AcousticBrainz you should show that you are able to set up the server software and understand how some of the infrastructure works. Here are some things that you could do to get familiar with the AcousticBrainz project and code:
- Install the server on your computer or use the Vagrant setup scripts to build a virtual machine
- Download the AcousticBrainz submission tool and configure it to compute features for some of your audio files and submit them to the local server that you configured
- Use your preferred programming language to access the API to download the data that you submitted to your server, or other data from the main AcousticBrainz server
- Create an oauth application on the MusicBrainz website and add the configuration information to your AcousticBrainz server. Use this to log in to your server with your MusicBrainz details
- Look at the system to build a Dataset (accessible from your profile page on the AcousticBrainz server) and try and build a simple dataset
Join in on development
We like it when potential students show initiative and make contributions to code without asking us what to do next. We have tagged tickets that we think are suitable for for new contributors with the "good-first-bug" label. Take a look at these tickets and see if any of them grab your interest. It's a good idea to talk to us before starting work on a ticket, to make sure that you understand what tasks are involved to finish the ticket, and to make sure that you're not duplicating any work which has already been done. To talk to us, join our IRC channel or post a message in the forums or on a ticket.
Here are some ideas for projects that we would like to complete in AcousticBrainz in the near future. They are a good size for a Summer of Code project, but are in no way a complete list of possible ideas. If you have other ideas that you think might be interesting for the project join us in IRC and talk to us about your ideas.
Statistics and data description
We have a lot of data in AcousticBrainz, but we don't know much about what this data looks like. This task involves looking at the data that we have and finding interesting ways to show this data to visitors to the AB website. Part of the proposal for this task would be to look at and understand the data and come up with a list of recommended visualisations/descriptions. For many of the types of statistics that we want to show, it is infeasible to compute the data at every page load, therefore part of this task is to also come up with an appropriate caching system.
Here are a few ideas for statistics that we have thought of so far:
Automatic updating statistics page, containing data about our submissions:
- Formats, year, reported genre, other tags (mood)?
- BPM analysis
- Compare audio content md5_encoded with mbids
- Use the musicbrainz mbid redirect tables to find more duplicates
- Lists of artists + albums/recordings for each artist
Visualize AB data - either a sub-dataset/list or all data in AB
- distribution plots for all low-level descriptors
- expectedness of features for each particular track (paper: Corpus Analysis Tools for Computational Hook Discover by Jan Van Balen)
2D visual maps
- Improving visualization of high-dimensional music similarity spaces (Flexter)
- 2d maps with t-Stochastic Neighbor Embedding (TSNE, but there are other approaches in the paper) with shared nearest neighbor distance normalization (against hubs)
New machine learning infrastructure
Skills: Python, C++, Machine learning, scikitlearn, postgres
We build what we call high level models in acousticbrainz, which are multiclass SVM models trained using libsvm wrapped in a custom library called gaia. Gaia performs its task well, but it is written in C++ and not easy to extend with different machine learning algorithms and new techniques like deep learning.
We would like to replace our model training infrastructure with scikit learn, which is widely known and contains a large number of machine learning algorithms
- Understand the existing gaia-based training process
- Reproduce the existing SVM model process using scikitlearn
- Replace the highlevel model training process with scikitlearn
- Perform an analysis of other ML algorithms in scikitlearn to see if they give better results than those that we currently have with SVM
Storage for AcousticBrainz v2 data
When we release a new version of the AcousticBrainz extractor tool we will want to store data for this new version in addition to data from the current version of the extractor that we provide.
This project needs to consider at least the following items:
- Update the database schema to include a data version field, and allow the Submit and Read methods to switch between them.
- Update the frontend including the dataset editor
- Update the client software to include a check where they announce to the server what version they are
- Update the client software to enhance the "already submitted" local database to allow data from the new version of the extractor
Skills: Python, Postgres, data crunching
AcousticBrainz contains acoustic information for a large number of recordings (music tracks). One of the important tools that we haven’t created yet is a tool to compare recordings against each other in order to determine how similar they are. The similarity of recordings is an important data set that we wish to include in future recommendation engines in the ListenBrainz project. This topic has been the focus of a previous thesis and there are pieces of this thesis that we can use to make this project considerably easier. The work in this thesis did not project a result that would work for a large number of recordings.
The project breaks into three parts:
- Take existing code from the thesis and to extract the feature vectors for each of the recordings in AcousticBrainz and then calculate their similarity in a reasonable amount of time. This portion of the project should mostly involve getting an existing piece of code working -- it does not involve selecting which features to use for our comparisons.
- Once the feature vectors have been extracted, we will want to compare the vectors for each recording against the vectors of all other recordings. This is an O(n^2) operation and must work on all of the unique recordings in AcousticBrainz -- some 4 million recordings. Clearly a complex O(n^2) operation on 4M tracks would require enormous computing power that we do not have at our disposal. We hope that the use of Annoy will allow us to run this algorithm on a large memory virtual machine and to compute the similarity values between all of the recordings in a reasonably short period of time. As part of this project you should install Annoy and write the necessary tools to feed the AB feature vectors to Annoy and to build Annoy indexes for all 4M tracks.
- Finally we will lookup each of the 4M recordings in order to compute the similarity between each of the recordings and create a data dump that contains the mapping between two recordings and their similarity score.