Updated
about 1 hour
ago.
1 hackers ♥️.
1 follower.
Description
Contrastive Language-Audio Pretraining (CLAP) is an opensoruce library that enable to train a neural network on both Audio and Text description enabling the possibility to search Audio from a Text input. On huggingface are also already present some pre-trained model to search on Songs.
Goals
Check how CLAP can be used for searching on song and which kind of query can give a good result by developing a Minimum Viable Product (MVP) in Python. Based on the result of this MVP, the future next step could be:
- Music Tagging;
- Free text search;
- Integration with an LLM model (for example with MCP or OpenAPI api) for music suggestions on your own library
The code of this project will be entirely written with AI to better explore and proof the AI capability.
Resources
- CLAP: the model object of the research;
- huggingface: pre-trained model for CLAP;
- Free Music Archive: creative common songs that can be used for testing;
- Colab: to be used as development envirorment.
Looking for hackers with the skills:
Nothing? Add some keywords!
This project is part of:
Hack Week 25
Activity
Comments
Be the first to comment!
Similar Projects
This project is one of its kind!