
Sound is more than just waves—it’s data. Whether it’s the rhythmic beat of a song or the subtle tremor in a nervous speaker's voice, audio carries a wealth of hidden information.
For my latest academic project, I wanted to bridge the gap between manipulating sound (DSP) and understanding it (Machine Learning). The result is Sentim, a Python-based web application that you can run directly in your browser.
The first part of the app is a custom Vocoder. It demonstrates how we can use mathematics to alter the physics of sound without destroying it. I used librosa and scipy to build tools that allow you to:
The second part answers a complex question: Can we quantify emotion?
Using the RAVDESS dataset (Ryerson Audio-Visual Database of Emotional Speech and Song), I trained a Random Forest Classifier. The app extracts acoustic features—like timbre (MFCCs), pitch, and spectral contrast—to predict if an uploaded audio clip conveys Joy, Anger, Sadness, or a Neutral state.
It also features a "Heuristic Mode," allowing you to compare modern Machine Learning predictions against a traditional rule-based baseline.
This project was built entirely in Python 3.9+, leveraging the power of open-source data science:

A 2D game developed from scratch, emphasizing robust Object-Oriented Programming (OOP) principles and optimized software architecture. This project demonstrates strong proficiency in Java and efficient performance across more than five levels. The game is published open-source on GitHub with full documentation.

A real-time fire and smoke detection system developed using a fine-tuned YOLO model. The project achieved 92% detection accuracy and was integrated with the drone navigation system , leading to a 30% reduction in false positives. This solution is designed to provide rapid and reliable information for emergency services.