SIVO
Sign to speech & speech to sign
01 / Overview
For my Final Year Project, I built an app called SIVO. Basically, I wanted to solve a real, everyday problem: the communication gap between the deaf community and hearing people.Instead of relying on a human interpreter, I built a two-way translator that fits right in your pocket. Here is how it works: you point your phone’s camera at someone using Pakistan Sign Language (PSL). The app tracks their hand and body movements, figures out what they are signing, and instantly speaks the words out loud. It also works in reverse—you can talk into the phone, and the app will translate your voice into sign language on the screen.
03 / The Hardest Path
Getting a heavy AI model to run in real-time on a mobile phone was honestly the hardest part. I built the mobile app using React Native, but I offloaded all the heavy AI processing—using tools like TensorFlow, MediaPipe, and OpenCV—to a custom Python cloud server.
04 / Challenges
If I’m being honest, getting an AI to recognize a single, isolated sign—like 'Hello'—is actually pretty straightforward. You show the model the gesture, it memorizes the hand landmarks, and you’re good to go.But real life isn’t a flashcard app. In the real world, people sign continuously and fluidly.When a user transitions from signing the word 'Boss' to the word 'Send,' their hands move through the air in a very unpredictable way. To a trained AI, that blurry, halfway-there hand movement looks like a completely different, random word. Instead of predicting my model was initially spitting out absolute gibberish like 'Boss... Make... Apple... Send.' The 'noise' between the signs was ruining the entire sentence.