Date: February 21, 2025
The Voice-to-Action App is a React Native application that allows users to record audio, transcribe it into text, and extract actionable insights such as meeting details, key points, decisions, and follow-up tasks. The app integrates with Google Speech-to-Text API for transcription and Perplexity AI for advanced analysis of the transcript.
This project is designed to streamline meeting management by providing users with:
- A transcript of the recording.
- Structured meeting summaries.
- Actionable to-do lists.
- Email integration for sharing meeting details.
- Record audio directly within the app using the device's microphone.
- Supports both iOS and Android platforms.
- Converts recorded audio into text using the Google Speech-to-Text API.
- Extracts structured data from transcripts using Perplexity AI, including:
- Meeting details (date, time, participants).
- Key discussion points.
- Decisions made.
- Action items with assignees and deadlines.
- Compose and send meeting summaries via email directly from the app.
- Modify meeting details and tasks through an intuitive edit modal.
- A visually appealing dark mode design for better user experience.
Here is my main link where I have helped people understand how my project works by giving a hands-on demo:
Watch Demo
Here is my second link where I have explained how the code works in the backend and frontend, line by line:
Watch Code Walkthrough
- React Native
- Expo AV (for audio recording)
- Axios (for backend communication)
- React Native Modal (for modals)
- React Native Linking (for email integration)
- Node.js with Express.js
- Google Speech-to-Text API
- Perplexity AI API
- Install Node.js and npm.
- Install Expo CLI for running the React Native app.
- Set up a Google Cloud account and enable the Speech-to-Text API.
- Obtain an API key from Perplexity AI.
- Clone the repository and navigate to the backend folder:
git clone cd backend - Install dependencies:
npm install - Add your Google Cloud credentials JSON file (
myjson.json) to the backend folder. - Replace
perplexityApikeyintranscriptionService.jswith your Perplexity AI API key. - Start the server:
node index.js
- Navigate to the frontend folder:
cd frontend - Install dependencies:
npm install - Update the
myipvariable inApp.jswith your local machine's IP address (e.g.,192.168.x.x). - Start the Expo development server:
expo start - Scan the QR code from Expo on your mobile device or run it on an emulator.
- Launch the app on your mobile device or emulator.
- Tap "π€ Start Recording" to begin recording audio.
- Tap "βΉοΈ Stop Recording" to stop recording and process the audio.
- View the generated transcript and structured insights (meeting details, tasks, etc.).
- Edit any details by tapping the edit icon βοΈ.
- Share meeting summaries via email by tapping the email icon π§.
project/
βββ backend/
β βββ index.js # Main server file
β βββ transcriptionService.js # Handles transcription and AI analysis
β βββ myjson.json # Google Cloud credentials file (not included in repo)
βββ frontend/
β βββ App.js # Main React Native app file
β βββ styles.js # Styling for UI components
β βββ assets/ # App assets (if any)
βββ README.md # Documentation file
| Home Screen | Transcript & Meeting Details |
|---|---|
![]() |
![]() |
- Add support for multiple languages in transcription.
- Integrate calendar APIs (e.g., Google Calendar) for automatic event creation.
- Enable offline transcription using local models.
For any questions or feedback, feel free to reach out:
Yasharth Bajpai
Email: yasharthbajpai0103@gmail.com

