Are you interested in adding audio recording and playback functionality to your React Native Expo app? With the rise of audio-based applications and the popularity of podcasts, adding audio capabilities to your app can enhance the user experience and provide new opportunities for engagement. In this tutorial, we will guide you through the process of recording and playing audio in a React Native Expo app, step-by-step. Whether you're building a language learning app, a music player, or a podcast platform, this tutorial will provide you with the skills you need to add audio functionality to your app. So let's get started!
Do not forget to like, comment, and subscribe to the channel before getting into it!
Step 1-) Initialize an Expo App
Make sure you have Node.js and npm installed on your machine. You can download them from the official website: https://nodejs.org/en/download/.
- Open your terminal or command prompt and run the following command to install the Expo CLI globally:
- npm install -g expo-cli
- Once the installation is complete, navigate to the directory where you want to create your app and run the following command:
- expo init my-new-app
- Replace
my-new-app
with the name of your app. This command will create a new directory with the same name as your app and initialize a new Expo project inside it. - Choose a template for your app from the list of available options. You can select a blank template or choose from one of the preconfigured templates that include common features such as navigation, authentication, and more.
- Once you've chosen a template, Expo will install all the necessary dependencies and set up your app. This may take a few minutes, depending on your internet connection.
Step 2-) Add the Following Code to your Component:
import { Text, TouchableOpacity, View, StyleSheet } from 'react-native'; import React, { useState, useEffect } from 'react'; import { Audio } from 'expo-av'; import * as FileSystem from 'expo-file-system'; import { FontAwesome } from '@expo/vector-icons'; export default function App() { const [recording, setRecording] = useState(null); const [recordingStatus, setRecordingStatus] = useState('idle'); const [audioPermission, setAudioPermission] = useState(null); useEffect(() => { // Simply get recording permission upon first render async function getPermission() { await Audio.requestPermissionsAsync().then((permission) => { console.log('Permission Granted: ' + permission.granted); setAudioPermission(permission.granted) }).catch(error => { console.log(error); }); } // Call function to get permission getPermission() // Cleanup upon first render return () => { if (recording) { stopRecording(); } }; }, []); async function startRecording() { try { // needed for IoS if (audioPermission) { await Audio.setAudioModeAsync({ allowsRecordingIOS: true, playsInSilentModeIOS: true }) } const newRecording = new Audio.Recording(); console.log('Starting Recording') await newRecording.prepareToRecordAsync(Audio.RECORDING_OPTIONS_PRESET_HIGH_QUALITY); await newRecording.startAsync(); setRecording(newRecording); setRecordingStatus('recording'); } catch (error) { console.error('Failed to start recording', error); } } async function stopRecording() { try { if (recordingStatus === 'recording') { console.log('Stopping Recording') await recording.stopAndUnloadAsync(); const recordingUri = recording.getURI(); // Create a file name for the recording const fileName = `recording-${Date.now()}.caf`; // Move the recording to the new directory with the new file name await FileSystem.makeDirectoryAsync(FileSystem.documentDirectory + 'recordings/', { intermediates: true }); await FileSystem.moveAsync({ from: recordingUri, to: FileSystem.documentDirectory + 'recordings/' + `${fileName}` }); // This is for simply playing the sound back const playbackObject = new Audio.Sound(); await playbackObject.loadAsync({ uri: FileSystem.documentDirectory + 'recordings/' + `${fileName}` }); await playbackObject.playAsync(); // resert our states to record again setRecording(null); setRecordingStatus('stopped'); } } catch (error) { console.error('Failed to stop recording', error); } } async function handleRecordButtonPress() { if (recording) { const audioUri = await stopRecording(recording); if (audioUri) { console.log('Saved audio file to', savedUri); } } else { await startRecording(); } } return ( <View style={styles.container}> <TouchableOpacity style={styles.button} onPress={handleRecordButtonPress}> <FontAwesome name={recording ? 'stop-circle' : 'circle'} size={64} color="white" /> </TouchableOpacity> <Text style={styles.recordingStatusText}>{`Recording status: ${recordingStatus}`}</Text> </View> ); } const styles = StyleSheet.create({ container: { flex: 1, alignItems: 'center', justifyContent: 'center', }, button: { alignItems: 'center', justifyContent: 'center', width: 128, height: 128, borderRadius: 64, backgroundColor: 'red', }, recordingStatusText: { marginTop: 16, }, });
- In the useEffect we simply ensure we have recording permission from the user, we use the Audio library to do that. We also clean up any existing recordings in the return function of the useEffect.
- startRecording()
- We use this function to start getting Audio from the user.
- We need setAudioModeAsync() to be able to record on IOS
- We initialize an Audio object, prepare, and begin recording all within this function
- stopRecording()
- This function is used to simply stop, save, and playback the recording to the user
- We use the FileSystem library to move the recording URI to the filesystem, and we initialize a Playback Object to play the audio itself
- handleRecordButtonPress()
- This function simply starts or stops a recording based on the state of a recording.
The rest of the App.js file is the html and styling, which you can copy or create your own style!
**Note that the expo library can be buggy with the simulator, so sometimes you may need to close and reopen it for it to work. Make sure you turn up the volume as well in the simulator.