Compare commits
38 Commits
dafbc486ad
...
LINXD-2270
Author | SHA1 | Date | |
---|---|---|---|
451fff0a6b | |||
a568e542a7 | |||
14af825eab | |||
f3ba6d37c2 | |||
a4356f06be | |||
c4d6cdd209 | |||
1520fb88ae | |||
da906ed4ba | |||
2dd9eb5eaf | |||
b4fccc4d4c | |||
88da70731f | |||
68c80d563f | |||
eb668e2500 | |||
e3536e87cd | |||
8634cb4b6e | |||
2a6f24b2bb | |||
381e665062 | |||
33e30339f2 | |||
968da6ea98 | |||
2ada7b66db | |||
28059144cf | |||
714fe0ec5e | |||
898cc0cbf2 | |||
6379e1ae34 | |||
dd264e39ea | |||
72136132ba | |||
4978e8d51f | |||
825ded83c2 | |||
b15357d089 | |||
bb4cd756f3 | |||
0731090b0e | |||
782683c3e2 | |||
dfa175d0c7 | |||
2c375c01ea | |||
2fbb355fea | |||
2ddeb4baaa | |||
a31e646e2b | |||
fe792f93b6 |
10
app.js
10
app.js
@ -428,7 +428,9 @@ peers.on('connection', async socket => {
|
||||
callback({ videoParams, audioParams });
|
||||
} else if (!canConsumeVideo && canConsumeAudio) {
|
||||
const audioParams = await consumeAudio(callId, socket.id, rtpCapabilities)
|
||||
callback({ videoParams: null, audioParams });
|
||||
const data = { videoParams: null, audioParams };
|
||||
console.log('-----------======= data', data);
|
||||
callback(data);
|
||||
} else {
|
||||
console.log(`[consume] Can't consume | callId ${callId}`);
|
||||
callback(null);
|
||||
@ -449,11 +451,13 @@ peers.on('connection', async socket => {
|
||||
console.log(`[consumer-resume] callId ${callId}`)
|
||||
|
||||
if (isInitiator(callId, socket.id)) {
|
||||
console.log(`[consumer-resume] isInitiator true`);
|
||||
await videoCalls[callId].initiatorConsumerVideo.resume();
|
||||
await videoCalls[callId].initiatorConsumerAudio.resume();
|
||||
} else {
|
||||
await videoCalls[callId].receiverConsumerVideo.resume();
|
||||
await videoCalls[callId].receiverConsumerAudio.resume();
|
||||
console.log(`[consumer-resume] isInitiator false`);
|
||||
(videoCalls[callId].receiverConsumerVideo) && await videoCalls[callId].receiverConsumerVideo.resume();
|
||||
(videoCalls[callId].receiverConsumerVideo) && await videoCalls[callId].receiverConsumerAudio.resume();
|
||||
}
|
||||
// await videoCalls[callId].consumerVideo.resume();
|
||||
// await videoCalls[callId].consumerAudio.resume();
|
||||
|
183
public/bundle.js
183
public/bundle.js
@ -20368,10 +20368,24 @@ const ASSET_NAME = urlParams.get('assetName') || null;
|
||||
const ASSET_TYPE = urlParams.get('assetType') || null;
|
||||
let callId = parseInt(urlParams.get('callId')) || null;
|
||||
const IS_PRODUCER = urlParams.get('producer') === 'true' ? true : false
|
||||
let remoteVideo = document.getElementById('remoteVideo')
|
||||
remoteVideo.defaultMuted = true
|
||||
let produceAudio = false
|
||||
|
||||
console.log('[URL] ASSET_ID', ASSET_ID, '| ACCOUNT_ID', ACCOUNT_ID, '| callId', callId, ' | IS_PRODUCER', IS_PRODUCER)
|
||||
|
||||
console.log('🟩 config', config)
|
||||
|
||||
produceAudioSelector = document.getElementById('produceAudio');
|
||||
produceAudioSelector.addEventListener('change', e => {
|
||||
if(e.target.checked) {
|
||||
produceAudio = true
|
||||
console.log('produce audio');
|
||||
} else {
|
||||
produceAudio = false
|
||||
}
|
||||
});
|
||||
|
||||
let socket, hub
|
||||
let device
|
||||
let rtpCapabilities
|
||||
@ -20381,6 +20395,21 @@ let producerVideo
|
||||
let producerAudio
|
||||
let consumer
|
||||
let originAssetId
|
||||
let consumerVideo // local consumer video(consumer not transport)
|
||||
let consumerAudio // local consumer audio(consumer not transport)
|
||||
|
||||
const remoteSoundControl = document.getElementById('remoteSoundControl');
|
||||
|
||||
remoteSoundControl.addEventListener('click', function handleClick() {
|
||||
console.log('remoteSoundControl.textContent', remoteSoundControl.textContent);
|
||||
if (remoteSoundControl.textContent === 'Unmute') {
|
||||
remoteVideo.muted = false
|
||||
remoteSoundControl.textContent = 'Mute';
|
||||
} else {
|
||||
remoteVideo.muted = true
|
||||
remoteSoundControl.textContent = 'Unmute';
|
||||
}
|
||||
});
|
||||
|
||||
// https://mediasoup.org/documentation/v3/mediasoup-client/api/#ProducerOptions
|
||||
// https://mediasoup.org/documentation/v3/mediasoup-client/api/#transport-produce
|
||||
@ -20502,7 +20531,7 @@ const streamSuccess = (stream) => {
|
||||
const getLocalStream = () => {
|
||||
console.log('[getLocalStream]');
|
||||
navigator.mediaDevices.getUserMedia({
|
||||
audio: true,
|
||||
audio: produceAudio ? true : false,
|
||||
video: {
|
||||
qvga : { width: { ideal: 320 }, height: { ideal: 240 } },
|
||||
vga : { width: { ideal: 640 }, height: { ideal: 480 } },
|
||||
@ -20651,34 +20680,36 @@ const connectSendTransport = async () => {
|
||||
// this action will trigger the 'connect' and 'produce' events above
|
||||
|
||||
// Produce video
|
||||
producerVideo = await producerTransport.produce(videoParams)
|
||||
let producerVideoHandler = await producerTransport.produce(videoParams)
|
||||
console.log('videoParams', videoParams);
|
||||
console.log('producerVideo', producerVideo);
|
||||
|
||||
producerVideo.on('trackended', () => {
|
||||
producerVideoHandler.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close video track
|
||||
})
|
||||
|
||||
producerVideo.on('transportclose', () => {
|
||||
producerVideoHandler.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close video track
|
||||
})
|
||||
|
||||
// Produce audio
|
||||
producerAudio = await producerTransport.produce(audioParams)
|
||||
console.log('audioParams', audioParams);
|
||||
console.log('producerAudio', producerAudio);
|
||||
|
||||
producerAudio.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close audio track
|
||||
})
|
||||
|
||||
producerAudio.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close audio track
|
||||
})
|
||||
if (produceAudio) {
|
||||
let producerAudioHandler = await producerTransport.produce(audioParams)
|
||||
console.log('audioParams', audioParams);
|
||||
console.log('producerAudio', producerAudio);
|
||||
|
||||
producerAudioHandler.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close audio track
|
||||
})
|
||||
|
||||
producerAudioHandler.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close audio track
|
||||
})
|
||||
}
|
||||
|
||||
const answer = {
|
||||
origin_asset_id: ASSET_ID,
|
||||
@ -20762,34 +20793,108 @@ const connectRecvTransport = async () => {
|
||||
await socket.emit('consume', {
|
||||
rtpCapabilities: device.rtpCapabilities,
|
||||
callId
|
||||
}, async ({ params }) => {
|
||||
if (params.error) {
|
||||
console.log('Cannot Consume')
|
||||
return
|
||||
}
|
||||
|
||||
// Then consume with the local consumer transport
|
||||
// which creates a consumer
|
||||
consumer = await consumerTransport.consume({
|
||||
id: params.id,
|
||||
producerId: params.producerId,
|
||||
kind: params.kind,
|
||||
rtpParameters: params.rtpParameters
|
||||
})
|
||||
|
||||
// destructure and retrieve the video track from the producer
|
||||
const { track } = consumer
|
||||
}, async ({videoParams, audioParams}) => {
|
||||
console.log(`[consume] 🟩 videoParams`, videoParams)
|
||||
console.log(`[consume] 🟩 audioParams`, audioParams)
|
||||
console.log('[consume] 🟩 consumerTransport', consumerTransport)
|
||||
|
||||
let stream = new MediaStream()
|
||||
stream.addTrack(track)
|
||||
// stream.removeTrack(track)
|
||||
remoteVideo.srcObject = stream
|
||||
socket.emit('consumer-resume')
|
||||
console.log('consumer', consumer);
|
||||
|
||||
// Maybe the unit does not produce video or audio, so we must only consume what is produced
|
||||
if (videoParams) {
|
||||
console.log('❗ Have VIDEO stream to consume');
|
||||
stream.addTrack(await getVideoTrask(videoParams))
|
||||
} else {
|
||||
console.log('❗ Don\'t have VIDEO stream to consume');
|
||||
}
|
||||
|
||||
if (audioParams) {
|
||||
console.log('❗ Have AUDIO stream to consume');
|
||||
let audioTrack = await getAudioTrask(audioParams)
|
||||
console.log('audioTrack', audioTrack);
|
||||
stream.addTrack(audioTrack)
|
||||
} else {
|
||||
console.log('❗ Don\'t have AUDIO stream to consume');
|
||||
}
|
||||
|
||||
console.log('----------stream', stream);
|
||||
console.log('stream.getAudioTracks()', stream.getAudioTracks());
|
||||
socket.emit('consumer-resume')
|
||||
|
||||
remoteVideo.srcObject = stream
|
||||
// remoteVideo.autoplay = true
|
||||
remoteVideo.controls = true;
|
||||
remoteVideo.muted = true;
|
||||
remoteVideo.loop = true;
|
||||
remoteVideo.setAttribute('playsinline', '');
|
||||
remoteVideo.src = stream;
|
||||
remoteVideo.volume = 1.0;
|
||||
|
||||
// window.localStream = stream; // A
|
||||
// window.localAudio.srcObject = stream; // B
|
||||
// window.localAudio.autoplay = true; // C
|
||||
|
||||
remoteVideo.play()
|
||||
.then(() => {
|
||||
console.log('remoteVideo PLAY')
|
||||
})
|
||||
.catch((error) => {
|
||||
displayError(`remoteVideo PLAY ERROR | ${error.message}`)
|
||||
})
|
||||
})
|
||||
}
|
||||
|
||||
const getVideoTrask = async (videoParams) => {
|
||||
consumerVideo = await consumerTransport.consume({
|
||||
id: videoParams.id,
|
||||
producerId: videoParams.producerId,
|
||||
kind: videoParams.kind,
|
||||
rtpParameters: videoParams.rtpParameters
|
||||
})
|
||||
|
||||
consumerVideo.on('transportclose', () => {
|
||||
console.log('transport closed so consumer closed')
|
||||
})
|
||||
|
||||
return consumerVideo.track
|
||||
}
|
||||
|
||||
const getAudioTrask = async (audioParams) => {
|
||||
consumerAudio = await consumerTransport.consume({
|
||||
id: audioParams.id,
|
||||
producerId: audioParams.producerId,
|
||||
kind: audioParams.kind,
|
||||
rtpParameters: audioParams.rtpParameters
|
||||
})
|
||||
|
||||
consumerAudio.on('transportclose', () => {
|
||||
console.log('transport closed so consumer closed')
|
||||
})
|
||||
|
||||
const audioTrack = consumerAudio.track
|
||||
|
||||
// audioTrack.applyConstraints({
|
||||
// audio: {
|
||||
// advanced: [
|
||||
// {
|
||||
// echoCancellation: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// autoGainControl: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// noiseSuppression: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// highpassFilter: {exact: true}
|
||||
// }
|
||||
// ]
|
||||
// }
|
||||
// })
|
||||
|
||||
return audioTrack
|
||||
}
|
||||
|
||||
const closeCall = () => {
|
||||
console.log('closeCall');
|
||||
|
||||
|
BIN
public/images/favicon.ico
Normal file
BIN
public/images/favicon.ico
Normal file
Binary file not shown.
After Width: | Height: | Size: 4.2 KiB |
BIN
public/images/volume-high.png
Normal file
BIN
public/images/volume-high.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 493 B |
BIN
public/images/volume-mute.png
Normal file
BIN
public/images/volume-mute.png
Normal file
Binary file not shown.
After Width: | Height: | Size: 418 B |
@ -34,6 +34,9 @@
|
||||
<body>
|
||||
<body>
|
||||
<div id="video">
|
||||
<legend>Client options:</legend>
|
||||
<input type="checkbox" id="produceAudio" name="produceAudio">
|
||||
<label for="produceAudio">Produce audio</label><br>
|
||||
<table>
|
||||
<thead>
|
||||
<th>Local Video</th>
|
||||
@ -43,12 +46,24 @@
|
||||
<tr>
|
||||
<td>
|
||||
<div id="sharedBtns">
|
||||
<video id="localVideo" autoplay class="video" muted></video>
|
||||
<video
|
||||
id="localVideo"
|
||||
class="video"
|
||||
autoplay
|
||||
muted
|
||||
playsinline
|
||||
></video>
|
||||
</div>
|
||||
</td>
|
||||
<td>
|
||||
<div id="sharedBtns">
|
||||
<video id="remoteVideo" autoplay class="video" ></video>
|
||||
<video
|
||||
id="remoteVideo"
|
||||
class="video"
|
||||
autoplay
|
||||
muted
|
||||
playsinline
|
||||
></video>
|
||||
</div>
|
||||
</td>
|
||||
</tr>
|
||||
@ -61,33 +76,10 @@
|
||||
<td>
|
||||
<div id="sharedBtns">
|
||||
<button id="btnRecvSendTransport">Consume</button>
|
||||
<button id="remoteSoundControl">Unmute</button>
|
||||
</div>
|
||||
</td>
|
||||
</tr>
|
||||
<!-- <tr>
|
||||
<td colspan="2">
|
||||
<div id="sharedBtns">
|
||||
<button id="btnRtpCapabilities">2. Get Rtp Capabilities</button>
|
||||
<br />
|
||||
<button id="btnDevice">3. Create Device</button>
|
||||
</div>
|
||||
</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td>
|
||||
<div id="sharedBtns">
|
||||
<button id="btnCreateSendTransport">4. Create Send Transport</button>
|
||||
<br />
|
||||
<button id="btnConnectSendTransport">5. Connect Send Transport & Produce</button></td>
|
||||
</div>
|
||||
<td>
|
||||
<div id="sharedBtns">
|
||||
<button id="btnRecvSendTransport">6. Create Recv Transport</button>
|
||||
<br />
|
||||
<button id="btnConnectRecvTransport">7. Connect Recv Transport & Consume</button>
|
||||
</div>
|
||||
</td>
|
||||
</tr> -->
|
||||
</tbody>
|
||||
</table>
|
||||
<div id="closeCallBtn">
|
||||
|
185
public/index.js
185
public/index.js
@ -10,10 +10,24 @@ const ASSET_NAME = urlParams.get('assetName') || null;
|
||||
const ASSET_TYPE = urlParams.get('assetType') || null;
|
||||
let callId = parseInt(urlParams.get('callId')) || null;
|
||||
const IS_PRODUCER = urlParams.get('producer') === 'true' ? true : false
|
||||
let remoteVideo = document.getElementById('remoteVideo')
|
||||
remoteVideo.defaultMuted = true
|
||||
let produceAudio = false
|
||||
|
||||
console.log('[URL] ASSET_ID', ASSET_ID, '| ACCOUNT_ID', ACCOUNT_ID, '| callId', callId, ' | IS_PRODUCER', IS_PRODUCER)
|
||||
|
||||
console.log('🟩 config', config)
|
||||
|
||||
produceAudioSelector = document.getElementById('produceAudio');
|
||||
produceAudioSelector.addEventListener('change', e => {
|
||||
if(e.target.checked) {
|
||||
produceAudio = true
|
||||
console.log('produce audio');
|
||||
} else {
|
||||
produceAudio = false
|
||||
}
|
||||
});
|
||||
|
||||
let socket, hub
|
||||
let device
|
||||
let rtpCapabilities
|
||||
@ -23,6 +37,21 @@ let producerVideo
|
||||
let producerAudio
|
||||
let consumer
|
||||
let originAssetId
|
||||
let consumerVideo // local consumer video(consumer not transport)
|
||||
let consumerAudio // local consumer audio(consumer not transport)
|
||||
|
||||
const remoteSoundControl = document.getElementById('remoteSoundControl');
|
||||
|
||||
remoteSoundControl.addEventListener('click', function handleClick() {
|
||||
console.log('remoteSoundControl.textContent', remoteSoundControl.textContent);
|
||||
if (remoteSoundControl.textContent === 'Unmute') {
|
||||
remoteVideo.muted = false
|
||||
remoteSoundControl.textContent = 'Mute';
|
||||
} else {
|
||||
remoteVideo.muted = true
|
||||
remoteSoundControl.textContent = 'Unmute';
|
||||
}
|
||||
});
|
||||
|
||||
// https://mediasoup.org/documentation/v3/mediasoup-client/api/#ProducerOptions
|
||||
// https://mediasoup.org/documentation/v3/mediasoup-client/api/#transport-produce
|
||||
@ -144,7 +173,7 @@ const streamSuccess = (stream) => {
|
||||
const getLocalStream = () => {
|
||||
console.log('[getLocalStream]');
|
||||
navigator.mediaDevices.getUserMedia({
|
||||
audio: true,
|
||||
audio: produceAudio ? true : false,
|
||||
video: {
|
||||
qvga : { width: { ideal: 320 }, height: { ideal: 240 } },
|
||||
vga : { width: { ideal: 640 }, height: { ideal: 480 } },
|
||||
@ -293,34 +322,36 @@ const connectSendTransport = async () => {
|
||||
// this action will trigger the 'connect' and 'produce' events above
|
||||
|
||||
// Produce video
|
||||
producerVideo = await producerTransport.produce(videoParams)
|
||||
let producerVideoHandler = await producerTransport.produce(videoParams)
|
||||
console.log('videoParams', videoParams);
|
||||
console.log('producerVideo', producerVideo);
|
||||
|
||||
producerVideo.on('trackended', () => {
|
||||
producerVideoHandler.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close video track
|
||||
})
|
||||
|
||||
producerVideo.on('transportclose', () => {
|
||||
producerVideoHandler.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close video track
|
||||
})
|
||||
|
||||
// Produce audio
|
||||
producerAudio = await producerTransport.produce(audioParams)
|
||||
console.log('audioParams', audioParams);
|
||||
console.log('producerAudio', producerAudio);
|
||||
|
||||
producerAudio.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close audio track
|
||||
})
|
||||
|
||||
producerAudio.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close audio track
|
||||
})
|
||||
if (produceAudio) {
|
||||
let producerAudioHandler = await producerTransport.produce(audioParams)
|
||||
console.log('audioParams', audioParams);
|
||||
console.log('producerAudio', producerAudio);
|
||||
|
||||
producerAudioHandler.on('trackended', () => {
|
||||
console.log('track ended')
|
||||
// close audio track
|
||||
})
|
||||
|
||||
producerAudioHandler.on('transportclose', () => {
|
||||
console.log('transport ended')
|
||||
// close audio track
|
||||
})
|
||||
}
|
||||
|
||||
const answer = {
|
||||
origin_asset_id: ASSET_ID,
|
||||
@ -404,36 +435,108 @@ const connectRecvTransport = async () => {
|
||||
await socket.emit('consume', {
|
||||
rtpCapabilities: device.rtpCapabilities,
|
||||
callId
|
||||
}, async ({ params }) => {
|
||||
if (params.error) {
|
||||
console.log('Cannot Consume')
|
||||
return
|
||||
}
|
||||
|
||||
console.log(`[connectRecvTransport] consume params ${params}`);
|
||||
|
||||
// Then consume with the local consumer transport
|
||||
// which creates a consumer
|
||||
consumer = await consumerTransport.consume({
|
||||
id: params.id,
|
||||
producerId: params.producerId,
|
||||
kind: params.kind,
|
||||
rtpParameters: params.rtpParameters
|
||||
})
|
||||
|
||||
// destructure and retrieve the video track from the producer
|
||||
const { track } = consumer
|
||||
}, async ({videoParams, audioParams}) => {
|
||||
console.log(`[consume] 🟩 videoParams`, videoParams)
|
||||
console.log(`[consume] 🟩 audioParams`, audioParams)
|
||||
console.log('[consume] 🟩 consumerTransport', consumerTransport)
|
||||
|
||||
let stream = new MediaStream()
|
||||
stream.addTrack(track)
|
||||
// stream.removeTrack(track)
|
||||
remoteVideo.srcObject = stream
|
||||
socket.emit('consumer-resume')
|
||||
console.log('consumer', consumer);
|
||||
|
||||
// Maybe the unit does not produce video or audio, so we must only consume what is produced
|
||||
if (videoParams) {
|
||||
console.log('❗ Have VIDEO stream to consume');
|
||||
stream.addTrack(await getVideoTrask(videoParams))
|
||||
} else {
|
||||
console.log('❗ Don\'t have VIDEO stream to consume');
|
||||
}
|
||||
|
||||
if (audioParams) {
|
||||
console.log('❗ Have AUDIO stream to consume');
|
||||
let audioTrack = await getAudioTrask(audioParams)
|
||||
console.log('audioTrack', audioTrack);
|
||||
stream.addTrack(audioTrack)
|
||||
} else {
|
||||
console.log('❗ Don\'t have AUDIO stream to consume');
|
||||
}
|
||||
|
||||
console.log('----------stream', stream);
|
||||
console.log('stream.getAudioTracks()', stream.getAudioTracks());
|
||||
socket.emit('consumer-resume')
|
||||
|
||||
remoteVideo.srcObject = stream
|
||||
// remoteVideo.autoplay = true
|
||||
remoteVideo.controls = true;
|
||||
remoteVideo.muted = true;
|
||||
remoteVideo.loop = true;
|
||||
remoteVideo.setAttribute('playsinline', '');
|
||||
remoteVideo.src = stream;
|
||||
remoteVideo.volume = 1.0;
|
||||
|
||||
// window.localStream = stream; // A
|
||||
// window.localAudio.srcObject = stream; // B
|
||||
// window.localAudio.autoplay = true; // C
|
||||
|
||||
remoteVideo.play()
|
||||
.then(() => {
|
||||
console.log('remoteVideo PLAY')
|
||||
})
|
||||
.catch((error) => {
|
||||
displayError(`remoteVideo PLAY ERROR | ${error.message}`)
|
||||
})
|
||||
})
|
||||
}
|
||||
|
||||
const getVideoTrask = async (videoParams) => {
|
||||
consumerVideo = await consumerTransport.consume({
|
||||
id: videoParams.id,
|
||||
producerId: videoParams.producerId,
|
||||
kind: videoParams.kind,
|
||||
rtpParameters: videoParams.rtpParameters
|
||||
})
|
||||
|
||||
consumerVideo.on('transportclose', () => {
|
||||
console.log('transport closed so consumer closed')
|
||||
})
|
||||
|
||||
return consumerVideo.track
|
||||
}
|
||||
|
||||
const getAudioTrask = async (audioParams) => {
|
||||
consumerAudio = await consumerTransport.consume({
|
||||
id: audioParams.id,
|
||||
producerId: audioParams.producerId,
|
||||
kind: audioParams.kind,
|
||||
rtpParameters: audioParams.rtpParameters
|
||||
})
|
||||
|
||||
consumerAudio.on('transportclose', () => {
|
||||
console.log('transport closed so consumer closed')
|
||||
})
|
||||
|
||||
const audioTrack = consumerAudio.track
|
||||
|
||||
// audioTrack.applyConstraints({
|
||||
// audio: {
|
||||
// advanced: [
|
||||
// {
|
||||
// echoCancellation: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// autoGainControl: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// noiseSuppression: {exact: true}
|
||||
// },
|
||||
// {
|
||||
// highpassFilter: {exact: true}
|
||||
// }
|
||||
// ]
|
||||
// }
|
||||
// })
|
||||
|
||||
return audioTrack
|
||||
}
|
||||
|
||||
const closeCall = () => {
|
||||
console.log('closeCall');
|
||||
|
||||
|
Reference in New Issue
Block a user