完成mp4

This commit is contained in:
youngq 2024-09-23 13:03:05 +08:00
parent 606715ced6
commit 1b7480cadc
3 changed files with 152 additions and 138 deletions

View File

@ -278,8 +278,8 @@ app.on('ready', () => {
return app.getPath('userData'); return app.getPath('userData');
}); });
// 用户数据目录路径 // 用户数据目录路径
global.userDataPath = app.getPath('userData'); // 全局变量 const userDataPath = app.getPath('userData'); // 全局变量
console.log('User Data Path:', global.userDataPath); console.log('User Data Path:', userDataPath);
// 检查并下载 ffmpeg // 检查并下载 ffmpeg
checkAndDownloadFFmpeg(userDataPath) checkAndDownloadFFmpeg(userDataPath)
.then(() => { .then(() => {

View File

@ -1,7 +1,7 @@
{ {
"name": "WGShare.Metting", "name": "WGShare.Metting",
"private": true, "private": true,
"version": "0.1.14", "version": "0.3.0",
"main": "main.js", "main": "main.js",
"authors": "yj", "authors": "yj",
"description": "智汇享", "description": "智汇享",

View File

@ -21,9 +21,10 @@ import StupWizard from '@/components/StupWizard';
import EquipmentManagement from '@/components/EquipmentManagement'; import EquipmentManagement from '@/components/EquipmentManagement';
import UserVideo from '@/components/UserVideo'; import UserVideo from '@/components/UserVideo';
import { role } from '@/config/role'; import { role } from '@/config/role';
import path from 'path';
const { confirm } = Modal;
const { ipcRenderer } = require('electron'); const { ipcRenderer } = require('electron');
import * as path from 'path';
const { confirm } = Modal;
const { exec } = require('child_process'); const { exec } = require('child_process');
const fs = require('fs').promises; const fs = require('fs').promises;
dayjs.extend(durationPlugin); dayjs.extend(durationPlugin);
@ -547,39 +548,38 @@ const Meeting: React.FC = () => {
}); });
const reader = new FileReader() as any; const reader = new FileReader() as any;
reader.onload = async () => { reader.onload = async () => {
// const setting = await JSON.parse(storage.getItem('setting') as string) try {
// const buffer = Buffer.from(reader.result); const userDataPath = await ipcRenderer.invoke('get-user-data-path');
// await fs.writeFile(`${setting.recordingFilesPath}会议录制_${state.roomName}_${state.channelId}_${+new Date()}.mp4`, buffer, {});
// 获取当前日期并格式化 // 获取当前日期并格式化
const date = new Date(); const date = new Date();
const year = date.getFullYear(); const year = date.getFullYear();
const month = date.getMonth() + 1; // JavaScript月份从0开始 const month = date.getMonth() + 1; // JavaScript月份从0开始
const day = date.getDate(); const day = date.getDate();
const hours = date.getHours(); const hours = date.getHours();
const minutes = date.getMinutes(); const minutes = date.getMinutes();
const formattedDate = `${year}${month}${day}${hours}${minutes}`; const formattedDate = `${year}${month}${day}${hours}${minutes}`;
const setting = await JSON.parse(storage.getItem('setting') as string) const setting = await JSON.parse(storage.getItem('setting') as string)
const buffer = Buffer.from(reader.result); const buffer = Buffer.from(reader.result);
const mp4Path = `${setting.recordingFilesPath}会议录制_${state.roomName}_${state.channelId}_${formattedDate}_beforehanlder.mp4`; const mp4Path=`${setting.recordingFilesPath}会议录制_${state.roomName}_${state.channelId}_${formattedDate}_beforehanlder.mp4`;
await fs.writeFile(mp4Path, buffer); await fs.writeFile(mp4Path, buffer);
const userDataPath = await ipcRenderer.invoke('get-user-data-path'); // 获取应用程序安装路径
// 获取应用程序安装路径
const ffmpegPath = path.join(userDataPath, "ffmpeg.exe"); const ffmpegPath = path.join(userDataPath, "ffmpeg.exe");
const inputFilePath = mp4Path; // 输入文件路径 const inputFilePath = mp4Path; // 输入文件路径
const outputFilePath = mp4Path.replace('_beforehanlder', ''); // 输出文件路径 const outputFilePath = mp4Path.replace('_beforehanlder',''); // 输出文件路径
const command = `${ffmpegPath} -i "${inputFilePath}" -vcodec copy -acodec copy "${outputFilePath}"`; const command = `${ffmpegPath} -i "${inputFilePath}" -vcodec copy -acodec copy "${outputFilePath}"`;
exec(command, (error, stdout, stderr) => { exec(command, (error, stdout, stderr) => {
if (error) { if (error) {
console.error('Error executing ffmpeg command:', error); console.error('Error executing ffmpeg command:', error);
return; return;
} }
// 删除输入文件
// 删除输入文件
fs.unlink(inputFilePath, (err) => { fs.unlink(inputFilePath, (err) => {
if (err) { if (err) {
console.error('Error deleting input file:', err); console.error('Error deleting input file:', err);
@ -589,27 +589,32 @@ const Meeting: React.FC = () => {
}); });
confirm({
title: '提示',
icon: <ExclamationCircleFilled />,
content: `录制成功!文件已保存至:${setting.recordingFilesPath}`,
centered: true,
okText: '打开文件夹',
cancelText: '关闭',
async onOk() {
await fs.access(setting.recordingFilesPath, fs.constants.F_OK);
if (process.platform === 'win32') {
exec(`explorer "${setting.recordingFilesPath}"`);
} else if (process.platform === 'darwin') {
exec(`open "${setting.recordingFilesPath}"`);
}
},
onCancel() {
}
})
});
} catch (err) {
console.error('处理录制时出错:', err);
}
}
confirm({
title: '提示',
icon: <ExclamationCircleFilled />,
content: `录制成功!文件已保存至:${setting.recordingFilesPath}`,
centered: true,
okText: '打开文件夹',
cancelText: '关闭',
async onOk() {
await fs.access(setting.recordingFilesPath, fs.constants.F_OK);
if (process.platform === 'win32') {
exec(`explorer "${setting.recordingFilesPath}"`);
} else if (process.platform === 'darwin') {
exec(`open "${setting.recordingFilesPath}"`);
}
},
onCancel() {
}
})
})
};
reader.readAsArrayBuffer(blob); reader.readAsArrayBuffer(blob);
} }
}; };
@ -627,32 +632,32 @@ const Meeting: React.FC = () => {
return () => clearTimeout(timer); return () => clearTimeout(timer);
}, [isClicked]); }, [isClicked]);
// useEffect(() => { useEffect(() => {
// const elements = document.querySelectorAll('.intersectionObserver-view'); const elements = document.querySelectorAll('.intersectionObserver-view');
// if (elements.length && currentVideoId) { if (elements.length && currentVideoId) {
// elements.forEach(element => { elements.forEach(element => {
// observer?.unobserve(element); observer?.unobserve(element);
// }); });
// const observerObject = new IntersectionObserver(async (entries: IntersectionObserverEntry[], _observer: IntersectionObserver) => { const observerObject = new IntersectionObserver(async (entries: IntersectionObserverEntry[], _observer: IntersectionObserver) => {
// entries.forEach(async (entry) => { entries.forEach(async (entry) => {
// if (entry.target.id !== user.uid) { if (entry.target.id !== user.uid) {
// await agora.muteRemoteVideoStreamEx(Number(entry.target.id), !entry.isIntersecting) await agora.muteRemoteVideoStreamEx(Number(entry.target.id), !entry.isIntersecting)
// } }
// }); });
// await agora.muteRemoteVideoStreamEx(Number(currentVideoId), false) await agora.muteRemoteVideoStreamEx(Number(currentVideoId), false)
// }, { threshold: 0, root: document.getElementById('videoView') }); }, { threshold: 0, root: document.getElementById('videoView') });
// setObserver(observerObject) setObserver(observerObject)
// elements.forEach(element => { elements.forEach(element => {
// observerObject.observe(element); observerObject.observe(element);
// }); });
// } }
// return () => { return () => {
// elements.forEach(element => { elements.forEach(element => {
// observer?.unobserve(element); observer?.unobserve(element);
// }); });
// observer?.disconnect(); observer?.disconnect();
// } }
// }, [roomUserList, currentVideoId]); }, [roomUserList, currentVideoId]);
// 声网初始化 // 声网初始化
const agoraInit = async () => { const agoraInit = async () => {
@ -1061,67 +1066,74 @@ const Meeting: React.FC = () => {
}) })
break; break;
case '录制': case '录制':
const setting = await JSON.parse(storage.getItem('setting') as string); const setting = await JSON.parse(storage.getItem('setting') as string);
try { try {
await fs.access(setting.recordingFilesPath, fs.constants.F_OK); await fs.access(setting.recordingFilesPath, fs.constants.F_OK);
footerListTemplate[itemIndex][rowIndex].title = '录制中'; footerListTemplate[itemIndex][rowIndex].title = '录制中';
footerListTemplate[itemIndex][rowIndex].active = true; footerListTemplate[itemIndex][rowIndex].active = true;
setFooterList(footerListTemplate); setFooterList(footerListTemplate);
window.electron.getSources().then(async (sources: any) => { window.electron.getSources().then(async (sources: any) => {
const screenId = sources[0].id; const screenId = sources[0].id;
const stream = await navigator.mediaDevices.getUserMedia({
audio: { const stream = await navigator.mediaDevices.getUserMedia({
mandatory: { audio: {
chromeMediaSource: 'desktop', mandatory: {
chromeMediaSourceId: screenId, chromeMediaSource: 'desktop',
} chromeMediaSourceId: screenId,
} as any, }
video: { },
mandatory: { video: {
chromeMediaSource: 'desktop', mandatory: {
chromeMediaSourceId: screenId, chromeMediaSource: 'desktop',
} chromeMediaSourceId: screenId,
} as any }
}); }
// 获取所有音频输入设备
const devices = await navigator.mediaDevices.enumerateDevices();
const audioInputDevices = devices.filter(device => device.kind === 'audioinput' &&
device.deviceId !== 'default' &&
device.deviceId !== 'communications');
// 使用Web Audio API来捕获系统声音和麦克风声音将它们合并到同一个MediaStream中。
const audioCtx = new (window.AudioContext || (window as any).webkitAudioContext)();
const systemSoundSource = audioCtx.createMediaStreamSource(stream);
const systemSoundDestination = audioCtx.createMediaStreamDestination();
systemSoundSource.connect(systemSoundDestination);
// 录制所有音频输入设备
audioInputDevices.forEach(async device => {
const micStream = await navigator.mediaDevices.getUserMedia({ audio: { deviceId: { exact: device.deviceId } } });
setMediaStream(micStream);
const micSoundSource = audioCtx.createMediaStreamSource(micStream);
micSoundSource.connect(systemSoundDestination);
})
// 合并音频流与视频流
const combinedSource = new MediaStream([...stream.getVideoTracks(), ...systemSoundDestination.stream.getAudioTracks()]);
// 开始录制
const mediaRecorder = new MediaRecorder(combinedSource, {
mimeType: 'video/webm;codecs=vp9,opus',
videoBitsPerSecond: 1.5e6,
});
setRecorder(mediaRecorder);
}); });
} catch (error: any) {
if (error.code === 'ENOENT') { // 获取所有音频输入设备
message.error({ const devices = await navigator.mediaDevices.enumerateDevices();
content: <div> <span style={{ color: '#606fc7', cursor: 'pointer' }} onClick={() => { const audioInputDevices = devices.filter(device => device.kind === 'audioinput' &&
stupWizardRef.current.changeModal(3); device.deviceId !== 'default' &&
}}></span></div> device.deviceId !== 'communications' );
});
return; // 使用Web Audio API来捕获系统声音和麦克风声音将它们合并到同一个MediaStream中。
} else { const audioCtx = new (window.AudioContext || window.webkitAudioContext)();
message.error(error); const systemSoundSource = audioCtx.createMediaStreamSource(stream);
} const systemSoundDestination = audioCtx.createMediaStreamDestination();
systemSoundSource.connect(systemSoundDestination);
// 录制所有音频输入设备
audioInputDevices.forEach( async device=>{
const micStream = await navigator.mediaDevices.getUserMedia({ audio: { deviceId: { exact: device.deviceId } }});
const micSoundSource = audioCtx.createMediaStreamSource(micStream);
micSoundSource.connect(systemSoundDestination);
})
// 合并音频流与视频流
const combinedSource = new MediaStream([...stream.getVideoTracks(), ...systemSoundDestination.stream.getAudioTracks()]);
// 开始录制
const recorder = new MediaRecorder(combinedSource, {
mimeType: 'video/webm;codecs=vp9,opus',
videoBitsPerSecond: 1.5e6,
});
setMediaStream(combinedSource);
setRecorder(recorder);
});
} catch (error: any) {
if (error.code === 'ENOENT') {
message.error({
content: <div> <span style={{ color: '#606fc7', cursor: 'pointer' }} onClick={() => {
stupWizardRef.current.changeModal(3);
}}></span></div>
});
return;
} else {
message.error(error);
} }
}
break; break;
case '录制中': case '录制中':
@ -1329,6 +1341,8 @@ const Meeting: React.FC = () => {
} }
}) })
} }
// 开关麦克风 // 开关麦克风
const postOpenMicrApi = async (enableMicr: boolean, uid: string, isAll: boolean, isMessage: boolean = false): Promise<void> => { const postOpenMicrApi = async (enableMicr: boolean, uid: string, isAll: boolean, isMessage: boolean = false): Promise<void> => {
if (isAll) { if (isAll) {