Extract Any Audio Pro for Mac - mp3格式转换器

ExtractAnyAudioProforMac是一款专业的音频转换软件,专注于将各种视频和音频文件中的音频导出并转换为.mp3格式。它支持包括.m4a、.wav、.wma、.mp4等多种音视频格式,操作简便,只需三步即可完成转换。

Extract Any Audio Pro for Mac是一款专业的mp3格式转换器,只需三步即可帮助用户快速从各种视频和音频文件中提取音频并转换成 .mp3格式,Extract Any Audio Pro mac版支持.m4a、.wav、.wma、.mp3、.flac、.mp4、.mov、.m4v、.mkv、.avi等各种音视频格式1

 Extract Any Audio Pro for Mac mac.orsoon.com/Mac/188249.html未来软件

import os import numpy as np import librosa import pywt import joblib # -------------------------- # 配置参数(需与训练时保持一致) # -------------------------- SAMPLE_RATE = 44100 # 与训练时相同的采样率 DURATION = 0.4 # 音频时长(秒),与训练时一致 WAVELET = "db4" # 小波函数,与训练时一致 DECOMPOSE_LEVEL = 3 # 小波分解层数,与训练时一致 CATEGORIES = ["normal", "crack"] # 分类标签,与训练时一致 # 模型和PCA转换器路径(根据实际保存路径修改) MODEL_PATH = "coin_classifier_model_optimized111.pkl" PCA_PATH = "pca_transformer_optimized111.pkl" # -------------------------- # 数据预处理函数(与训练时完全一致) # -------------------------- def load_audio(file_path): """加载音频并统一长度""" signal, sr = librosa.load(file_path, sr=SAMPLE_RATE, duration=DURATION) target_length = int(SAMPLE_RATE * DURATION) if len(signal) < target_length: signal = np.pad(signal, (0, target_length - len(signal)), mode='constant') else: signal = signal[:target_length] return signal, sr def trim_silence(signal): """动态阈值裁剪静音段""" energy = np.square(signal) threshold = np.percentile(energy, 50) * 0.1 # 与训练时相同的阈值策略 non_silent = np.where(energy > threshold)[0] if len(non_silent) == 0: return signal return signal[non_silent[0]:non_silent[-1]] def wavelet_denoise(signal): """小波去噪""" coeffs = pywt.wavedec(signal, wavelet=WAVELET, level=2) sigma = np.median(np.abs(coeffs[-1])) / 0.6745 threshold = sigma * np.sqrt(2 * np.log(len(signal))) * 0.5 # 与训练时相同的阈值 def modified_soft(coeff): mask = np.abs(coeff) > threshold coeff[mask] = coeff[mask] - np.sign(coeff[mask]) * threshold coeff[~mask] *= 0.5 return coeff coeffs[1:] = [modified_soft(c) for c in coeffs[1:]] denoised = pywt.waverec(coeffs, wavelet=WAVELET) return denoised[:len(signal)] def preprocess_signal(signal): """完整预处理流程""" trimmed = trim_silence(signal) denoised = wavelet_denoise(trimmed) target_length = int(SAMPLE_RATE * DURATION * 0.8) # 与训练时一致的长度 if len(denoised) < target_length: denoised = np.pad(denoised, (0, target_length - len(denoised)), mode='constant') else: denoised = denoised[:target_length] return denoised # -------------------------- # 特征提取函数(与训练时完全一致) # -------------------------- def extract_time_domain_features(signal): peak_amplitude = np.max(np.abs(signal)) rms = np.sqrt(np.mean(np.square(signal))) zero_crossing = librosa.feature.zero_crossing_rate(signal).mean() kurtosis = np.mean(np.power(signal / (rms + 1e-10), 4)) if rms != 0 else 0 return [peak_amplitude, rms, zero_crossing, kurtosis] def extract_transient_features(signal): onset_strength = librosa.onset.onset_strength(y=signal, sr=SAMPLE_RATE) onset_frames = librosa.onset.onset_detect(onset_envelope=onset_strength, sr=SAMPLE_RATE) onset_count = len(onset_frames) max_onset_strength = np.max(onset_strength) if len(onset_strength) > 0 else 0 mean_onset_strength = np.mean(onset_strength) if len(onset_strength) > 0 else 0 hop_length = 512 onset_samples = librosa.frames_to_samples(onset_frames, hop_length=hop_length) high_energy_at_onset = 0 for s in onset_samples: window = int(SAMPLE_RATE * 0.01) start = max(0, s - window) end = min(len(signal), s + window) segment = signal[start:end] stft = np.abs(librosa.stft(segment, hop_length=hop_length // 2)) freq = librosa.fft_frequencies(sr=SAMPLE_RATE, n_fft=stft.shape[0] * 2 - 1) high_idx = np.where(freq > 20000)[0] high_energy_at_onset += np.sum(stft[high_idx, :]) if len(high_idx) > 0 else 0 mean_high_energy = high_energy_at_onset / (onset_count + 1e-10) return [onset_count, max_onset_strength, mean_onset_strength, mean_high_energy] def extract_freq_domain_features(signal): mfcc = librosa.feature.mfcc(y=signal, sr=SAMPLE_RATE, n_mfcc=20) mfcc_stats = np.concatenate([np.mean(mfcc, axis=1), np.std(mfcc, axis=1)]) n_fft = 1024 hop_length = 256 stft = np.abs(librosa.stft(signal, n_fft=n_fft, hop_length=hop_length)) freq_bins = librosa.fft_frequencies(sr=SAMPLE_RATE, n_fft=n_fft) bands = [ (0, 5000), (5000, 10000), (10000, 15000), (15000, 20000), (20000, 25000), (25000, SAMPLE_RATE // 2) ] band_energies = [] total_energy = np.sum(stft) + 1e-10 for (low, high) in bands: idx = np.where((freq_bins >= low) & (freq_bins < high))[0] energy = np.sum(stft[idx, :]) if len(idx) > 0 else 0 band_energies.append(energy / total_energy) spectral_entropy = -np.sum((stft / total_energy) * np.log2(stft / total_energy + 1e-10)) high_idx = np.where(freq_bins > 15000)[0] high_energy = stft[high_idx, :].flatten() if len(high_idx) > 0 else np.array([]) high_fluctuation = np.std(high_energy) if len(high_energy) > 0 else 0 return np.concatenate([mfcc_stats, band_energies, [spectral_entropy, high_fluctuation]]) def extract_wpd_features(signal): wp = pywt.WaveletPacket(data=signal, wavelet=WAVELET, mode='symmetric', maxlevel=DECOMPOSE_LEVEL) nodes = [node.path for node in wp.get_level(DECOMPOSE_LEVEL, 'natural')] subband_features = [] for node in nodes: data = wp[node].data subband_features.append(np.sum(np.square(data))) subband_features.append(np.mean(np.abs(data))) subband_features.append(np.std(data)) total_energy = np.sum(subband_features[::3]) + 1e-10 subband_features = [f / total_energy if i % 3 == 0 else f for i, f in enumerate(subband_features)] return subband_features def extract_features(signal): """组合所有特征""" time_feats = extract_time_domain_features(signal) transient_feats = extract_transient_features(signal) freq_feats = extract_freq_domain_features(signal) wpd_feats = extract_wpd_features(signal) return np.array(time_feats + transient_feats + list(freq_feats) + wpd_feats) # -------------------------- # 模型加载与预测函数 # -------------------------- def load_model(): """加载训练好的模型和PCA转换器""" try: model = joblib.load(MODEL_PATH) pca = joblib.load(PCA_PATH) print("模型和PCA转换器加载成功") return model, pca except Exception as e: print(f"模型加载失败:{str(e)}") return None, None def predict_audio(model, pca, audio_path): """预测单个音频文件""" if not os.path.exists(audio_path): return "文件不存在", 0.0 try: # 加载并预处理音频 signal, _ = load_audio(audio_path) processed_signal = preprocess_signal(signal) # 提取特征 features = extract_features(processed_signal).reshape(1, -1) # PCA降维 features_pca = pca.transform(features) # 预测 pred_label = model.predict(features_pca)[0] pred_prob = model.predict_proba(features_pca)[0][pred_label] return CATEGORIES[pred_label], pred_prob except Exception as e: return f"预测出错:{str(e)}", 0.0 # -------------------------- # 批量测试函数 # -------------------------- def batch_predict(audio_folder): """批量预测文件夹中的所有音频文件""" model, pca = load_model() if model is None or pca is None: return print(f"\n开始批量预测文件夹:{audio_folder}") for file in os.listdir(audio_folder): if file.endswith(".wav"): file_path = os.path.join(audio_folder, file) result, prob = predict_audio(model, pca, file_path) print(f"{file} → {result}(置信度:{prob:.2f})") # -------------------------- # 主函数(可直接运行测试) # -------------------------- if __name__ == "__main__": # 加载模型 model, pca = load_model() if model and pca: # 示例1:预测单个音频文件(修改为你的测试文件路径) test_file = "F:\zhuanli_jianbao\jiequ\\normal\AI1-01_20250805202742.wav" # 替换为实际测试文件 if os.path.exists(test_file): result, prob = predict_audio(model, pca, test_file) print(f"\n单个文件预测结果:{os.path.basename(test_file)} → {result}(置信度:{prob:.2f})") # 示例2:批量预测文件夹(修改为你的测试文件夹路径) # test_folder = "F:\zhuanli_jianbao\jiequ\\normal" # 替换为实际测试文件夹 # batch_predict(test_folder)这是我的python代码,改成matlab2021可以运行的代码,实现所有功能,并给我讲解
10-05
检查代码错误并优化,以减少内存资源,但需要保证原始代码的一致性(包括输出、核心内容): class EnhancedAudioProcessor: SUPPORTED_FORMATS = ('.mp3', '.wav', '.amr', '.m4a') MAX_SEGMENT_DURATION = 5 * 60 * 1000 # 5分钟分段限制 # 新增配置参数 ENHANCEMENT_CONFIG = { 'enable_voice_extraction': True, 'enable_telephone_enhancement': True, 'noise_sample_duration': 0.5, # 噪声采样时长(秒) 'telephone_filter_range': (300, 3400), # 电话频段范围(Hz) 'compression_threshold': -25.0, # 压缩阈值(dBFS) 'compression_ratio': 3.0 # 压缩比 } @staticmethod def check_dependencies(): """检查音频处理所需的依赖""" # 新增检查noisereduce依赖 try: import noisereduce as nr return True, "依赖检查通过" except ImportError: return False, "缺少noisereduce库,请执行: pip install noisereduce" @staticmethod def convert_to_wav(input_path: str, temp_dir: str) -> Optional[List[str]]: # 先检查ffmpeg是否可用 ffmpeg_available, ffmpeg_msg = check_ffmpeg_available() if not ffmpeg_available: logger.error(f"ffmpeg错误: {ffmpeg_msg}") return None try: os.makedirs(temp_dir, exist_ok=True) ext = os.path.splitext(input_path)[1].lower() if ext not in EnhancedAudioProcessor.SUPPORTED_FORMATS: raise ValueError( f"不支持的音频格式: {ext},支持的格式为: {', '.join(EnhancedAudioProcessor.SUPPORTED_FORMATS)}") # 加载原始音频文件 try: audio = AudioSegment.from_file(input_path) except Exception as e: raise RuntimeError(f"无法加载音频文件: {str(e)}") # ============== 新增: 音频增强处理流程 ============== if EnhancedAudioProcessor._should_enhance_audio(input_path): logger.info(f"开始增强处理音频: {os.path.basename(input_path)}") # 步骤1: 主要说话人提取 audio = EnhancedAudioProcessor._extract_main_voice(audio) # 步骤2: 电话质量增强 if EnhancedAudioProcessor.ENHANCEMENT_CONFIG['enable_telephone_enhancement']: audio = EnhancedAudioProcessor._enhance_telephone_quality(audio) # ================================================ # 检查时长并决定是否需要分段 max_duration = ConfigManager().get("max_audio_duration", 3600) * 1000 if len(audio) > max_duration: return EnhancedAudioProcessor._split_long_audio(audio, input_path, temp_dir) return EnhancedAudioProcessor._convert_single_audio(audio, input_path, temp_dir) except Exception as e: logger.error(f"音频处理失败: {str(e)}", exc_info=True) return None # ============== 新增音频增强方法 ============== @staticmethod def _should_enhance_audio(file_path: str) -> bool: """判断是否需要应用增强处理""" # 基于配置和文件扩展名判断 config = ConfigManager().get("audio_enhancement", {}) enable = config.get("enable", True) # 特殊处理: 电话录音文件通常需要增强 filename = os.path.basename(file_path).lower() if "phone" in filename or "tel" in filename: return True return enable and EnhancedAudioProcessor.ENHANCEMENT_CONFIG['enable_voice_extraction'] @staticmethod def _extract_main_voice(audio: AudioSegment) -> AudioSegment: """核心方法: 提取主要说话人声音""" try: # 获取采样率和音频数据 samples = np.array(audio.get_array_of_samples()) sr = audio.frame_rate # 噪声采样 (取前0.5秒作为噪声参考) noise_duration = int( sr * EnhancedAudioProcessor.ENHANCEMENT_CONFIG['noise_sample_duration'] ) noise_sample = samples[:min(noise_duration, len(samples))] # 应用降噪处理 import noisereduce as nr reduced_noise = nr.reduce_noise( y=samples.astype(np.float32), sr=sr, y_noise=noise_sample.astype(np.float32), prop_decrease=0.8, stationary=True, n_std_thresh_stationary=1.5 ) # 转换为16位PCM格式 processed = reduced_noise.astype(np.int16) # 创建新的AudioSegment对象 return AudioSegment( processed.tobytes(), frame_rate=sr, sample_width=audio.sample_width, channels=1 ) except Exception as e: logger.warning(f"说话人提取失败, 使用原始音频: {str(e)}") return audio @staticmethod def _enhance_telephone_quality(audio: AudioSegment) -> AudioSegment: """核心方法: 增强电话录音质量""" try: # 获取滤波器范围 low_cut, high_cut = EnhancedAudioProcessor.ENHANCEMENT_CONFIG['telephone_filter_range'] # 应用电话频段滤波 audio = audio.high_pass_filter(low_cut).low_pass_filter(high_cut) # 动态范围压缩 compression_params = { 'threshold': EnhancedAudioProcessor.ENHANCEMENT_CONFIG['compression_threshold'], 'ratio': EnhancedAudioProcessor.ENHANCEMENT_CONFIG['compression_ratio'] } audio = audio.compress_dynamic_range(**compression_params) # 音量标准化 audio = effects.normalize(audio) return audio except Exception as e: logger.warning(f"电话质量增强失败, 使用原始音频: {str(e)}") return audio # ============== 保留原始分段处理方法 ============== @staticmethod def _split_long_audio(audio: AudioSegment, input_path: str, temp_dir: str) -> List[str]: chunks = split_on_silence( audio, min_silence_len=ConfigManager().get("min_silence_len", 1000), silence_thresh=ConfigManager().get("silence_thresh", -40), keep_silence=500 ) merged_chunks = [] current_chunk = AudioSegment.empty() for chunk in chunks: if len(current_chunk) + len(chunk) < EnhancedAudioProcessor.MAX_SEGMENT_DURATION: current_chunk += chunk else: if len(current_chunk) > 0: merged_chunks.append(current_chunk) current_chunk = chunk if len(current_chunk) > 0: merged_chunks.append(current_chunk) # 确保每个分段不超过最大时长 final_chunks = [] for chunk in merged_chunks: if len(chunk) <= EnhancedAudioProcessor.MAX_SEGMENT_DURATION: final_chunks.append(chunk) else: # 强制分段 subchunks = make_chunks(chunk, EnhancedAudioProcessor.MAX_SEGMENT_DURATION) final_chunks.extend(subchunks) wav_paths = [] sample_rate = ConfigManager().get("sample_rate", 16000) for i, chunk in enumerate(final_chunks): chunk = chunk.set_frame_rate(sample_rate).set_channels(1) chunk_path = os.path.join(temp_dir, f"{os.path.splitext(os.path.basename(input_path))[0]}_part{i + 1}.wav") chunk.export(chunk_path, format="wav") wav_paths.append(chunk_path) return wav_paths @staticmethod def _convert_single_audio(audio: AudioSegment, input_path: str, temp_dir: str) -> List[str]: sample_rate = ConfigManager().get("sample_rate", 16000) audio = audio.set_frame_rate(sample_rate).set_channels(1) wav_path = os.path.join(temp_dir, os.path.splitext(os.path.basename(input_path))[0] + ".wav") audio.export(wav_path, format="wav") return [wav_path] def extract_features_from_audio(y: np.ndarray, sr: int) -> Dict[str, float]: try: duration = librosa.get_duration(y=y, sr=sr) segment_length = 60 total_segments = max(1, int(np.ceil(duration / segment_length))) syllable_rates, volume_stabilities = [], [] total_samples = len(y) samples_per_segment = int(segment_length * sr) for i in range(total_segments): start = i * samples_per_segment end = min((i + 1) * samples_per_segment, total_samples) y_segment = y[start:end] if len(y_segment) == 0: continue intervals = librosa.effects.split(y_segment, top_db=20) speech_samples = sum(end - start for start, end in intervals) speech_duration = speech_samples / sr syllable_rates.append(len(intervals) / speech_duration if speech_duration > 0.1 else 0) rms = librosa.feature.rms(y=y_segment, frame_length=2048, hop_length=512)[0] if len(rms) > 0 and np.mean(rms) > 0: volume_stabilities.append(np.std(rms) / np.mean(rms)) return { "duration": duration, "syllable_rate": round(np.mean([r for r in syllable_rates if r > 0]) if syllable_rates else 0, 2), "volume_stability": round(np.mean(volume_stabilities) if volume_stabilities else 0, 4) } except Exception as e: logger.error(f"特征提取错误: {str(e)}") return {"duration": 0, "syllable_rate": 0, "volume_stability": 0}
09-04
检查代码是否合理是否错误,并评价代码,计算运行峰值、准确率、效率速度,已两人对话十分钟为例。 import os import sys import re import json import gc import time import concurrent.futures import traceback import numpy as np import librosa import torch import psutil from typing import List, Dict, Tuple, Optional from threading import RLock, Semaphore from pydub import AudioSegment from pydub.silence import split_on_silence from modelscope.pipelines import pipeline from modelscope.utils.constant import Tasks from transformers import AutoModelForSequenceClassification, AutoTokenizer from torch.utils.data import TensorDataset, DataLoader from PyQt5.QtWidgets import (QApplication, QMainWindow, QWidget, QVBoxLayout, QHBoxLayout, QPushButton, QLabel, QLineEdit, QTextEdit, QFileDialog, QProgressBar, QGroupBox, QMessageBox, QListWidget, QSplitter, QTabWidget, QTableWidget, QTableWidgetItem, QHeaderView, QAction, QMenu, QToolBar, QComboBox, QSpinBox, QDialog, QDialogButtonBox) from PyQt5.QtCore import QThread, pyqtSignal, Qt from PyQt5.QtGui import QFont, QColor, QIcon # ====================== 资源监控器 ====================== class ResourceMonitor: def __init__(self): self.gpu_available = torch.cuda.is_available() def memory_percent(self) -> Dict[str, float]: try: result = {"cpu": psutil.virtual_memory().percent} if self.gpu_available: allocated = torch.cuda.memory_allocated() / (1024 ** 3) total = torch.cuda.get_device_properties(0).total_memory / (1024 ** 3) result["gpu"] = (allocated / total) * 100 if total > 0 else 0 return result except Exception as e: print(f"内存监控失败: {str(e)}") return {"cpu": 0, "gpu": 0} # ====================== 方言处理器(简化版) ====================== class DialectProcessor: # 合并贵州方言和普通话关键词 KEYWORDS = { "opening": ["您好", "很高兴为您服务", "请问有什么可以帮您", "麻烦您喽", "请问搞哪样", "有咋个可以帮您", "多谢喽"], "closing": ["感谢来电", "祝您生活愉快", "再见", "搞归一喽", "麻烦您喽", "再见喽", "慢走喽"], "forbidden": ["不知道", "没办法", "你投诉吧", "随便你", "搞不成", "没得法", "随便你喽", "你投诉吧喽"], "salutation": ["先生", "女士", "小姐", "老师", "师傅", "哥", "姐", "兄弟", "妹儿"], "reassurance": ["非常抱歉", "请不要着急", "我们会尽快处理", "理解您的心情", "实在对不住", "莫急哈", "马上帮您整", "理解您得很"] } # 贵州方言到普通话的固定映射 DIALECT_MAPPING = { "恼火得很": "非常生气", "鬼火戳": "很愤怒", "搞不成": "无法完成", "没得": "没有", "搞哪样嘛": "做什么呢", "归一喽": "完成了", "咋个": "怎么", "克哪点": "去哪里", "麻烦您喽": "麻烦您了", "多谢喽": "多谢了", "憨包": "傻瓜", "归一": "结束", "板扎": "很好", "鬼火冒": "非常生气", "背时": "倒霉", "吃豁皮": "占便宜" } # Trie树根节点 _trie_root = None class TrieNode: def __init__(self): self.children = {} self.is_end = False self.value = "" @classmethod def build_dialect_trie(cls): """构建方言转换的Trie树""" if cls._trie_root is not None: return cls._trie_root root = cls.TrieNode() # 按长度降序排序,确保最长匹配优先 for dialect, standard in sorted(cls.DIALECT_MAPPING.items(), key=lambda x: len(x[0]), reverse=True): node = root for char in dialect: if char not in node.children: node.children[char] = cls.TrieNode() node = node.children[char] node.is_end = True node.value = standard cls._trie_root = root return root @classmethod def preprocess_text(cls, texts: List[str]) -> List[str]: """使用Trie树进行方言转换""" if cls._trie_root is None: cls.build_dialect_trie() processed_texts = [] for text in texts: processed = [] i = 0 n = len(text) while i < n: node = cls._trie_root j = i found = False # 在Trie树中查找最长匹配 while j < n and text[j] in node.children: node = node.children[text[j]] j += 1 if node.is_end: # 找到完整匹配 processed.append(node.value) i = j found = True break if not found: # 无匹配 processed.append(text[i]) i += 1 processed_texts.append(''.join(processed)) return processed_texts # ====================== 系统配置管理器 ====================== class ConfigManager: _instance = None def __new__(cls): if cls._instance is None: cls._instance = super().__new__(cls) cls._instance._init_config() return cls._instance def _init_config(self): self.config = { "model_paths": { "asr": "./models/iic-speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn", "sentiment": "./models/IDEA-CCNL-Erlangshen-Roberta-110M-Sentiment" }, "sample_rate": 16000, "silence_thresh": -40, "min_silence_len": 1000, "max_concurrent": 1, "max_audio_duration": 3600 # 移除了方言配置 } self.load_config() def load_config(self): try: if os.path.exists("config.json"): with open("config.json", "r") as f: self.config.update(json.load(f)) except: pass def save_config(self): try: with open("config.json", "w") as f: json.dump(self.config, f, indent=2) except: pass def get(self, key: str, default=None): return self.config.get(key, default) def set(self, key: str, value): self.config[key] = value self.save_config() # ====================== 音频处理工具 ====================== class AudioProcessor: SUPPORTED_FORMATS = ('.mp3', '.wav', '.amr', '.m4a') @staticmethod def convert_to_wav(input_path: str, temp_dir: str) -> Optional[List[str]]: try: os.makedirs(temp_dir, exist_ok=True) if not any(input_path.lower().endswith(ext) for ext in AudioProcessor.SUPPORTED_FORMATS): raise ValueError(f"不支持的音频格式: {os.path.splitext(input_path)[1]}") if input_path.lower().endswith('.wav'): return [input_path] audio = AudioSegment.from_file(input_path) max_duration = ConfigManager().get("max_audio_duration", 3600) * 1000 if len(audio) > max_duration: return AudioProcessor._split_long_audio(audio, input_path, temp_dir) return AudioProcessor._convert_single_audio(audio, input_path, temp_dir) except Exception as e: print(f"格式转换失败: {str(e)}") return None @staticmethod def _split_long_audio(audio: AudioSegment, input_path: str, temp_dir: str) -> List[str]: chunks = split_on_silence( audio, min_silence_len=ConfigManager().get("min_silence_len", 1000), silence_thresh=ConfigManager().get("silence_thresh", -40), keep_silence=500 ) merged_chunks = [] current_chunk = AudioSegment.empty() for chunk in chunks: if len(current_chunk) + len(chunk) < 5 * 60 * 1000: current_chunk += chunk else: if len(current_chunk) > 0: merged_chunks.append(current_chunk) current_chunk = chunk if len(current_chunk) > 0: merged_chunks.append(current_chunk) wav_paths = [] sample_rate = ConfigManager().get("sample_rate", 16000) for i, chunk in enumerate(merged_chunks): chunk = chunk.set_frame_rate(sample_rate).set_channels(1) chunk_path = os.path.join(temp_dir, f"{os.path.splitext(os.path.basename(input_path))[0]}_part{i + 1}.wav") chunk.export(chunk_path, format="wav") wav_paths.append(chunk_path) return wav_paths @staticmethod def _convert_single_audio(audio: AudioSegment, input_path: str, temp_dir: str) -> List[str]: sample_rate = ConfigManager().get("sample_rate", 16000) audio = audio.set_frame_rate(sample_rate).set_channels(1) wav_path = os.path.join(temp_dir, os.path.splitext(os.path.basename(input_path))[0] + ".wav") audio.export(wav_path, format="wav") return [wav_path] @staticmethod def extract_features_from_audio(y: np.ndarray, sr: int) -> Dict[str, float]: try: duration = librosa.get_duration(y=y, sr=sr) segment_length = 60 total_segments = max(1, int(np.ceil(duration / segment_length))) syllable_rates, volume_stabilities = [], [] total_samples = len(y) samples_per_segment = int(segment_length * sr) for i in range(total_segments): start = i * samples_per_segment end = min((i + 1) * samples_per_segment, total_samples) y_segment = y[start:end] if len(y_segment) == 0: continue intervals = librosa.effects.split(y_segment, top_db=20) speech_samples = sum(end - start for start, end in intervals) speech_duration = speech_samples / sr syllable_rates.append(len(intervals) / speech_duration if speech_duration > 0.1 else 0) rms = librosa.feature.rms(y=y_segment, frame_length=2048, hop_length=512)[0] if len(rms) > 0 and np.mean(rms) > 0: volume_stabilities.append(np.std(rms) / np.mean(rms)) return { "duration": duration, "syllable_rate": round(np.mean([r for r in syllable_rates if r > 0]) if syllable_rates else 0, 2), "volume_stability": round(np.mean(volume_stabilities) if volume_stabilities else 0, 4) } except Exception as e: print(f"特征提取错误: {str(e)}") return {"duration": 0, "syllable_rate": 0, "volume_stability": 0} # ====================== 模型加载器 ====================== class ModelLoader: asr_pipeline = None sentiment_model = None sentiment_tokenizer = None model_lock = RLock() models_loaded = False @classmethod def load_models(cls): config = ConfigManager() if not cls.asr_pipeline: with cls.model_lock: if not cls.asr_pipeline: cls._load_asr_model(config.get("model_paths")["asr"]) if not cls.sentiment_model: with cls.model_lock: if not cls.sentiment_model: cls._load_sentiment_model(config.get("model_paths")["sentiment"]) cls.models_loaded = True @classmethod def reload_models(cls): with cls.model_lock: cls.asr_pipeline = None cls.sentiment_model = None cls.sentiment_tokenizer = None gc.collect() if torch.cuda.is_available(): torch.cuda.empty_cache() cls.load_models() @classmethod def _load_asr_model(cls, model_path: str): try: if not os.path.exists(model_path): raise FileNotFoundError(f"ASR模型路径不存在: {model_path}") asr_kwargs = {'quantize': 'int8'} if hasattr(torch, 'quantization') else {} cls.asr_pipeline = pipeline( task=Tasks.auto_speech_recognition, model=model_path, device='cuda' if torch.cuda.is_available() else 'cpu', **asr_kwargs ) except Exception as e: print(f"加载ASR模型失败: {str(e)}") raise @classmethod def _load_sentiment_model(cls, model_path: str): try: if not os.path.exists(model_path): raise FileNotFoundError(f"情感分析模型路径不存在: {model_path}") cls.sentiment_model = AutoModelForSequenceClassification.from_pretrained(model_path) cls.sentiment_tokenizer = AutoTokenizer.from_pretrained(model_path) if torch.cuda.is_available(): cls.sentiment_model = cls.sentiment_model.cuda() except Exception as e: print(f"加载情感分析模型失败: {str(e)}") raise # ====================== 核心分析线程(简化版) ====================== class AnalysisThread(QThread): progress_updated = pyqtSignal(int, str, str) result_ready = pyqtSignal(dict) finished_all = pyqtSignal() error_occurred = pyqtSignal(str, str) memory_warning = pyqtSignal() resource_cleanup = pyqtSignal() def __init__(self, audio_paths: List[str], temp_dir: str = "temp_wav"): super().__init__() self.audio_paths = audio_paths self.temp_dir = temp_dir self.is_running = True self.current_file = "" self.max_concurrent = min(ConfigManager().get("max_concurrent", 1), self._get_max_concurrent_tasks()) self.resource_monitor = ResourceMonitor() self.semaphore = Semaphore(self.max_concurrent) os.makedirs(temp_dir, exist_ok=True) def run(self): try: if not ModelLoader.models_loaded: self.error_occurred.emit("模型未加载", "请等待模型加载完成后再开始分析") return self.progress_updated.emit(0, f"最大并行任务数: {self.max_concurrent}", "") with concurrent.futures.ThreadPoolExecutor(max_workers=self.max_concurrent) as executor: future_to_path = {} for path in self.audio_paths: if not self.is_running: break self.semaphore.acquire() future = executor.submit(self.analyze_audio, path, self._get_available_batch_size()) future_to_path[future] = path future.add_done_callback(lambda f: self.semaphore.release()) for i, future in enumerate(concurrent.futures.as_completed(future_to_path)): if not self.is_running: break path = future_to_path[future] self.current_file = os.path.basename(path) if self._check_memory_usage(): self.memory_warning.emit() self.is_running = False break try: result = future.result() if result: self.result_ready.emit(result) progress = int((i + 1) / len(self.audio_paths) * 100) self.progress_updated.emit(progress, f"完成: {self.current_file} ({i + 1}/{len(self.audio_paths)})", self.current_file) except Exception as e: result = {"file_name": self.current_file, "status": "error", "error": f"分析失败: {str(e)}"} self.result_ready.emit(result) if self.is_running: self.finished_all.emit() except Exception as e: self.error_occurred.emit("系统错误", str(e)) traceback.print_exc() finally: self.resource_cleanup.emit() self._cleanup_resources() def analyze_audio(self, audio_path: str, batch_size: int) -> Dict: result = {"file_name": os.path.basename(audio_path), "status": "processing"} wav_paths = [] try: wav_paths = AudioProcessor.convert_to_wav(audio_path, self.temp_dir) if not wav_paths: result["error"] = "格式转换失败" result["status"] = "error" return result audio_features = self._extract_audio_features(wav_paths) result.update(audio_features) result["duration_str"] = self._format_duration(audio_features["duration"]) all_segments, full_text = self._process_asr_segments(wav_paths) agent_segments, customer_segments = self._identify_speakers(all_segments) result["asr_text"] = self._generate_labeled_text(all_segments, agent_segments, customer_segments).strip() text_analysis = self._analyze_text(agent_segments, customer_segments, batch_size) result.update(text_analysis) service_check = self._check_service_rules(agent_segments) result.update(service_check) result["issue_resolved"] = self._check_issue_resolution(customer_segments, agent_segments) result["status"] = "success" except Exception as e: result["error"] = f"分析失败: {str(e)}" result["status"] = "error" finally: self._cleanup_temp_files(wav_paths) self._cleanup_resources() return result def _identify_speakers(self, segments: List[Dict]) -> Tuple[List[Dict], List[Dict]]: """使用四层逻辑识别客服""" if not segments: return [], [] # 逻辑1:前三片段开场白关键词 agent_id = self._identify_by_opening(segments) # 逻辑2:后三片段结束语关键词 if agent_id is None: agent_id = self._identify_by_closing(segments) # 逻辑3:称呼与敬语关键词 if agent_id is None: agent_id = self._identify_by_salutation(segments) # 逻辑4:安抚语关键词 if agent_id is None: agent_id = self._identify_by_reassurance(segments) # 后备策略:说话模式识别 if agent_id is None and len(segments) >= 4: agent_id = self._identify_by_speech_patterns(segments) if agent_id is None: # 最后手段:选择说话最多的说话人 spk_counts = {} for seg in segments: spk_id = seg["spk_id"] spk_counts[spk_id] = spk_counts.get(spk_id, 0) + 1 agent_id = max(spk_counts, key=spk_counts.get) if spk_counts else None if agent_id is None: return [], [] return ( [seg for seg in segments if seg["spk_id"] == agent_id], [seg for seg in segments if seg["spk_id"] != agent_id] ) def _identify_by_opening(self, segments: List[Dict]) -> Optional[str]: """逻辑1:前三片段开场白关键词""" keywords = DialectProcessor.KEYWORDS["opening"] for seg in segments[:3]: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_closing(self, segments: List[Dict]) -> Optional[str]: """逻辑2:后三片段结束语关键词""" keywords = DialectProcessor.KEYWORDS["closing"] last_segments = segments[-3:] if len(segments) >= 3 else segments for seg in reversed(last_segments): if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_salutation(self, segments: List[Dict]) -> Optional[str]: """逻辑3:称呼与敬语关键词""" keywords = DialectProcessor.KEYWORDS["salutation"] for seg in segments: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_reassurance(self, segments: List[Dict]) -> Optional[str]: """逻辑4:安抚语关键词""" keywords = DialectProcessor.KEYWORDS["reassurance"] for seg in segments: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_speech_patterns(self, segments: List[Dict]) -> Optional[str]: """后备策略:说话模式识别""" speaker_features = {} for seg in segments: spk_id = seg["spk_id"] if spk_id not in speaker_features: speaker_features[spk_id] = {"total_duration": 0.0, "turn_count": 0, "question_count": 0} features = speaker_features[spk_id] features["total_duration"] += (seg["end"] - seg["start"]) features["turn_count"] += 1 if any(q_word in seg["text"] for q_word in ["吗", "呢", "?", "?", "如何", "怎样"]): features["question_count"] += 1 if speaker_features: max_duration = max(f["total_duration"] for f in speaker_features.values()) question_rates = {spk_id: f["question_count"] / f["turn_count"] for spk_id, f in speaker_features.items()} candidates = [] for spk_id, features in speaker_features.items(): score = (0.6 * (features["total_duration"] / max_duration) + 0.4 * question_rates[spk_id]) candidates.append((spk_id, score)) return max(candidates, key=lambda x: x[1])[0] return None def _analyze_text(self, agent_segments: List[Dict], customer_segments: List[Dict], batch_size: int) -> Dict: """优化情感分析方法""" def split_long_sentences(texts: List[str]) -> List[str]: splitted = [] for text in texts: if len(text) > 128: parts = re.split(r'(?<=[。!?;,])', text) current = "" for part in parts: if len(current) + len(part) < 128: current += part else: if current: splitted.append(current) current = part if current: splitted.append(current) else: splitted.append(text) return splitted def enhance_with_keywords(texts: List[str]) -> List[str]: enhanced = [] emotion_keywords = { "positive": ["满意", "高兴", "感谢", "专业", "解决", "帮助", "谢谢", "很好", "不错"], "negative": ["生气", "愤怒", "不满", "投诉", "问题", "失望", "差劲", "糟糕", "投诉"], "neutral": ["了解", "明白", "知道", "确认", "查询", "记录", "需要", "提供"] } for text in texts: found_emotion = None for emotion, keywords in emotion_keywords.items(): if any(kw in text for kw in keywords): found_emotion = emotion break if found_emotion: enhanced.append(f"[{found_emotion}] {text}") else: enhanced.append(text) return enhanced # 分析单个说话者 def analyze_speaker(segments: List[Dict], speaker_type: str) -> Dict: if not segments: return { f"{speaker_type}_negative": 0.0, f"{speaker_type}_neutral": 1.0, f"{speaker_type}_positive": 0.0, f"{speaker_type}_emotions": "无" } texts = [seg["text"] for seg in segments] processed_texts = DialectProcessor.preprocess_text(texts) splitted_texts = split_long_sentences(processed_texts) enhanced_texts = enhance_with_keywords(splitted_texts) with ModelLoader.model_lock: inputs = ModelLoader.sentiment_tokenizer( enhanced_texts, padding=True, truncation=True, max_length=128, return_tensors="pt" ) dataset = TensorDataset(inputs['input_ids'], inputs['attention_mask']) dataloader = DataLoader(dataset, batch_size=batch_size, shuffle=False) device = "cuda" if torch.cuda.is_available() else "cpu" sentiment_dist = [] emotions = [] for batch in dataloader: input_ids, attention_mask = batch inputs = {'input_ids': input_ids.to(device), 'attention_mask': attention_mask.to(device)} with torch.no_grad(): outputs = ModelLoader.sentiment_model(**inputs) batch_probs = torch.nn.functional.softmax(outputs.logits, dim=-1) sentiment_dist.append(batch_probs.cpu()) emotion_keywords = ["愤怒", "生气", "鬼火", "不耐烦", "搞哪样嘛", "恼火", "背时", "失望", "不满"] for text in enhanced_texts: if any(kw in text for kw in emotion_keywords): if any(kw in text for kw in ["愤怒", "生气", "鬼火", "恼火"]): emotions.append("愤怒") elif any(kw in text for kw in ["不耐烦", "搞哪样嘛"]): emotions.append("不耐烦") elif "背时" in text: emotions.append("沮丧") elif any(kw in text for kw in ["失望", "不满"]): emotions.append("失望") if sentiment_dist: all_probs = torch.cat(sentiment_dist, dim=0) avg_sentiment = torch.mean(all_probs, dim=0).tolist() else: avg_sentiment = [0.0, 1.0, 0.0] return { f"{speaker_type}_negative": round(avg_sentiment[0], 4), f"{speaker_type}_neutral": round(avg_sentiment[1], 4), f"{speaker_type}_positive": round(avg_sentiment[2], 4), f"{speaker_type}_emotions": ",".join(set(emotions)) if emotions else "无" } return { **analyze_speaker(agent_segments, "agent"), **analyze_speaker(customer_segments, "customer") } def _check_service_rules(self, agent_segments: List[Dict]) -> Dict: keywords = DialectProcessor.KEYWORDS found_forbidden = [] found_opening = any(kw in seg["text"] for seg in agent_segments[:3] for kw in keywords["opening"]) found_closing = any( kw in seg["text"] for seg in (agent_segments[-3:] if len(agent_segments) >= 3 else agent_segments) for kw in keywords["closing"]) for seg in agent_segments: for kw in keywords["forbidden"]: if kw in seg["text"]: found_forbidden.append(kw) break return { "opening_found": found_opening, "closing_found": found_closing, "forbidden_words": ", ".join(set(found_forbidden)) if found_forbidden else "无" } def _check_issue_resolution(self, customer_segments: List[Dict], agent_segments: List[Dict]) -> bool: if not customer_segments or not agent_segments: return False resolution_keywords = ["解决", "处理", "完成", "已", "好了", "可以了", "没问题", "明白", "清楚", "满意", "行"] unresolved_keywords = ["没解决", "不行", "不对", "还是", "仍然", "再", "未", "无法", "不能", "不行", "不满意"] negation_words = ["不", "没", "未", "非", "无"] gratitude_keywords = ["谢谢", "感谢", "多谢", "麻烦", "辛苦", "有劳"] full_conversation = " ".join(seg["text"] for seg in customer_segments + agent_segments) last_customer_text = customer_segments[-1]["text"] for kw in unresolved_keywords: if kw in full_conversation: negation_context = re.search(rf".{{0,5}}{kw}", full_conversation) if negation_context: context = negation_context.group(0) if not any(neg in context for neg in negation_words): return False else: return False if any(kw in last_customer_text for kw in gratitude_keywords): if not any(neg + kw in last_customer_text for neg in negation_words): return True for agent_text in [seg["text"] for seg in agent_segments[-3:]]: if any(kw in agent_text for kw in resolution_keywords): if not any(neg in agent_text for neg in negation_words): return True for cust_seg in customer_segments[-2:]: if any(kw in cust_seg["text"] for kw in ["好", "行", "可以", "明白"]): if not any(neg in cust_seg["text"] for neg in negation_words): return True if any("?" in seg["text"] or "?" in seg["text"] for seg in customer_segments[-2:]): return False return False # ====================== 辅助方法 ====================== def _get_available_batch_size(self) -> int: if not torch.cuda.is_available(): return 4 total_mem = torch.cuda.get_device_properties(0).total_memory / (1024 ** 3) per_task_mem = total_mem / self.max_concurrent return 2 if per_task_mem < 2 else 4 if per_task_mem < 4 else 8 def _get_max_concurrent_tasks(self) -> int: if torch.cuda.is_available(): total_mem = torch.cuda.get_device_properties(0).total_memory / (1024 ** 3) return 1 if total_mem < 6 else 2 if total_mem < 12 else 3 return max(1, os.cpu_count() // 2) def _check_memory_usage(self) -> bool: try: mem_percent = self.resource_monitor.memory_percent() return mem_percent.get("cpu", 0) > 85 or mem_percent.get("gpu", 0) > 85 except: return False def _extract_audio_features(self, wav_paths: List[str]) -> Dict[str, float]: combined_y = np.array([], dtype=np.float32) sr = ConfigManager().get("sample_rate", 16000) for path in wav_paths: y, _ = librosa.load(path, sr=sr) combined_y = np.concatenate((combined_y, y)) return AudioProcessor.extract_features_from_audio(combined_y, sr) def _process_asr_segments(self, wav_paths: List[str]) -> Tuple[List[Dict], str]: segments = [] full_text = "" batch_size = min(4, len(wav_paths), self._get_available_batch_size()) for i in range(0, len(wav_paths), batch_size): if not self.is_running: break batch_paths = wav_paths[i:i + batch_size] try: results = ModelLoader.asr_pipeline(batch_paths, output_dir=None, batch_size=batch_size) for result in results: for seg in result[0]["sentences"]: segments.append({ "start": seg["start"], "end": seg["end"], "text": seg["text"], "spk_id": seg.get("spk_id", "0") }) full_text += seg["text"] + " " except Exception as e: print(f"ASR批处理错误: {str(e)}") for path in batch_paths: try: result = ModelLoader.asr_pipeline(path, output_dir=None) for seg in result[0]["sentences"]: segments.append({ "start": seg["start"], "end": seg["end"], "text": seg["text"], "spk_id": seg.get("spk_id", "0") }) full_text += seg["text"] + " " except: continue return segments, full_text.strip() def _generate_labeled_text(self, all_segments: List[Dict], agent_segments: List[Dict], customer_segments: List[Dict]) -> str: agent_spk_id = agent_segments[0]["spk_id"] if agent_segments else None customer_spk_id = customer_segments[0]["spk_id"] if customer_segments else None labeled_text = [] for seg in all_segments: if seg["spk_id"] == agent_spk_id: speaker = "客服" elif seg["spk_id"] == customer_spk_id: speaker = "客户" else: speaker = f"说话人{seg['spk_id']}" labeled_text.append(f"[{speaker}]: {seg['text']}") return "\n".join(labeled_text) def _cleanup_temp_files(self, paths: List[str]): def safe_remove(path): if os.path.exists(path): try: os.remove(path) except: pass for path in paths: safe_remove(path) now = time.time() for file in os.listdir(self.temp_dir): file_path = os.path.join(self.temp_dir, file) if os.path.isfile(file_path) and (now - os.path.getmtime(file_path)) > 3600: safe_remove(file_path) def _format_duration(self, seconds: float) -> str: minutes, seconds = divmod(int(seconds), 60) hours, minutes = divmod(minutes, 60) return f"{hours:02d}:{minutes:02d}:{seconds:02d}" def _cleanup_resources(self): gc.collect() if torch.cuda.is_available(): torch.cuda.empty_cache() def stop(self): self.is_running = False # ====================== 模型加载线程 ====================== class ModelLoadThread(QThread): progress_updated = pyqtSignal(int, str) finished = pyqtSignal(bool, str) def run(self): try: config = ConfigManager().get("model_paths") if not os.path.exists(config["asr"]): self.finished.emit(False, "ASR模型路径不存在") return if not os.path.exists(config["sentiment"]): self.finished.emit(False, "情感分析模型路径不存在") return self.progress_updated.emit(20, "加载语音识别模型...") ModelLoader._load_asr_model(config["asr"]) self.progress_updated.emit(60, "加载情感分析模型...") ModelLoader._load_sentiment_model(config["sentiment"]) self.progress_updated.emit(100, "模型加载完成") self.finished.emit(True, "模型加载成功") except Exception as e: self.finished.emit(False, f"模型加载失败: {str(e)}") # ====================== GUI主界面(简化版) ====================== class MainWindow(QMainWindow): def __init__(self): super().__init__() self.setWindowTitle("贵州方言客服质检系统") self.setGeometry(100, 100, 1200, 800) self.setup_ui() self.setup_menu() self.analysis_thread = None self.model_load_thread = None self.temp_dir = "temp_wav" os.makedirs(self.temp_dir, exist_ok=True) self.model_loaded = False def setup_ui(self): main_widget = QWidget() main_layout = QVBoxLayout() main_widget.setLayout(main_layout) self.setCentralWidget(main_widget) toolbar = QToolBar("主工具栏") self.addToolBar(toolbar) actions = [ ("添加文件", "icons/add.png", self.add_files), ("开始分析", "icons/start.png", self.start_analysis), ("停止分析", "icons/stop.png", self.stop_analysis), ("设置", "icons/settings.png", self.open_settings) ] for name, icon, func in actions: action = QAction(QIcon(icon), name, self) action.triggered.connect(func) toolbar.addAction(action) splitter = QSplitter(Qt.Horizontal) main_layout.addWidget(splitter) left_widget = QWidget() left_layout = QVBoxLayout() left_widget.setLayout(left_layout) left_layout.addWidget(QLabel("待分析文件列表")) self.file_list = QListWidget() self.file_list.setSelectionMode(QListWidget.ExtendedSelection) left_layout.addWidget(self.file_list) right_widget = QWidget() right_layout = QVBoxLayout() right_widget.setLayout(right_layout) right_layout.addWidget(QLabel("分析进度")) self.progress_bar = QProgressBar() self.progress_bar.setRange(0, 100) right_layout.addWidget(self.progress_bar) self.current_file_label = QLabel("当前文件: 无") right_layout.addWidget(self.current_file_label) self.tab_widget = QTabWidget() right_layout.addWidget(self.tab_widget, 1) text_tab = QWidget() text_layout = QVBoxLayout() text_tab.setLayout(text_layout) self.text_result = QTextEdit() self.text_result.setReadOnly(True) text_layout.addWidget(self.text_result) self.tab_widget.addTab(text_tab, "文本结果") detail_tab = QWidget() detail_layout = QVBoxLayout() detail_tab.setLayout(detail_layout) self.result_table = QTableWidget() self.result_table.setColumnCount(10) self.result_table.setHorizontalHeaderLabels([ "文件名", "时长", "语速", "音量稳定性", "客服情感", "客户情感", "开场白", "结束语", "禁用词", "问题解决" ]) self.result_table.horizontalHeader().setSectionResizeMode(QHeaderView.Stretch) detail_layout.addWidget(self.result_table) self.tab_widget.addTab(detail_tab, "详细结果") splitter.addWidget(left_widget) splitter.addWidget(right_widget) splitter.setSizes([300, 900]) def setup_menu(self): menu_bar = self.menuBar() file_menu = menu_bar.addMenu("文件") file_actions = [ ("添加文件", self.add_files), ("导出结果", self.export_results), ("退出", self.close) ] for name, func in file_actions: action = QAction(name, self) action.triggered.connect(func) file_menu.addAction(action) analysis_menu = menu_bar.addMenu("分析") analysis_actions = [ ("开始分析", self.start_analysis), ("停止分析", self.stop_analysis) ] for name, func in analysis_actions: action = QAction(name, self) action.triggered.connect(func) analysis_menu.addAction(action) settings_menu = menu_bar.addMenu("设置") settings_actions = [ ("系统配置", self.open_settings), ("加载模型", self.load_models) ] for name, func in settings_actions: action = QAction(name, self) action.triggered.connect(func) settings_menu.addAction(action) def add_files(self): files, _ = QFileDialog.getOpenFileNames( self, "选择音频文件", "", "音频文件 (*.mp3 *.wav *.amr *.m4a)" ) for file in files: self.file_list.addItem(file) def start_analysis(self): if self.file_list.count() == 0: QMessageBox.warning(self, "警告", "请先添加要分析的音频文件") return if not self.model_loaded: QMessageBox.warning(self, "警告", "模型未加载,请先加载模型") return audio_paths = [self.file_list.item(i).text() for i in range(self.file_list.count())] self.text_result.clear() self.result_table.setRowCount(0) self.analysis_thread = AnalysisThread(audio_paths, self.temp_dir) self.analysis_thread.progress_updated.connect(self.update_progress) self.analysis_thread.result_ready.connect(self.handle_result) self.analysis_thread.finished_all.connect(self.analysis_finished) self.analysis_thread.error_occurred.connect(self.show_error) self.analysis_thread.memory_warning.connect(self.handle_memory_warning) self.analysis_thread.start() def stop_analysis(self): if self.analysis_thread and self.analysis_thread.isRunning(): self.analysis_thread.stop() self.analysis_thread.wait() QMessageBox.information(self, "信息", "分析已停止") def load_models(self): if self.model_load_thread and self.model_load_thread.isRunning(): return self.model_load_thread = ModelLoadThread() self.model_load_thread.progress_updated.connect(lambda value, _: self.progress_bar.setValue(value)) self.model_load_thread.finished.connect(self.handle_model_load_result) self.model_load_thread.start() def update_progress(self, progress: int, message: str, current_file: str): self.progress_bar.setValue(progress) self.current_file_label.setText(f"当前文件: {current_file}") def handle_result(self, result: Dict): if result["status"] == "success": self.text_result.append( f"文件: {result['file_name']}\n状态: {result['status']}\n时长: {result['duration_str']}") self.text_result.append( f"语速: {result['syllable_rate']} 音节/秒\n音量稳定性: {result['volume_stability']}") self.text_result.append( f"客服情感: 负面({result['agent_negative']:.2%}) 中性({result['agent_neutral']:.2%}) 正面({result['agent_positive']:.2%})") self.text_result.append(f"客服情绪: {result['agent_emotions']}") self.text_result.append( f"客户情感: 负面({result['customer_negative']:.2%}) 中性({result['customer_neutral']:.2%}) 正面({result['customer_positive']:.2%})") self.text_result.append(f"客户情绪: {result['customer_emotions']}") self.text_result.append( f"开场白: {'有' if result['opening_found'] else '无'}\n结束语: {'有' if result['closing_found'] else '无'}") self.text_result.append( f"禁用词: {result['forbidden_words']}\n问题解决: {'是' if result['issue_resolved'] else '否'}") self.text_result.append("\n=== 对话文本 ===\n" + result["asr_text"] + "\n" + "=" * 50 + "\n") row = self.result_table.rowCount() self.result_table.insertRow(row) items = [ result["file_name"], result["duration_str"], str(result["syllable_rate"]), str(result["volume_stability"]), f"负:{result['agent_negative']:.2f} 中:{result['agent_neutral']:.2f} 正:{result['agent_positive']:.2f}", f"负:{result['customer_negative']:.2f} 中:{result['customer_neutral']:.2f} 正:{result['customer_positive']:.2f}", "是" if result["opening_found"] else "否", "是" if result["closing_found"] else "否", result["forbidden_words"], "是" if result["issue_resolved"] else "否" ] for col, text in enumerate(items): item = QTableWidgetItem(text) if col in [6, 7] and text == "否": item.setBackground(QColor(255, 200, 200)) if col == 8 and text != "无": item.setBackground(QColor(255, 200, 200)) if col == 9 and text == "否": item.setBackground(QColor(255, 200, 200)) self.result_table.setItem(row, col, item) def analysis_finished(self): QMessageBox.information(self, "完成", "所有音频分析完成") self.progress_bar.setValue(100) def show_error(self, title: str, message: str): QMessageBox.critical(self, title, message) def handle_memory_warning(self): QMessageBox.warning(self, "内存警告", "内存使用过高,分析已停止") def handle_model_load_result(self, success: bool, message: str): if success: self.model_loaded = True QMessageBox.information(self, "成功", message) else: QMessageBox.critical(self, "错误", message) def open_settings(self): settings_dialog = QDialog(self) settings_dialog.setWindowTitle("系统设置") settings_dialog.setFixedSize(500, 300) # 高度减少 layout = QVBoxLayout() config = ConfigManager().get("model_paths") settings = [ ("ASR模型路径:", config["asr"], self.browse_directory), ("情感模型路径:", config["sentiment"], self.browse_directory) ] for label, value, func in settings: h_layout = QHBoxLayout() h_layout.addWidget(QLabel(label)) line_edit = QLineEdit(value) browse_btn = QPushButton("浏览...") browse_btn.clicked.connect(lambda _, le=line_edit: func(le)) h_layout.addWidget(line_edit) h_layout.addWidget(browse_btn) layout.addLayout(h_layout) spin_settings = [ ("最大并发任务:", "max_concurrent", 1, 8), ("最大音频时长(秒):", "max_audio_duration", 60, 86400) ] for label, key, min_val, max_val in spin_settings: h_layout = QHBoxLayout() h_layout.addWidget(QLabel(label)) spin_box = QSpinBox() spin_box.setRange(min_val, max_val) spin_box.setValue(ConfigManager().get(key, min_val)) h_layout.addWidget(spin_box) layout.addLayout(h_layout) button_box = QDialogButtonBox(QDialogButtonBox.Ok | QDialogButtonBox.Cancel) button_box.accepted.connect(settings_dialog.accept) button_box.rejected.connect(settings_dialog.reject) layout.addWidget(button_box) settings_dialog.setLayout(layout) if settings_dialog.exec_() == QDialog.Accepted: ConfigManager().set("model_paths", { "asr": layout.itemAt(0).layout().itemAt(1).widget().text(), "sentiment": layout.itemAt(1).layout().itemAt(1).widget().text() }) ConfigManager().set("max_concurrent", layout.itemAt(2).layout().itemAt(1).widget().value()) ConfigManager().set("max_audio_duration", layout.itemAt(3).layout().itemAt(1).widget().value()) ModelLoader.reload_models() def browse_directory(self, line_edit): path = QFileDialog.getExistingDirectory(self, "选择目录") if path: line_edit.setText(path) def export_results(self): if self.result_table.rowCount() == 0: QMessageBox.warning(self, "警告", "没有可导出的结果") return path, _ = QFileDialog.getSaveFileName(self, "保存结果", "", "CSV文件 (*.csv)") if not path: return try: with open(path, "w", encoding="utf-8") as f: headers = [self.result_table.horizontalHeaderItem(col).text() for col in range(self.result_table.columnCount())] f.write(",".join(headers) + "\n") for row in range(self.result_table.rowCount()): row_data = [self.result_table.item(row, col).text() for col in range(self.result_table.columnCount())] f.write(",".join(row_data) + "\n") QMessageBox.information(self, "成功", f"结果已导出到: {path}") except Exception as e: QMessageBox.critical(self, "错误", f"导出失败: {str(e)}") def closeEvent(self, event): if self.analysis_thread and self.analysis_thread.isRunning(): self.analysis_thread.stop() self.analysis_thread.wait() try: for file in os.listdir(self.temp_dir): file_path = os.path.join(self.temp_dir, file) if os.path.isfile(file_path): for _ in range(3): try: os.remove(file_path); break except: time.sleep(0.1) os.rmdir(self.temp_dir) except: pass event.accept() # ====================== 程序入口 ====================== if __name__ == "__main__": torch.set_num_threads(4) app = QApplication(sys.argv) app.setStyle('Fusion') window = MainWindow() window.show() sys.exit(app.exec_())
08-05
检查这段代码,是否存在错误,是否可优化降低内存资源在保持核心功能及输出一致性的情况下,给出修改后的所有完整代码(包括未改变的): class EnhancedAnalysisThread(QThread): progress_updated = pyqtSignal(int, str, str) result_ready = pyqtSignal(dict) finished_all = pyqtSignal() error_occurred = pyqtSignal(str, str) memory_warning = pyqtSignal() resource_cleanup = pyqtSignal() def __init__(self, audio_paths: List[str], temp_dir: str = None): super().__init__() self.audio_paths = audio_paths # 创建唯一临时目录 if temp_dir is None: timestamp = int(time.time() * 1000) self.temp_dir = f"temp_wav_{os.getpid()}_{timestamp}" else: self.temp_dir = temp_dir os.makedirs(self.temp_dir, exist_ok=True) self.is_running = True self.current_file = "" self.max_concurrent = min(ConfigManager().get("max_concurrent", 1), self._get_max_concurrent_tasks()) self.resource_monitor = EnhancedResourceMonitor() self.semaphore = Semaphore(self.max_concurrent) # 创建语音识别处理器 self.asr_processor = ASRProcessor(EnhancedModelLoader, self.resource_monitor) def run(self): try: # 检查ffmpeg是否可用 ffmpeg_available, ffmpeg_msg = check_ffmpeg_available() if not ffmpeg_available: self.error_occurred.emit("音频处理依赖缺失", f"无法处理音频: {ffmpeg_msg}\n\n请安装ffmpeg并确保其在系统PATH中。\nWindows用户可从https://ffmpeg.org/download.html下载并添加到环境变量。") return if not EnhancedModelLoader.models_loaded: self.error_occurred.emit("模型未加载", "请等待模型加载完成后再开始分析") return self.progress_updated.emit(0, f"最大并行任务数: {self.max_concurrent}", "") with concurrent.futures.ThreadPoolExecutor(max_workers=self.max_concurrent) as executor: future_to_path = {} for path in self.audio_paths: if not self.is_running: break self.semaphore.acquire() future = executor.submit(self.analyze_audio, path, self._get_available_batch_size()) future_to_path[future] = path future.add_done_callback(lambda f: self.semaphore.release()) for i, future in enumerate(concurrent.futures.as_completed(future_to_path)): if not self.is_running: break path = future_to_path[future] self.current_file = os.path.basename(path) if self._check_memory_usage(): self.memory_warning.emit() self.is_running = False break try: result = future.result() if result: self.result_ready.emit(result) progress = int((i + 1) / len(self.audio_paths) * 100) self.progress_updated.emit(progress, f"完成: {self.current_file} ({i + 1}/{len(self.audio_paths)})", self.current_file) except Exception as e: result = {"file_name": self.current_file, "status": "error", "error": f"分析失败: {str(e)}"} self.result_ready.emit(result) if self.is_running: self.finished_all.emit() except Exception as e: self.error_occurred.emit("系统错误", str(e)) traceback.print_exc() finally: self.resource_cleanup.emit() self._cleanup_resources() def analyze_audio(self, audio_path: str, batch_size: int) -> Dict: result = {"file_name": os.path.basename(audio_path), "status": "processing"} wav_paths = [] try: wav_paths = EnhancedAudioProcessor.convert_to_wav(audio_path, self.temp_dir) if not wav_paths: result["error"] = "格式转换失败,请检查文件是否损坏或格式是否支持" result["status"] = "error" return result audio_features = self._extract_audio_features(wav_paths) result.update(audio_features) result["duration_str"] = self._format_duration(audio_features["duration"]) all_segments, full_text = self._process_asr_segments(wav_paths) agent_segments, customer_segments = self._identify_speakers(all_segments) result["asr_text"] = self._generate_labeled_text(all_segments, agent_segments, customer_segments).strip() text_analysis = self._analyze_text(agent_segments, customer_segments, batch_size) result.update(text_analysis) service_check = self._check_service_rules(agent_segments) result.update(service_check) result["issue_resolved"] = self._check_issue_resolution(customer_segments, agent_segments) result["status"] = "success" except Exception as e: result["error"] = f"分析失败: {str(e)}" result["status"] = "error" finally: self._cleanup_temp_files(wav_paths) self._cleanup_resources() return result def _identify_speakers(self, segments: List[Dict]) -> Tuple[List[Dict], List[Dict]]: """使用五层逻辑识别客服(增加混合策略)""" if not segments: return [], [] # 逻辑1:前三片段开场白关键词 agent_id = self._identify_by_opening(segments) # 逻辑2:后三片段结束语关键词 if agent_id is None: agent_id = self._identify_by_closing(segments) # 逻辑3:称呼与敬语关键词 if agent_id is None: agent_id = self._identify_by_salutation(segments) # 逻辑4:安抚语关键词 if agent_id is None: agent_id = self._identify_by_reassurance(segments) # 后备策略:说话模式识别 if agent_id is None and len(segments) >= 4: agent_id = self._identify_by_speech_patterns(segments) # 第五层:混合策略(加权评分) if agent_id is None and segments: scores = defaultdict(float) for seg in segments: text = seg["text"] # 开场白权重 (前三段) if seg in segments[:3]: if any(kw in text for kw in EnhancedDialectProcessor.KEYWORDS["opening"]): scores[seg["spk_id"]] += 1.5 # 结束语权重 (后三段) if seg in (segments[-3:] if len(segments) >= 3 else segments): if any(kw in text for kw in EnhancedDialectProcessor.KEYWORDS["closing"]): scores[seg["spk_id"]] += 1.5 # 安抚语权重 if any(kw in text for kw in EnhancedDialectProcessor.KEYWORDS["reassurance"]): scores[seg["spk_id"]] += 1.0 # 称呼权重 if any(kw in text for kw in EnhancedDialectProcessor.KEYWORDS["salutation"]): scores[seg["spk_id"]] += 0.8 if scores: agent_id = max(scores, key=scores.get) if agent_id is None: # 最后手段:选择说话最多的说话人 spk_counts = {} for seg in segments: spk_id = seg["spk_id"] spk_counts[spk_id] = spk_counts.get(spk_id, 0) + 1 agent_id = max(spk_counts, key=spk_counts.get) if spk_counts else None if agent_id is None: return [], [] return ( [seg for seg in segments if seg["spk_id"] == agent_id], [seg for seg in segments if seg["spk_id"] != agent_id] ) def _identify_by_opening(self, segments: List[Dict]) -> Optional[str]: """逻辑1:前三片段开场白关键词""" keywords = EnhancedDialectProcessor.KEYWORDS["opening"] for seg in segments[:3]: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_closing(self, segments: List[Dict]) -> Optional[str]: """逻辑2:后三片段结束语关键词""" keywords = EnhancedDialectProcessor.KEYWORDS["closing"] last_segments = segments[-3:] if len(segments) >= 3 else segments for seg in reversed(last_segments): if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_salutation(self, segments: List[Dict]) -> Optional[str]: """逻辑3:称呼与敬语关键词""" keywords = EnhancedDialectProcessor.KEYWORDS["salutation"] for seg in segments: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_reassurance(self, segments: List[Dict]) -> Optional[str]: """逻辑4:安抚语关键词""" keywords = EnhancedDialectProcessor.KEYWORDS["reassurance"] for seg in segments: if any(kw in seg["text"] for kw in keywords): return seg["spk_id"] return None def _identify_by_speech_patterns(self, segments: List[Dict]) -> Optional[str]: """后备策略:说话模式识别""" speaker_features = {} for seg in segments: spk_id = seg["spk_id"] if spk_id not in speaker_features: speaker_features[spk_id] = {"total_duration": 0.0, "turn_count": 0, "question_count": 0} features = speaker_features[spk_id] features["total_duration"] += (seg["end"] - seg["start"]) features["turn_count"] += 1 if any(q_word in seg["text"] for q_word in ["吗", "呢", "?", "?", "如何", "怎样"]): features["question_count"] += 1 if speaker_features: max_duration = max(f["total_duration"] for f in speaker_features.values()) question_rates = {spk_id: f["question_count"] / f["turn_count"] for spk_id, f in speaker_features.items()} candidates = [] for spk_id, features in speaker_features.items(): score = (0.6 * (features["total_duration"] / max_duration) + 0.4 * question_rates[spk_id]) candidates.append((spk_id, score)) return max(candidates, key=lambda x: x[1])[0] return None def _analyze_text(self, agent_segments: List[Dict], customer_segments: List[Dict], batch_size: int) -> Dict: """情感分析方法""" def split_long_sentences(texts: List[str]) -> List[str]: splitted = [] for text in texts: if len(text) > 128: parts = re.split(r'(?<=[。!?;,])', text) current = "" for part in parts: if len(current) + len(part) < 128: current += part else: if current: splitted.append(current) current = part if current: splitted.append(current) else: splitted.append(text) return splitted def enhance_with_keywords(texts: List[str]) -> List[str]: enhanced = [] emotion_keywords = { "positive": ["满意", "高兴", "感谢", "专业", "解决", "帮助", "谢谢", "很好", "不错"], "negative": ["生气", "愤怒", "不满", "投诉", "问题", "失望", "差劲", "糟糕", "投诉"], "neutral": ["了解", "明白", "知道", "确认", "查询", "记录", "需要", "提供"] } for text in texts: found_emotion = None for emotion, keywords in emotion_keywords.items(): if any(kw in text for kw in keywords): found_emotion = emotion break if found_emotion: enhanced.append(f"[{found_emotion}] {text}") else: enhanced.append(text) return enhanced def analyze_emotion_intensity(texts: List[str]) -> List[str]: """分析情绪强度""" intensity_keywords = { "high": ["非常", "极其", "特别", "十分", "太"], "medium": ["比较", "相当", "挺"], "low": ["有点", "稍微", "些许"] } enhanced = [] for text in texts: intensity = "" for level, keywords in intensity_keywords.items(): if any(kw in text for kw in keywords): intensity = f"[{level}-intensity]" break enhanced.append(f"{intensity} {text}" if intensity else text) return enhanced # 分析单个说话者 def analyze_speaker(segments: List[Dict], speaker_type: str) -> Dict: if not segments: return { f"{speaker_type}_negative": 0.0, f"{speaker_type}_neutral": 1.0, f"{speaker_type}_positive": 0.0, f"{speaker_type}_emotions": "无" } texts = [seg["text"] for seg in segments] processed_texts = EnhancedDialectProcessor.preprocess_text(texts) splitted_texts = split_long_sentences(processed_texts) enhanced_texts = enhance_with_keywords(splitted_texts) # 增加情绪强度分析 enhanced_texts = analyze_emotion_intensity(enhanced_texts) with EnhancedModelLoader.model_lock: inputs = EnhancedModelLoader.sentiment_tokenizer( enhanced_texts, padding=True, truncation=True, max_length=128, return_tensors="pt" ) dataset = TensorDataset(inputs['input_ids'], inputs['attention_mask']) dataloader = DataLoader(dataset, batch_size=batch_size, shuffle=False) device = "cuda" if torch.cuda.is_available() else "cpu" sentiment_dist = [] emotions = [] for batch in dataloader: input_ids, attention_mask = batch inputs = {'input_ids': input_ids.to(device), 'attention_mask': attention_mask.to(device)} with torch.no_grad(): outputs = EnhancedModelLoader.sentiment_model(**inputs) batch_probs = torch.nn.functional.softmax(outputs.logits, dim=-1) sentiment_dist.append(batch_probs.cpu()) emotion_keywords = ["愤怒", "生气", "鬼火", "不耐烦", "搞哪样嘛", "恼火", "背时", "失望", "不满"] for text in enhanced_texts: if any(kw in text for kw in emotion_keywords): if any(kw in text for kw in ["愤怒", "生气", "鬼火", "恼火"]): emotions.append("愤怒") elif any(kw in text for kw in ["不耐烦", "搞哪样嘛"]): emotions.append("不耐烦") elif "背时" in text: emotions.append("沮丧") elif any(kw in text for kw in ["失望", "不满"]): emotions.append("失望") if sentiment_dist: all_probs = torch.cat(sentiment_dist, dim=0) avg_sentiment = torch.mean(all_probs, dim=0).tolist() else: avg_sentiment = [0.0, 1.0, 0.0] return { f"{speaker_type}_negative": round(avg_sentiment[0], 4), f"{speaker_type}_neutral": round(avg_sentiment[1], 4), f"{speaker_type}_positive": round(avg_sentiment[2], 4), f"{speaker_type}_emotions": ",".join(set(emotions)) if emotions else "无" } return {**analyze_speaker(agent_segments, "agent"), **analyze_speaker(customer_segments, "customer") } def _check_service_rules(self, agent_segments: List[Dict]) -> Dict: keywords = EnhancedDialectProcessor.KEYWORDS found_forbidden = [] found_opening = any(kw in seg["text"] for seg in agent_segments[:3] for kw in keywords["opening"]) found_closing = any( kw in seg["text"] for seg in (agent_segments[-3:] if len(agent_segments) >= 3 else agent_segments) for kw in keywords["closing"]) for seg in agent_segments: for kw in keywords["forbidden"]: if kw in seg["text"]: found_forbidden.append(kw) break return { "opening_found": found_opening, "closing_found": found_closing, "forbidden_words": ", ".join(set(found_forbidden)) if found_forbidden else "无" } def _check_issue_resolution(self, customer_segments: List[Dict], agent_segments: List[Dict]) -> bool: if not customer_segments or not agent_segments: return False resolution_keywords = ["解决", "处理", "完成", "已", "好了", "可以了", "没问题", "明白", "清楚", "满意", "行"] unresolved_keywords = ["没解决", "不行", "不对", "还是", "仍然", "再", "未", "无法", "不能", "不行", "不满意"] negation_words = ["不", "没", "未", "非", "无"] gratitude_keywords = ["谢谢", "感谢", "多谢", "麻烦", "辛苦", "有劳"] full_conversation = " ".join(seg["text"] for seg in customer_segments + agent_segments) last_customer_text = customer_segments[-1]["text"] for kw in unresolved_keywords: if kw in full_conversation: negation_context = re.search(rf".{{0,5}}{kw}", full_conversation) if negation_context: context = negation_context.group(0) if not any(neg in context for neg in negation_words): return False else: return False if any(kw in last_customer_text for kw in gratitude_keywords): if not any(neg + kw in last_customer_text for neg in negation_words): return True for agent_text in [seg["text"] for seg in agent_segments[-3:]]: if any(kw in agent_text for kw in resolution_keywords): if not any(neg in agent_text for neg in negation_words): return True for cust_seg in customer_segments[-2:]: if any(kw in cust_seg["text"] for kw in ["好", "行", "可以", "明白"]): if not any(neg in cust_seg["text"] for neg in negation_words): return True if any("?" in seg["text"] or "?" in seg["text"] for seg in customer_segments[-2:]): return False return False # ====================== 辅助方法 ====================== def _get_available_batch_size(self) -> int: if not torch.cuda.is_available(): return 4 # 根据内存使用趋势动态调整批次大小 mem_trend = self.resource_monitor.get_usage_trend() gpu_usage = mem_trend.get("gpu", 0) if gpu_usage > 70: return 2 elif gpu_usage < 30: return 8 else: return 4 def _get_max_concurrent_tasks(self) -> int: if torch.cuda.is_available(): total_mem = torch.cuda.get_device_properties(0).total_memory / (1024 ** 3) return 1 if total_mem < 6 else 2 if total_mem < 12 else 3 return max(1, os.cpu_count() // 2) def _process_asr_segments(self, wav_paths: List[str]) -> Tuple[List[Dict], str]: """使用ASR处理器处理音频分段""" return self.asr_processor.process_audio_segments(wav_paths, lambda: self.is_running) def _check_memory_usage(self) -> bool: try: return self.resource_monitor.is_under_heavy_load() except: return False def _extract_audio_features(self, wav_paths: List[str]) -> Dict[str, float]: combined_y = np.array([], dtype=np.float32) sr = ConfigManager().get("sample_rate", 16000) for path in wav_paths: y, _ = librosa.load(path, sr=sr) combined_y = np.concatenate((combined_y, y)) return EnhancedAudioProcessor.extract_features_from_audio(combined_y, sr) def _generate_labeled_text(self, all_segments: List[Dict], agent_segments: List[Dict], customer_segments: List[Dict]) -> str: agent_spk_id = agent_segments[0]["spk_id"] if agent_segments else None customer_spk_id = customer_segments[0]["spk_id"] if customer_segments else None labeled_text = [] for seg in all_segments: if seg["spk_id"] == agent_spk_id: speaker = "客服" elif seg["spk_id"] == customer_spk_id: speaker = "客户" else: speaker = f"说话人{seg['spk_id']}" labeled_text.append(f"[{speaker}]: {seg['text']}") return "\n".join(labeled_text) def _cleanup_temp_files(self, paths: List[str]): def safe_remove(path): if os.path.exists(path): try: os.remove(path) except: pass for path in paths: safe_remove(path) now = time.time() for file in os.listdir(self.temp_dir): file_path = os.path.join(self.temp_dir, file) if os.path.isfile(file_path) and (now - os.path.getmtime(file_path)) > 3600: safe_remove(file_path) def _format_duration(self, seconds: float) -> str: minutes, seconds = divmod(int(seconds), 60) hours, minutes = divmod(minutes, 60) return f"{hours:02d}:{minutes:02d}:{seconds:02d}" def _cleanup_resources(self): gc.collect() if torch.cuda.is_available(): torch.cuda.empty_cache() def stop(self): self.is_running = False
09-06
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值