{"id":31016,"date":"2025-04-21T07:07:19","date_gmt":"2025-04-20T23:07:19","guid":{"rendered":"https:\/\/www.wsisp.com\/helps\/31016.html"},"modified":"2025-04-21T07:07:19","modified_gmt":"2025-04-20T23:07:19","slug":"python%e5%9f%ba%e4%ba%8e%e6%b7%b1%e5%ba%a6%e5%ad%a6%e4%b9%a0%e7%9a%84%e5%a4%9a%e6%a8%a1%e6%80%81%e4%ba%ba%e8%84%b8%e6%83%85%e7%bb%aa%e8%af%86%e5%88%ab%e7%a0%94%e7%a9%b6%e4%b8%8e%e5%ae%9e%e7%8e%b0","status":"publish","type":"post","link":"https:\/\/www.wsisp.com\/helps\/31016.html","title":{"rendered":"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0"},"content":{"rendered":"<p>\u4e00\u3001\u7cfb\u7edf\u67b6\u6784\u8bbe\u8ba1<\/p>\n<p>A[\u6570\u636e\u91c7\u96c6] &#8211;&gt; B[\u9884\u5904\u7406\u6a21\u5757]<\/p>\n<p>B &#8211;&gt; C[\u7279\u5f81\u63d0\u53d6]<\/p>\n<p>C &#8211;&gt; D[\u591a\u6a21\u6001\u878d\u5408]<\/p>\n<p>D &#8211;&gt; E[\u60c5\u7eea\u5206\u7c7b]<\/p>\n<p>E &#8211;&gt; F[\u7cfb\u7edf\u90e8\u7f72]<\/p>\n<p>F &#8211;&gt; G[\u7528\u6237\u754c\u9762]<\/p>\n<p>\u4e8c\u3001\u6570\u636e\u51c6\u5907\u4e0e\u5904\u7406<\/p>\n<p>1. \u6570\u636e\u6536\u96c6<\/p>\n<p>&#8211; \u89c6\u9891\u6570\u636e&#xff1a;FER2013&#xff08;\u9759\u6001\u56fe\u50cf&#xff09;\u3001RAVDESS&#xff08;\u52a8\u6001\u89c6\u9891&#xff09;<\/p>\n<p>&#8211; \u97f3\u9891\u6570\u636e&#xff1a;CREMA-D\u3001IEMOCAP<\/p>\n<p>&#8211; \u81ea\u5b9a\u4e49\u91c7\u96c6&#xff1a;\u4f7f\u7528OpenCV&#043;PyAudio\u5b9e\u73b0\u540c\u6b65\u91c7\u96c6<\/p>\n<p>2. \u6570\u636e\u9884\u5904\u7406<\/p>\n<p>\u89c6\u9891\u5904\u7406&#xff1a;<\/p>\n<p>import cv2<\/p>\n<p>def process_video(video_path):<\/p>\n<p>\u00a0 \u00a0 cap &#061; cv2.VideoCapture(video_path)<\/p>\n<p>\u00a0 \u00a0 frames &#061; []<\/p>\n<p>\u00a0 \u00a0 while cap.isOpened():<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 ret, frame &#061; cap.read()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 if not ret: break<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # \u4eba\u8138\u68c0\u6d4b<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 face &#061; cv2.CascadeClassifier(&#039;haarcascade_frontalface_default.xml&#039;)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 gray &#061; cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 faces &#061; face.detectMultiScale(gray, 1.3, 5)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # \u88c1\u526a\u548c\u5f52\u4e00\u5316<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 if len(faces) &gt; 0:<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 (x,y,w,h) &#061; faces[0]<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 roi &#061; cv2.resize(gray[y:y&#043;h, x:x&#043;w], (128,128))<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 frames.append(roi)<\/p>\n<p>\u00a0 \u00a0 return np.array(frames)<\/p>\n<p>\u00a0<\/p>\n<p>\u97f3\u9891\u5904\u7406&#xff1a;<\/p>\n<p>import librosa<\/p>\n<p>def extract_audio_features(audio_path):<\/p>\n<p>\u00a0 \u00a0 y, sr &#061; librosa.load(audio_path, sr&#061;16000)<\/p>\n<p>\u00a0 \u00a0 # \u5206\u5e27\u5904\u7406&#xff08;30ms\u7a97\u53e3&#xff09;<\/p>\n<p>\u00a0 \u00a0 frames &#061; librosa.util.frame(y, frame_length&#061;480, hop_length&#061;160)<\/p>\n<p>\u00a0 \u00a0 # \u63d0\u53d6MFCC\u7279\u5f81<\/p>\n<p>\u00a0 \u00a0 mfcc &#061; librosa.feature.mfcc(y&#061;y, sr&#061;sr, n_mfcc&#061;40)<\/p>\n<p>\u00a0 \u00a0 # \u52a8\u6001\u7279\u5f81\u62fc\u63a5<\/p>\n<p>\u00a0 \u00a0 delta &#061; librosa.feature.delta(mfcc)<\/p>\n<p>\u00a0 \u00a0 ddelta &#061; librosa.feature.delta(mfcc, order&#061;2)<\/p>\n<p>\u00a0 \u00a0 return np.concatenate([mfcc, delta, ddelta], axis&#061;0)<\/p>\n<p>\u00a0<\/p>\n<p>3. \u6570\u636e\u540c\u6b65\u7b56\u7565<\/p>\n<p>&#8211; \u4f7f\u7528FFmpeg\u63d0\u53d6\u89c6\u9891\u65f6\u95f4\u6233<\/p>\n<p>&#8211; \u52a8\u6001\u65f6\u95f4\u89c4\u6574&#xff08;DTW&#xff09;\u5bf9\u9f50\u97f3\u89c6\u9891\u5e8f\u5217<\/p>\n<p>&#8211; \u521b\u5efa\u65f6\u95f4\u5bf9\u9f50\u7684\u5143\u6570\u636e\u6587\u4ef6<\/p>\n<p>\u00a0<\/p>\n<p>\u4e09\u3001\u6a21\u578b\u8bbe\u8ba1\u4e0e\u8bad\u7ec3<\/p>\n<p>1. \u89c6\u89c9\u5206\u652f&#xff08;PyTorch\u5b9e\u73b0&#xff09;<\/p>\n<p>import torch<\/p>\n<p>from torchvision.models import resnet34<\/p>\n<p>\u00a0<\/p>\n<p>class VisualNet(nn.Module):<\/p>\n<p>\u00a0 \u00a0 def __init__(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 super().__init__()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.base &#061; resnet34(pretrained&#061;True)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.base.fc &#061; nn.Identity() # \u79fb\u9664\u5168\u8fde\u63a5\u5c42<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.temporal &#061; nn.LSTM(512, 256, bidirectional&#061;True)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def forward(self, x):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # x: (B, T, C, H, W)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 B, T &#061; x.shape[:2]<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 x &#061; x.view(B*T, *x.shape[2:])<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 features &#061; self.base(x) # (B*T, 512)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 features &#061; features.view(B, T, -1)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 out, _ &#061; self.temporal(features)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 return out[:, -1] # \u53d6\u6700\u540e\u65f6\u523b\u8f93\u51fa<\/p>\n<p>\u00a0<\/p>\n<p>2. \u97f3\u9891\u5206\u652f<\/p>\n<p>class AudioNet(nn.Module):<\/p>\n<p>\u00a0 \u00a0 def __init__(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 super().__init__()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.conv &#061; nn.Sequential(<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 nn.Conv1d(120, 64, 3, padding&#061;1),<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 nn.BatchNorm1d(64),<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 nn.ReLU(),<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 nn.MaxPool1d(2))<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.lstm &#061; nn.LSTM(64, 128, bidirectional&#061;True)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def forward(self, x):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # x: (B, T, Features)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 x &#061; x.permute(0,2,1) # (B, Features, T)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 x &#061; self.conv(x)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 x &#061; x.permute(2,0,1) # (T, B, Features)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 out, _ &#061; self.lstm(x)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 return out[-1]<\/p>\n<p>3. \u591a\u6a21\u6001\u878d\u5408<\/p>\n<p>\u6ce8\u610f\u529b\u878d\u5408\u5c42&#xff1a;<\/p>\n<p>class FusionModule(nn.Module):<\/p>\n<p>\u00a0 \u00a0 def __init__(self, v_dim, a_dim):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 super().__init__()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.v_proj &#061; nn.Linear(v_dim, 256)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.a_proj &#061; nn.Linear(a_dim, 256)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.attention &#061; nn.MultiheadAttention(256, 4)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def forward(self, v_feat, a_feat):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 v &#061; self.v_proj(v_feat).unsqueeze(0) # (1,B,256)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 a &#061; self.a_proj(a_feat).unsqueeze(0)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 combined &#061; torch.cat([v, a], dim&#061;0) # (2,B,256)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 attn_out, _ &#061; self.attention(combined, combined, combined)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 return attn_out.mean(dim&#061;0)<\/p>\n<p>\u00a0<\/p>\n<p>\u00a0\u56db\u3001\u8bad\u7ec3\u7b56\u7565<\/p>\n<p>1. \u635f\u5931\u51fd\u6570\u8bbe\u8ba1<\/p>\n<p>class MultimodalLoss(nn.Module):<\/p>\n<p>\u00a0 \u00a0 def __init__(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 super().__init__()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.ce &#061; nn.CrossEntropyLoss()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.kl &#061; nn.KLDivLoss()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def forward(self, pred, label, v_out, a_out):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # \u4e3b\u635f\u5931<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 main_loss &#061; self.ce(pred, label)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 # \u6a21\u6001\u4e00\u81f4\u6027\u635f\u5931<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 p_v &#061; F.log_softmax(v_out, dim&#061;1)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 p_a &#061; F.softmax(a_out, dim&#061;1)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 consistency_loss &#061; self.kl(p_v, p_a.detach())<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 return main_loss &#043; 0.5 * consistency_loss<\/p>\n<p>\u00a0<\/p>\n<p>2. \u8bad\u7ec3\u6280\u5de7<\/p>\n<p>&#8211; \u5206\u9636\u6bb5\u8bad\u7ec3&#xff1a;\u5148\u5355\u6a21\u6001\u9884\u8bad\u7ec3&#xff0c;\u518d\u8054\u5408\u5fae\u8c03<\/p>\n<p>&#8211; \u6570\u636e\u589e\u5f3a\u7b56\u7565&#xff1a;<\/p>\n<p>\u00a0 &#8211; \u89c6\u89c9&#xff1a;\u968f\u673a\u906e\u6321\u3001\u8272\u5f69\u6296\u52a8<\/p>\n<p>\u00a0 &#8211; \u97f3\u9891&#xff1a;\u6dfb\u52a0\u566a\u58f0\u3001\u65f6\u79fb\u53d8\u6362<\/p>\n<p>&#8211; \u4f18\u5316\u5668\u914d\u7f6e&#xff1a;<\/p>\n<p>\u00a0 optimizer &#061; torch.optim.AdamW([<\/p>\n<p>\u00a0 \u00a0 \u00a0 {&#039;params&#039;: visual_net.parameters(), &#039;lr&#039;: 1e-4},<\/p>\n<p>\u00a0 \u00a0 \u00a0 {&#039;params&#039;: audio_net.parameters(), &#039;lr&#039;: 3e-4},<\/p>\n<p>\u00a0 \u00a0 \u00a0 {&#039;params&#039;: fusion_module.parameters(), &#039;lr&#039;: 5e-4}<\/p>\n<p>\u00a0 ], weight_decay&#061;1e-5)<\/p>\n<p>\u4e94\u3001\u5b9e\u65f6\u5904\u7406\u4e0e\u90e8\u7f72<\/p>\n<p>1. \u5b9e\u65f6\u5904\u7406\u67b6\u6784<\/p>\n<p>import queue<\/p>\n<p>from threading import Thread<\/p>\n<p>\u00a0<\/p>\n<p>class RealTimeProcessor:<\/p>\n<p>\u00a0 \u00a0 def __init__(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.video_queue &#061; queue.Queue(maxsize&#061;30)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 self.audio_queue &#061; queue.Queue(maxsize&#061;100)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def video_capture(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 cap &#061; cv2.VideoCapture(0)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 while True:<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 ret, frame &#061; cap.read()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 processed &#061; process_frame(frame)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 self.video_queue.put(processed)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def audio_capture(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 p &#061; pyaudio.PyAudio()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 stream &#061; p.open(format&#061;pyaudio.paInt16, channels&#061;1,<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 rate&#061;16000, input&#061;True,<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 frames_per_buffer&#061;1024)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 while True:<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 data &#061; stream.read(1024)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 features &#061; extract_features(data)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 self.audio_queue.put(features)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 def sync_processor(self):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 while True:<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 # \u52a8\u6001\u65f6\u95f4\u5bf9\u9f50\u7b97\u6cd5<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 video_batch &#061; self.get_video_window()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 audio_batch &#061; self.get_audio_window()<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 aligned_data &#061; dtw_align(video_batch, audio_batch)<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 yield aligned_data<\/p>\n<p>\u00a0<\/p>\n<p>2. \u90e8\u7f72\u4f18\u5316\u65b9\u6848<\/p>\n<p>&#8211; \u4f7f\u7528TensorRT\u8fdb\u884c\u6a21\u578b\u91cf\u5316&#xff1a;<\/p>\n<p>\u00a0 trtexec &#8211;onnx&#061;model.onnx &#8211;saveEngine&#061;model.engine \\\\<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0&#8211;fp16 &#8211;workspace&#061;2048<\/p>\n<p>&#8211; \u8fb9\u7f18\u8bbe\u5907\u4f18\u5316&#xff1a;<\/p>\n<p>\u00a0 import torch_tensorrt<\/p>\n<p>\u00a0 traced_model &#061; torch.jit.trace(model, example_input)<\/p>\n<p>\u00a0 trt_model &#061; torch_tensorrt.compile(traced_model,<\/p>\n<p>\u00a0 \u00a0 \u00a0 inputs&#061; [torch_tensorrt.Input((1, 3, 128, 128),\u00a0<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0torch_tensorrt.Input((1, 100, 120))],<\/p>\n<p>\u00a0 \u00a0 \u00a0 enabled_precisions&#061; {torch.float16})<\/p>\n<p>\u00a0<\/p>\n<p>\u516d\u3001\u8bc4\u4f30\u4e0e\u8c03\u4f18<\/p>\n<p>\u00a01. \u8bc4\u4f30\u6307\u6807<\/p>\n<p>from sklearn.metrics import f1_score, confusion_matrix<\/p>\n<p>\u00a0<\/p>\n<p>def evaluate(y_true, y_pred):<\/p>\n<p>\u00a0 \u00a0 acc &#061; (y_true &#061;&#061; y_pred).mean()<\/p>\n<p>\u00a0 \u00a0 f1 &#061; f1_score(y_true, y_pred, average&#061;&#039;macro&#039;)<\/p>\n<p>\u00a0 \u00a0 cm &#061; confusion_matrix(y_true, y_pred)<\/p>\n<p>\u00a0 \u00a0 return {&#039;accuracy&#039;: acc, &#039;f1&#039;: f1, &#039;confusion_matrix&#039;: cm}<\/p>\n<p>\u00a0<\/p>\n<p>2. \u6a21\u578b\u5206\u6790\u5de5\u5177<\/p>\n<p>import shap<\/p>\n<p>\u00a0<\/p>\n<p>def explain_sample(video, audio):<\/p>\n<p>\u00a0 \u00a0 explainer &#061; shap.DeepExplainer(model)<\/p>\n<p>\u00a0 \u00a0 shap_values &#061; explainer.shap_values([video, audio])<\/p>\n<p>\u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0 # \u53ef\u89c6\u5316\u5404\u6a21\u6001\u8d21\u732e\u5ea6<\/p>\n<p>\u00a0 \u00a0 shap.image_plot(shap_values[0], video)<\/p>\n<p>\u00a0 \u00a0 shap.summary_plot(shap_values[1], audio)<\/p>\n<p>\u00a0<\/p>\n<p>\u4e03\u3001\u7cfb\u7edf\u96c6\u6210\u65b9\u6848<\/p>\n<p>1. \u670d\u52a1\u7aef\u67b6\u6784<\/p>\n<p>from fastapi import FastAPI<\/p>\n<p>from pydantic import BaseModel<\/p>\n<p>app &#061; FastAPI()<\/p>\n<p>class Request(BaseModel):<\/p>\n<p>\u00a0 \u00a0 video_url: str<\/p>\n<p>\u00a0 \u00a0 audio_url: str<\/p>\n<p>&#064;app.post(&#034;\/analyze&#034;)<\/p>\n<p>async def analyze(data: Request):<\/p>\n<p>\u00a0 \u00a0 video &#061; download_and_process(data.video_url)<\/p>\n<p>\u00a0 \u00a0 audio &#061; process_audio(data.audio_url)<\/p>\n<p>\u00a0 \u00a0 with torch.no_grad():<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 prediction &#061; model(video, audio)<\/p>\n<p>\u00a0 \u00a0 return {&#034;emotion&#034;: class_names[prediction.argmax()]}<\/p>\n<p>\u00a02. \u524d\u7aef\u754c\u9762\u793a\u4f8b<\/p>\n<p>\/\/ React\u7ec4\u4ef6\u793a\u4f8b<\/p>\n<p>function EmotionDetector() {<\/p>\n<p>\u00a0 const [result, setResult] &#061; useState(null);<\/p>\n<p>\u00a0 const handleUpload &#061; async (files) &#061;&gt; {<\/p>\n<p>\u00a0 \u00a0 const formData &#061; new FormData();<\/p>\n<p>\u00a0 \u00a0 formData.append(&#039;video&#039;, files[0]);<\/p>\n<p>\u00a0 \u00a0 formData.append(&#039;audio&#039;, files[1]);<\/p>\n<p>\u00a0 \u00a0 const res &#061; await fetch(&#039;\/analyze&#039;, {<\/p>\n<p>\u00a0 \u00a0 \u00a0 method: &#039;POST&#039;,<\/p>\n<p>\u00a0 \u00a0 \u00a0 body: formData<\/p>\n<p>\u00a0 \u00a0 });<\/p>\n<p>\u00a0 \u00a0 setResult(await res.json());<\/p>\n<p>\u00a0 };<\/p>\n<p>\u00a0 return (<\/p>\n<p>\u00a0 \u00a0 &lt;div&gt;<\/p>\n<p>\u00a0 \u00a0 \u00a0 &lt;input type&#061;&#034;file&#034; onChange&#061;{e &#061;&gt; handleUpload(e.target.files)} \/&gt;<\/p>\n<p>\u00a0 \u00a0 \u00a0 {result &amp;&amp; &lt;EmotionChart data&#061;{result}\/&gt;}<\/p>\n<p>\u00a0 \u00a0 &lt;\/div&gt;<\/p>\n<p>\u00a0 );<\/p>\n<p>}<\/p>\n<p>\u516b\u3001\u6311\u6218\u89e3\u51b3\u65b9\u6848<\/p>\n<p>1. \u6a21\u6001\u5f02\u6b65\u95ee\u9898&#xff1a;<\/p>\n<p>\u00a0 \u00a0&#8211; \u91c7\u7528\u53cc\u7f13\u51b2\u961f\u5217&#043;\u52a8\u6001\u65f6\u95f4\u89c4\u6574<\/p>\n<p>\u00a0 \u00a0&#8211; \u8bbe\u7f6e\u6700\u5927\u7b49\u5f85\u65f6\u5ef6&#xff08;200ms&#xff09;&#xff0c;\u8d85\u65f6\u4f7f\u7528\u63d2\u503c\u8865\u507f<\/p>\n<p>2. \u566a\u58f0\u5904\u7406&#xff1a;<\/p>\n<p>\u00a0 \u00a0def denoise_audio(audio):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0return nr.reduce_noise(y&#061;audio, sr&#061;16000,\u00a0<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0stationary&#061;True,<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0prop_decrease&#061;0.8)\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0\u00a0<\/p>\n<p>\u00a0 \u00a0def enhance_video(frame):<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0clahe &#061; cv2.createCLAHE(clipLimit&#061;3.0, tileGridSize&#061;(8,8))<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0return clahe.apply(frame)<\/p>\n<p>3. \u8d44\u6e90\u4f18\u5316&#xff1a;<\/p>\n<p>\u00a0 \u00a0&#8211; \u4f7f\u7528\u6a21\u578b\u84b8\u998f\u6280\u672f&#xff1a;<\/p>\n<p>\u00a0 \u00a0distiller &#061; Distiller(teacher&#061;teacher_model, student&#061;student_model)<\/p>\n<p>\u00a0 \u00a0distiller.train_with_distillation(train_loader,\u00a0<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0alpha&#061;0.3,\u00a0<\/p>\n<p>\u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0 \u00a0temperature&#061;4)<\/p>\n<p>\u603b\u7ed3&#xff1a;<\/p>\n<p>\u8be5\u65b9\u6848\u5b8c\u6574\u8986\u76d6\u4e86\u4ece\u6570\u636e\u91c7\u96c6\u5230\u90e8\u7f72\u7684\u5168\u6d41\u7a0b&#xff0c;\u91cd\u70b9\u89e3\u51b3\u4e86\u591a\u6a21\u6001\u7cfb\u7edf\u4e2d\u7684\u5173\u952e\u6311\u6218\u3002\u5b9e\u9645\u90e8\u7f72\u65f6\u53ef\u6839\u636e\u786c\u4ef6\u8d44\u6e90\u8c03\u6574\u6a21\u578b\u590d\u6742\u5ea6&#xff0c;\u63a8\u8350\u4f7f\u7528NVIDIA Jetson\u7cfb\u5217\u8bbe\u5907\u8fdb\u884c\u8fb9\u7f18\u90e8\u7f72\u3002<\/p>\n","protected":false},"excerpt":{"rendered":"<p>\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb1.8k\u6b21\uff0c\u70b9\u8d5e10\u6b21\uff0c\u6536\u85cf14\u6b21\u3002return out[:, -1] # \u53d6\u6700\u540e\u65f6\u523b\u8f93\u51fa\u3002- \u89c6\u9891\u6570\u636e\uff1aFER2013\uff08\u9759\u6001\u56fe\u50cf\uff09\u3001RAVDESS\uff08\u52a8\u6001\u89c6\u9891\uff09- \u81ea\u5b9a\u4e49\u91c7\u96c6\uff1a\u4f7f\u7528OpenCV+PyAudio\u5b9e\u73b0\u540c\u6b65\u91c7\u96c6\u3002- \u8bbe\u7f6e\u6700\u5927\u7b49\u5f85\u65f6\u5ef6\uff08200ms\uff09\uff0c\u8d85\u65f6\u4f7f\u7528\u63d2\u503c\u8865\u507f\u3002- \u97f3\u9891\u6570\u636e\uff1aCREMA-D\u3001IEMOCAP\u3002- \u5206\u9636\u6bb5\u8bad\u7ec3\uff1a\u5148\u5355\u6a21\u6001\u9884\u8bad\u7ec3\uff0c\u518d\u8054\u5408\u5fae\u8c03\u3002A[\u6570\u636e\u91c7\u96c6] &#8211;&gt; B[\u9884\u5904\u7406\u6a21\u5757]- \u52a8\u6001\u65f6\u95f4\u89c4\u6574\uff08DTW\uff09\u5bf9\u9f50\u97f3\u89c6\u9891\u5e8f\u5217\u3002- \u91c7\u7528\u53cc\u7f13\u51b2\u961f\u5217+\u52a8\u6001\u65f6\u95f4\u89c4\u6574\u3002- \u89c6\u89c9\uff1a\u968f\u673a\u906e\u6321\u3001\u8272\u5f69\u6296\u52a8\u3002- \u97f3\u9891\uff1a\u6dfb\u52a0\u566a\u58f0\u3001\u65f6\u79fb\u53d8\u6362\u3002_python \u591a\u6a21\u6001\u8bc6\u522b\u7684\u7b56\u7565<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[81,1395,214],"topic":[],"class_list":["post-31016","post","type-post","status-publish","format-standard","hentry","category-server","tag-python","tag-1395","tag-214"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.wsisp.com\/helps\/31016.html\" \/>\n<meta property=\"og:locale\" content=\"zh_CN\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"og:description\" content=\"\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb1.8k\u6b21\uff0c\u70b9\u8d5e10\u6b21\uff0c\u6536\u85cf14\u6b21\u3002return out[:, -1] # \u53d6\u6700\u540e\u65f6\u523b\u8f93\u51fa\u3002- \u89c6\u9891\u6570\u636e\uff1aFER2013\uff08\u9759\u6001\u56fe\u50cf\uff09\u3001RAVDESS\uff08\u52a8\u6001\u89c6\u9891\uff09- \u81ea\u5b9a\u4e49\u91c7\u96c6\uff1a\u4f7f\u7528OpenCV+PyAudio\u5b9e\u73b0\u540c\u6b65\u91c7\u96c6\u3002- \u8bbe\u7f6e\u6700\u5927\u7b49\u5f85\u65f6\u5ef6\uff08200ms\uff09\uff0c\u8d85\u65f6\u4f7f\u7528\u63d2\u503c\u8865\u507f\u3002- \u97f3\u9891\u6570\u636e\uff1aCREMA-D\u3001IEMOCAP\u3002- \u5206\u9636\u6bb5\u8bad\u7ec3\uff1a\u5148\u5355\u6a21\u6001\u9884\u8bad\u7ec3\uff0c\u518d\u8054\u5408\u5fae\u8c03\u3002A[\u6570\u636e\u91c7\u96c6] --&gt; B[\u9884\u5904\u7406\u6a21\u5757]- \u52a8\u6001\u65f6\u95f4\u89c4\u6574\uff08DTW\uff09\u5bf9\u9f50\u97f3\u89c6\u9891\u5e8f\u5217\u3002- \u91c7\u7528\u53cc\u7f13\u51b2\u961f\u5217+\u52a8\u6001\u65f6\u95f4\u89c4\u6574\u3002- \u89c6\u89c9\uff1a\u968f\u673a\u906e\u6321\u3001\u8272\u5f69\u6296\u52a8\u3002- \u97f3\u9891\uff1a\u6dfb\u52a0\u566a\u58f0\u3001\u65f6\u79fb\u53d8\u6362\u3002_python \u591a\u6a21\u6001\u8bc6\u522b\u7684\u7b56\u7565\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.wsisp.com\/helps\/31016.html\" \/>\n<meta property=\"og:site_name\" content=\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"article:published_time\" content=\"2025-04-20T23:07:19+00:00\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u4f5c\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4\" \/>\n\t<meta name=\"twitter:data2\" content=\"5 \u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/31016.html\",\"url\":\"https:\/\/www.wsisp.com\/helps\/31016.html\",\"name\":\"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"isPartOf\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\"},\"datePublished\":\"2025-04-20T23:07:19+00:00\",\"dateModified\":\"2025-04-20T23:07:19+00:00\",\"author\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\"},\"breadcrumb\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/31016.html#breadcrumb\"},\"inLanguage\":\"zh-Hans\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.wsisp.com\/helps\/31016.html\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/31016.html#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"\u9996\u9875\",\"item\":\"https:\/\/www.wsisp.com\/helps\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\",\"url\":\"https:\/\/www.wsisp.com\/helps\/\",\"name\":\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"description\":\"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"zh-Hans\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"contentUrl\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"caption\":\"admin\"},\"sameAs\":[\"http:\/\/wp.wsisp.com\"],\"url\":\"https:\/\/www.wsisp.com\/helps\/author\/admin\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.wsisp.com\/helps\/31016.html","og_locale":"zh_CN","og_type":"article","og_title":"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","og_description":"\u6587\u7ae0\u6d4f\u89c8\u9605\u8bfb1.8k\u6b21\uff0c\u70b9\u8d5e10\u6b21\uff0c\u6536\u85cf14\u6b21\u3002return out[:, -1] # \u53d6\u6700\u540e\u65f6\u523b\u8f93\u51fa\u3002- \u89c6\u9891\u6570\u636e\uff1aFER2013\uff08\u9759\u6001\u56fe\u50cf\uff09\u3001RAVDESS\uff08\u52a8\u6001\u89c6\u9891\uff09- \u81ea\u5b9a\u4e49\u91c7\u96c6\uff1a\u4f7f\u7528OpenCV+PyAudio\u5b9e\u73b0\u540c\u6b65\u91c7\u96c6\u3002- \u8bbe\u7f6e\u6700\u5927\u7b49\u5f85\u65f6\u5ef6\uff08200ms\uff09\uff0c\u8d85\u65f6\u4f7f\u7528\u63d2\u503c\u8865\u507f\u3002- \u97f3\u9891\u6570\u636e\uff1aCREMA-D\u3001IEMOCAP\u3002- \u5206\u9636\u6bb5\u8bad\u7ec3\uff1a\u5148\u5355\u6a21\u6001\u9884\u8bad\u7ec3\uff0c\u518d\u8054\u5408\u5fae\u8c03\u3002A[\u6570\u636e\u91c7\u96c6] --&gt; B[\u9884\u5904\u7406\u6a21\u5757]- \u52a8\u6001\u65f6\u95f4\u89c4\u6574\uff08DTW\uff09\u5bf9\u9f50\u97f3\u89c6\u9891\u5e8f\u5217\u3002- \u91c7\u7528\u53cc\u7f13\u51b2\u961f\u5217+\u52a8\u6001\u65f6\u95f4\u89c4\u6574\u3002- \u89c6\u89c9\uff1a\u968f\u673a\u906e\u6321\u3001\u8272\u5f69\u6296\u52a8\u3002- \u97f3\u9891\uff1a\u6dfb\u52a0\u566a\u58f0\u3001\u65f6\u79fb\u53d8\u6362\u3002_python \u591a\u6a21\u6001\u8bc6\u522b\u7684\u7b56\u7565","og_url":"https:\/\/www.wsisp.com\/helps\/31016.html","og_site_name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","article_published_time":"2025-04-20T23:07:19+00:00","author":"admin","twitter_card":"summary_large_image","twitter_misc":{"\u4f5c\u8005":"admin","\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4":"5 \u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/www.wsisp.com\/helps\/31016.html","url":"https:\/\/www.wsisp.com\/helps\/31016.html","name":"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0 - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","isPartOf":{"@id":"https:\/\/www.wsisp.com\/helps\/#website"},"datePublished":"2025-04-20T23:07:19+00:00","dateModified":"2025-04-20T23:07:19+00:00","author":{"@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41"},"breadcrumb":{"@id":"https:\/\/www.wsisp.com\/helps\/31016.html#breadcrumb"},"inLanguage":"zh-Hans","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.wsisp.com\/helps\/31016.html"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/www.wsisp.com\/helps\/31016.html#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"\u9996\u9875","item":"https:\/\/www.wsisp.com\/helps"},{"@type":"ListItem","position":2,"name":"Python\u57fa\u4e8e\u6df1\u5ea6\u5b66\u4e60\u7684\u591a\u6a21\u6001\u4eba\u8138\u60c5\u7eea\u8bc6\u522b\u7814\u7a76\u4e0e\u5b9e\u73b0"}]},{"@type":"WebSite","@id":"https:\/\/www.wsisp.com\/helps\/#website","url":"https:\/\/www.wsisp.com\/helps\/","name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","description":"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"zh-Hans"},{"@type":"Person","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41","name":"admin","image":{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/","url":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","contentUrl":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","caption":"admin"},"sameAs":["http:\/\/wp.wsisp.com"],"url":"https:\/\/www.wsisp.com\/helps\/author\/admin"}]}},"_links":{"self":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/31016","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/comments?post=31016"}],"version-history":[{"count":0,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/31016\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/media?parent=31016"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/categories?post=31016"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/tags?post=31016"},{"taxonomy":"topic","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/topic?post=31016"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}