{"id":61423,"date":"2026-01-17T15:50:04","date_gmt":"2026-01-17T07:50:04","guid":{"rendered":"https:\/\/www.wsisp.com\/helps\/61423.html"},"modified":"2026-01-17T15:50:04","modified_gmt":"2026-01-17T07:50:04","slug":"yolo%e7%b3%bb%e5%88%97%e7%ae%97%e6%b3%95%e6%94%b9%e8%bf%9b-%e8%bd%bb%e9%87%8f%e5%8c%96irmb%e6%b3%a8%e6%84%8f%e5%8a%9b%e6%9c%ba%e5%88%b6-%e5%90%abc3k2_irmb%e4%b8%8ec2psa_irmb","status":"publish","type":"post","link":"https:\/\/www.wsisp.com\/helps\/61423.html","title":{"rendered":"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB"},"content":{"rendered":"<h2 id=\"iRMB%E5%80%92%E7%BD%AE%E6%AE%8B%E5%B7%AE%E5%9D%97%E6%B3%A8%E6%84%8F%E5%8A%9B%E6%9C%BA%E5%88%B6%E7%AE%80%E4%BB%8B\">iRMB\u5012\u7f6e\u6b8b\u5dee\u5757\u6ce8\u610f\u529b\u673a\u5236\u7b80\u4ecb<\/h2>\n<p>iRMB&#xff08;Inverted Residual Mobile Block&#xff09;\u7684\u6846\u67b6\u539f\u7406&#xff0c;\u662f\u4e00\u79cd\u7ed3\u5408\u8f7b\u91cf\u7ea7CNN\u548c\u6ce8\u610f\u529b\u673a\u5236\u7684\u65b9\u6cd5&#xff0c;\u7528\u4e8e\u6539\u8fdb\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u76ee\u6807\u68c0\u6d4b\u6a21\u578b\u3002IRMB\u901a\u8fc7\u5012\u7f6e\u6b8b\u5dee\u5757\u548c\u5143\u79fb\u52a8\u5757\u5b9e\u73b0\u9ad8\u6548\u4fe1\u606f\u5904\u7406&#xff0c;\u540c\u65f6\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u5316\u3002\u672c\u6587\u4e2d\u63d0\u51fa\u4e00\u4e2a\u65b0\u7684\u4e3b\u5e72\u7f51\u7edcEMO&#xff0c;\u4e3b\u8981\u601d\u60f3\u662f\u5c06\u8f7b\u91cf\u7ea7\u7684CNN\u67b6\u6784\u4e0e\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u7ed3\u6784\u76f8\u7ed3\u5408\u3002<\/p>\n<p id=\"main-toc\">\u76ee\u5f55<\/p>\n<p id=\"1.%20%E7%AE%80%E4%BB%8B-toc\" style=\"margin-left:0px\">1. \u7b80\u4ecb<\/p>\n<p id=\"2.%20iRMB%E4%B8%BB%E8%A6%81%E6%80%9D%E6%83%B3-toc\" style=\"margin-left:0px\">2. iRMB\u4e3b\u8981\u601d\u60f3<\/p>\n<p id=\"3.%20iRMB%E7%BB%93%E6%9E%84-toc\" style=\"margin-left:0px\">3. iRMB\u7ed3\u6784<\/p>\n<hr id=\"hr-toc\" \/>\n<p><img loading=\"lazy\" decoding=\"async\" alt=\"\" height=\"179\" src=\"https:\/\/www.wsisp.com\/helps\/wp-content\/uploads\/2026\/01\/20260117075002-696b3f2ad59d4.png\" width=\"745\" \/><\/p>\n<p>\u539f\u59cb\u8bba\u6587&#xff1a;https:\/\/arxiv.org\/pdf\/2301.01146<\/p>\n<p>\u4ee3\u7801\u5730\u5740&#xff1a;https:\/\/github.com\/zhangzjn\/EMO<\/p>\n<h2 id=\"\"><\/h2>\n<h2 id=\"1.%20%E7%AE%80%E4%BB%8B\">1. \u7b80\u4ecb<\/h2>\n<p>\u5012\u7f6e\u6b8b\u5dee\u5757&#xff08;IRB&#xff09;\u4f5c\u4e3a\u8f7b\u91cf\u7ea7CNN\u7684\u57fa\u7840\u8bbe\u65bd&#xff0c;\u5728\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u7814\u7a76\u4e2d\u5c1a\u672a\u6709\u5bf9\u5e94\u7684\u90e8\u5206\u3002\u672c\u6587\u4ece\u7edf\u4e00\u7684\u89c6\u89d2\u91cd\u65b0\u8003\u8651\u4e86\u57fa\u4e8e\u9ad8\u6548IRB\u548cTransformer\u6709\u6548\u7ec4\u4ef6\u7684\u8f7b\u91cf\u7ea7\u57fa\u7840\u8bbe\u65bd&#xff0c;\u5c06\u57fa\u4e8eCNN\u7684IRB\u6269\u5c55\u5230\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b&#xff0c;\u5e76\u62bd\u8c61\u51fa\u4e00\u4e2a\u6b8b\u5dee\u7684\u5143\u79fb\u52a8\u5757&#xff08;MMB&#xff09;\u7528\u4e8e\u8f7b\u91cf\u7ea7\u6a21\u578b\u8bbe\u8ba1\u3002\u9075\u5faa\u7b80\u5355\u4f46\u6709\u6548\u7684\u8bbe\u8ba1\u51c6\u5219&#xff0c;\u672c\u6587\u63a8\u5bfc\u51fa\u4e86\u73b0\u4ee3\u5316\u7684\u5012\u7f6e\u6b8b\u5dee\u79fb\u52a8\u5757&#xff08;IRMB&#xff09;&#xff0c;\u5e76\u4ee5\u6b64\u6784\u5efa\u4e86\u7c7b\u4f3cResNet\u7684\u9ad8\u6548\u6a21\u578b&#xff08;EMO&#xff09;\u7528\u4e8e\u4e0b\u6e38\u4efb\u52a1\u3002<\/p>\n<h2 id=\"2.%20iRMB%E4%B8%BB%E8%A6%81%E6%80%9D%E6%83%B3\">2. iRMB\u4e3b\u8981\u601d\u60f3<\/h2>\n<p>iRMB&#xff08;Inverted Residual Mobile Block&#xff09;\u7684\u4e3b\u8981\u601d\u60f3\u662f\u5c06\u8f7b\u91cf\u7ea7\u7684CNN\u67b6\u6784\u4e0e\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u7ed3\u6784\u76f8\u7ed3\u5408&#xff08;\u4f18\u70b9\u7c7b\u4f3c\u4e8eACmix&#xff09;&#xff0c;\u4ee5\u521b\u5efa\u9ad8\u6548\u7684\u79fb\u52a8\u7f51\u7edc\u3002Irmb\u901a\u8fc7\u91cd\u65b0\u8003\u8651\u5bfc\u81f4\u6b8b\u5dee&#xff08;IRB&#xff09;\u548cTransformer\u7684\u6709\u6548\u7ec4\u4ef6&#xff0c;\u5b9e\u73b0\u4e86\u4e00\u79cd\u7edf\u4e00\u7684\u89c6\u89d2&#xff0c;\u4ece\u800c\u6269\u5c55\u4e86CNN\u7684IRB\u5230\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u3002Irmb\u7684\u8bbe\u8ba1\u76ee\u6807\u662f\u5728\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u7ea7\u7684\u540c\u65f6&#xff0c;\u5b9e\u73b0\u5bf9\u8ba1\u7b97\u8d44\u6e90\u7684\u6709\u6548\u5229\u7528\u548c\u9ad8\u51c6\u786e\u7387\u3002\u8fd9\u4e00\u65b9\u6cd5\u901a\u8fc7\u5728\u4e0b\u6e38\u4efb\u52a1\u4e0a\u7684\u5e7f\u6cdb\u5b9e\u9a8c\u5f97\u5230\u9a8c\u8bc1&#xff0c;\u5c55\u793a\u51fa\u5176\u5728\u8f7b\u91cf\u7ea7\u6a21\u578b\u9886\u57df\u7684\u4f18\u8d8a\u6027\u80fd\u3002<\/p>\n<p>iRMB\u7684\u4e3b\u8981\u521b\u65b0\u70b9\u5728\u4e8e\u4ee5\u4e0b\u4e09\u4e2a\u70b9&#xff1a;<\/p>\n<li>\n<p>\u7ed3\u5408CNN\u7684\u8f7b\u91cf\u7ea7\u7279\u5f81\u548cTransformer\u7684\u52a8\u6001\u6a21\u578b\u80fd\u529b&#xff0c;\u521b\u65b0\u63d0\u51fa\u4e86iRMB\u7ed3\u6784&#xff0c;\u9002\u7528\u4e8e\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u5bc6\u96c6\u9884\u6d4b\u4efb\u52a1<\/p>\n<\/li>\n<li>\n<p>\u4f7f\u7528\u5012\u7f6e\u6b8b\u5dee\u5757\u8bbe\u8ba1&#xff0c;\u6269\u5c55\u4e86\u4f20\u7edfCNN\u7684IRB\u5230\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b&#xff0c;\u589e\u5f3a\u4e86\u6a21\u578b\u5904\u7406\u957f\u8ddd\u79bb\u4fe1\u606f\u7684\u80fd\u529b<\/p>\n<\/li>\n<li>\n<p>\u63d0\u51fa\u4e86\u5143\u79fb\u52a8\u5757&#xff08;Meta-Mobile Block&#xff09;&#xff0c;\u901a\u8fc7\u4e0d\u540c\u7684\u6269\u5c55\u6bd4\u7387\u548c\u9ad8\u6548\u64cd\u4f5c\u7b26&#xff0c;\u5b9e\u73b0\u4e86\u6a21\u578b\u7684\u6a21\u5757\u5316\u8bbe\u8ba1&#xff0c;\u4f7f\u5f97\u6a21\u578b\u66f4\u52a0\u7075\u6d3b\u548c\u9ad8\u6548<\/p>\n<\/li>\n<h2>3. IRMB\u7ed3\u6784<\/h2>\n<p>iRMB\u7ed3\u6784\u7684\u4e3b\u8981\u521b\u65b0\u70b9\u662f\u5b83\u7ed3\u5408\u4e86\u5377\u79ef\u795e\u7ecf\u7f51\u7edc\u7684\u8f7b\u91cf\u7ea7\u7279\u6027\u548cTransformer\u6a21\u578b\u7684\u52a8\u6001\u5904\u7406\u80fd\u529b\u3002\u8fd9\u79cd\u7ed3\u6784\u7279\u522b\u9002\u7528\u4e8e\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u5bc6\u96c6\u9884\u6d4b\u4efb\u52a1&#xff0c;\u56e0\u4e3a\u5b83\u65e8\u5728\u8ba1\u7b97\u80fd\u529b\u6709\u9650\u7684\u73af\u5883\u4e2d\u63d0\u4f9b\u9ad8\u6548\u7684\u6027\u80fd\u3002IRMB\u901a\u8fc7\u5176\u5012\u7f6e\u6b8b\u5dee\u8bbe\u8ba1\u6539\u8fdb\u4e86\u4fe1\u606f\u6d41\u7684\u5904\u7406&#xff0c;\u5141\u8bb8\u5728\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u7684\u540c\u65f6\u6355\u6349\u548c\u5229\u7528\u957f\u8ddd\u79bb\u4f9d\u8d56&#xff0c;\u8fd9\u5bf9\u4e8e\u56fe\u50cf\u5206\u7c7b&#xff0c;\u5bf9\u8c61\u68c0\u6d4b\u548c\u8bed\u4e49\u5206\u5272\u7b49\u4efb\u52a1\u81f3\u5173\u91cd\u8981\u3002\u8fd9\u79cd\u8bbe\u8ba1\u4f7f\u5f97\u6a21\u578b\u5728\u8d44\u6e90\u53d7\u9650\u7684\u8bbe\u5907\u4e0a\u4e5f\u80fd\u591f\u9ad8\u6548\u8fd0\u884c&#xff0c;\u540c\u65f6\u4fdd\u6301\u6216\u63d0\u9ad8\u9884\u6d4b\u51c6\u786e\u6027\u3002<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" alt=\"\" height=\"405\" src=\"https:\/\/www.wsisp.com\/helps\/wp-content\/uploads\/2026\/01\/20260117075002-696b3f2ae6050.png\" width=\"1049\" \/><\/p>\n<h2>4.iRMB\u6a21\u5757\u4ee3\u7801\u5b9e\u73b0<\/h2>\n<p>import math<br \/>\nimport torch<br \/>\nimport torch.nn as nn<br \/>\nimport torch.nn.functional as F<br \/>\nfrom functools import partial<br \/>\nfrom einops import rearrange<br \/>\nfrom timm.models._efficientnet_blocks import SqueezeExcite<br \/>\nfrom timm.models.layers import DropPath<\/p>\n<p>__all__ &#061; [&#039;iRMB&#039;, &#039;C2PSA_iRMB&#039;, &#039;C3k2_iRMB&#039;]<\/p>\n<p>inplace &#061; True  # \u5168\u5c40\u53d8\u91cf<\/p>\n<p>class LayerNorm2d(nn.Module):<\/p>\n<p>    def __init__(self, normalized_shape, eps&#061;1e-6, elementwise_affine&#061;True):<br \/>\n        super().__init__()<br \/>\n        self.norm &#061; nn.LayerNorm(normalized_shape, eps, elementwise_affine)<\/p>\n<p>    def forward(self, x):<br \/>\n        x &#061; rearrange(x, &#039;b c h w -&gt; b h w c&#039;).contiguous()<br \/>\n        x &#061; self.norm(x)<br \/>\n        x &#061; rearrange(x, &#039;b h w c -&gt; b c h w&#039;).contiguous()<br \/>\n        return x<\/p>\n<p>def get_norm(norm_layer&#061;&#039;in_1d&#039;):<br \/>\n    eps &#061; 1e-6<br \/>\n    norm_dict &#061; {<br \/>\n        &#039;none&#039;: nn.Identity,<br \/>\n        &#039;in_1d&#039;: partial(nn.InstanceNorm1d, eps&#061;eps),<br \/>\n        &#039;in_2d&#039;: partial(nn.InstanceNorm2d, eps&#061;eps),<br \/>\n        &#039;in_3d&#039;: partial(nn.InstanceNorm3d, eps&#061;eps),<br \/>\n        &#039;bn_1d&#039;: partial(nn.BatchNorm1d, eps&#061;eps),<br \/>\n        &#039;bn_2d&#039;: partial(nn.BatchNorm2d, eps&#061;eps),<br \/>\n        # &#039;bn_2d&#039;: partial(nn.SyncBatchNorm, eps&#061;eps),<br \/>\n        &#039;bn_3d&#039;: partial(nn.BatchNorm3d, eps&#061;eps),<br \/>\n        &#039;gn&#039;: partial(nn.GroupNorm, eps&#061;eps),<br \/>\n        &#039;ln_1d&#039;: partial(nn.LayerNorm, eps&#061;eps),<br \/>\n        &#039;ln_2d&#039;: partial(LayerNorm2d, eps&#061;eps),<br \/>\n    }<br \/>\n    return norm_dict[norm_layer]<\/p>\n<p>def get_act(act_layer&#061;&#039;relu&#039;):<br \/>\n    act_dict &#061; {<br \/>\n        &#039;none&#039;: nn.Identity,<br \/>\n        &#039;relu&#039;: nn.ReLU,<br \/>\n        &#039;relu6&#039;: nn.ReLU6,<br \/>\n        &#039;silu&#039;: nn.SiLU<br \/>\n    }<br \/>\n    return act_dict[act_layer]<\/p>\n<p>class ConvNormAct(nn.Module):<\/p>\n<p>    def __init__(self, dim_in, dim_out, kernel_size, stride&#061;1, dilation&#061;1, groups&#061;1, bias&#061;False,<br \/>\n                 skip&#061;False, norm_layer&#061;&#039;bn_2d&#039;, act_layer&#061;&#039;relu&#039;, inplace&#061;True, drop_path_rate&#061;0.):<br \/>\n        super(ConvNormAct, self).__init__()<br \/>\n        self.has_skip &#061; skip and dim_in &#061;&#061; dim_out<br \/>\n        padding &#061; math.ceil((kernel_size &#8211; stride) \/ 2)<br \/>\n        self.conv &#061; nn.Conv2d(dim_in, dim_out, kernel_size, stride, padding, dilation, groups, bias)<br \/>\n        self.norm &#061; get_norm(norm_layer)(dim_out)<br \/>\n        self.act &#061; get_act(act_layer)(inplace&#061;inplace)<br \/>\n        self.drop_path &#061; DropPath(drop_path_rate) if drop_path_rate else nn.Identity()<\/p>\n<p>    def forward(self, x):<br \/>\n        shortcut &#061; x<br \/>\n        x &#061; self.conv(x)<br \/>\n        x &#061; self.norm(x)<br \/>\n        x &#061; self.act(x)<br \/>\n        if self.has_skip:<br \/>\n            x &#061; self.drop_path(x) &#043; shortcut<br \/>\n        return x<\/p>\n<p>class iRMB(nn.Module):<\/p>\n<p>    def __init__(self, dim_in,  norm_in&#061;True, has_skip&#061;True, exp_ratio&#061;1.0, norm_layer&#061;&#039;bn_2d&#039;,<br \/>\n                 act_layer&#061;&#039;relu&#039;, v_proj&#061;True, dw_ks&#061;3, stride&#061;1, dilation&#061;1, se_ratio&#061;0.0, dim_head&#061;8, window_size&#061;7,<br \/>\n                 attn_s&#061;True, qkv_bias&#061;False, attn_drop&#061;0., drop&#061;0., drop_path&#061;0., v_group&#061;False, attn_pre&#061;False):<br \/>\n        super().__init__()<br \/>\n        dim_out &#061; dim_in<br \/>\n        self.norm &#061; get_norm(norm_layer)(dim_in) if norm_in else nn.Identity()<br \/>\n        dim_mid &#061; int(dim_in * exp_ratio)<br \/>\n        self.has_skip &#061; (dim_in &#061;&#061; dim_out and stride &#061;&#061; 1) and has_skip<br \/>\n        self.attn_s &#061; attn_s<br \/>\n        if self.attn_s:<br \/>\n            assert dim_in % dim_head &#061;&#061; 0, &#039;dim should be divisible by num_heads&#039;<br \/>\n            self.dim_head &#061; dim_head<br \/>\n            self.window_size &#061; window_size<br \/>\n            self.num_head &#061; dim_in \/\/ dim_head<br \/>\n            self.scale &#061; self.dim_head ** -0.5<br \/>\n            self.attn_pre &#061; attn_pre<br \/>\n            self.qk &#061; ConvNormAct(dim_in, int(dim_in * 2), kernel_size&#061;1, bias&#061;qkv_bias, norm_layer&#061;&#039;none&#039;,<br \/>\n                                  act_layer&#061;&#039;none&#039;)<br \/>\n            self.v &#061; ConvNormAct(dim_in, dim_mid, kernel_size&#061;1, groups&#061;self.num_head if v_group else 1, bias&#061;qkv_bias,<br \/>\n                                 norm_layer&#061;&#039;none&#039;, act_layer&#061;act_layer, inplace&#061;inplace)<br \/>\n            self.attn_drop &#061; nn.Dropout(attn_drop)<br \/>\n        else:<br \/>\n            if v_proj:<br \/>\n                self.v &#061; ConvNormAct(dim_in, dim_mid, kernel_size&#061;1, bias&#061;qkv_bias, norm_layer&#061;&#039;none&#039;,<br \/>\n                                     act_layer&#061;act_layer, inplace&#061;inplace)<br \/>\n            else:<br \/>\n                self.v &#061; nn.Identity()<br \/>\n        self.conv_local &#061; ConvNormAct(dim_mid, dim_mid, kernel_size&#061;dw_ks, stride&#061;stride, dilation&#061;dilation,<br \/>\n                                      groups&#061;dim_mid, norm_layer&#061;&#039;bn_2d&#039;, act_layer&#061;&#039;silu&#039;, inplace&#061;inplace)<br \/>\n        self.se &#061; SqueezeExcite(dim_mid, rd_ratio&#061;se_ratio, act_layer&#061;get_act(act_layer)) if se_ratio &gt; 0.0 else nn.Identity()<\/p>\n<p>        self.proj_drop &#061; nn.Dropout(drop)<br \/>\n        self.proj &#061; ConvNormAct(dim_mid, dim_out, kernel_size&#061;1, norm_layer&#061;&#039;none&#039;, act_layer&#061;&#039;none&#039;, inplace&#061;inplace)<br \/>\n        self.drop_path &#061; DropPath(drop_path) if drop_path else nn.Identity()<\/p>\n<p>    def forward(self, x):<br \/>\n        shortcut &#061; x<br \/>\n        x &#061; self.norm(x)<br \/>\n        B, C, H, W &#061; x.shape<br \/>\n        if self.attn_s:<br \/>\n            # padding<br \/>\n            if self.window_size &lt;&#061; 0:<br \/>\n                window_size_W, window_size_H &#061; W, H<br \/>\n            else:<br \/>\n                window_size_W, window_size_H &#061; self.window_size, self.window_size<br \/>\n            pad_l, pad_t &#061; 0, 0<br \/>\n            pad_r &#061; (window_size_W &#8211; W % window_size_W) % window_size_W<br \/>\n            pad_b &#061; (window_size_H &#8211; H % window_size_H) % window_size_H<br \/>\n            x &#061; F.pad(x, (pad_l, pad_r, pad_t, pad_b, 0, 0,))<br \/>\n            n1, n2 &#061; (H &#043; pad_b) \/\/ window_size_H, (W &#043; pad_r) \/\/ window_size_W<br \/>\n            x &#061; rearrange(x, &#039;b c (h1 n1) (w1 n2) -&gt; (b n1 n2) c h1 w1&#039;, n1&#061;n1, n2&#061;n2).contiguous()<br \/>\n            # attention<br \/>\n            b, c, h, w &#061; x.shape<br \/>\n            qk &#061; self.qk(x)<br \/>\n            qk &#061; rearrange(qk, &#039;b (qk heads dim_head) h w -&gt; qk b heads (h w) dim_head&#039;, qk&#061;2, heads&#061;self.num_head,<br \/>\n                           dim_head&#061;self.dim_head).contiguous()<br \/>\n            q, k &#061; qk[0], qk[1]<br \/>\n            attn_spa &#061; (q &#064; k.transpose(-2, -1)) * self.scale<br \/>\n            attn_spa &#061; attn_spa.softmax(dim&#061;-1)<br \/>\n            attn_spa &#061; self.attn_drop(attn_spa)<br \/>\n            if self.attn_pre:<br \/>\n                x &#061; rearrange(x, &#039;b (heads dim_head) h w -&gt; b heads (h w) dim_head&#039;, heads&#061;self.num_head).contiguous()<br \/>\n                x_spa &#061; attn_spa &#064; x<br \/>\n                x_spa &#061; rearrange(x_spa, &#039;b heads (h w) dim_head -&gt; b (heads dim_head) h w&#039;, heads&#061;self.num_head, h&#061;h,<br \/>\n                                  w&#061;w).contiguous()<br \/>\n                x_spa &#061; self.v(x_spa)<br \/>\n            else:<br \/>\n                v &#061; self.v(x)<br \/>\n                v &#061; rearrange(v, &#039;b (heads dim_head) h w -&gt; b heads (h w) dim_head&#039;, heads&#061;self.num_head).contiguous()<br \/>\n                x_spa &#061; attn_spa &#064; v<br \/>\n                x_spa &#061; rearrange(x_spa, &#039;b heads (h w) dim_head -&gt; b (heads dim_head) h w&#039;, heads&#061;self.num_head, h&#061;h,<br \/>\n                                  w&#061;w).contiguous()<br \/>\n            # unpadding<br \/>\n            x &#061; rearrange(x_spa, &#039;(b n1 n2) c h1 w1 -&gt; b c (h1 n1) (w1 n2)&#039;, n1&#061;n1, n2&#061;n2).contiguous()<br \/>\n            if pad_r &gt; 0 or pad_b &gt; 0:<br \/>\n                x &#061; x[:, :, :H, :W].contiguous()<br \/>\n        else:<br \/>\n            x &#061; self.v(x)<\/p>\n<p>        x &#061; x &#043; self.se(self.conv_local(x)) if self.has_skip else self.se(self.conv_local(x))<\/p>\n<p>        x &#061; self.proj_drop(x)<br \/>\n        x &#061; self.proj(x)<\/p>\n<p>        x &#061; (shortcut &#043; self.drop_path(x)) if self.has_skip else x<br \/>\n        return x<\/p>\n<p>def autopad(k, p&#061;None, d&#061;1):  # kernel, padding, dilation<br \/>\n    &#034;&#034;&#034;Pad to &#039;same&#039; shape outputs.&#034;&#034;&#034;<br \/>\n    if d &gt; 1:<br \/>\n        k &#061; d * (k &#8211; 1) &#043; 1 if isinstance(k, int) else [d * (x &#8211; 1) &#043; 1 for x in k]  # actual kernel-size<br \/>\n    if p is None:<br \/>\n        p &#061; k \/\/ 2 if isinstance(k, int) else [x \/\/ 2 for x in k]  # auto-pad<br \/>\n    return p<\/p>\n<p>class Conv(nn.Module):<br \/>\n    &#034;&#034;&#034;Standard convolution with args(ch_in, ch_out, kernel, stride, padding, groups, dilation, activation).&#034;&#034;&#034;<\/p>\n<p>    default_act &#061; nn.SiLU()  # default activation<\/p>\n<p>    def __init__(self, c1, c2, k&#061;1, s&#061;1, p&#061;None, g&#061;1, d&#061;1, act&#061;True):<br \/>\n        &#034;&#034;&#034;Initialize Conv layer with given arguments including activation.&#034;&#034;&#034;<br \/>\n        super().__init__()<br \/>\n        self.conv &#061; nn.Conv2d(c1, c2, k, s, autopad(k, p, d), groups&#061;g, dilation&#061;d, bias&#061;False)<br \/>\n        self.bn &#061; nn.BatchNorm2d(c2)<br \/>\n        self.act &#061; self.default_act if act is True else act if isinstance(act, nn.Module) else nn.Identity()<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Apply convolution, batch normalization and activation to input tensor.&#034;&#034;&#034;<br \/>\n        return self.act(self.bn(self.conv(x)))<\/p>\n<p>    def forward_fuse(self, x):<br \/>\n        &#034;&#034;&#034;Perform transposed convolution of 2D data.&#034;&#034;&#034;<br \/>\n        return self.act(self.conv(x))<\/p>\n<p>class PSABlock(nn.Module):<br \/>\n    &#034;&#034;&#034;<br \/>\n    PSABlock class implementing a Position-Sensitive Attention block for neural networks.<br \/>\n    This class encapsulates the functionality for applying multi-head attention and feed-forward neural network layers<br \/>\n    with optional shortcut connections.<br \/>\n    Attributes:<br \/>\n        attn (Attention): Multi-head attention module.<br \/>\n        ffn (nn.Sequential): Feed-forward neural network module.<br \/>\n        add (bool): Flag indicating whether to add shortcut connections.<br \/>\n    Methods:<br \/>\n        forward: Performs a forward pass through the PSABlock, applying attention and feed-forward layers.<br \/>\n    Examples:<br \/>\n        Create a PSABlock and perform a forward pass<br \/>\n        &gt;&gt;&gt; psablock &#061; PSABlock(c&#061;128, attn_ratio&#061;0.5, num_heads&#061;4, shortcut&#061;True)<br \/>\n        &gt;&gt;&gt; input_tensor &#061; torch.randn(1, 128, 32, 32)<br \/>\n        &gt;&gt;&gt; output_tensor &#061; psablock(input_tensor)<br \/>\n    &#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c, attn_ratio&#061;0.5, num_heads&#061;4, shortcut&#061;True) -&gt; None:<br \/>\n        &#034;&#034;&#034;Initializes the PSABlock with attention and feed-forward layers for enhanced feature extraction.&#034;&#034;&#034;<br \/>\n        super().__init__()<\/p>\n<p>        self.attn &#061; iRMB(c)<br \/>\n        self.ffn &#061; nn.Sequential(Conv(c, c * 2, 1), Conv(c * 2, c, 1, act&#061;False))<br \/>\n        self.add &#061; shortcut<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Executes a forward pass through PSABlock, applying attention and feed-forward layers to the input tensor.&#034;&#034;&#034;<br \/>\n        x &#061; x &#043; self.attn(x) if self.add else self.attn(x)<br \/>\n        x &#061; x &#043; self.ffn(x) if self.add else self.ffn(x)<br \/>\n        return x<\/p>\n<p>class C2PSA_iRMB(nn.Module):<br \/>\n    &#034;&#034;&#034;<br \/>\n    C2PSA module with attention mechanism for enhanced feature extraction and processing.<br \/>\n    This module implements a convolutional block with attention mechanisms to enhance feature extraction and processing<br \/>\n    capabilities. It includes a series of PSABlock modules for self-attention and feed-forward operations.<br \/>\n    Attributes:<br \/>\n        c (int): Number of hidden channels.<br \/>\n        cv1 (Conv): 1&#215;1 convolution layer to reduce the number of input channels to 2*c.<br \/>\n        cv2 (Conv): 1&#215;1 convolution layer to reduce the number of output channels to c.<br \/>\n        m (nn.Sequential): Sequential container of PSABlock modules for attention and feed-forward operations.<br \/>\n    Methods:<br \/>\n        forward: Performs a forward pass through the C2PSA module, applying attention and feed-forward operations.<br \/>\n    Notes:<br \/>\n        This module essentially is the same as PSA module, but refactored to allow stacking more PSABlock modules.<br \/>\n    Examples:<br \/>\n        &gt;&gt;&gt; c2psa &#061; C2PSA(c1&#061;256, c2&#061;256, n&#061;3, e&#061;0.5)<br \/>\n        &gt;&gt;&gt; input_tensor &#061; torch.randn(1, 256, 64, 64)<br \/>\n        &gt;&gt;&gt; output_tensor &#061; c2psa(input_tensor)<br \/>\n    &#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, n&#061;1, e&#061;0.5):<br \/>\n        &#034;&#034;&#034;Initializes the C2PSA module with specified input\/output channels, number of layers, and expansion ratio.&#034;&#034;&#034;<br \/>\n        super().__init__()<br \/>\n        assert c1 &#061;&#061; c2<br \/>\n        self.c &#061; int(c1 * e)<br \/>\n        self.cv1 &#061; Conv(c1, 2 * self.c, 1, 1)<br \/>\n        self.cv2 &#061; Conv(2 * self.c, c1, 1)<\/p>\n<p>        self.m &#061; nn.Sequential(*(PSABlock(self.c, attn_ratio&#061;0.5, num_heads&#061;self.c \/\/ 64) for _ in range(n)))<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Processes the input tensor &#039;x&#039; through a series of PSA blocks and returns the transformed tensor.&#034;&#034;&#034;<br \/>\n        a, b &#061; self.cv1(x).split((self.c, self.c), dim&#061;1)<br \/>\n        b &#061; self.m(b)<br \/>\n        return self.cv2(torch.cat((a, b), 1))<\/p>\n<p>class Bottleneck(nn.Module):<br \/>\n    &#034;&#034;&#034;Standard bottleneck.&#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, shortcut&#061;True, g&#061;1, k&#061;(3, 3), e&#061;0.5):<br \/>\n        &#034;&#034;&#034;Initializes a standard bottleneck module with optional shortcut connection and configurable parameters.&#034;&#034;&#034;<br \/>\n        super().__init__()<br \/>\n        c_ &#061; int(c2 * e)  # hidden channels<br \/>\n        self.cv1 &#061; Conv(c1, c_, k[0], 1)<br \/>\n        self.cv2 &#061; Conv(c_, c2, k[1], 1, g&#061;g)<br \/>\n        self.add &#061; shortcut and c1 &#061;&#061; c2<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Applies the YOLO FPN to input data.&#034;&#034;&#034;<br \/>\n        return x &#043; self.cv2(self.cv1(x)) if self.add else self.cv2(self.cv1(x))<\/p>\n<p>class C3(nn.Module):<br \/>\n    &#034;&#034;&#034;CSP Bottleneck with 3 convolutions.&#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, n&#061;1, shortcut&#061;True, g&#061;1, e&#061;0.5):<br \/>\n        &#034;&#034;&#034;Initialize the CSP Bottleneck with given channels, number, shortcut, groups, and expansion values.&#034;&#034;&#034;<br \/>\n        super().__init__()<br \/>\n        c_ &#061; int(c2 * e)  # hidden channels<br \/>\n        self.cv1 &#061; Conv(c1, c_, 1, 1)<br \/>\n        self.cv2 &#061; Conv(c1, c_, 1, 1)<br \/>\n        self.cv3 &#061; Conv(2 * c_, c2, 1)  # optional act&#061;FReLU(c2)<br \/>\n        self.m &#061; nn.Sequential(*(Bottleneck(c_, c_, shortcut, g, k&#061;((1, 1), (3, 3)), e&#061;1.0) for _ in range(n)))<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Forward pass through the CSP bottleneck with 2 convolutions.&#034;&#034;&#034;<br \/>\n        return self.cv3(torch.cat((self.m(self.cv1(x)), self.cv2(x)), 1))<\/p>\n<p>class C3k(C3):<br \/>\n    &#034;&#034;&#034;C3k is a CSP bottleneck module with customizable kernel sizes for feature extraction in neural networks.&#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, n&#061;1, shortcut&#061;True, g&#061;1, e&#061;0.5, k&#061;3):<br \/>\n        &#034;&#034;&#034;Initializes the C3k module with specified channels, number of layers, and configurations.&#034;&#034;&#034;<br \/>\n        super().__init__(c1, c2, n, shortcut, g, e)<br \/>\n        c_ &#061; int(c2 * e)  # hidden channels<br \/>\n        # self.m &#061; nn.Sequential(*(RepBottleneck(c_, c_, shortcut, g, k&#061;(k, k), e&#061;1.0) for _ in range(n)))<br \/>\n        self.m &#061; nn.Sequential(*(Bottleneck(c_, c_, shortcut, g, k&#061;(k, k), e&#061;1.0) for _ in range(n)))<\/p>\n<p>class C2f(nn.Module):<br \/>\n    &#034;&#034;&#034;Faster Implementation of CSP Bottleneck with 2 convolutions.&#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, n&#061;1, shortcut&#061;False, g&#061;1, e&#061;0.5):<br \/>\n        &#034;&#034;&#034;Initializes a CSP bottleneck with 2 convolutions and n Bottleneck blocks for faster processing.&#034;&#034;&#034;<br \/>\n        super().__init__()<br \/>\n        self.c &#061; int(c2 * e)  # hidden channels<br \/>\n        self.cv1 &#061; Conv(c1, 2 * self.c, 1, 1)<br \/>\n        self.cv2 &#061; Conv((2 &#043; n) * self.c, c2, 1)  # optional act&#061;FReLU(c2)<br \/>\n        self.m &#061; nn.ModuleList(Bottleneck(self.c, self.c, shortcut, g, k&#061;((3, 3), (3, 3)), e&#061;1.0) for _ in range(n))<\/p>\n<p>    def forward(self, x):<br \/>\n        &#034;&#034;&#034;Forward pass through C2f layer.&#034;&#034;&#034;<br \/>\n        y &#061; list(self.cv1(x).chunk(2, 1))<br \/>\n        y.extend(m(y[-1]) for m in self.m)<br \/>\n        return self.cv2(torch.cat(y, 1))<\/p>\n<p>    def forward_split(self, x):<br \/>\n        &#034;&#034;&#034;Forward pass using split() instead of chunk().&#034;&#034;&#034;<br \/>\n        y &#061; list(self.cv1(x).split((self.c, self.c), 1))<br \/>\n        y.extend(m(y[-1]) for m in self.m)<br \/>\n        return self.cv2(torch.cat(y, 1))<\/p>\n<p>class C3k2(C2f):<br \/>\n    &#034;&#034;&#034;Faster Implementation of CSP Bottleneck with 2 convolutions.&#034;&#034;&#034;<\/p>\n<p>    def __init__(self, c1, c2, n&#061;1, c3k&#061;False, e&#061;0.5, g&#061;1, shortcut&#061;True):<br \/>\n        &#034;&#034;&#034;Initializes the C3k2 module, a faster CSP Bottleneck with 2 convolutions and optional C3k blocks.&#034;&#034;&#034;<br \/>\n        super().__init__(c1, c2, n, shortcut, g, e)<br \/>\n        self.m &#061; nn.ModuleList(<br \/>\n            C3k(self.c, self.c, 2, shortcut, g) if c3k else Bottleneck(self.c, self.c, shortcut, g) for _ in range(n)<br \/>\n        )<\/p>\n<p>class C3k_iRMB(C3k):<br \/>\n    def __init__(self, c1, c2, n&#061;1, shortcut&#061;False, g&#061;1, e&#061;0.5, k&#061;3):<br \/>\n        super().__init__(c1, c2, n, shortcut, g, e, k)<br \/>\n        c_ &#061; int(c2 * e)  # hidden channels<br \/>\n        self.m &#061; nn.Sequential(*(iRMB(c_, c_) for _ in range(n)))<\/p>\n<p>class C3k2_iRMB(C3k2):<br \/>\n    def __init__(self, c1, c2, n&#061;1, c3k&#061;False, e&#061;0.5, g&#061;1, shortcut&#061;True):<br \/>\n        super().__init__(c1, c2, n, c3k, e, g, shortcut)<br \/>\n        self.m &#061; nn.ModuleList(C3k_iRMB(self.c, self.c, 2, shortcut, g) if c3k else iRMB(self.c, self.c) for _ in range(n)) <\/p>\n<h2>5.\u5177\u4f53\u6539\u8fdb\u64cd\u9aa4<\/h2>\n<h3>5.1 \u5728ultralytics\/nn\/modules\u4e0b\u65b0\u5efaiRMB.py\u6587\u4ef6<\/h3>\n<p><img loading=\"lazy\" decoding=\"async\" alt=\"\" height=\"440\" src=\"https:\/\/www.wsisp.com\/helps\/wp-content\/uploads\/2026\/01\/20260117075003-696b3f2b2616a.png\" width=\"346\" \/><\/p>\n<p>\u5728iRMB.py\u6587\u4ef6\u91cc\u6dfb\u52a0\u7ed9\u51fa\u7684iRMB\u4ee3\u7801<\/p>\n<p>\u6dfb\u52a0\u5b8ciRMB\u4ee3\u7801\u540e&#xff0c;\u5728ultralytics\/nn\/modules\/__init__.py\u6587\u4ef6\u4e2d\u5f15\u7528<\/p>\n<p>from .iRMB import * <\/p>\n<p>\u7136\u540e&#xff0c;\u5728ultralytics\/nn\/tasks.py\u91cc\u5f15\u7528<\/p>\n<p>from .modules import * <\/p>\n<h3>5.2 \u5728ultralytics\/nn\/tasks.py\u4fee\u6539<\/h3>\n<p>\u5728tasks.py\u627e\u5230parse_model\u51fd\u6570&#xff08;\u53ef\u4ee5ctrl&#043;f \u76f4\u63a5\u641c\u7d22parse_model\u4f4d\u7f6e&#xff09;\u6dfb\u52a0&#xff1a;<\/p>\n<p>        elif m in {iRMB}:<br \/>\n            c2 &#061; ch[f]<br \/>\n            args &#061; [c2, *args] <\/p>\n<h2>6.\u521b\u5efaYAML\u914d\u7f6e\u6587\u4ef6<\/h2>\n<p>\u6b64\u5904\u4ee5YOLOv11\u4ee3\u7801\u4e3a\u4f8b&#xff0c;\u521b\u5efaYOLO11_iRMB.yaml<\/p>\n<p># Ultralytics YOLO &#x1f680;, AGPL-3.0 license<br \/>\n# YOLO11 object detection model with P3-P5 outputs. For Usage examples see https:\/\/docs.ultralytics.com\/tasks\/detect<\/p>\n<p># Parameters<br \/>\nnc: 80 # number of classes<br \/>\nscales: # model compound scaling constants, i.e. &#039;model&#061;yolo11n.yaml&#039; will call yolo11.yaml with scale &#039;n&#039;<br \/>\n  # [depth, width, max_channels]<br \/>\n  n: [0.50, 0.25, 1024] # summary: 319 layers, 2624080 parameters, 2624064 gradients, 6.6 GFLOPs<br \/>\n  s: [0.50, 0.50, 1024] # summary: 319 layers, 9458752 parameters, 9458736 gradients, 21.7 GFLOPs<br \/>\n  m: [0.50, 1.00, 512] # summary: 409 layers, 20114688 parameters, 20114672 gradients, 68.5 GFLOPs<br \/>\n  l: [1.00, 1.00, 512] # summary: 631 layers, 25372160 parameters, 25372144 gradients, 87.6 GFLOPs<br \/>\n  x: [1.00, 1.50, 512] # summary: 631 layers, 56966176 parameters, 56966160 gradients, 196.0 GFLOPs<\/p>\n<p># YOLO11n backbone<br \/>\nbackbone:<br \/>\n  # [from, repeats, module, args]<br \/>\n  &#8211; [-1, 1, Conv, [64, 3, 2]] # 0-P1\/2<br \/>\n  &#8211; [-1, 1, Conv, [128, 3, 2]] # 1-P2\/4<br \/>\n  &#8211; [-1, 2, C3k2, [256, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [256, 3, 2]] # 3-P3\/8<br \/>\n  &#8211; [-1, 2, C3k2, [512, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [512, 3, 2]] # 5-P4\/16<br \/>\n  &#8211; [-1, 2, C3k2, [512, True]]<br \/>\n  &#8211; [-1, 1, Conv, [1024, 3, 2]] # 7-P5\/32<br \/>\n  &#8211; [-1, 2, C3k2, [1024, True]]<br \/>\n  &#8211; [-1, 1, SPPF, [1024, 5]] # 9<br \/>\n  &#8211; [-1, 2, C2PSA, [1024]] # 10<\/p>\n<p># YOLO11n head<br \/>\nhead:<br \/>\n  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 6], 1, Concat, [1]] # cat backbone P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 13<\/p>\n<p>  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 4], 1, Concat, [1]] # cat backbone P3<br \/>\n  &#8211; [-1, 2, C3k2, [256, False]] # 16<br \/>\n  &#8211; [-1, 1, iRMB, []] # 17     \u5c0f\u76ee\u6807\u68c0\u6d4b\u5c42\u589e\u52a0\u6ce8\u610f\u529b\u673a\u5236<\/p>\n<p>  &#8211; [-1, 1, Conv, [256, 3, 2]]<br \/>\n  &#8211; [[-1, 13], 1, Concat, [1]] # cat head P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 20 (P4\/16-medium)<br \/>\n  &#8211; [-1, 1, iRMB, []] # 21     \u4e2d\u76ee\u6807\u68c0\u6d4b\u5c42\u589e\u52a0\u6ce8\u610f\u529b\u673a\u5236<\/p>\n<p>  &#8211; [-1, 1, Conv, [512, 3, 2]]<br \/>\n  &#8211; [[-1, 10], 1, Concat, [1]] # cat head P5<br \/>\n  &#8211; [-1, 2, C3k2, [1024, True]] # 24<br \/>\n  &#8211; [-1, 1, iRMB, []] # 25     \u5927\u76ee\u6807\u68c0\u6d4b\u5c42\u589e\u52a0\u6ce8\u610f\u529b\u673a\u5236<\/p>\n<p>  &#8211; [[17, 21, 25], 1, Detect, [nc]] # Detect(P3, P4, P5) <\/p>\n<p>\u4ee5\u4e0a\u662f\u4e00\u79cd\u6539\u8fdb\u65b9\u6cd5&#xff0c;\u4e5f\u53ef\u4ee5\u5728C3k2\u6a21\u5757\u3001C2PSA\u6a21\u5757\u7b49\u5904\u8fdb\u884c\u6539\u8fdb&#xff0c;\u6784\u5efa\u65b0\u7684\u7684C3k2_iRMB\u3001C2PSA_iRMB\u6a21\u5757\u3002<\/p>\n<p>YOLO11_C2PSA_iRMB.yaml<\/p>\n<p># Ultralytics YOLO &#x1f680;, AGPL-3.0 license<br \/>\n# YOLO11 object detection model with P3-P5 outputs. For Usage examples see https:\/\/docs.ultralytics.com\/tasks\/detect<\/p>\n<p># Parameters<br \/>\nnc: 80 # number of classes<br \/>\nscales: # model compound scaling constants, i.e. &#039;model&#061;yolo11n.yaml&#039; will call yolo11.yaml with scale &#039;n&#039;<br \/>\n  # [depth, width, max_channels]<br \/>\n  n: [0.50, 0.25, 1024] # summary: 319 layers, 2624080 parameters, 2624064 gradients, 6.6 GFLOPs<br \/>\n  s: [0.50, 0.50, 1024] # summary: 319 layers, 9458752 parameters, 9458736 gradients, 21.7 GFLOPs<br \/>\n  m: [0.50, 1.00, 512] # summary: 409 layers, 20114688 parameters, 20114672 gradients, 68.5 GFLOPs<br \/>\n  l: [1.00, 1.00, 512] # summary: 631 layers, 25372160 parameters, 25372144 gradients, 87.6 GFLOPs<br \/>\n  x: [1.00, 1.50, 512] # summary: 631 layers, 56966176 parameters, 56966160 gradients, 196.0 GFLOPs<\/p>\n<p># YOLO11n backbone<br \/>\nbackbone:<br \/>\n  # [from, repeats, module, args]<br \/>\n  &#8211; [-1, 1, Conv, [64, 3, 2]] # 0-P1\/2<br \/>\n  &#8211; [-1, 1, Conv, [128, 3, 2]] # 1-P2\/4<br \/>\n  &#8211; [-1, 2, C3k2, [256, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [256, 3, 2]] # 3-P3\/8<br \/>\n  &#8211; [-1, 2, C3k2, [512, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [512, 3, 2]] # 5-P4\/16<br \/>\n  &#8211; [-1, 2, C3k2, [512, True]]<br \/>\n  &#8211; [-1, 1, Conv, [1024, 3, 2]] # 7-P5\/32<br \/>\n  &#8211; [-1, 2, C3k2, [1024, True]]<br \/>\n  &#8211; [-1, 1, SPPF, [1024, 5]] # 9<br \/>\n  &#8211; [-1, 2, C2PSA_iRMB, [1024]] # 10<\/p>\n<p># YOLO11n head<br \/>\nhead:<br \/>\n  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 6], 1, Concat, [1]] # cat backbone P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 13<\/p>\n<p>  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 4], 1, Concat, [1]] # cat backbone P3<br \/>\n  &#8211; [-1, 2, C3k2, [256, False]] # 16 (P3\/8-small)<\/p>\n<p>  &#8211; [-1, 1, Conv, [256, 3, 2]]<br \/>\n  &#8211; [[-1, 13], 1, Concat, [1]] # cat head P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 19 (P4\/16-medium)<\/p>\n<p>  &#8211; [-1, 1, Conv, [512, 3, 2]]<br \/>\n  &#8211; [[-1, 10], 1, Concat, [1]] # cat head P5<br \/>\n  &#8211; [-1, 2, C3k2, [1024, True]] # 22 (P5\/32-large)<\/p>\n<p>  &#8211; [[16, 19, 22], 1, Detect, [nc]] # Detect(P3, P4, P5) <\/p>\n<p>YOLO11_C3k2_iRMB.yaml<\/p>\n<p># Ultralytics YOLO &#x1f680;, AGPL-3.0 license<br \/>\n# YOLO11 object detection model with P3-P5 outputs. For Usage examples see https:\/\/docs.ultralytics.com\/tasks\/detect<\/p>\n<p># Parameters<br \/>\nnc: 80 # number of classes<br \/>\nscales: # model compound scaling constants, i.e. &#039;model&#061;yolo11n.yaml&#039; will call yolo11.yaml with scale &#039;n&#039;<br \/>\n  # [depth, width, max_channels]<br \/>\n  n: [0.50, 0.25, 1024] # summary: 319 layers, 2624080 parameters, 2624064 gradients, 6.6 GFLOPs<br \/>\n  s: [0.50, 0.50, 1024] # summary: 319 layers, 9458752 parameters, 9458736 gradients, 21.7 GFLOPs<br \/>\n  m: [0.50, 1.00, 512] # summary: 409 layers, 20114688 parameters, 20114672 gradients, 68.5 GFLOPs<br \/>\n  l: [1.00, 1.00, 512] # summary: 631 layers, 25372160 parameters, 25372144 gradients, 87.6 GFLOPs<br \/>\n  x: [1.00, 1.50, 512] # summary: 631 layers, 56966176 parameters, 56966160 gradients, 196.0 GFLOPs<\/p>\n<p># YOLO11n backbone<br \/>\nbackbone:<br \/>\n  # [from, repeats, module, args]<br \/>\n  &#8211; [-1, 1, Conv, [64, 3, 2]] # 0-P1\/2<br \/>\n  &#8211; [-1, 1, Conv, [128, 3, 2]] # 1-P2\/4<br \/>\n  &#8211; [-1, 2, C3k2_iRMB, [256, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [256, 3, 2]] # 3-P3\/8<br \/>\n  &#8211; [-1, 2, C3k2_iRMB, [512, False, 0.25]]<br \/>\n  &#8211; [-1, 1, Conv, [512, 3, 2]] # 5-P4\/16<br \/>\n  &#8211; [-1, 2, C3k2_iRMB, [512, True]]<br \/>\n  &#8211; [-1, 1, Conv, [1024, 3, 2]] # 7-P5\/32<br \/>\n  &#8211; [-1, 2, C3k2_iRMB, [1024, True]]<br \/>\n  &#8211; [-1, 1, SPPF, [1024, 5]] # 9<br \/>\n  &#8211; [-1, 2, C2PSA, [1024]] # 10<\/p>\n<p># YOLO11n head<br \/>\nhead:<br \/>\n  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 6], 1, Concat, [1]] # cat backbone P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 13<\/p>\n<p>  &#8211; [-1, 1, nn.Upsample, [None, 2, &#034;nearest&#034;]]<br \/>\n  &#8211; [[-1, 4], 1, Concat, [1]] # cat backbone P3<br \/>\n  &#8211; [-1, 2, C3k2, [256, False]] # 16 (P3\/8-small)<\/p>\n<p>  &#8211; [-1, 1, Conv, [256, 3, 2]]<br \/>\n  &#8211; [[-1, 13], 1, Concat, [1]] # cat head P4<br \/>\n  &#8211; [-1, 2, C3k2, [512, False]] # 19 (P4\/16-medium)<\/p>\n<p>  &#8211; [-1, 1, Conv, [512, 3, 2]]<br \/>\n  &#8211; [[-1, 10], 1, Concat, [1]] # cat head P5<br \/>\n  &#8211; [-1, 2, C3k2, [1024, True]] # 22 (P5\/32-large)<\/p>\n<p>  &#8211; [[16, 19, 22], 1, Detect, [nc]] # Detect(P3, P4, P5) <\/p>\n<h2>7. \u6a21\u578b\u8bad\u7ec3<\/h2>\n<p>import warnings<br \/>\nwarnings.filterwarnings(&#039;ignore&#039;)<br \/>\nfrom ultralytics import YOLO<\/p>\n<p>if __name__ &#061;&#061; &#039;__main__&#039;:<br \/>\n    model &#061; YOLO(&#039;YOLO11_iRMB.yaml&#039;)<br \/>\n    # model.load(&#039;yolo11n.pt&#039;) # loading pretrain weights<br \/>\n    model.train(data&#061;&#039;dataset\/data.yaml&#039;,<br \/>\n                cache&#061;False,<br \/>\n                imgsz&#061;640,<br \/>\n                epochs&#061;300,<br \/>\n                batch&#061;32,<br \/>\n                close_mosaic&#061;0,<br \/>\n                workers&#061;4, # Windows\u4e0b\u51fa\u73b0\u83ab\u540d\u5176\u5999\u5361\u4e3b\u7684\u60c5\u51b5\u53ef\u4ee5\u5c1d\u8bd5\u628aworkers\u8bbe\u7f6e\u4e3a0<br \/>\n                # device&#061;&#039;0&#039;,<br \/>\n                optimizer&#061;&#039;SGD&#039;, # using SGD<br \/>\n                # patience&#061;0, # set 0 to close earlystop.<br \/>\n                # resume&#061;True, # \u65ad\u70b9\u7eed\u8bad,YOLO\u521d\u59cb\u5316\u65f6\u9009\u62e9last.pt<br \/>\n                # amp&#061;False, # close amp<br \/>\n                # fraction&#061;0.2,<br \/>\n                project&#061;&#039;runs\/train&#039;,<br \/>\n                name&#061;&#039;exp&#039;,<br \/>\n                ) <\/p>\n","protected":false},"excerpt":{"rendered":"<p>iRMB\u5012\u7f6e\u6b8b\u5dee\u5757\u6ce8\u610f\u529b\u673a\u5236\u7b80\u4ecb<br \/>\niRMB&#xff08;Inverted Residual Mobile Block&#xff09;\u7684\u6846\u67b6\u539f\u7406&#xff0c;\u662f\u4e00\u79cd\u7ed3\u5408\u8f7b\u91cf\u7ea7CNN\u548c\u6ce8\u610f\u529b\u673a\u5236\u7684\u65b9\u6cd5&#xff0c;\u7528\u4e8e\u6539\u8fdb\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u76ee\u6807\u68c0\u6d4b\u6a21\u578b\u3002IRMB\u901a\u8fc7\u5012\u7f6e\u6b8b\u5dee\u5757\u548c\u5143\u79fb\u52a8\u5757\u5b9e\u73b0\u9ad8\u6548\u4fe1\u606f\u5904\u7406&#xff0c;\u540c\u65f6\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u5316\u3002\u672c\u6587\u4e2d\u63d0\u51fa\u4e00\u4e2a\u65b0\u7684\u4e3b\u5e72\u7f51\u7edcEMO&#xff0c;\u4e3b\u8981\u601d\u60f3\u662f\u5c06\u8f7b\u91cf\u7ea7\u7684CNN\u67b6\u6784\u4e0e\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u7ed3\u6784\u76f8\u7ed3\u5408\u3002<br \/>\n\u76ee\u5f55<br \/>\n1. \u7b80\u4ecb<br \/>\n2. iR<\/p>\n","protected":false},"author":2,"featured_media":61420,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1],"tags":[6492,152,156,50,1665,1326,523],"topic":[],"class_list":["post-61423","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-server","tag-irmb","tag-pytorch","tag-yolo","tag-50","tag-1665","tag-1326","tag-523"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.3 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/www.wsisp.com\/helps\/61423.html\" \/>\n<meta property=\"og:locale\" content=\"zh_CN\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"og:description\" content=\"iRMB\u5012\u7f6e\u6b8b\u5dee\u5757\u6ce8\u610f\u529b\u673a\u5236\u7b80\u4ecb iRMB&#xff08;Inverted Residual Mobile Block&#xff09;\u7684\u6846\u67b6\u539f\u7406&#xff0c;\u662f\u4e00\u79cd\u7ed3\u5408\u8f7b\u91cf\u7ea7CNN\u548c\u6ce8\u610f\u529b\u673a\u5236\u7684\u65b9\u6cd5&#xff0c;\u7528\u4e8e\u6539\u8fdb\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u76ee\u6807\u68c0\u6d4b\u6a21\u578b\u3002IRMB\u901a\u8fc7\u5012\u7f6e\u6b8b\u5dee\u5757\u548c\u5143\u79fb\u52a8\u5757\u5b9e\u73b0\u9ad8\u6548\u4fe1\u606f\u5904\u7406&#xff0c;\u540c\u65f6\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u5316\u3002\u672c\u6587\u4e2d\u63d0\u51fa\u4e00\u4e2a\u65b0\u7684\u4e3b\u5e72\u7f51\u7edcEMO&#xff0c;\u4e3b\u8981\u601d\u60f3\u662f\u5c06\u8f7b\u91cf\u7ea7\u7684CNN\u67b6\u6784\u4e0e\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u7ed3\u6784\u76f8\u7ed3\u5408\u3002 \u76ee\u5f55 1. \u7b80\u4ecb 2. iR\" \/>\n<meta property=\"og:url\" content=\"https:\/\/www.wsisp.com\/helps\/61423.html\" \/>\n<meta property=\"og:site_name\" content=\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\" \/>\n<meta property=\"article:published_time\" content=\"2026-01-17T07:50:04+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/www.wsisp.com\/helps\/wp-content\/uploads\/2026\/01\/20260117075002-696b3f2ad59d4.png\" \/>\n<meta name=\"author\" content=\"admin\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"\u4f5c\u8005\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin\" \/>\n\t<meta name=\"twitter:label2\" content=\"\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4\" \/>\n\t<meta name=\"twitter:data2\" content=\"16 \u5206\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/61423.html\",\"url\":\"https:\/\/www.wsisp.com\/helps\/61423.html\",\"name\":\"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"isPartOf\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\"},\"datePublished\":\"2026-01-17T07:50:04+00:00\",\"dateModified\":\"2026-01-17T07:50:04+00:00\",\"author\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\"},\"breadcrumb\":{\"@id\":\"https:\/\/www.wsisp.com\/helps\/61423.html#breadcrumb\"},\"inLanguage\":\"zh-Hans\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/www.wsisp.com\/helps\/61423.html\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/61423.html#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"\u9996\u9875\",\"item\":\"https:\/\/www.wsisp.com\/helps\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#website\",\"url\":\"https:\/\/www.wsisp.com\/helps\/\",\"name\":\"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3\",\"description\":\"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"zh-Hans\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41\",\"name\":\"admin\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"zh-Hans\",\"@id\":\"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"contentUrl\":\"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery\",\"caption\":\"admin\"},\"sameAs\":[\"http:\/\/wp.wsisp.com\"],\"url\":\"https:\/\/www.wsisp.com\/helps\/author\/admin\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/www.wsisp.com\/helps\/61423.html","og_locale":"zh_CN","og_type":"article","og_title":"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","og_description":"iRMB\u5012\u7f6e\u6b8b\u5dee\u5757\u6ce8\u610f\u529b\u673a\u5236\u7b80\u4ecb iRMB&#xff08;Inverted Residual Mobile Block&#xff09;\u7684\u6846\u67b6\u539f\u7406&#xff0c;\u662f\u4e00\u79cd\u7ed3\u5408\u8f7b\u91cf\u7ea7CNN\u548c\u6ce8\u610f\u529b\u673a\u5236\u7684\u65b9\u6cd5&#xff0c;\u7528\u4e8e\u6539\u8fdb\u79fb\u52a8\u8bbe\u5907\u4e0a\u7684\u76ee\u6807\u68c0\u6d4b\u6a21\u578b\u3002IRMB\u901a\u8fc7\u5012\u7f6e\u6b8b\u5dee\u5757\u548c\u5143\u79fb\u52a8\u5757\u5b9e\u73b0\u9ad8\u6548\u4fe1\u606f\u5904\u7406&#xff0c;\u540c\u65f6\u4fdd\u6301\u6a21\u578b\u8f7b\u91cf\u5316\u3002\u672c\u6587\u4e2d\u63d0\u51fa\u4e00\u4e2a\u65b0\u7684\u4e3b\u5e72\u7f51\u7edcEMO&#xff0c;\u4e3b\u8981\u601d\u60f3\u662f\u5c06\u8f7b\u91cf\u7ea7\u7684CNN\u67b6\u6784\u4e0e\u57fa\u4e8e\u6ce8\u610f\u529b\u7684\u6a21\u578b\u7ed3\u6784\u76f8\u7ed3\u5408\u3002 \u76ee\u5f55 1. \u7b80\u4ecb 2. iR","og_url":"https:\/\/www.wsisp.com\/helps\/61423.html","og_site_name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","article_published_time":"2026-01-17T07:50:04+00:00","og_image":[{"url":"https:\/\/www.wsisp.com\/helps\/wp-content\/uploads\/2026\/01\/20260117075002-696b3f2ad59d4.png"}],"author":"admin","twitter_card":"summary_large_image","twitter_misc":{"\u4f5c\u8005":"admin","\u9884\u8ba1\u9605\u8bfb\u65f6\u95f4":"16 \u5206"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/www.wsisp.com\/helps\/61423.html","url":"https:\/\/www.wsisp.com\/helps\/61423.html","name":"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB - \u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","isPartOf":{"@id":"https:\/\/www.wsisp.com\/helps\/#website"},"datePublished":"2026-01-17T07:50:04+00:00","dateModified":"2026-01-17T07:50:04+00:00","author":{"@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41"},"breadcrumb":{"@id":"https:\/\/www.wsisp.com\/helps\/61423.html#breadcrumb"},"inLanguage":"zh-Hans","potentialAction":[{"@type":"ReadAction","target":["https:\/\/www.wsisp.com\/helps\/61423.html"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/www.wsisp.com\/helps\/61423.html#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"\u9996\u9875","item":"https:\/\/www.wsisp.com\/helps"},{"@type":"ListItem","position":2,"name":"YOLO\u7cfb\u5217\u7b97\u6cd5\u6539\u8fdb | \u8f7b\u91cf\u5316iRMB\u6ce8\u610f\u529b\u673a\u5236 | \u542bC3k2_iRMB\u4e0eC2PSA_iRMB"}]},{"@type":"WebSite","@id":"https:\/\/www.wsisp.com\/helps\/#website","url":"https:\/\/www.wsisp.com\/helps\/","name":"\u7f51\u7855\u4e92\u8054\u5e2e\u52a9\u4e2d\u5fc3","description":"\u9999\u6e2f\u670d\u52a1\u5668_\u9999\u6e2f\u4e91\u670d\u52a1\u5668\u8d44\u8baf_\u670d\u52a1\u5668\u5e2e\u52a9\u6587\u6863_\u670d\u52a1\u5668\u6559\u7a0b","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/www.wsisp.com\/helps\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"zh-Hans"},{"@type":"Person","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/358e386c577a3ab51c4493330a20ad41","name":"admin","image":{"@type":"ImageObject","inLanguage":"zh-Hans","@id":"https:\/\/www.wsisp.com\/helps\/#\/schema\/person\/image\/","url":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","contentUrl":"https:\/\/gravatar.wp-china-yes.net\/avatar\/?s=96&d=mystery","caption":"admin"},"sameAs":["http:\/\/wp.wsisp.com"],"url":"https:\/\/www.wsisp.com\/helps\/author\/admin"}]}},"_links":{"self":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/61423","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/comments?post=61423"}],"version-history":[{"count":0,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/posts\/61423\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/media\/61420"}],"wp:attachment":[{"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/media?parent=61423"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/categories?post=61423"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/tags?post=61423"},{"taxonomy":"topic","embeddable":true,"href":"https:\/\/www.wsisp.com\/helps\/wp-json\/wp\/v2\/topic?post=61423"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}