Search alternatives:
significant linear » significant clinical (Expand Search), significant gender (Expand Search), significant level (Expand Search)
greatest decrease » treatment decreased (Expand Search), greater increase (Expand Search)
linear decrease » linear increase (Expand Search)
significant linear » significant clinical (Expand Search), significant gender (Expand Search), significant level (Expand Search)
greatest decrease » treatment decreased (Expand Search), greater increase (Expand Search)
linear decrease » linear increase (Expand Search)
-
2281
The structure of attention gate block [31].
Published 2025“…The actual accuracy and mean latency time of the model were 92.43% and 260ms, respectively. …”
-
2282
DSC block and its application network structure.
Published 2025“…The actual accuracy and mean latency time of the model were 92.43% and 260ms, respectively. …”
-
2283
The structure of multi-scale residual block [30].
Published 2025“…The actual accuracy and mean latency time of the model were 92.43% and 260ms, respectively. …”
-
2284
The structure of IRAU and Res2Net+ block [22].
Published 2025“…The actual accuracy and mean latency time of the model were 92.43% and 260ms, respectively. …”
-
2285
-
2286
-
2287
Prediction of transition readiness.
Published 2025“…In most transition domains, help needed did not decrease with age and was not affected by function. …”
-
2288
Dataset visualization diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2289
Dataset sample images.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2290
Performance comparison of different models.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2291
C2f and BC2f module structure diagrams.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2292
YOLOv8n detection results diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2293
YOLOv8n-BWG model structure diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2294
BiFormer structure diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2295
YOLOv8n-BWG detection results diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2296
GSConv module structure diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2297
Performance comparison of three loss functions.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2298
mAP0.5 Curves of various models.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2299
Network loss function change diagram.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”
-
2300
Comparative diagrams of different indicators.
Published 2025“…Results on a specialized dataset reveal that YOLOv8n-BWG outperforms YOLOv8n by increasing the mean Average Precision (mAP) by 4.2%, boosting recognition speed by 21.3% per second, and decreasing both the number of floating-point operations (FLOPs) by 28.9% and model size by 26.3%. …”