Showing 10,621 - 10,640 results of 21,342 for search '(( significant ((teer decrease) OR (mean decrease)) ) OR ( significant decrease decrease ))', query time: 0.53s Refine Results
  1. 10621

    The schematic diagram of free-field boundary. by Jie Zhao (49409)

    Published 2025
    “…Results demonstrate that unreinforced foundations exhibit systematic residual deformation due to liquefaction-induced sand flow, which is significantly reduced by gravel pile reinforcement. Both excess pore water pressure and pore pressure ratio decrease markedly after reinforcement. …”
  2. 10622

    Residual deformation parameters of soil. by Jie Zhao (49409)

    Published 2025
    “…Results demonstrate that unreinforced foundations exhibit systematic residual deformation due to liquefaction-induced sand flow, which is significantly reduced by gravel pile reinforcement. Both excess pore water pressure and pore pressure ratio decrease markedly after reinforcement. …”
  3. 10623

    Soil fluid and liquefaction parameters. by Jie Zhao (49409)

    Published 2025
    “…Results demonstrate that unreinforced foundations exhibit systematic residual deformation due to liquefaction-induced sand flow, which is significantly reduced by gravel pile reinforcement. Both excess pore water pressure and pore pressure ratio decrease markedly after reinforcement. …”
  4. 10624

    PL-Finn model development procedure diagram. by Jie Zhao (49409)

    Published 2025
    “…Results demonstrate that unreinforced foundations exhibit systematic residual deformation due to liquefaction-induced sand flow, which is significantly reduced by gravel pile reinforcement. Both excess pore water pressure and pore pressure ratio decrease markedly after reinforcement. …”
  5. 10625

    Key modeling details for CoVPF and controls. by Zhong-yi Lei (22552944)

    Published 2025
    “…Furthermore, we found that accounting for epistasis was critical, as ignoring epistasis led to a 43% decrease in forecasting accuracy. Case studies showed that CoVPF delivered more accurate and timely forecasts for lineage expansions and inflections such as EG.5.1 and XBB.1.5. …”
  6. 10626

    LSTM model. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  7. 10627

    CNN model. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  8. 10628

    Ceramic bearings. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  9. 10629

    Geometric contact arc length model. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  10. 10630

    Indentation fracture mechanics model. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  11. 10631

    Grinding particle cutting machining model. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  12. 10632

    Three stages of abrasive cutting process. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  13. 10633

    CNN-LSTM action recognition process. by Longfei Gao (698900)

    Published 2025
    “…According to the experimental results, when the grinding depth increases to 21 μm, the average training loss of the model further decreases to 0.03622, and the surface roughness Ra value significantly decreases to 0.1624 μm. …”
  14. 10634
  15. 10635

    Overall model framework. by Ke Yan (331581)

    Published 2024
    “…The results show that: (1) From the experimental data of word sense disambiguation, the accuracy of the SMOSS-LSTM model proposed in this paper is the lowest when the context range is "3+3", then it rises in turn at "5+5" and "7+7", reaches the highest at "7+7", and then begins to decrease at "10+10"; (2) Compared with the control group, the accuracy of syntactic analysis in the experimental group reached 89.5%, while that in the control group was only 73.2%. (3) In the aspect of English text error detection, the detection accuracy of the proposed model in the experimental group is as high as 94.8%, which is significantly better than the traditional SMOSS-based text error detection method, and its accuracy is only 68.3%. (4) Compared with other existing researches, although it is slightly inferior to Bidirectional Encoder Representations from Transformers (BERT) in word sense disambiguation, this proposed model performs well in syntactic analysis and English text error detection, and its comprehensive performance is excellent. …”
  16. 10636

    Key parameters of LSTM training model. by Ke Yan (331581)

    Published 2024
    “…The results show that: (1) From the experimental data of word sense disambiguation, the accuracy of the SMOSS-LSTM model proposed in this paper is the lowest when the context range is "3+3", then it rises in turn at "5+5" and "7+7", reaches the highest at "7+7", and then begins to decrease at "10+10"; (2) Compared with the control group, the accuracy of syntactic analysis in the experimental group reached 89.5%, while that in the control group was only 73.2%. (3) In the aspect of English text error detection, the detection accuracy of the proposed model in the experimental group is as high as 94.8%, which is significantly better than the traditional SMOSS-based text error detection method, and its accuracy is only 68.3%. (4) Compared with other existing researches, although it is slightly inferior to Bidirectional Encoder Representations from Transformers (BERT) in word sense disambiguation, this proposed model performs well in syntactic analysis and English text error detection, and its comprehensive performance is excellent. …”
  17. 10637
  18. 10638

    Comparison chart of model evaluation results. by Ke Yan (331581)

    Published 2024
    “…The results show that: (1) From the experimental data of word sense disambiguation, the accuracy of the SMOSS-LSTM model proposed in this paper is the lowest when the context range is "3+3", then it rises in turn at "5+5" and "7+7", reaches the highest at "7+7", and then begins to decrease at "10+10"; (2) Compared with the control group, the accuracy of syntactic analysis in the experimental group reached 89.5%, while that in the control group was only 73.2%. (3) In the aspect of English text error detection, the detection accuracy of the proposed model in the experimental group is as high as 94.8%, which is significantly better than the traditional SMOSS-based text error detection method, and its accuracy is only 68.3%. (4) Compared with other existing researches, although it is slightly inferior to Bidirectional Encoder Representations from Transformers (BERT) in word sense disambiguation, this proposed model performs well in syntactic analysis and English text error detection, and its comprehensive performance is excellent. …”
  19. 10639

    Model performance evaluation results. by Ke Yan (331581)

    Published 2024
    “…The results show that: (1) From the experimental data of word sense disambiguation, the accuracy of the SMOSS-LSTM model proposed in this paper is the lowest when the context range is "3+3", then it rises in turn at "5+5" and "7+7", reaches the highest at "7+7", and then begins to decrease at "10+10"; (2) Compared with the control group, the accuracy of syntactic analysis in the experimental group reached 89.5%, while that in the control group was only 73.2%. (3) In the aspect of English text error detection, the detection accuracy of the proposed model in the experimental group is as high as 94.8%, which is significantly better than the traditional SMOSS-based text error detection method, and its accuracy is only 68.3%. (4) Compared with other existing researches, although it is slightly inferior to Bidirectional Encoder Representations from Transformers (BERT) in word sense disambiguation, this proposed model performs well in syntactic analysis and English text error detection, and its comprehensive performance is excellent. …”
  20. 10640

    The result compared with other existing methods. by Ke Yan (331581)

    Published 2024
    “…The results show that: (1) From the experimental data of word sense disambiguation, the accuracy of the SMOSS-LSTM model proposed in this paper is the lowest when the context range is "3+3", then it rises in turn at "5+5" and "7+7", reaches the highest at "7+7", and then begins to decrease at "10+10"; (2) Compared with the control group, the accuracy of syntactic analysis in the experimental group reached 89.5%, while that in the control group was only 73.2%. (3) In the aspect of English text error detection, the detection accuracy of the proposed model in the experimental group is as high as 94.8%, which is significantly better than the traditional SMOSS-based text error detection method, and its accuracy is only 68.3%. (4) Compared with other existing researches, although it is slightly inferior to Bidirectional Encoder Representations from Transformers (BERT) in word sense disambiguation, this proposed model performs well in syntactic analysis and English text error detection, and its comprehensive performance is excellent. …”