يعرض 61 - 80 نتائج من 748 نتيجة بحث عن '(( ((python models) OR (python code)) representing ) OR ( python tool implementation ))', وقت الاستعلام: 0.64s تنقيح النتائج
  1. 61

    A high-performance and highly reusable fast multipole method library and its application to solvation energy calculations at virus-scale حسب Tingyu Wang (12342757)

    منشور في 2022
    "…It empowers researchers to compute solvation energy at the scale of viruses interactively via easy-to-use Python interfaces. The 1/N convergence rate observed in mesh-refinement studies confirms code correctness, and result comparison with other trusted PB software shows agreement using a wide range of proteins.Performance results report timings and breakdowns with between 8,000 and 2 million boundary elements and confirm the linear complexity in both time and space. …"
  2. 62
  3. 63

    Datasets To EVAL. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  4. 64

    Statistical significance test results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  5. 65

    How RAG work. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  6. 66

    OpenBookQA experimental results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  7. 67

    AI2_ARC experimental results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  8. 68

    TQA experimental results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  9. 69

    E-EVAL experimental results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  10. 70

    TQA Accuracy Comparison Chart on different LLM. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  11. 71

    ScienceQA experimental results. حسب Jin Lu (428513)

    منشور في 2025
    "…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …"
  12. 72

    Example of model modularity. حسب Adam Husar (18437351)

    منشور في 2024
    الموضوعات:
  13. 73
  14. 74
  15. 75
  16. 76
  17. 77
  18. 78
  19. 79

    Local Python Code Protector Script: A Tool for Source Code Protection and Secure Code Sharing حسب Pavel Izosimov (20096259)

    منشور في 2024
    "…<p dir="ltr">Local Python Code Protector Script: Advanced Tool for Python Code Protection and Secure Sharing</p><h2>Introduction</h2><p dir="ltr">The <b>Local Python Code Protector Script</b> is a powerful command-line tool designed to provide <a href="https://xn--mxac.net/local-python-code-protector.html" target="_blank"><b>source code protection</b></a> and <a href="https://xn--mxac.net/local-python-code-protector.html" target="_blank"><b>secure code sharing</b></a> for Python scripts. …"
  20. 80

    Validation of model of the SNARE complex. حسب Adam Husar (18437351)

    منشور في 2024
    الموضوعات: