Search alternatives:
code implementing » model implementing (Expand Search), consider implementing (Expand Search), _ implementing (Expand Search)
python model » python tool (Expand Search), action model (Expand Search), motion model (Expand Search)
representing » represent (Expand Search), represents (Expand Search)
code implementing » model implementing (Expand Search), consider implementing (Expand Search), _ implementing (Expand Search)
python model » python tool (Expand Search), action model (Expand Search), motion model (Expand Search)
representing » represent (Expand Search), represents (Expand Search)
-
101
Code for High-quality Human Activity Intensity Maps in China from 2000-2020
Published 2025“…<p dir="ltr">Code and remote sensing images and interpretation results of the samples for uncertainty analysis for "High-quality Human Activity Intensity Maps in China from 2000-2020"</p><p dir="ltr">“Mapping_HAI.py”:We generated the HAI maps using ArcGIS 10.8, and the geoprocessing tasks were implemented using Python 2.7 with the ArcPy library (ArcGIS 10.8 + Python 2.7 environment). …”
-
102
The codes and data for "Lane Extraction from Trajectories at Road Intersections Based on Graph Transformer Network"
Published 2024“…Each lane includes 'geometry' and 'inter_id' attributes.</li></ul><h2>Codes</h2><p dir="ltr">This repository contains the following Python codes:</p><ul><li>`data_processing.py`: Contains the implementation of data processing and feature extraction. …”
-
103
MATH_code : False Data Injection Attack Detection in Smart Grids based on Reservoir Computing
Published 2025“…</li><li><b>3_literature_analysis_and_mapping.ipynb</b><br>Contains the Python code used for executing the systematic mapping study (SMS), including automated processing of literature data and thematic clustering.…”
-
104
-
105
Evaluation and Statistical Analysis Code for "Multi-Task Learning for Joint Fisheye Compression and Perception for Autonomous Driving"
Published 2025“…</li></ul><p dir="ltr">These scripts are implemented in Python using the PyTorch framework and are provided to ensure the reproducibility of the experimental results presented in the manuscript.…”
-
106
Monte Carlo Simulation Code for Evaluating Cognitive Biases in Penalty Shootouts Using ABAB and ABBA Formats
Published 2024“…<p dir="ltr">This Python code implements a Monte Carlo simulation to evaluate the impact of cognitive biases on penalty shootouts under two formats: ABAB (alternating shots) and ABBA (similar to tennis tiebreak format). …”
-
107
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The <b>innovations</b> and <b>steps</b> in Case 3, including data download, sample generation, and parallel computation optimization, were independently developed and are not dependent on the GeoCube’s code.</p><h2>Requirements</h2><p dir="ltr">The codes use the following dependencies with Python 3.8</p><ul><li>torch==2.0.0</li><li>torch_geometric==2.5.3</li><li>networkx==2.6.3</li><li>pyshp==2.3.1</li><li>tensorrt==8.6.1</li><li>matplotlib==3.7.2</li><li>scipy==1.10.1</li><li>scikit-learn==1.3.0</li><li>geopandas==0.13.2</li></ul><p><br></p>…”
-
108
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The <b>innovations</b> and <b>steps</b> in Case 3, including data download, sample generation, and parallel computation optimization, were independently developed and are not dependent on the GeoCube’s code.</p><h2>Requirements</h2><p dir="ltr">The codes use the following dependencies with Python 3.8</p><ul><li>torch==2.0.0</li><li>torch_geometric==2.5.3</li><li>networkx==2.6.3</li><li>pyshp==2.3.1</li><li>tensorrt==8.6.1</li><li>matplotlib==3.7.2</li><li>scipy==1.10.1</li><li>scikit-learn==1.3.0</li><li>geopandas==0.13.2</li></ul><p><br></p>…”
-
109
Data features examined for potential biases.
Published 2025“…Representativeness of the population, differences in calibration and model performance among groups, and differences in performance across hospital settings were identified as possible sources of bias.…”
-
110
Analysis topics.
Published 2025“…Representativeness of the population, differences in calibration and model performance among groups, and differences in performance across hospital settings were identified as possible sources of bias.…”
-
111
<b>Code and derived data for</b><b>Training Sample Location Matters: Accuracy Impacts in LULC Classification</b>
Published 2025“…</li><li>Python/Kaggle notebooks (<code>.ipynb</code>): reproducibility pipeline for accuracy metrics and statistical analysis.…”
-
112
<b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b>
Published 2025“…<p dir="ltr"><b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b></p><p dir="ltr">The code was developed in the Google Collaboratory environment, using Python version 3.7.13, with TensorFlow 2.8.2. …”
-
113
JASPEX model
Published 2025“…</p><p dir="ltr">We wrote new sets of python codes and developed python programming codes to rework on the map to generate the coloured map of Southwest Nigeria from the map of Nigeria (which represented the region of our study). …”
-
114
Datasets To EVAL.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
115
Statistical significance test results.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
116
How RAG work.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
117
OpenBookQA experimental results.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
118
AI2_ARC experimental results.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
119
TQA experimental results.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”
-
120
E-EVAL experimental results.
Published 2025“…We evaluated our proposed system on five educational datasets—AI2_ARC, OpenBookQA, E-EVAL, TQA, and ScienceQA—which represent diverse question types and domains. Compared to vanilla Large Language Models (LLMs), our approach combining Retrieval-Augmented Generation (RAG) with Code Interpreters achieved an average accuracy improvement of 10−15 percentage points. …”