Search alternatives:
code implementing » model implementing (Expand Search), consider implementing (Expand Search), _ implementing (Expand Search)
from implementing » after implementing (Expand Search), _ implementing (Expand Search)
code implementing » model implementing (Expand Search), consider implementing (Expand Search), _ implementing (Expand Search)
from implementing » after implementing (Expand Search), _ implementing (Expand Search)
-
101
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
-
102
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
-
103
Evaluation and Statistical Analysis Code for "Multi-Task Learning for Joint Fisheye Compression and Perception for Autonomous Driving"
Published 2025“…</li></ul><p dir="ltr">These scripts are implemented in Python using the PyTorch framework and are provided to ensure the reproducibility of the experimental results presented in the manuscript.…”
-
104
Monte Carlo Simulation Code for Evaluating Cognitive Biases in Penalty Shootouts Using ABAB and ABBA Formats
Published 2024“…<p dir="ltr">This Python code implements a Monte Carlo simulation to evaluate the impact of cognitive biases on penalty shootouts under two formats: ABAB (alternating shots) and ABBA (similar to tennis tiebreak format). …”
-
105
<b>Code and derived data for</b><b>Training Sample Location Matters: Accuracy Impacts in LULC Classification</b>
Published 2025“…</li><li>Python/Kaggle notebooks (<code>.ipynb</code>): reproducibility pipeline for accuracy metrics and statistical analysis.…”
-
106
<b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b>
Published 2025“…<p dir="ltr"><b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b></p><p dir="ltr">The code was developed in the Google Collaboratory environment, using Python version 3.7.13, with TensorFlow 2.8.2. …”
-
107
Workflow of a typical Epydemix run.
Published 2025“…By lowering the barrier for the implementation of computational and inference approaches, Epydemix makes epidemic modeling more accessible to a wider range of users, from academic researchers to public health professionals.…”
-
108
Data and some code used in the paper:<b>Expansion quantization network: A micro-emotion detection and annotation framework</b>
Published 2025“…</p><p dir="ltr">GPU:NVIDIA GeForce RTX 3090 GPU</p><p dir="ltr">Bert-base-cased pre-trained model: https://huggingface.co/google-bert/bert-base-cased</p><p dir="ltr">python=3.7,pytorch=1.9.0,cudatoolkit=11.3.1,cudnn=8.9.7.29.…”
-
109
Number of tweets collected over time.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
110
Descriptive measures of the dataset.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
111
Media information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
112
Table of the database statistical measures.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
113
Tweets information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
114
Examples of tweets texts (Portuguese).
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
115
Methodological flowchart.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
116
Number of tweets collected per query and type.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
117
Examples of tweets texts (English).
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
118
Users information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
119
High-Throughput Mass Spectral Library Searching of Small Molecules in R with NIST MSPepSearch
Published 2025“…Despite the availability of numerous library search algorithms, those developed by NIST and implemented in MS Search remain predominant, partly because commercial databases (e.g., NIST, Wiley) are distributed in proprietary formats inaccessible to custom code. …”
-
120
Comparison data 7 for <i>Lamprologus ocellatus</i>.
Published 2024“…TIBA accepts data outputs from popular logging software and is implemented in Python and JavaScript, with all current browsers supported. …”