Search alternatives:
from implementing » after implementing (Expand Search), _ implementing (Expand Search)
from implementing » after implementing (Expand Search), _ implementing (Expand Search)
-
121
-
122
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
-
123
The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation"
Published 2025“…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
-
124
Evaluation and Statistical Analysis Code for "Multi-Task Learning for Joint Fisheye Compression and Perception for Autonomous Driving"
Published 2025“…</li></ul><p dir="ltr">These scripts are implemented in Python using the PyTorch framework and are provided to ensure the reproducibility of the experimental results presented in the manuscript.…”
-
125
Monte Carlo Simulation Code for Evaluating Cognitive Biases in Penalty Shootouts Using ABAB and ABBA Formats
Published 2024“…<p dir="ltr">This Python code implements a Monte Carlo simulation to evaluate the impact of cognitive biases on penalty shootouts under two formats: ABAB (alternating shots) and ABBA (similar to tennis tiebreak format). …”
-
126
<b>Code and derived data for</b><b>Training Sample Location Matters: Accuracy Impacts in LULC Classification</b>
Published 2025“…</li><li>Python/Kaggle notebooks (<code>.ipynb</code>): reproducibility pipeline for accuracy metrics and statistical analysis.…”
-
127
<b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b>
Published 2025“…<p dir="ltr"><b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b></p><p dir="ltr">The code was developed in the Google Collaboratory environment, using Python version 3.7.13, with TensorFlow 2.8.2. …”
-
128
Data and some code used in the paper:<b>Expansion quantization network: A micro-emotion detection and annotation framework</b>
Published 2025“…</p><p dir="ltr">GPU:NVIDIA GeForce RTX 3090 GPU</p><p dir="ltr">Bert-base-cased pre-trained model: https://huggingface.co/google-bert/bert-base-cased</p><p dir="ltr">python=3.7,pytorch=1.9.0,cudatoolkit=11.3.1,cudnn=8.9.7.29.…”
-
129
Workflow of a typical Epydemix run.
Published 2025“…By lowering the barrier for the implementation of computational and inference approaches, Epydemix makes epidemic modeling more accessible to a wider range of users, from academic researchers to public health professionals.…”
-
130
Number of tweets collected over time.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
131
Descriptive measures of the dataset.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
132
Media information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
133
Table of the database statistical measures.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
134
Tweets information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
135
Examples of tweets texts (Portuguese).
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
136
Methodological flowchart.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
137
Number of tweets collected per query and type.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
138
Examples of tweets texts (English).
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
139
Users information.
Published 2025“…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
-
140
Data Sheet 1_COCαDA - a fast and scalable algorithm for interatomic contact detection in proteins using Cα distance matrices.pdf
Published 2025“…Here, we introduce COCαDA (COntact search pruning by Cα Distance Analysis), a Python-based command-line tool for improving search pruning in large-scale interatomic protein contact analysis using alpha-carbon (Cα) distance matrices. …”