Search alternatives:
from implementing » after implementing (Expand Search), _ implementing (Expand Search)
Showing 121 - 140 results of 238 for search '(( ((python tool) OR (python code)) implementation ) OR ( python from implementing ))', query time: 0.33s Refine Results
  1. 121
  2. 122

    The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation" by FirstName LastName (20554465)

    Published 2025
    “…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
  3. 123

    The codes and data for "A Graph Convolutional Neural Network-based Method for Predicting Computational Intensity of Geocomputation" by FirstName LastName (20554465)

    Published 2025
    “…The model results are saved in <code>1point2dem/SampleGeneration/result</code>, and the results for <b>Table 3</b> in the paper are derived from this output.…”
  4. 124

    Evaluation and Statistical Analysis Code for "Multi-Task Learning for Joint Fisheye Compression and Perception for Autonomous Driving" by Basem Ahmed (18127861)

    Published 2025
    “…</li></ul><p dir="ltr">These scripts are implemented in Python using the PyTorch framework and are provided to ensure the reproducibility of the experimental results presented in the manuscript.…”
  5. 125

    Monte Carlo Simulation Code for Evaluating Cognitive Biases in Penalty Shootouts Using ABAB and ABBA Formats by Raul MATSUSHITA (10276562)

    Published 2024
    “…<p dir="ltr">This Python code implements a Monte Carlo simulation to evaluate the impact of cognitive biases on penalty shootouts under two formats: ABAB (alternating shots) and ABBA (similar to tennis tiebreak format). …”
  6. 126

    <b>Code and derived data for</b><b>Training Sample Location Matters: Accuracy Impacts in LULC Classification</b> by Pajtim Zariqi (22155799)

    Published 2025
    “…</li><li>Python/Kaggle notebooks (<code>.ipynb</code>): reproducibility pipeline for accuracy metrics and statistical analysis.…”
  7. 127

    <b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b> by Marly G F Costa (19812192)

    Published 2025
    “…<p dir="ltr"><b>Use case codes of the DDS3 and DDS4 datasets for bacillus segmentation and tuberculosis diagnosis, respectively</b></p><p dir="ltr">The code was developed in the Google Collaboratory environment, using Python version 3.7.13, with TensorFlow 2.8.2. …”
  8. 128

    Data and some code used in the paper:<b>Expansion quantization network: A micro-emotion detection and annotation framework</b> by Zhou (20184816)

    Published 2025
    “…</p><p dir="ltr">GPU:NVIDIA GeForce RTX 3090 GPU</p><p dir="ltr">Bert-base-cased pre-trained model: https://huggingface.co/google-bert/bert-base-cased</p><p dir="ltr">python=3.7,pytorch=1.9.0,cudatoolkit=11.3.1,cudnn=8.9.7.29.…”
  9. 129

    Workflow of a typical Epydemix run. by Nicolò Gozzi (8837522)

    Published 2025
    “…By lowering the barrier for the implementation of computational and inference approaches, Epydemix makes epidemic modeling more accessible to a wider range of users, from academic researchers to public health professionals.…”
  10. 130

    Number of tweets collected over time. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  11. 131

    Descriptive measures of the dataset. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  12. 132

    Media information. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  13. 133

    Table of the database statistical measures. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  14. 134

    Tweets information. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  15. 135

    Examples of tweets texts (Portuguese). by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  16. 136

    Methodological flowchart. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  17. 137

    Number of tweets collected per query and type. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  18. 138

    Examples of tweets texts (English). by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  19. 139

    Users information. by Sylvia Iasulaitis (8301189)

    Published 2025
    “…The process of collecting and creating the database for this study went through three major stages, subdivided into several processes: (1) A preliminary analysis of the platform and its operation; (2) Contextual analysis, creation of the conceptual model, and definition of Keywords and (3) Implementation of the Data Collection Strategy. Python algorithms were developed to model each primary collection type. …”
  20. 140

    Data Sheet 1_COCαDA - a fast and scalable algorithm for interatomic contact detection in proteins using Cα distance matrices.pdf by Rafael Pereira Lemos (9104911)

    Published 2025
    “…Here, we introduce COCαDA (COntact search pruning by Cα Distance Analysis), a Python-based command-line tool for improving search pruning in large-scale interatomic protein contact analysis using alpha-carbon (Cα) distance matrices. …”