Showing 1 - 15 results of 15 for search '(( binary data driven optimization algorithm ) OR ( library based gpu optimization algorithm ))', query time: 0.56s Refine Results
  1. 1
  2. 2

    Event-driven data flow processing. by Yixian Wen (12201388)

    Published 2025
    “…Subsequently, we implement an optimal binary tree decision-making algorithm, grounded in dynamic programming, to achieve precise allocation of elastic resources within data streams, significantly bolstering resource utilization. …”
  3. 3
  4. 4

    Flow diagram of the proposed model. by Uğur Ejder (22683228)

    Published 2025
    “…<div><p>Machine learning models are increasingly applied to assisted reproductive technologies (ART), yet most studies rely on conventional algorithms with limited optimization. This proof-of-concept study investigates whether a hybrid Logistic Regression–Artificial Bee Colony (LR–ABC) framework can enhance predictive performance in in vitro fertilization (IVF) outcomes while producing interpretable, hypothesis-driven associations with nutritional and pharmaceutical supplement use. …”
  5. 5
  6. 6

    Data_Sheet_1_Fast Simulation of a Multi-Area Spiking Network Model of Macaque Cortex on an MPI-GPU Cluster.PDF by Gianmarco Tiddia (10824118)

    Published 2022
    “…NEST GPU is a GPU library written in CUDA-C/C++ for large-scale simulations of spiking neural networks, which was recently extended with a novel algorithm for remote spike communication through MPI on a GPU cluster. …”
  7. 7

    Data_Sheet_1_A Data-Driven Framework for Identifying Intensive Care Unit Admissions Colonized With Multidrug-Resistant Organisms.docx by Çaǧlar Çaǧlayan (12253934)

    Published 2022
    “…</p>Materials and Methods<p>Leveraging data from electronic healthcare records and a unique MDRO universal screening program, we developed a data-driven modeling framework to predict MRSA, VRE, and CRE colonization upon intensive care unit (ICU) admission, and identified the associated socio-demographic and clinical factors using logistic regression (LR), random forest (RF), and XGBoost algorithms. …”
  8. 8

    Confusion matrix. by Yixian Wen (12201388)

    Published 2025
    “…Subsequently, we implement an optimal binary tree decision-making algorithm, grounded in dynamic programming, to achieve precise allocation of elastic resources within data streams, significantly bolstering resource utilization. …”
  9. 9

    Parameter settings. by Yixian Wen (12201388)

    Published 2025
    “…Subsequently, we implement an optimal binary tree decision-making algorithm, grounded in dynamic programming, to achieve precise allocation of elastic resources within data streams, significantly bolstering resource utilization. …”
  10. 10

    Dynamic resource allocation process. by Yixian Wen (12201388)

    Published 2025
    “…Subsequently, we implement an optimal binary tree decision-making algorithm, grounded in dynamic programming, to achieve precise allocation of elastic resources within data streams, significantly bolstering resource utilization. …”
  11. 11

    Thesis-RAMIS-Figs_Slides by Felipe Santibañez-Leal (10967991)

    Published 2024
    “…In this direction, the option of estimating the statistics of the model directly from the training image (performing a refined pattern search instead of simulating data) is a very promising.<br><br>Finally, although the developed concepts, ideas and algorithms have been developed for inverse problems in geostatistics, the results are applicable to a wide range of disciplines where similar sampling problems need to be faced, included but not limited to design of communication networks, optimal integration and communication of swarms of robots and drones, remote sensing.…”
  12. 12

    Image 1_A multimodal AI-driven framework for cardiovascular screening and risk assessment in diverse athletic populations: innovations in sports cardiology.png by Minjin Guo (22751300)

    Published 2025
    “…RSEE projects heterogeneous input data into an exertion-conditioned latent space, aligning model predictions with observed physiological variance and mitigating false positives by explicitly modeling the overlap between athletic remodeling and subclinical pathology.…”
  13. 13

    An Ecological Benchmark of Photo Editing Software: A Comparative Analysis of Local vs. Cloud Workflows by Pierre-Alexis DELAROCHE (22092572)

    Published 2025
    “…Technical Architecture Overview Computational Environment Specifications Our experimental infrastructure leverages a heterogeneous multi-node computational topology encompassing three distinct hardware abstraction layers: Node Configuration Alpha (Intel-NVIDIA Heterogeneous Architecture) Processor: Intel Core i7-12700K (Alder Lake microarchitecture) - 12-core hybrid architecture (8 P-cores + 4 E-cores) - Base frequency: 3.6 GHz, Max turbo: 5.0 GHz - Cache hierarchy: 32KB L1I + 48KB L1D per P-core, 12MB L3 shared - Instruction set extensions: AVX2, AVX-512, SSE4.2 - Thermal design power: 125W (PL1), 190W (PL2) Memory Subsystem: 32GB DDR4-3200 JEDEC-compliant DIMM - Dual-channel configuration, ECC-disabled - Memory controller integrated within CPU die - Peak theoretical bandwidth: 51.2 GB/s GPU Accelerator: NVIDIA GeForce RTX 3070 (GA104 silicon) - CUDA compute capability: 8.6 - RT cores: 46 (2nd gen), Tensor cores: 184 (3rd gen) - Memory: 8GB GDDR6 @ 448 GB/s bandwidth - PCIe 4.0 x16 interface with GPU Direct RDMA support Node Configuration Beta (AMD Zen3+ Architecture) Processor: AMD Ryzen 7 5800X (Zen 3 microarchitecture) - 8-core monolithic design, simultaneous multithreading enabled - Base frequency: 3.8 GHz, Max boost: 4.7 GHz - Cache hierarchy: 32KB L1I + 32KB L1D per core, 32MB L3 shared - Infinity Fabric interconnect @ 1800 MHz - Thermal design power: 105W Memory Subsystem: 16GB DDR4-3600 overclocked configuration - Dual-channel with optimized subtimings (CL16-19-19-39) - Memory controller frequency: 1800 MHz (1:1 FCLK ratio) GPU Accelerator: NVIDIA GeForce GTX 1660 (TU116 silicon) - CUDA compute capability: 7.5 - Memory: 6GB GDDR5 @ 192 GB/s bandwidth - Turing shader architecture without RT/Tensor cores Node Configuration Gamma (Intel Raptor Lake High-Performance) Processor: Intel Core i9-13900K (Raptor Lake microarchitecture) - 24-core hybrid topology (8 P-cores + 16 E-cores) - P-core frequency: 3.0 GHz base, 5.8 GHz max turbo - E-core frequency: 2.2 GHz base, 4.3 GHz max turbo - Cache hierarchy: 36MB L3 shared, Intel Smart Cache technology - Thermal velocity boost with thermal monitoring Memory Subsystem: 64GB DDR5-5600 high-bandwidth configuration - Quad-channel topology with advanced error correction - Peak theoretical bandwidth: 89.6 GB/s GPU Accelerator: NVIDIA GeForce RTX 4080 (AD103 silicon) - Ada Lovelace architecture, CUDA compute capability: 8.9 - RT cores: 76 (3rd gen), Tensor cores: 304 (4th gen) - Memory: 16GB GDDR6X @ 716.8 GB/s bandwidth - PCIe 4.0 x16 with NVLink-ready topology Instrumentation and Telemetry Framework Power Consumption Monitoring Infrastructure Our energy profiling subsystem employs a multi-layered approach to capture granular power consumption metrics across the entire computational stack: Hardware Performance Counters (HPC): Intel RAPL (Running Average Power Limit) interface for CPU package power measurement with sub-millisecond resolution GPU Telemetry: NVIDIA Management Library (NVML) API for real-time GPU power draw monitoring via PCIe sideband signaling System-level PMU: Performance Monitoring Unit instrumentation leveraging MSR (Model Specific Register) access for architectural event sampling Network Interface Telemetry: SNMP-based monitoring of NIC power consumption during cloud upload/download phases Temporal Synchronization Protocol All measurement vectors utilize high-resolution performance counters (HPET) with nanosecond precision timestamps, synchronized via Network Time Protocol (NTP) to ensure temporal coherence across distributed measurement points. …”
  14. 14

    Aluminum alloy industrial materials defect by Ying Han (20349093)

    Published 2024
    “…</p><p dir="ltr">Install PyTorch based on your system:</p><p dir="ltr">For Windows/Linux users with a CUDA GPU: bash conda install pytorch==1.10.0 torchvision==0.11.0 torchaudio==0.10.0 cudatoolkit=11.3 -c pytorch -c conda-forge</p><p dir="ltr">Install some necessary libraries:</p><p dir="ltr">Install scikit-learn with the command: conda install anaconda scikit-learn=0.24.1</p><p dir="ltr">Install astropy with: conda install astropy=4.2.1</p><p dir="ltr">Install pandas using: conda install anaconda pandas=1.2.4</p><p dir="ltr">Install Matplotlib with: conda install conda-forge matplotlib=3.5.3</p><p dir="ltr">Install scipy by entering: conda install scipy=1.10.1</p><h4><b>Repeatability</b></h4><p dir="ltr">For PyTorch, it's a well-known fact:</p><p dir="ltr">There is no guarantee of fully reproducible results between PyTorch versions, individual commits, or different platforms. …”
  15. 15

    LinearSolve.jl: because A\b is not good enough by Christopher Rackauckas (9197216)

    Published 2022
    “…Short list: LU, QR, SVD, RecursiveFactorization.jl (pure Julia, and the fastest?), GPU-offload LU, UMFPACK, KLU, CG, GMRES, Pardiso, ...…”