Search alternatives:
processing optimisation » process optimisation (Expand Search), process optimization (Expand Search), prescribing optimisation (Expand Search)
optimisation algorithm » optimization algorithms (Expand Search), maximization algorithm (Expand Search), identification algorithm (Expand Search)
d optimization » _ optimization (Expand Search), b optimization (Expand Search), led optimization (Expand Search)
library based » laboratory based (Expand Search)
based d » based 3d (Expand Search), based _ (Expand Search), based 2 (Expand Search)
processing optimisation » process optimisation (Expand Search), process optimization (Expand Search), prescribing optimisation (Expand Search)
optimisation algorithm » optimization algorithms (Expand Search), maximization algorithm (Expand Search), identification algorithm (Expand Search)
d optimization » _ optimization (Expand Search), b optimization (Expand Search), led optimization (Expand Search)
library based » laboratory based (Expand Search)
based d » based 3d (Expand Search), based _ (Expand Search), based 2 (Expand Search)
-
1
-
2
-
3
-
4
-
5
-
6
DataSheet1_Large-Scale Heterogeneous Computing for 3D Deterministic Particle Transport on Tianhe-2A Supercomputer.zip
Published 2021“…We use OpenMP instructions to exploit the parallelism of threads based on Matrix 2000. The test results show that the optimization of applying OpenMP on particle transport algorithm modified by our method can get 11.3 times acceleration at most. …”
-
7
-
8
Distribution of Bound Conformations in Conformational Ensembles for X‑ray Ligands Predicted by the ANI-2X Machine Learning Potential
Published 2023“…This information is useful to guide the construction of libraries for shape-based virtual screening and to improve the docking algorithm to efficiently sample bound conformations.…”
-
9
Distribution of Bound Conformations in Conformational Ensembles for X‑ray Ligands Predicted by the ANI-2X Machine Learning Potential
Published 2023“…This information is useful to guide the construction of libraries for shape-based virtual screening and to improve the docking algorithm to efficiently sample bound conformations.…”
-
10
-
11
Otago's Network for Engagement and Research: Mapping Academic Expertise and Connections
Published 2020“…<br></div><div><br></div><div>In the next stage of the project, we will develop further the data integration schemes, enhance our algorithm to infer expertise based on this data, and update the interactive visualisation to reflect these inferences. …”
-
12
An Ecological Benchmark of Photo Editing Software: A Comparative Analysis of Local vs. Cloud Workflows
Published 2025“…Technical Architecture Overview Computational Environment Specifications Our experimental infrastructure leverages a heterogeneous multi-node computational topology encompassing three distinct hardware abstraction layers: Node Configuration Alpha (Intel-NVIDIA Heterogeneous Architecture) Processor: Intel Core i7-12700K (Alder Lake microarchitecture) - 12-core hybrid architecture (8 P-cores + 4 E-cores) - Base frequency: 3.6 GHz, Max turbo: 5.0 GHz - Cache hierarchy: 32KB L1I + 48KB L1D per P-core, 12MB L3 shared - Instruction set extensions: AVX2, AVX-512, SSE4.2 - Thermal design power: 125W (PL1), 190W (PL2) Memory Subsystem: 32GB DDR4-3200 JEDEC-compliant DIMM - Dual-channel configuration, ECC-disabled - Memory controller integrated within CPU die - Peak theoretical bandwidth: 51.2 GB/s GPU Accelerator: NVIDIA GeForce RTX 3070 (GA104 silicon) - CUDA compute capability: 8.6 - RT cores: 46 (2nd gen), Tensor cores: 184 (3rd gen) - Memory: 8GB GDDR6 @ 448 GB/s bandwidth - PCIe 4.0 x16 interface with GPU Direct RDMA support Node Configuration Beta (AMD Zen3+ Architecture) Processor: AMD Ryzen 7 5800X (Zen 3 microarchitecture) - 8-core monolithic design, simultaneous multithreading enabled - Base frequency: 3.8 GHz, Max boost: 4.7 GHz - Cache hierarchy: 32KB L1I + 32KB L1D per core, 32MB L3 shared - Infinity Fabric interconnect @ 1800 MHz - Thermal design power: 105W Memory Subsystem: 16GB DDR4-3600 overclocked configuration - Dual-channel with optimized subtimings (CL16-19-19-39) - Memory controller frequency: 1800 MHz (1:1 FCLK ratio) GPU Accelerator: NVIDIA GeForce GTX 1660 (TU116 silicon) - CUDA compute capability: 7.5 - Memory: 6GB GDDR5 @ 192 GB/s bandwidth - Turing shader architecture without RT/Tensor cores Node Configuration Gamma (Intel Raptor Lake High-Performance) Processor: Intel Core i9-13900K (Raptor Lake microarchitecture) - 24-core hybrid topology (8 P-cores + 16 E-cores) - P-core frequency: 3.0 GHz base, 5.8 GHz max turbo - E-core frequency: 2.2 GHz base, 4.3 GHz max turbo - Cache hierarchy: 36MB L3 shared, Intel Smart Cache technology - Thermal velocity boost with thermal monitoring Memory Subsystem: 64GB DDR5-5600 high-bandwidth configuration - Quad-channel topology with advanced error correction - Peak theoretical bandwidth: 89.6 GB/s GPU Accelerator: NVIDIA GeForce RTX 4080 (AD103 silicon) - Ada Lovelace architecture, CUDA compute capability: 8.9 - RT cores: 76 (3rd gen), Tensor cores: 304 (4th gen) - Memory: 16GB GDDR6X @ 716.8 GB/s bandwidth - PCIe 4.0 x16 with NVLink-ready topology Instrumentation and Telemetry Framework Power Consumption Monitoring Infrastructure Our energy profiling subsystem employs a multi-layered approach to capture granular power consumption metrics across the entire computational stack: Hardware Performance Counters (HPC): Intel RAPL (Running Average Power Limit) interface for CPU package power measurement with sub-millisecond resolution GPU Telemetry: NVIDIA Management Library (NVML) API for real-time GPU power draw monitoring via PCIe sideband signaling System-level PMU: Performance Monitoring Unit instrumentation leveraging MSR (Model Specific Register) access for architectural event sampling Network Interface Telemetry: SNMP-based monitoring of NIC power consumption during cloud upload/download phases Temporal Synchronization Protocol All measurement vectors utilize high-resolution performance counters (HPET) with nanosecond precision timestamps, synchronized via Network Time Protocol (NTP) to ensure temporal coherence across distributed measurement points. …”
-
13
Code
Published 2025“…</p><p><br></p><p dir="ltr">This architecture was implemented using the PyTorch library and trained using cross-entropy loss. The model was optimized to classify RNA sequences, achieving robust performance across multiple test sets.…”
-
14
Core data
Published 2025“…</p><p><br></p><p dir="ltr">This architecture was implemented using the PyTorch library and trained using cross-entropy loss. The model was optimized to classify RNA sequences, achieving robust performance across multiple test sets.…”
-
15
<b>AI for imaging plant stress in invasive species </b>(dataset from the article https://doi.org/10.1093/aob/mcaf043)
Published 2025“…The described extracted features were used to predict leaf betalain content (µg per FW) using multiple machine learning regression algorithms (Linear regression, Ridge regression, Gradient boosting, Decision tree, Random forest and Support vector machine) using the <i>Scikit-learn</i> 1.2.1 library in Python (v.3.10.1) (list of hyperparameters used is given in <a href="#sup1" target="_blank">Supplementary Data S5</a>). …”