Showing 481 - 490 results of 490 for search '(( python model implementation ) OR ( ((python model) OR (python code)) predicted ))', query time: 0.46s Refine Results
  1. 481

    RNACOREX package workflow. by Aitor Oviedo-Madrid (22553399)

    Published 2025
    “…These sets are then used to build probabilistic models based on Conditional Linear Gaussian (CLG) classifiers, which allow both prediction on new samples and validation of the inferred networks.…”
  2. 482

    Performance. by Aitor Oviedo-Madrid (22553399)

    Published 2025
    “…These sets are then used to build probabilistic models based on Conditional Linear Gaussian (CLG) classifiers, which allow both prediction on new samples and validation of the inferred networks.…”
  3. 483

    Nucleotide analogue tolerant synthetic RdRp mutant construct for Surveillance and Therapeutic Resistance Monitoring in SARS-CoV-2 by Tahir Bhatti (20961974)

    Published 2025
    “…</p><p dir="ltr">Using the output of this locally trained AI model a synthetic RdRp region was generated incorporating those predicted mutations and was inserted into the consensus backbone.…”
  4. 484

    MSc Personalised Medicine at Ulster University by Steven Watterson (100045)

    Published 2025
    “…This includes the economic models that underpin big pharma as well the importance of entrepreneurship and small medium enterprises in driving forward healthcare innovation.…”
  5. 485

    Microscopic Detection and Quantification of Microplastic Particles in Environmental Water Samples by Derek Lam (11944213)

    Published 2025
    “…Image processing algorithms, implemented in Python using adaptive thresholding techniques, were applied to segment particles from the background. …”
  6. 486

    Comprehensive Fluid and Gravitational Dynamics Script for General Symbolic Navier-Stokes Calculations and Validation by Stylianos Touloumidis (19938747)

    Published 2024
    “…It provides a flexible foundation on which theoretical assumptions can be validated, and practical calculations performed. Implemented in Python with symbolic calculations, the script facilitates in-depth analysis of complex flow patterns and makes advanced mathematical computations more accessible. …”
  7. 487

    AGU24 - EP11D-1300 - Revisiting Megacusp Embayment Occurrence in Monterey Bay and Beyond: High Spatiotemporal Resolution Satellite Imagery Provides New Insight into the Wave Condit... by Kelby Kramer (20706767)

    Published 2025
    “…Previous studies using both site observation and numerical models have yielded rough characterization of the wave conditions necessary for MCE formation, including wave energy and direction. …”
  8. 488

    Supplementary Data: Biodiversity and Energy System Planning - Queensland 2025 by Andrew Rogers (17623239)

    Published 2025
    “…</p><h2>Software and Spatial Resolution</h2><p dir="ltr">The VRE siting model is implemented using Python and relies heavily on ArcGIS for comprehensive spatial data handling and analysis.…”
  9. 489

    An Ecological Benchmark of Photo Editing Software: A Comparative Analysis of Local vs. Cloud Workflows by Pierre-Alexis DELAROCHE (22092572)

    Published 2025
    “…Reproducibility Framework Container Orchestration # Kubernetes deployment manifest for reproducible environment apiVersion: apps/v1 kind: Deployment metadata: name: energy-benchmark-pod spec: replicas: 1 selector: matchLabels: app: benchmark-runner template: metadata: labels: app: benchmark-runner spec: nodeSelector: hardware.profile: "high-performance" containers: - name: benchmark-container image: albumforge/energy-benchmark:v2.1.3 resources: requests: cpu: "8000m" memory: "16Gi" nvidia.com/gpu: 1 limits: cpu: "16000m" memory: "32Gi" env: - name: MEASUREMENT_PRECISION value: "high" - name: POWER_SAMPLING_RATE value: "1000" # 1kHz sampling Dependency Management FROM ubuntu:22.04-cuda11.8-devel RUN apt-get update && apt-get install -y \ perf-tools \ powertop \ intel-gpu-tools \ nvidia-smi \ cpupower \ msr-tools \ && rm -rf /var/lib/apt/lists/* COPY requirements.txt /opt/ RUN pip install -r /opt/requirements.txt Usage Examples and API Documentation Python Data Analysis Interface import pandas as pd import numpy as np from scipy import stats import matplotlib.pyplot as plt import seaborn as sns # Load dataset with optimized dtypes for memory efficiency df = pd.read_csv('ecological_benchmark_dataset.csv', dtype={'hardware_config': 'category', 'test_type': 'category'}) # Compute energy efficiency metrics df['energy_per_photo'] = df['energy_consumption_kwh'] / df['photo_count'] df['co2_per_gigabyte'] = df['co2_equivalent_g'] / df['total_volume_gb'] # Statistical analysis with confidence intervals local_energy = df[df['test_type'] == 'local_processing']['energy_consumption_kwh'] cloud_energy = df[df['test_type'] == 'cloud_processing']['energy_consumption_kwh'] t_stat, p_value = stats.ttest_ind(local_energy, cloud_energy) effect_size = (cloud_energy.mean() - local_energy.mean()) / np.sqrt((cloud_energy.var() + local_energy.var()) / 2) print(f"Statistical significance: p = {p_value:.2e}") print(f"Cohen's d effect size: {effect_size:.3f}") R Statistical Computing Environment library(tidyverse) library(lme4) # Linear mixed-effects models library(ggplot2) library(corrplot) # Load and preprocess data df <- read_csv("ecological_benchmark_dataset.csv") %>% mutate( test_type = factor(test_type), hardware_config = factor(hardware_config), log_energy = log(energy_consumption_kwh), efficiency_ratio = energy_consumption_kwh / processing_time_sec ) # Mixed-effects regression model accounting for hardware heterogeneity model <- lmer(log_energy ~ test_type + log(photo_count) + (1|hardware_config), data = df) # Extract model coefficients with confidence intervals summary(model) confint(model, method = "Wald") Advanced Analytics and Machine Learning Integration Predictive Modeling Framework from sklearn.ensemble import RandomForestRegressor, GradientBoostingRegressor from sklearn.model_selection import cross_val_score, GridSearchCV from sklearn.preprocessing import StandardScaler, LabelEncoder from sklearn.metrics import mean_absolute_error, r2_score # Feature engineering pipeline def create_feature_matrix(df): features = df[['photo_count', 'avg_file_size_mb', 'total_volume_gb']].copy() # Polynomial features for capturing non-linear relationships features['photo_count_squared'] = features['photo_count'] ** 2 features['size_volume_interaction'] = features['avg_file_size_mb'] * features['total_volume_gb'] # Hardware configuration encoding le = LabelEncoder() features['hardware_encoded'] = le.fit_transform(df['hardware_config']) return features # Energy consumption prediction model X = create_feature_matrix(df) y = df['energy_consumption_kwh'] # Hyperparameter optimization param_grid = { 'n_estimators': [100, 200, 500], 'max_depth': [10, 20, None], 'min_samples_split': [2, 5, 10], 'min_samples_leaf': [1, 2, 4] } rf_model = RandomForestRegressor(random_state=42) grid_search = GridSearchCV(rf_model, param_grid, cv=5, scoring='neg_mean_absolute_error') grid_search.fit(X, y) print(f"Best cross-validation score: {-grid_search.best_score_:.6f}") print(f"Optimal hyperparameters: {grid_search.best_params_}") Carbon Footprint Calculation Methodology Emission Factor Coefficients Carbon intensity calculations employ region-specific emission factors from the International Energy Agency (IEA) database: EMISSION_FACTORS = { 'EU_AVERAGE': 0.276, # kg CO₂/kWh (European Union average 2024) 'FRANCE': 0.057, # kg CO₂/kWh (Nuclear-dominant grid) 'GERMANY': 0.485, # kg CO₂/kWh (Coal transition period) 'NORWAY': 0.013, # kg CO₂/kWh (Hydroelectric dominant) 'GLOBAL_AVERAGE': 0.475 # kg CO₂/kWh (Global weighted average) } def calculate_carbon_footprint(energy_kwh: float, region: str = 'EU_AVERAGE') -> float: """ Calculate CO₂ equivalent emissions using lifecycle assessment methodology Args: energy_kwh: Energy consumption in kilowatt-hours region: Geographic region for emission factor selection Returns: CO₂ equivalent emissions in grams """ emission_factor = EMISSION_FACTORS.get(region, EMISSION_FACTORS['GLOBAL_AVERAGE']) co2_kg = energy_kwh * emission_factor return co2_kg * 1000 # Convert to grams Citation and Attribution This dataset is released under Creative Commons Attribution 4.0 International (CC BY 4.0) license. …”
  10. 490

    <b>Historical Nifty 50 Constituent Weights (Rolling 20-Year Window)</b> by Sukrit Bera (22314922)

    Published 2025
    “…</li><li>Building features for quantitative models that aim to predict market movements.</li><li>Backtesting investment strategies benchmarked against the Nifty 50.…”