يعرض 1 - 8 نتائج من 8 نتيجة بحث عن '(( ((algorithm against) OR (algorithm i)) function ) OR ( algorithm python function ))~', وقت الاستعلام: 0.31s تنقيح النتائج
  1. 1

    S1 File - حسب Yuh-Chin T. Huang (17867207)

    منشور في 2024
    "…In this study, we developed a computerized algorithm using the python package (pdfplumber) and validated against clinicians’ interpretation. …"
  2. 2

    S1 Dataset - حسب Yuh-Chin T. Huang (17867207)

    منشور في 2024
    "…In this study, we developed a computerized algorithm using the python package (pdfplumber) and validated against clinicians’ interpretation. …"
  3. 3

    Algoritmo de clasificación de expresiones de odio por tipos en español (Algorithm for classifying hate expressions by type in Spanish) حسب Daniel Pérez Palau (11097348)

    منشور في 2024
    "…</p><h2>Model Architecture</h2><p dir="ltr">The model is based on <code>pysentimiento/robertuito-base-uncased</code> with the following modifications:</p><ul><li>A dense classification layer was added over the base model</li><li>Uses input IDs and attention masks as inputs</li><li>Generates a multi-class classification with 5 hate categories</li></ul><h2>Dataset</h2><p dir="ltr"><b>HATEMEDIA Dataset</b>: Custom hate speech dataset with categorization by type:</p><ul><li><b>Labels</b>: 5 hate type categories (0-4)</li><li><b>Preprocessing</b>:</li><li>Null values ​​removed from text and labels</li><li>Reindexing and relabeling (original labels are adjusted by subtracting 1)</li><li>Exclusion of category 2 during training</li><li>Conversion of category 5 to category 2</li></ul><h2>Training Process</h2><h3>Configuration</h3><ul><li><b>Batch size</b>: 128</li><li><b>Epoches</b>: 5</li><li><b>Learning rate</b>: 2e-5 with 10% warmup steps</li><li><b>Early stopping</b> with patience=2</li><li><b>Class weights</b>: Balanced to handle class imbalance</li></ul><h3>Custom Metrics</h3><ul><li>Recall for specific classes (focus on class 2)</li><li>Precision for specific classes (focus on class 3)</li><li>F1-score (weighted)</li><li>AUC-PR</li><li>Recall at precision=0.6 (class 3)</li><li>Precision at recall=0.6 (class 2)</li></ul><h2>Evaluation Metrics</h2><p dir="ltr">The model is evaluated using:</p><ul><li>Macro recall, precision, and F1-score</li><li>One-vs-Rest AUC</li><li>Accuracy</li><li>Per-class metrics</li><li>Confusion matrix</li><li>Full classification report</li></ul><h2>Technical Features</h2><h3>Data Preprocessing</h3><ul><li><b>Tokenization</b>: Maximum length of 128 tokens (truncation and padding)</li><li><b>Encoding of labels</b>: One-hot encoding for multi-class classification</li><li><b>Data split</b>: 80% training, 10% validation, 10% testing</li></ul><h3>Optimization</h3><ul><li><b>Optimizer</b>: Adam with linear warmup scheduling</li><li><b>Loss function</b>: Categorical Crossentropy (from_logits=True)</li><li><b>Imbalance handling</b>: Class weights computed automatically</li></ul><h2>Requirements</h2><p dir="ltr">The following Python packages are required:</p><ul><li>TensorFlow</li><li>Transformers</li><li>scikit-learn</li><li>pandas</li><li>datasets</li><li>matplotlib</li><li>seaborn</li><li>numpy</li></ul><h2>Usage</h2><ol><li><b>Data format</b>:</li></ol><ul><li>CSV file or Pandas DataFrame</li><li>Required column name: <code>text</code> (string type)</li><li>Required column name: Data type label (integer type, 0-4) - optional for evaluation</li></ul><ol><li><b>Text preprocessing</b>:</li></ol><ul><li>Automatic tokenization with a maximum length of 128 tokens</li><li>Long texts will be automatically truncated</li><li>Handling of special characters, URLs, and emojis included</li></ul><ol><li><b>Label encoding</b>:</li></ol><ul><li>The model classifies hate speech into 5 categories (0-4)</li><li><code>0</code>: Political hatred: Expressions directed against individuals or groups based on political orientation.…"
  4. 4
  5. 5
  6. 6

    Decoding fairness motivations - repository حسب Sebastian Speer (6489207)

    منشور في 2020
    "…All analysis were conducted in Python 2.7.</div><div><br></div><div><b>Behavioral Data:</b><br></div><div><br></div><div><u>Files:</u> </div><div><br></div><div><i>DifffereceOffers.csv </i>- Offers made by participants in Study 1</div><div><i>Diffs_W.csv</i> - Offers made by participants in Study 2</div><div><br></div><div><i>Individual-differences-in-offers2.png</i> - Plot of individual differences as illustrated in the paper</div><div><i>Individual-differences-MeanOffers.png </i>- Individual differences in mean offers in both games as illustrated in the Appendix</div><div><i>SocialvsNonSocial2.png </i>- Difference in Offers between Selfish and strategic players when playing against humans and computers</div><div><br></div><div>Behavioral Data, specfically Ultimatum Game and Dictator Game Offers and Plots resulting from behavioral analysis reported in the following paper:</div><div><br></div><div>S.P.H. …"
  7. 7

    Landscape17 حسب Vlad Carare (22092515)

    منشور في 2025
    "…We validated the convergence, grid, and spin settings against published data from rMD17, using the appropriate functional and basis set: PBE/def2-SVP. …"
  8. 8

    An Ecological Benchmark of Photo Editing Software: A Comparative Analysis of Local vs. Cloud Workflows حسب Pierre-Alexis DELAROCHE (22092572)

    منشور في 2025
    "…Performance Profiling Algorithms Energy Measurement Methodology # Pseudo-algorithmic representation of measurement protocol def capture_energy_metrics(workflow_type: WorkflowEnum, asset_vector: List[PhotoAsset]) -> EnergyProfile: baseline_power = sample_idle_power_draw(duration=30) with PowerMonitoringContext() as pmc: start_timestamp = rdtsc() # Read time-stamp counter if workflow_type == WorkflowEnum.LOCAL: result = execute_local_pipeline(asset_vector) elif workflow_type == WorkflowEnum.CLOUD: result = execute_cloud_pipeline(asset_vector) end_timestamp = rdtsc() energy_profile = EnergyProfile( duration=cycles_to_seconds(end_timestamp - start_timestamp), peak_power=pmc.get_peak_consumption(), average_power=pmc.get_mean_consumption(), total_energy=integrate_power_curve(pmc.get_power_trace()) ) return energy_profile Statistical Analysis Framework Our analytical pipeline employs advanced statistical methodologies including: Variance Decomposition: ANOVA with nested factors for hardware configuration effects Regression Analysis: Generalized Linear Models (GLM) with log-link functions for energy modeling Temporal Analysis: Fourier transform-based frequency domain analysis of power consumption patterns Cluster Analysis: K-means clustering with Euclidean distance metrics for workflow classification Data Validation and Quality Assurance Measurement Uncertainty Quantification All energy measurements incorporate systematic and random error propagation analysis: Instrument Precision: ±0.1W for CPU power, ±0.5W for GPU power Temporal Resolution: 1ms sampling with Nyquist frequency considerations Calibration Protocol: NIST-traceable power standards with periodic recalibration Environmental Controls: Temperature-compensated measurements in climate-controlled facility Outlier Detection Algorithms Statistical outliers are identified using the Interquartile Range (IQR) method with Tukey's fence criteria (Q₁ - 1.5×IQR, Q₃ + 1.5×IQR). …"