Search alternatives:
feature optimization » resource optimization (Expand Search), feature elimination (Expand Search), structure optimization (Expand Search)
codon optimization » wolf optimization (Expand Search)
data feature » data figure (Expand Search), each feature (Expand Search), a feature (Expand Search)
binary data » primary data (Expand Search), dietary data (Expand Search)
feature optimization » resource optimization (Expand Search), feature elimination (Expand Search), structure optimization (Expand Search)
codon optimization » wolf optimization (Expand Search)
data feature » data figure (Expand Search), each feature (Expand Search), a feature (Expand Search)
binary data » primary data (Expand Search), dietary data (Expand Search)
-
101
DataSheet_1_Raman Spectroscopic Differentiation of Streptococcus pneumoniae From Other Streptococci Using Laboratory Strains and Clinical Isolates.pdf
Published 2022“…Improvement of the classification rate is expected with optimized model parameters and algorithms as well as with a larger spectral data base for training.…”
-
102
-
103
GSE96058 information.
Published 2024“…Subsequently, feature selection was conducted using ANOVA and binary Particle Swarm Optimization (PSO). …”
-
104
The performance of classifiers.
Published 2024“…Subsequently, feature selection was conducted using ANOVA and binary Particle Swarm Optimization (PSO). …”
-
105
Algoritmo de clasificación de expresiones de odio por tipos en español (Algorithm for classifying hate expressions by type in Spanish)
Published 2024“…</p><h2>Model Architecture</h2><p dir="ltr">The model is based on <code>pysentimiento/robertuito-base-uncased</code> with the following modifications:</p><ul><li>A dense classification layer was added over the base model</li><li>Uses input IDs and attention masks as inputs</li><li>Generates a multi-class classification with 5 hate categories</li></ul><h2>Dataset</h2><p dir="ltr"><b>HATEMEDIA Dataset</b>: Custom hate speech dataset with categorization by type:</p><ul><li><b>Labels</b>: 5 hate type categories (0-4)</li><li><b>Preprocessing</b>:</li><li>Null values removed from text and labels</li><li>Reindexing and relabeling (original labels are adjusted by subtracting 1)</li><li>Exclusion of category 2 during training</li><li>Conversion of category 5 to category 2</li></ul><h2>Training Process</h2><h3>Configuration</h3><ul><li><b>Batch size</b>: 128</li><li><b>Epoches</b>: 5</li><li><b>Learning rate</b>: 2e-5 with 10% warmup steps</li><li><b>Early stopping</b> with patience=2</li><li><b>Class weights</b>: Balanced to handle class imbalance</li></ul><h3>Custom Metrics</h3><ul><li>Recall for specific classes (focus on class 2)</li><li>Precision for specific classes (focus on class 3)</li><li>F1-score (weighted)</li><li>AUC-PR</li><li>Recall at precision=0.6 (class 3)</li><li>Precision at recall=0.6 (class 2)</li></ul><h2>Evaluation Metrics</h2><p dir="ltr">The model is evaluated using:</p><ul><li>Macro recall, precision, and F1-score</li><li>One-vs-Rest AUC</li><li>Accuracy</li><li>Per-class metrics</li><li>Confusion matrix</li><li>Full classification report</li></ul><h2>Technical Features</h2><h3>Data Preprocessing</h3><ul><li><b>Tokenization</b>: Maximum length of 128 tokens (truncation and padding)</li><li><b>Encoding of labels</b>: One-hot encoding for multi-class classification</li><li><b>Data split</b>: 80% training, 10% validation, 10% testing</li></ul><h3>Optimization</h3><ul><li><b>Optimizer</b>: Adam with linear warmup scheduling</li><li><b>Loss function</b>: Categorical Crossentropy (from_logits=True)</li><li><b>Imbalance handling</b>: Class weights computed automatically</li></ul><h2>Requirements</h2><p dir="ltr">The following Python packages are required:</p><ul><li>TensorFlow</li><li>Transformers</li><li>scikit-learn</li><li>pandas</li><li>datasets</li><li>matplotlib</li><li>seaborn</li><li>numpy</li></ul><h2>Usage</h2><ol><li><b>Data format</b>:</li></ol><ul><li>CSV file or Pandas DataFrame</li><li>Required column name: <code>text</code> (string type)</li><li>Required column name: Data type label (integer type, 0-4) - optional for evaluation</li></ul><ol><li><b>Text preprocessing</b>:</li></ol><ul><li>Automatic tokenization with a maximum length of 128 tokens</li><li>Long texts will be automatically truncated</li><li>Handling of special characters, URLs, and emojis included</li></ul><ol><li><b>Label encoding</b>:</li></ol><ul><li>The model classifies hate speech into 5 categories (0-4)</li><li><code>0</code>: Political hatred: Expressions directed against individuals or groups based on political orientation.…”
-
106
Flowchart scheme of the ML-based model.
Published 2024“…<b>G)</b> Deep feature extraction using VGG16. <b>H)</b> Training data comprising 80% of the dataset. …”
-
107
Solubility Prediction of Different Forms of Pharmaceuticals in Single and Mixed Solvents Using Symmetric Electrolyte Nonrandom Two-Liquid Segment Activity Coefficient Model
Published 2019“…The methodology incorporates key features of the symmetric eNRTL-SAC model structure to reduce the number of parameters and uses a hybrid of global search algorithms for parameter estimation. …”
-
108
-
109
Data_Sheet_1_Alzheimer’s Disease Diagnosis and Biomarker Analysis Using Resting-State Functional MRI Functional Brain Network With Multi-Measures Features and Hippocampal Subfield...
Published 2022“…Finally, we implemented and compared the different feature selection algorithms to integrate the structural features, brain networks, and voxel features to optimize the diagnostic identifications of AD using support vector machine (SVM) classifiers. …”
-
110
Natural language processing for automated quantification of bone metastases reported in free-text bone scintigraphy reports
Published 2020“…<p> The widespread use of electronic patient-generated health data has led to unprecedented opportunities for automated extraction of clinical features from free-text medical notes. …”
-
111
Contextual Dynamic Pricing with Strategic Buyers
Published 2024“…In this process, buyers can also strategically manipulate their feature data to obtain a lower price, incurring certain manipulation costs. …”
-
112
Supplementary Material 8
Published 2025“…</li><li><b>XGboost: </b>An optimized gradient boosting algorithm that efficiently handles large genomic datasets, commonly used for high-accuracy predictions in <i>E. coli</i> classification.…”
-
113
PathOlOgics_RBCs Python Scripts.zip
Published 2023“…</p><p dir="ltr">In terms of classification, a second algorithm was developed and employed to preliminary sort or group the individual cells (after excluding the overlapping cells manually) into different categories using five geometric measurements applied to the extracted contour from each binary image mask (see PathOlOgics_script_2; preliminary shape measurements). …”
-
114
-
115
DataSheet_1_Near infrared spectroscopy for cooking time classification of cassava genotypes.docx
Published 2024“…Cooking data were classified into binary and multiclass variables (CT4C and CT6C). …”
-
116
-
117
-
118
Models and Dataset
Published 2025“…<p dir="ltr"><b>P3DE (Parameter-less Population Pyramid with Deep Ensemble):</b><br>P3DE is a hybrid feature selection framework that combines the Parameter-less Population Pyramid (P3) metaheuristic optimization algorithm with a deep ensemble of autoencoders. …”
-
119
Flowchart of the entire pipeline.
Published 2024“…Next, the general protein dataset is converted to a specific transporter dataset according to specified parameters (see Section Dataset creation pipeline). Then, the protein feature generation algorithms described in our previous study [<a href="http://www.plosone.org/article/info:doi/10.1371/journal.pone.0315330#pone.0315330.ref022" target="_blank">22</a>] are applied to the data, and pairwise ML models are trained and evaluated (see Section Evaluation of pairwise machine learning models). …”
-
120
Table_1_Near infrared spectroscopy for cooking time classification of cassava genotypes.docx
Published 2024“…Cooking data were classified into binary and multiclass variables (CT4C and CT6C). …”