Science topic

Classification - Science topic

The systematic arrangement of entities in any field into categories classes based on common characteristics such as properties, morphology, subject matter, etc.
Questions related to Classification
  • asked a question related to Classification
Question
6 answers
Keeping Murphy's Law, the KISS principle, and Popper's Logic, right, left, and center, respectively, as well as the all-encompassing muses of insight, innovation, intuition, imagination, and insurrection (the 5 I's that holistically, through immersive-integrative multi-disciplinary contemplative approach identifies the noise or separates the wheat-from-the-chaff at the intersection of fact and fiction), and importantly and synergistically compose the whole that is greater than the sum of its parts -- in true Aristotelian fashion -- that govern progress and advance in human thinking through the synapse in all human endeavor, scientific and non-scientific.
I will put exactly 50-years of my part in one of the greatest mysteries ever faced by humans, and that will follow this species indefinitely to perpetuity, but with secure and fearless knowledge through application of principles or laws of theory and therapy, elimination of canonical or Institutional myths and assumptions, with a complete unwinding of this humungous Gordian knot of neuro-ophthalmology.
Da Vinci guarded against excessive use of words to describe any entity or anything. Migraine is an entity of excess -- incidence, words, data, statistics, analyses, meta-analyses, hypotheses, viewpoints, perspectives, Editorials, Medical Conferencing Abstracts, invited Lectures, hyper-splitting of nosology, and Letters-to-the-Editor, all claiming to know a slice of truth or presumed truth about migraine with a hyper-exponential absolutely unlimited untrammeled expansion. Quo vadis is not even a remote concept.
I, in the Third Millennium, describe the 'what' of migraine in 6-10 words, a definition that will last to perpetuity:
Migraine is the delayed outcome of an oculo-cephalic autonomic storm (causing the non-homonymous scintillating scotoma as well as the lateralizing headache). More succinctly, migraine is an oculo-cephalic autonomic storm.
Nothing is static. No theory or therapy cannot be improved. The core of migraine is here.
With the cause-effect mechanisms in migraine pathophysiology fully described, what has been missing for 6 millennia or more is presented right here and now.
The doors of perception for cluster headache and other indomethacin-responsive headaches are now open.
Reversal of the hyper-split classification of primary headaches is imminent, leading to a holistic comprehensive understand of a large section of medicine and neuroscience.
06-MAY-2024
New Delhi
ORCID iD: 0000-0002-6770-5916
Relevant answer
Answer
Nature is a jealous mistress, guarding its secrets with persistence and with an overpowering devastating diligence till the Time is ripe for butterfly wings to cause a hurricane. That time is upon science of migraine / primary headaches, straddling the 20th/21st Centuries, with the floodgates of intuition and insight and imagination, an immersive insurrective release from ~8 millennia of ignornance, myths, mysticism, serendipity, assumptions, technology, and of course, an explosive exponential amassing of data that have no central framework. Migraine / primary headaches showcase, as nothing else in medicine, the limitations of a madding collection of diverse disparate contrary controversial data that do not promote critical or abstract thinking. Mathematical statistics and the specialty of Neurology, has for all purposes, kidnapped migraine / primary headache pathophysiology and management with an opioid crisis and extreme empiricism, lending wings to speculation in science of a kind that prove horribly incorrect gor generations of migraineurs, current and future. The butterfly wings will raise the insurrection, the hurricane, the nihilism, and the despondency that surrounds migraine / primary headache research, with the clarity, the certitude of a thousand Suns in the Solar system. In fact, such insurrection has progressed silently in last 5 decades, leaving the contemporary cohort breathless and trying their best to stymie the tide of absolute progress. The butterfly wings will, of course, prevail despite the dog-in-manger opposition.
  • asked a question related to Classification
  • asked a question related to Classification
Question
1 answer
Pediatric supra condylar fractures sometimes show rotation after fixation. It may be significant in some cases. What is the normal range. What is significance of gordons classification.
Relevant answer
Answer
The acceptable amount of rotation in pediatric supracondylar fractures varies depending on the severity of the fracture and the age of the child. Generally, up to 10-15 degrees of rotational malalignment is considered acceptable in pediatric supracondylar fractures.
  • asked a question related to Classification
Question
6 answers
Within a project about geographical traceability of horticultural products, we would like to apply classification models to our data set (e.g. LDA) to predict if it is possible to correctly classify samples according to their origin and based on the results of 20-25 different chemical variables.
We identified 5 cultivation areas and selected 41 orchards (experimental units) in total. In each orchard, 10 samples were collected (each sample from a different tree). The samples were analyzed separately. So, at the end, we have the results for 410 samples.
The question is: the 10 samples per orchard have to be considered pseudoreplicates since they belong to the same experimental unit (even if collected from indepedent trees)? Should the LDA be performed considering 41 replicates (the 41 orchards, taking the average of the 10 samples) or should we run it for the whole dataset?
Thank you for your help.
Relevant answer
Answer
In general, dealing with false copies in linear discriminant analysis depends on a good understanding of the data and applying the necessary procedures to correctly identify and treat these copies.
  • asked a question related to Classification
Question
6 answers
Please provide an explanation according to the classification of primary and secondary uranium ore
Relevant answer
Answer
The so-called "fertile" granites which had relatively high background U contents, such as 10-20 ppm.
For example, in the case of the Beaverlodge uranium province of Saskatchewan the granitic rocks "were relatively enriched in uranium".
  • asked a question related to Classification
Question
1 answer
Dear researchers,
a variability in the taxonomy classification of microbial communities when using different primer pairs (e.g. for 16S rDNA) is commonly known. However, the mismatches to these primers are not described as the major reason for this bias. My question is: what are other possible causes of this bias and which is now supposed to be the major one?
Thank you for your contribution. Lucie
Relevant answer
Answer
Some variable regions are simply not suitable for accurate identification, see eg
  • asked a question related to Classification
Question
1 answer
I am inclined to research on EEG classification using ML/DL. The research area seems saturated. Hence, I am confused as to where I can contribute.
Relevant answer
Answer
First of all, I want to encourage you not to give up on an area just because there are a lot of researchers in it. People should follow their interests if they are capable of managing the task and are interested in them. It's not only that EEG research is a promising field, but it's also interesting to classify EEG data using machine learning or deep learning approaches. It's okay if it seems saturated to you. Improving already completed work is always a way to contribute. There are many ways to propose improved algorithms and models if you have an interest for mathematical modelling. Remember that even in well explored research fields, there is always space for creativity and advancement of interest.
It's better to start with a review paper on the latest research article in this field. In one paper (latest review paper), you can gain a clear idea of the work that has been done and the suggestions put forward by the authors (researchers) based on their investigation. This approach helps you understand the current state of the field and identify potential gaps or areas for further exploration.
In the biomedical field, preference should be given to applications that demonstrate effectiveness in promoting health and safety.
1. And, I would like to suggest that you integrate ML/DL techniques for EEG classification along with IoT or some real-time device, such as Jetson Nano or an equivalent.
2. EEG signals should have noise and limited spatial resolution. Maybe you can investigate.
3. Left and right hand movements generate distinct EEG signals. If you can collect a real dataset from reputable medical resources, you could investigate EEG signals in paralyzed individuals and analyze them.
I am sharing here some of the article maybe you can have a look, i feels that could help you better:
*) Current Status, Challenges, and Possible Solutions of EEG-Based Brain-Computer Interface: A Comprehensive Review.
*) A review on analysis of EEG signals.
*) Deep Learning Algorithm for Brain-Computer Interface.
*)Encyclopedia of Clinical Neuropsychology.
Finally, as this is your graduation thesis, it's important to have a backup plan. During research, numerous byproducts are often produced, many of which hold value. I hope you will successfully reach your final destination with this research. However, it's essential to keep proper track of your byproducts. They may prove invaluable in shaping your thesis and ensuring you graduate on time. Furthermore, even after graduation, consider continuing your research if possible.
  • asked a question related to Classification
Question
3 answers
During my research for my bachelors thesis into the classification of mining in the Reichsgruppe under the Nazi regime, I came across two different classifications:
  1. On the one hand, Boris Gehlen in his chapter "3.10 Energy industry" (in https://www.degruyter.com/document/doi/10.1515/9783110796353-016/html?lang=de) assigned mining to the Reichsgruppe "Energy Industry", which sounds logical to me.
  2. On the other hand, I have found some works that explain that mining belonged to the Reichsgruppe "Industry": https://www.degruyter.com/document/doi/10.1524/jbwg.1980.21.1.177/html
So which is correct?
I can only explain this apparent double classification by the central importance of mining for both economic sectors: On the one hand as a supplier of raw materials for industrial production and on the other as a key sector for energy supply.
Is it possible that in different sources and at different times the emphasis was placed more on one or the other affiliation, depending on which aspect of coal mining was in the foreground, so that it can be said that mining belonged to both groups?
Relevant answer
Answer
Chapter 3.10 by Gehlen deals with energy industry in general, not specifically with the Reichsgruppe "Energy Industry". Of course, energy industry is connected with mining. However, mining was organized as a "Wirtschaftsgruppe" (industry group) in "Hauptabteilung" (main department) I of the "Reichsgruppe Industrie" (see https://archivfuehrer-kolonialzeit.de/reichsgruppe-industrie-bestand?sf_culture=en and https://de.wikipedia.org/wiki/Reichsgruppe_Industrie). The Reichsgruppe "Industry" had 32 groups, whereas the much smaller Reichsgruppe "Energy Industry" had only two groups for "Electricity supply" and "Gas and water supply" (see https://de.wikipedia.org/wiki/Reichsgruppe_Energiewirtschaft).
  • asked a question related to Classification
Question
2 answers
What is scope of the implementing LIS classification and cataloguing in different field ?
Relevant answer
Answer
  • William Badke thanks for reply, I need to learn about the implementation of classification and cataloguing in different facets other than book. I have a vision adopting in other field, please have look and reply 'Adoption of classification and cataloging techniques in the Maintenance Industry' 10.18231/j.ijlsit.2023.022
  • asked a question related to Classification
Question
1 answer
Should I use the traditional UHI index classification (<0, 0-0.005, 0.005-0.010, 0.010-0.015, 0.015-0.020) where 99% of my area falls under UHI effect, or would a natural classification or classification like this (<0, 0-0.05, 0.05-0.10, 0.10-0.15, 0.15-0.20, >0.20) be more suitable for studying the urban heat island effect in our tropical region?
Relevant answer
Answer
Choosing between the traditional UHI index classification and the alternative classification (<0, 0-0.05, 0.05-0.10, 0.10-0.15, 0.15-0.20, >0.20) for studying the urban heat island effect in a tropical region depends on factors such as data resolution, local climate, research objectives, and the need for comparative analysis. Higher resolution data and focusing on nuanced temperature variations may favor the traditional classification. At the same time, broader patterns of UHI across more significant regions might be better captured by the alternative classification with wider intervals. Considering the study area's specific research goals and characteristics is crucial in selecting the most suitable classification scheme to provide meaningful insights into the UHI phenomenon.
  • asked a question related to Classification
Question
2 answers
ما هي معايير التصنيف الأكاديمي للجامعات؟
Relevant answer
Answer
Universities are classified based on various factors, but some common ones include:
Research: How much and how good is the research they do?
Teaching: How well do they teach?
Reputation: How well-known are they?
Students: How successful are their students?
Campus: What are the facilities like?
Remember, different organizations might focus on different aspects, so explore universities that fit your priorities.
  • asked a question related to Classification
Question
2 answers
I have trained all Convolutional Neural Networks (CNNs) from the LeNet-5 model to the EfficientNet model for Benign Tumors and Malignant Tumors for breast masses with large data. The data was Mammogram Images(MI). All of these models give me a test accuracy of 50 %. Why did most journals publish fake results?
Relevant answer
Answer
Fake results are never published in journals (In some cases, certain journal websites may be fake and their their documents are available). However, certain details are often omitted or only vaguely presented in research papers when researchers conduct experiments based on specific datasets and models. These undisclosed variables may include precise preprocessing procedures, hyperparameter adjustments such as learning rate modifications, and L1 and L2 regularization strategies. It's important to note that researchers actively strive to find optimal solutions before reaching their expected accuracy levels.
Before training the deep learning model, it is essential to split the dataset using proper split scripts into train, validation, and test sets (There can be a different range of amounts of split data that give the optimal solution). If you observe that the accuracy is not good enough, then you should consider improving the network architecture. If the same dataset or paper showed good accuracy with a particular network architecture, but you are unable to achieve similar results, you may consult with the authors to understand how they attained optimal accuracy. This can be done via email or another platform where they are accessible.
You can also follow books, reference papers, or online articles that precisely mention the kinds of aspects that can provide you with optimal solutions to your current problems. Some recommended resources include:
  1. "Deep Learning for Computer Vision" by Rajalingappaa Shanmugamani.
  2. The article "Improving Performance of Convolutional Neural Networks" available at https://medium.com/@dipti.rohan.pawar/improving-performance-of-convolutional-neural-network-2ecfe0207de7.
  3. Forums dedicated to TensorFlow or PyTorch, where you can find more details about your coding part.
  4. The blog post "37 Reasons Why Your Neural Network Is Not Working" by Slav Ivanov, accessible at https://blog.slavv.com/37-reasons-why-your-neural-network-is-not-working-4020854bd607.
  • asked a question related to Classification
Question
5 answers
These beautiful specimens of Cerambyx were photographed by me, (not captured)
in Abruzzo (Central Italy) loc. Sulmona. I was able to identify the genus but not the species.
Entomologists help me, I am looking for an exact classification.
Relevant answer
Answer
E' un piacere aiutare, per quello che mi è dato conoscere. Auguri per un nuovo anno ricco di soddisfazioni!
  • asked a question related to Classification
Question
3 answers
..
Relevant answer
Answer
Dear Doctor
"Compared to its predecessors, the main advantage of CNN is that it automatically detects the important features without any human supervision. This is why CNN would be an ideal solution to computer vision and image classification problems."
  • asked a question related to Classification
Question
7 answers
Supervised Learning
In supervised learning, the dataset is labeled, meaning each input has an associated output or target variable. For instance, if you're working on a classification problem to predict whether an email is spam or not, each email in the dataset would be labeled as either spam or not spam. Algorithms in supervised learning are trained using this labeled data. They learn the relationship between the input variables and the output by being guided or supervised by this known information. The ultimate goal is to develop a model that can accurately map inputs to outputs by learning from the labeled dataset. Common tasks include classification, regression, and ranking.
Unsupervised Learning
Unsupervised learning deals with unlabeled data, where the information does not have corresponding output labels. There's no specific target variable for the algorithm to predict. Algorithms in unsupervised learning aim to find patterns, structures, or relationships within the data without explicit guidance. For instance, clustering algorithms group similar data points together based on some similarity or distance measure. The primary goal is to explore and extract insights from the data, uncover hidden structures, detect anomalies, or reduce the dimensionality of the dataset without any predefined outcomes. Supervised learning uses labeled data with known outcomes to train models for prediction or classification tasks, while unsupervised learning works with unlabeled data to explore and discover inherent patterns or structures within the dataset without explicit guidance on the expected output. Both have distinct applications and are used in different scenarios based on the nature of the dataset and the desired outcomes.
Relevant answer
Answer
In the realm of machine learning, the main distinction between supervised and unsupervised learning lies in the nature of the dataset used for training.
Supervised Learning Dataset:
In supervised learning, the dataset consists of labeled examples, where each data instance is associated with a corresponding target or output value. The dataset includes both input features and the desired output or target variable. The aim of supervised learning is to learn a mapping function that can accurately predict the target variable based on the input features. The model is trained using labeled examples, allowing it to generalize and make predictions on unseen data. Common examples of supervised learning algorithms include linear regression, decision trees, and support vector machines.
Unsupervised Learning Dataset:
On the other hand, unsupervised learning involves unlabeled datasets, meaning they do not have corresponding target values. In this scenario, the model learns patterns, structures, or relationships within the data based solely on the input features. The objective of unsupervised learning is to discover inherent patterns or groupings within the data without prior knowledge of the desired output. Common unsupervised learning algorithms include clustering algorithms such as k-means clustering and dimensionality reduction techniques like principal component analysis (PCA).
  • asked a question related to Classification
Question
4 answers
I am looking for a high-resolution dataset (alternative to ImageNet) that has classes with sub-groups. I need this dataset for the domain transfer experiment. Basically, I will be using DNN pre-trained on ImageNet to extract features. Example: in CIFAR100, 100 classes are grouped into 20 super classes, such that each super classes have some sub-classes. I need similar dataset, but it has to be high resolution. Can you suggest any suggest one?
Relevant answer
Answer
FGVC-Aircraft, iNaturalist 2018, 2019, and 2021, WebVision-1000, PASS (ImageNet without Human)
  • asked a question related to Classification
Question
5 answers
I am using Google Earth Engine for LULC classification map. For this purpose I have used smile random forest classifier to classify the Landsat 7 Top of Atmosphere data. Now could you please tell me how can I validate the LULC classification map?
Relevant answer
Answer
Great job, you can validate in two ways:
1. use Google earth or
2. Ground truthing (actual fieldwork visit with selected georeferenced points)
  • asked a question related to Classification
Question
5 answers
What is the pixel classification for different land use in an NDBI map for hilly areas?
Relevant answer
Answer
Kindly review the literature to gain an understanding of the different threshold values, which may not necessarily apply to your study area.
It is also recommended that you apply NDBI on your imagery and verify the classes against values using Google Earth images and/or field verification.
  • asked a question related to Classification
Question
16 answers
Can i cluster documents to label them as a first step. Then in the second step, can I use the labelled documents to apply a classification method such as svm, knn, etc.?
Relevant answer
Answer
Classification and clustering are two methods of pattern identification used in machine learning. Although both techniques have certain similarities, the difference lies in the fact that classification uses predefined classes in which objects are assigned, while clustering identifies similarities between objects, which it groups according to those characteristics in common and which differentiate them from other groups of objects. These groups are known as "clusters".
Regards,
Shafagat
  • asked a question related to Classification
Question
1 answer
What is the most appropriate classification for the 12 identified compounds from the essential oil of a medicinal plant, which include:
1. Camphene,
2.Para-Cymene,
3. 1-Limonene,
4.Gamma-Terpinen,
5.Trans-Decalone,
6. Cuminic Aldehyde,
7. Cyclopentanone,
8. Acetyl phenyl carbinol,
9. 1-amino-1-o,
10. 5-Methyl-2-Phenyl indolizine,
11. Silicic acid,
12. 5-nitrobenzofuran-2-carboxylic
Is the categorization in the attached image correct?
Relevant answer
Answer
The categorization in the attached image is correct. The 12 identified compounds can be classified into the following groups:
Monoterpenes: Camphene, Para-Cymene, 1-Limonene, Gamma-Terpinen
Sesquiterpenes: Trans-Decalone
Aromatic aldehydes: Cuminic Aldehyde
Cyclic ketones: Cyclopentanone
Aromatic alcohols: Acetyl phenyl carbinol
Amines: 1-amino-1-o
Heterocycles: 5-Methyl-2-Phenyl indolizine, 5-nitrobenzofuran-2-carboxylic
Inorganic acids: Silicic acid
Monoterpenes and sesquiterpenes are the largest groups of compounds in the essential oil. They are hydrocarbons that are derived from isoprene, a five-carbon molecule. Monoterpenes have ten carbon atoms, while sesquiterpenes have fifteen carbon atoms.
Aromatic aldehydes are a group of compounds that contain an aldehyde group (-CHO) attached to an aromatic ring. They are often used as fragrances in perfumes and cosmetics.
Cyclic ketones are a group of compounds that contain a ketone group (=O) in a cyclic ring. They are often used as solvents and flavorings.
Aromatic alcohols are a group of compounds that contain an alcohol group (-OH) attached to an aromatic ring. They are often used as fragrances and flavorings.
Amines are a group of compounds that contain an amino group (-NH2). They are often used as solvents and pharmaceuticals.
Heterocycles are a group of compounds that contain a ring with at least one atom that is not carbon. They are often used as pharmaceuticals and agrochemicals.
Inorganic acids are a group of compounds that contain a hydrogen atom that is ionizable in water. They are often used as solvents and catalysts.
The classification of these compounds is important because it can help to predict their properties and applications. For example, monoterpenes are often volatile and have a strong odor, while sesquiterpenes are less volatile and have a more subtle odor. Aromatic aldehydes are often used as fragrances, while cyclic ketones are often used as solvents. Amines are often used as pharmaceuticals, while heterocycles are often used as agrochemicals. Inorganic acids are often used as solvents and catalysts. Mehdi Aghighi Shahverdi
  • asked a question related to Classification
Question
1 answer
Need help with an unsupervised deep image stacking project. Image stacking is a commonly used technique in astrophotography and other areas to improve the signal-to-noise ratio of images. The process works by first aligning a large number of short exposure images and then averaging them which reduces the noise variance of individual pixels. I have to do this process with neural networks by predicting a distortion field for each image and using a consistency objective that tries to maximize the coherence between the undistorted images in the stack and the final output. I need some learning materials for performing image stacking preferably in python and make a neural network. I already have experiences with training object classification and detection models and have worked on different YOLO models.
Relevant answer
Answer
Due to the specific characteristics and complicated contents of remote sensing (RS) images, remote sensing image retrieval (RSIR) is always an open and tough research topic in the RS community. There are two basic blocks in RSIR, including feature learning and similarity matching. In this paper, we focus on developing an effective feature learning method for RSIR. With the help of the deep learning technique, the proposed feature learning method is designed under the bag-of-words (BOW) paradigm.
Regards,
Shafagat
  • asked a question related to Classification
Question
2 answers
Hyperspectral Imaging, Hyperspectral Classification, Statistical Test
Relevant answer
Answer
Hi
There are several reasons why statistical tests might not be applied:
1. Sample Size and Variability.
2. Marginal Improvements.
3. Computational Complexity.
4. Focus on Other Metrics.
5. Methodological Preference.
6. Lack of Standardization.
generally considered good practice to apply statistical tests in such scenarios to rigorously validate the results
  • asked a question related to Classification
Question
2 answers
Hello everyone,
I have applied 1D CNN on speech emotion recognition, when I shuffled columns I got diffrent results, for example, using matrix(:,[1 2 3]) gives different classification results than matrix(:,[2 3 1]) which should be the same, I have tried rng("default") but it hasn't resolved the issue. Can someone please assist me with this?
Thank you in advance!
Relevant answer
Answer
Hamza Roubhi I appreciate your dedication to applying 1D CNN in the domain of speech emotion recognition and your commitment to addressing the issue concerning the variation in classification results when shuffling columns. As a fellow researcher with a background in signal processing, I understand the significance of consistency and reliability in the outcomes of such analyses.
When encountering discrepancies in results due to column shuffling, it is essential to ensure that the underlying data preprocessing and feature extraction methods remain consistent across different column arrangements. Validating the integrity of the data and confirming that the shuffling process does not introduce unintended variations can help maintain the robustness and reliability of the classification results.
Additionally, verifying the implementation of the CNN architecture, including the configuration of the layers, activation functions, and training parameters, is crucial in ensuring reproducibility and consistency in the classification outcomes. Reviewing the model's initialization procedures and ensuring that the randomization process aligns with the desired standards of consistency can potentially address the issue you are facing.
Furthermore, conducting thorough checks on the data partitioning and validation procedures, such as cross-validation techniques, can help identify any potential sources of variability that may arise due to the shuffled column arrangements. Ensuring that the model training and testing processes maintain consistency and adhere to standardized protocols can contribute to the stability and reliability of the classification results.
I recommend conducting a systematic evaluation of the data preprocessing, model architecture, and training protocols to pinpoint any potential sources of variation that may arise during the column shuffling process. By adhering to rigorous validation practices and ensuring the reproducibility of results across different column permutations, you can enhance the reliability and robustness of your CNN-based classification framework for speech emotion recognition.
  • asked a question related to Classification
Question
10 answers
Hi. I am currently working on two deep learning research as final semester undergraduate student. In order to ensure the quality and acceptability of my work in a well known journal paper can someone provide me any guideline, hints and tips. As for my previous experience i have worked in multiple conference papers.
Tips and hints can include this or anything else that you think is necessary:
1. What are the common mistakes I should avoid?
2. What should i always include and not include?
3. How do i choose a good journal to publish my work?
and etc etc.
Relevant answer
Answer
When writing a journal paper about deep learning and image classification, there are several key aspects to focus on to ensure acceptability and quality:
  1. Clear Problem Statement: Start by clearly defining the problem you're addressing in the context of image classification. Explain the significance of the problem and why it matters.
  2. Literature Review: Provide a comprehensive review of existing literature related to deep learning in image classification. This demonstrates your knowledge of the field and the context for your work.
  3. Methodology: Describe in detail the deep learning techniques, models, and algorithms you are using. Justify your choices and explain how they are suitable for the specific problem you are addressing.
  4. Data Collection and Preprocessing: Explain how you collected and preprocessed your image data. Discuss any data augmentation techniques, labeling processes, and data splitting for training and testing.
  5. Model Training: Describe the training process, including hyperparameter tuning, optimization algorithms, and any regularization techniques employed. Mention the hardware and software tools you used.
  6. Experiments and Results: Present your experimental setup, including metrics used for evaluation. Report results clearly, using tables, figures, and charts. Discuss not only the accuracy but also other relevant metrics like precision, recall, F1-score, etc.
  7. Discussion: Interpret the results and discuss their implications. Compare your findings to prior research and explain any differences. Address any unexpected results.
  8. Contributions: Highlight the novel contributions of your work. What sets your approach apart from previous work in the field?
  9. Challenges and Limitations: Acknowledge any limitations and challenges faced during the research. Discuss how these limitations might affect the generalizability of your findings.
  10. Ethical Considerations: If applicable, discuss the ethical considerations of using deep learning for image classification, such as bias, privacy, and transparency.
  11. Conclusion: Summarize the key findings and their implications. State the practical relevance of your work.
  12. References: Ensure proper citation of all relevant sources. Follow the citation style recommended by the journal.
  13. Clarity and Structure: Ensure the paper is well-structured with clear headings and subheadings. Use proper grammar, spelling, and formatting.
  14. Peer Review and Feedback: Seek feedback from peers and mentors. Peer review is crucial for identifying potential improvements.
  15. Originality and Plagiarism: Ensure your work is original and properly cited. Plagiarism is unacceptable in academic writing.
  16. Contribution to the Field: Emphasize how your research contributes to the broader field of deep learning and image classification. What new insights or applications does it offer?
  17. Readability and Flow: Write in a clear, concise, and organized manner. Ensure the paper flows logically from one section to the next.
By addressing these aspects and paying attention to detail, you can increase the acceptability and quality of your journal paper on deep learning and image classification. Additionally, it's important to follow the specific guidelines and requirements of the journal you plan to submit to, as they may have their own formatting and submission guidelines.
  • asked a question related to Classification
Question
2 answers
Dear Antonella Petrillo, Valerio Antonio Pamplona Salomon, Claudemir Leif Tramarico
I read your paper
State-of-the-Art Review on the Analytic Hierarchy Process with Benefits, Opportunities, Costs, and Risks
These are my comments
1- In the abstract you say “Aggregation approaches and outranking approaches are better classifications”
I agree with this classification better that “American vs. European schools”. For your information there are methods that apply both.
2- In page 2 “The choice of an MCDM method should be based on characteristics of the decision problem”
I also agree with this, but unfortunately, practically in all MCDM methods, some characteristics are ignored in the modelling due to the inability of methods to cope with them. For instance, resources and limitations, inclusive and exclusive alternatives, precedence, time, binary variables, etc.
In my opinion, the choice of a method is simply: Choose the MCDM method that best adjust to the characteristics of your problem.
3- “One main reason for the AHP’s leadership in MCDM is its solid mathematical foundation”
This is inexact. AHP does not have any mathematical foundation, except in the use of Eigen values.
Let’s see, why I say this. Do you think that there is mathematical foundation by:
a) Using pair-wise comparisons? No mathematical supportand it is a highly criticized procedure.
b) Assigning values to criteria based on intuition? Is this scientific, and what happens if other DM thinks different?
c) Accepting that the final decision of the DM is controlled by a formula, and forcing the DM to correct her/his own estimates? So, a formula, to get transitivity, supersedes the honest findings of the DM,
d) Assuming that criterion trade-offs are equivalent to criteria weights? These are two different concepts
e) Assumming that what is in the mind of the DM is applicable to the real-life, and thus accepting that it is also transitive? What kind of mathematics supports this?
f) Using a logarithmic table, the ‘Fundamental scale’, based on the Weber and Fechter laws, on stimulus and results, and then AHP comparing invented weights to stimulus?
The Dictionary defines stimulus as “Physiology, Medicine/Medical. something that excites an organism or part to functional activity”
Not even a remote relationship with the ‘weight’ concept.
g) AHP is unable to deal with complex scenarios, because its rigid lineal hierarchical structure that cannot represent transversal relationships.
Some AHP drawbacks were refuted by Dyer in the 90s. and that Saaty responded, but nothing can be extracted from those rebuttals. To be fair, rank reversal was discovered in AHP, but it is present in all MCDM methods, not only in AHP
4- You talk about BOCR as it were something new, when it started in the 50s, when the old C/B analysis was considered no longer appropriate.
Why the four criteria BOCR are mutually exclusive? Normally they are considered in the set of criteria. MCDM is not looking for optimality,since normally, it does not exist. All MCDM methods look for a balance between opposite criteria like B and C.
Exclusivity means that BOCR cannot be together and this is not realistic, nor practical, because it is a common feature in most scenarios. If you want more information, I will be glad to supply examples, albeit not using AHP
You are mistaken. A criterion can be used twice, for instance, a criterion asking for minimization, and the same criterion, with the same values, asking for minimization. I use it frequently. The software must find the equilibrium between those extreme values.
You talk about ‘important criteria’? And how do you select those criteria? Just by the weights values? There is not a mathematical support for that. It is intuitive, no more than that.
And what if there is redundancy? Which is the effect? From the mathematical point of view, none.
You are referring to AHP but at the same time make references to ANP.
There is a large difference, since the ANP structure is able to handle complex scenarios because it works with a network. Probably Saaty developed it reckoning the limitations of AHP.
5- Page 3 Figure 1. Sorry, but you cannot apply AHP to this problem; AHP theory explicitly says that all criteria MUST be independent, which is not the case in your example, quite the opposite, there are many transversal interrelationships.
6- In page 4 “First, this alternative may be too risky compared to alternatives one and two
Obviously, you do not consider that an alternative may be too risky, but also it may have some properties that compensate for this risk.
I am not referring to the compensation issue used in weights. The problem with AHP and other methods, is that elements of the decision matrix are considered in isolation, when in reality, according to System’s Theory and reasoning, they have to be considered as a whole, holistically. For instance, you can reduce risk by increasing costs or/and decreasing benefits. Therefore, you have to consider both at the same time.
I hope my comments may help
Nolberto Munier
Relevant answer
Answer
Dear Qamar
NM- On May 11, 2023 I started reading and commenting RG published papers on MCDM, and today or tomorrow I expect to publish in RG the hundredth analysis and comments. All of them are in RG.
It has been a very nice surprise to receive you comments, and I thank you for that.
Do you know how many responses I have received out of 99? Only 2
The first, months ago when the author withdrew the paper based on my observations.
The second? YOU.
Refusals and feedback from authors? None
Of course, there is no obligation to write or feed me back, but what really surprises me is that NOBODY defended or backed what he/she wrote, even when normally there are several authors. I believe that if a person writes something and somebody else refutes the writer, the normal procedure would be to answer and refute it or not. This happened in the 90s when Dyer commented on AHP, and Saaty and Harker published their refusal. The famous letters from both sides are in the Web
Nobody asked me to write and publish my comments on work of others, and I don’t ask or expect anything, just my desire to collaborate in the improvement of this discipline. I do it because I have dedicated the last 20+ years in studying and analysing MCDM, a fascinating subject for me, and in my opinion, this discipline is not pursued correctly. There are more that 200 MCDM methods, and all of them, are unable to address complex scenarios, with the possible exception of the grand dad of all methods, Linear Programming, created about 1940.
Qamar UI-Islam- Thank you for your detailed feedback and insightful observations on the "State-of-the-Art Review on the Analytic Hierarchy Process with Benefits, Opportunities, Costs, and Risks" by Antonella Petrillo, Valerio Antonio Pamplona Salomon, and Claudemir Leif Tramarico. Your perspective sheds light on critical aspects of the Analytic Hierarchy Process (AHP) and the broader context of Multiple Criteria Decision Making (MCDM) methodologies. Allow me to provide some clarifications and responses to the points you raised:
1- Classification Approach: I appreciate your agreement with the aggregation and outranking approach classification, emphasizing their relevance in the context of AHP and MCDM methodologies.
NM - AHP was a method created in the 50s., using a concept never employed before in decision-making, and that at said time worked well, because it uses the hierarchical structure that was been employed by enterprises since the Egyptians, or before. That is: “Do as I say, don’t argue”
Saaty developed his method when he was working for the US Army, and possibly adopted its hierarchical lineal and rigid structure for AHP.
Unfortunately, the method incorporates aspects that border the absurd when applied to MCDM, like the pair-wise comparison, developed by psychometrician L. L. Thurstone, in 1927.
You can apply it to decide which restaurant to go for dinner, or the place for your vacations, or, in the case of the military, considering different strategies and war equipment, but not in a problem like selecting the best location for an industry. Here, there are multiple aspects, all of them related, and on what the DM does not have any control, let alone authority.
Qamar UI-Islam- Choice of Method: Your insight into the selection of an appropriate method based on the problem's characteristics aligns with the practical considerations of matching the method to the specific nature of the decision problem.
NM - It is not mathematics, only common sense, like you don’t buy a suit if it does not fit you
Qamar UI-Islam- Mathematical Foundation of AHP: Your critical assessment of the mathematical underpinnings of AHP raises pertinent questions regarding the method's reliance on subjective pair-wise comparisons and the practical implementation of criteria trade-offs.
NM - I already commented about pair-wise comparisons. Regarding trade-offs, considered as criteria weights, it was only a Saaty assumption,as he described it in one of his writings. But trade-offs and weights have completely different meanings and use. The first one is a balance, while the second is a quantification of criteria. The fact that these weights, DETERMINED WITHOUT CONSIDERING THE ALTERNATIVES, is a fallacy. It can be easily demonstrated using entropy weights.
1. Qamar UI-Islam- BOCR Criteria: Your emphasis on the dynamic nature of criteria relationships and their potential for non-mutually exclusive interactions contributes to the discussion around the practical applicability of BOCR criteria and their role in decision-making processes.
NM- Thank you for your appraisal
2. Qamar UI-Islam- AHP Application Challenges: Your detailed evaluation of the constraints and limitations of applying AHP to complex decision problems underscores the need for a more comprehensive and adaptable approach, especially in scenarios with interconnected criteria.
NM- In reality, there are more than 100 researchers that addressed these issues in AHP, who I mention and identify by name in one of my books, and only a handful that are in favor of the method, also identified; therefore, they are not only my ideas; what I did is to enlarge and illustrate these fallacies
Regarding criteria interconnections, there are very few scenarios where criteria are independent. What really surprises me, after reading hundreds of published papers that ‘solve’ this type of problems using AHP, even in reputable journals, is that reviewers accept them. WHY?
By ignorance, negligence, or vested interests, because some reviewers are only interested in being mentioned as reviewers of a journal and mentioned in ‘Publons’, which is good for their CV?
It is impossible to ignore essential points like a primordial condition in AHP, clearly established by Saaty, about the necessity to have only independent criteria, or trade-offs equivalent to weights, and other evident misconceptions.
3- Qamar UI-Islam- Risk-Benefit Trade-Offs: Your emphasis on the interconnectedness of risks, benefits, and compensatory mechanisms within decision matrices highlights the significance of holistic and systems-based considerations in decision-making frameworks.
NM- Again, it is only common sense. We are not talking on Quantum Mechanics
Qamar UI-Islam -Your insights provide valuable perspectives for further examination and exploration of the Analytic Hierarchy Process and its implications within the realm of Multiple Criteria Decision Making. I appreciate your engagement with the subject matter and your dedication to advancing the understanding of decision-making methodologies.
NM- Thank you for your words, which I deeply appreciate. Maybe this communication may encourage other researchers to publish and discuss their concerns and why not, support for AHP and other MCDM methods.
Regards
Nolberto
  • asked a question related to Classification
Question
4 answers
Discover ways to convert
Using artificial intelligence teaching
From ad hoc to automatic classification
How can teachers?
Integrating artificial intelligence into their classrooms.
Relevant answer
Answer
The advancements in the development of artificial intelligence spread all over the world at a tremendous speed and create an incredible hype increasing our expectations. As a matter of fact, it is rather difficult to disappoint a user in an entertaining domain: an introduction of AI and neural networks is instantly gaining immense popularity (Prisma and FaceApp applications are good examples of that). In this article, we have compiled 9 ways to use artificial intelligence in education.
Regards,
Shafagat
  • asked a question related to Classification
Question
3 answers
Where can I find and download the following statistical data?
  1. 2012-2022, China, statistical data related to "dishonest persons subject to enforcement."
  2. 2012-2022, China, institutional documents related to the establishment of the credit system and punitive mechanisms.
  3. 2012-2022, China, classification statistical data for various types of patents.
  4. 2012-2022, China, classification statistical data for the publication of academic research papers.
  5. 2012-2022, China, classification statistical data for business registration.
  6. 2012-2022, China, statistical data related to innovation and entrepreneurship.
  7. 2012-2022, China, statistical data related to the population involved in innovation and entrepreneurship.
Relevant answer
Answer
Finding and downloading specific statistical data for such a wide range of topics and over a specific timeframe may not be straightforward, especially when dealing with data from China, which can have various restrictions and sources. However, I can provide you with some general guidance on where you might start your search for this type of data:
  1. National Bureau of Statistics of China (NBS): The NBS is the primary source for official statistical data in China. You can visit their website and search their databases or contact them directly to inquire about the specific data you need.
  2. Chinese Government Websites: Many Chinese government agencies publish statistical reports and data on their official websites. Try searching for relevant government departments, such as the State Intellectual Property Office (for patent data) or the Ministry of Science and Technology (for innovation and entrepreneurship data).
  3. Academic Databases: For academic research paper data, you can explore Chinese academic databases like CNKI (China National Knowledge Infrastructure) or Wanfang Data, which may provide access to academic publications.
  4. Business Registration Data: For business registration data, you may want to look into the China Enterprise Credit Information Publicity System or local government websites, as business registration data is often handled at the provincial or municipal level.
  5. Academic Institutions and Research Centers: Sometimes, universities and research centers in China may have collected and published data on innovation, entrepreneurship, and related topics. Explore the websites of these institutions.
  6. Third-Party Data Providers: Some companies and research organizations collect and provide data on various topics, including those related to China. These sources may offer subscription-based services for detailed data.
  7. International Organizations: International organizations like the World Bank, IMF, or the United Nations may have statistical data related to China. They often provide access to global statistics and data for various years.
Remember to respect copyright and licensing restrictions when using and sharing any data you obtain. Additionally, data availability and accessibility can vary, so you may need to contact the respective organizations or institutions for more specific and detailed data requests.
  • asked a question related to Classification
Question
1 answer
Classify different regions or countries based on their SOC levels and trends, and provide an analysis of how these classifications relate to the specific challenges and priorities they face in achieving the SDGs, particularly in combating land and soil degradation.
Relevant answer
Answer
Interesting topic! Classifying countries by Soil Organic Carbon (SOC) levels is key for addressing SDG challenges, especially in combating land and soil degradation. High SOC regions may focus on preserving soil quality, while low SOC areas might prioritize restoration efforts. These classifications help tailor interventions for specific challenges each region faces in achieving SDGs. Thoughts?
  • asked a question related to Classification
Question
5 answers
Hello researchers!
I am currently working on the classification of objects in an image, I want to use the multiLabel classification method with matlab.
So for 7 classes (groups) of objects I would like someone to offer me a code.
Relevant answer
Answer
The Multi-Label Image Classification focuses on predicting labels for images in a multi-class classification problem where each image may belong to more than one class.
Regards,
Shafagat
  • asked a question related to Classification
Question
2 answers
Hello researchers!
I am currently working on the classification of objects in the same image, I want to use the multiLabel classification method with matlab.
So for 7 classes (groups) of objects I would like to be offered a code to generate an internal table which displays "1" for an object present in the image and "0" otherwise. Likewise for multiple objects in the same image
Relevant answer
Answer
n multilabel classification, in contrast to binary and multiclass classification, the deep learning model predicts the probability of each class.
Regards,
Shafagat
  • asked a question related to Classification
Question
3 answers
Can any one let me know Classifications of Optimization Techniques are useful for Current research trends? How to choose our suitable Optimization Technique for our research problem.
Relevant answer
Thank you Hamza Magaji Makarfi and Alfonsas Misevicius ....still any other...
  • asked a question related to Classification
Question
1 answer
brain tumor detection and classification
Relevant answer
Answer
ok
  • asked a question related to Classification
Question
1 answer
I am looking for research focused on classifying Arabic text into ( Verb / Noun/ Letter )
most of what I found is stemming, deep learning stuff but not word classification.
Any help please ?!
Relevant answer
Answer
There is a lot of work being done, check out the survey by Meshrif Alruily "Classification or Arabic tweets a review"
  • asked a question related to Classification
Question
3 answers
I am currently working on a prediction-project where I am using machine learning classification techniques to do this.
I have already computed various classification metrics like accuracy, precision, recall, AUC-ROC, and F1 score. What I am struggling with is how to (objectively) interpret these metrics in terms of their quality. For instance, in frequentist statistics, there are established ranges for interpreting effect sizes (e.g., small, medium, and large effect sizes).
Is there a similar set of guidelines or conventions with a citable source for interpreting classification metrics? I am particularly interested in categories like "poor", "sufficient", "satisfactory", "good", and "excellent" or similar.
I understand that the context and the specific task are crucial for any interpretation, but I still need a citable source that provides general guidelines, especially because this involves an educational context.
Thank you in advance!
Relevant answer
Answer
It is difficult to find papers and reviews that justify the threshold of criteria to assess models there are some rules of thumb with range values, for example, for accuracy 50 % is not good it means that your values were predicted randomly, more the percentage is close to 100% better the accuracy is however around 100% meaning that your model overfits, for RMSE MSE lower the value is better your model is because it is translated that your model doesn't have errors. Each model is different according to the included variables. You can run many models changing some parameters and comparing the values between them acknowledging each specificity of the model ensemble learning method will be more accurate than a simple model. You need to select your model according to your research question.
  • asked a question related to Classification
Question
3 answers
Im trying to create an image classification model that classifies plants from an image dataset made up of 33 classes, the total amount of images is 41,808, the images are unbalanced but that is something me and my thesis team will work on using Kfold; but going back to the main problem.
The VGG16 model itself is from a pre-trained model from keras
My source code should be attached in this question (paste_1292099)
The results of a 15-epoch run is also attached as well
what I have done so far is changing the optimizers from SGD to Adam, but the results are generally the same.
Am I doing something wrong or is there anything I can do to improve on this model to get it to atleast be in a "working" state, regardless if its overfitting or the like as that can be fixed later.
This is also the link to our dataset:
It is specifically a dataset consisting of Medicinal Plants and Herbs in our region with their augmentations. The are not yet resized and normalized in the dataset.
Relevant answer
Answer
To enhance the performance of your VGG16 model during training and validation, you can start by applying data augmentation techniques to increase dataset diversity and reduce overfitting. It's crucial to ensure the dataset's cleanliness, correct labeling, and appropriate division into training, validation, and test subsets. Experiment with different learning rates and optimizers, and consider using learning rate schedulers if necessary. Employ regularization methods like dropout and L2 regularization to tackle overfitting issues. Keep a close eye on the training process, implement early stopping, and adjust the batch size as needed. You might also want to explore alternative model architectures or smaller models that could better suit your dataset. Lastly, make sure your hardware resources are utilized effectively, and explore ensemble methods to potentially enhance model performance. These strategies should help you overcome the low accuracy challenge with your VGG16 model.
  • asked a question related to Classification
Question
4 answers
Dear colleagues
Good day to you all. One of the most famous debates in Igneous Petrology is the relation between the diamondiferous rocks (i.e. lamprophyres, lamproites, orangeites and kimberlites). In 1991, the late geologist Nick Rock considered them to have similar petrological and geochemical signatures and were included in one group named the "Lamprophyre Clan". Recent publications have shown that relations do exist (e.g. see The "Lamprophyre Clan" Revisited 2022 paper in ResearchGate). The Version of Record is available online at: https://doi.org/10.1007/s12594-022-2153-4. One can also read the Version of Record through the Springer SharedIt link https://rdcu.be/cVljF (note that you need to use Wi-Fi in order to open the Springer SharedIt link).
On the other hand igneous petrologist Roger Mitchell, who disagreed with the idea, proposed the "Lamprophyre facies" concept which includes rocks that formed under volatile-rich conditions. Which one is correct? GPT-4 was also asked. The answer was that both terms can be correct, but they represent different perspectives in the study of these rocks. What is your opinion? Please comment.
Best regards
Ioannis Kamvisis
Relevant answer
Answer
Thank you for your answer Saleh.
Best regards
Ioannis
  • asked a question related to Classification
Question
2 answers
I would like to ask for help with the classification of native or invasive species, especially those that are classified as NATIVE in Brazil but are classified as INVASIVE in other countries.
For example: Euphorbia prostrata is a species native to Brazil, but considered invasive in other countries. What would be the correct classification for this species: NATIVE; INVASIVE; or NATIVE/INVASIVE?
This raises a question because the species is both native and invasive at the same time.
Relevant answer
Answer
The classification of a species as native or invasive is region-specific. For species like Euphorbia prostrata, which is native to Brazil but invasive elsewhere, a dual classification "NATIVE/INVASIVE" is appropriate. It's crucial to specify the geographical context: a species can be 'native' in its home region and 'invasive' when it proliferates beyond its natural range. To clarify, you can note: "Euphorbia prostrata - NATIVE in Brazil but INVASIVE in [specific countries]". This dual nature underscores the importance of tailored biodiversity management strategies for each region.
  • asked a question related to Classification
Question
3 answers
Hello:
I have a question about the definition of adaptive control, since I´m researching about making a model-free adaptive control system. I will appreciate your help.
The definition I found says that an adaptive control modifies its parameters or structure in order to achieve a performance index. Reading about the subject in different sources I noticed that when they refer to an adaptive controller, it always has a model of the plant and implies an adaptation law which is usually obtained by taking the model and manipulating expressions.
My deduction is that when these sources refers to adaptive control, is about a kind of this instead. Am I right? I will really appreciate your support on this.
Thanks.
Pablo.
Relevant answer
Answer
People do many things without adaptive control.
However, at some point, we may understand that the performance may not be what we were looking for. Either the plant parameters were not exactly what we thought, or they might have changed from what they were in the beginning. Moreover, the object of control could be time-varying, nonlinear, etc.
This is what led people to think about “adapting” the control parameter to the actual situation.
A first idea was to only change the main control loop gain. This was called the MIT rule and it worked nicely and was even implemented on a plane that flew nicely…until it crashed. Comes out that working with time-varying control parameters is not the same as fixed parameters.
This made people reluctant to the idea of adaptive control.
Then, some names, such as Prof. Kumpati (Bob) Narendra from Yale, contributed a lot to the theory of stability and also came up with their own idea of adaptive control. This appears in literature as Model Reference Adaptive Control (MRAC) and this is what you usually see in literature.
It requires the plant to behave like some ideal model of the same order and indeed, it tries to modify the parameters of the big plant.
As it happens, by 1980, after some engineering experience, I came to RPI (Rensselaer) for a PhD. The “problem” was that the Professor’s grant was for Large Flexible Structures (LFS). They were not only large plants but also MIMO, so there was not much we could do with the classical MRAC.
They started using another idea, which was called Command Generator Tracking (CGT). You don’t try to make all state variables of the big plant follow all state variables of some big model. Instead, you define some low-order model that only defines the trajectory that you would like your plant output to track. They had some publications, yet the idea was negatively received, so they were almost on their way out. However, the grant forced my advisor to let me “try to do something.”
As I was coming from the industry, I liked this idea of a simplified controller with varying parameters, and I even “felt” that it should work. The plant can be of order 20 or 50 and yet the model and the controller can be of order 5.
What can I say? I ended up with a low-order adaptive controller that managed to control the pretty big LFS. It was interesting that people at JPL liked the idea and even dared to implement it on their large flexible antenna.
After playing with it, I decided to call our approach “Simple Adaptive Control (SAC).”
I won’t make this message a dissertation, but if you write Simple Adaptive Control on Google, you can see quite a bit, of publications from all around the world.
Also, please always feel free to come back with any questions.
  • asked a question related to Classification
Question
3 answers
What are the applications of an embedded system and what is classification and requirements of embedded system?
Relevant answer
Answer
Dr Aparna Sathya Murthy thank you for your contribution to the discussion
  • asked a question related to Classification
Question
7 answers
Does the large number of comorbidities in current mental illness suggest that the current classification of illness is problematic? What diagnostic classification criteria do psychiatrists need to identify and treat disorders?
Relevant answer
Answer
Our project: Quantifying Mental health/// is based on mood science and the pathological chronification of moods as, for example, represented in the Plutchik’s Wheel of Emotions, dear Cai Jinping This approach works by ontological engineering, forensic simulation and key AI technologies.
cc
  • asked a question related to Classification
Question
1 answer
Classification of rosehip (Rosa canina L.) genotypes according to different usage purposes and further breeding objectives
  • July 2023
  • DOI:
  • 10.21203/rs.3.rs-3174428/v1
  • License
  • CC BY 4.0
  • Melekber Sulusoglu Durul
  • Kerem Mertoglu
  • Nazan Korkmaz
  • Show all 5 authors
  • İbrahim Bulduk
Dear Editor,
This article was rejected by this journal and we need to remove itfor re-load another journal. We tried to remove this pre-print from our pages. Unfortunately it seems again and again. Please could you help us to remove it completely?
My best regards..
Melekber SULUSOGLU DURUL
Relevant answer
Answer
See the end of the page https://explore.researchgate.net/display/support/How+to+make+content+private+or+remove+it. You should contact the ResearchGate's at www.researchgate.net/contact asking to delete this page.
  • asked a question related to Classification
Question
26 answers
Why CNN is better than SVM for image classification and which is better for image classification machine learning or deep learning?
Relevant answer
Answer
CNNs (Convolutional Neural Networks) are generally preferred over SVMs (Support Vector Machines) for image classification tasks due to their ability to automatically learn hierarchical features from images. Here's why CNNs are often considered better for image classification:
  1. Hierarchical Feature Learning: CNNs are designed to mimic the visual processing in the human brain, where early layers capture low-level features like edges and textures, and deeper layers learn more complex and abstract features. This hierarchical feature learning allows CNNs to adaptively extract relevant information from images, making them highly effective for capturing intricate patterns that are crucial for image classification.
  2. Spatial Hierarchy: CNNs utilize convolutional and pooling layers to maintain the spatial hierarchy of features, which is crucial for recognizing objects regardless of their position or orientation within an image. SVMs treat all features as independent, which can result in losing spatial relationships.
  3. Translation Invariance: CNNs inherently possess translation invariance, meaning they can recognize objects regardless of their position in the image. SVMs require explicit feature engineering to achieve similar invariance.
  4. End-to-End Learning: CNNs learn features directly from raw pixel values, eliminating the need for manual feature extraction. SVMs often require careful feature engineering to achieve good results.
  5. Scale and Complexity: CNNs can handle a wide range of image scales and complexities, making them suitable for tasks with varying levels of detail and object sizes. SVMs might struggle with complex images or those with high dimensionality.
Regarding whether machine learning or deep learning is better for image classification, the answer depends on several factors:
  1. Amount of Data: Deep learning models, particularly CNNs, tend to require a large amount of labeled data to generalize effectively. If you have a limited dataset, traditional machine learning algorithms might perform better due to their ability to work well with smaller datasets.
  2. Feature Complexity: Deep learning excels at automatically learning intricate and complex features from data, which is essential for tasks like image classification. If your task involves capturing subtle visual patterns, deep learning is often the better choice.
  3. Computational Resources: Deep learning models are computationally intensive and often require significant computational resources, especially during training. If computational resources are limited, traditional machine learning algorithms might be more feasible.
  4. State-of-the-Art Performance: Deep learning, particularly CNNs, has achieved state-of-the-art performance in various image classification benchmarks. If achieving the highest accuracy is a priority and you have the necessary resources, deep learning is likely the better choice.
In summary, CNNs are generally favored over SVMs for image classification due to their ability to automatically learn relevant features. Whether to choose machine learning or deep learning for image classification depends on factors like data availability, feature complexity, computational resources, and the desired level of performance. In many cases, deep learning, especially CNNs, has demonstrated superior performance on large and complex image datasets.
  • asked a question related to Classification
Question
7 answers
What deep learning algorithms are used for image processing and which CNN algorithm is used for image classification?
Relevant answer
Answer
Deep learning has revolutionized image processing, and Convolutional Neural Networks (CNNs) are the most commonly used algorithms for image classification tasks. CNNs are a class of deep neural networks specifically designed to process and analyze visual data, making them highly effective for image-related tasks. CNNs leverage the concept of convolution to automatically learn hierarchical features from images, capturing patterns and structures at various levels of abstraction.
Some popular CNN architectures used for image classification include:
  1. LeNet-5: One of the earliest CNN architectures, designed for handwritten digit recognition.
  2. AlexNet: Introduced in the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) in 2012, it demonstrated the power of deep learning on large-scale image datasets.
  3. VGG (Visual Geometry Group) Networks: Known for their simplicity and uniform architecture, VGG networks have various depths, with VGG16 and VGG19 being common variants.
  4. GoogLeNet (Inception): Introduced the concept of "inception modules" that allow the network to learn features at multiple scales simultaneously.
  5. ResNet (Residual Network): Addresses the vanishing gradient problem by introducing residual connections, enabling training of extremely deep networks.
  6. DenseNet: Each layer is connected to every other layer in a feed-forward fashion, promoting feature reuse and encouraging more efficient parameter utilization.
  7. MobileNet: Designed for mobile and embedded vision applications, it uses depth-wise separable convolutions to reduce computational complexity.
  8. EfficientNet: A family of models designed to achieve better accuracy and efficiency by optimizing model scale and resolution.
  9. Xception: An extension of the Inception architecture that replaces standard convolutions with depth-wise separable convolutions.
  10. SqueezeNet: Focuses on reducing model size while maintaining accuracy, making it suitable for resource-constrained environments.
These are just a few examples, and there are many other CNN architectures tailored for different tasks, including object detection, image segmentation, and more.
For image classification specifically, the choice of CNN architecture often depends on the complexity of the problem, available computing resources, and dataset size. More recent architectures like ResNet, DenseNet, and EfficientNet have demonstrated superior performance on large-scale image classification challenges due to their ability to handle deep networks and capture intricate image features.
  • asked a question related to Classification
Question
3 answers
What are the classifications of agrivoltaics and how Agrivoltaics: A smart farming strategy to boost farmers’ income?
Relevant answer
Answer
Agrivoltaics, also known as AgroPhotoVoltaics (APV), is a concept of co-locating agricultural activities and solar energy production in the same land area. In other words, it is the practice of integrating solar panels into agricultural land to generate renewable energy while also utilising the land for farming. Agrivoltaic farming is the simultaneous use of areas of land for both solar photovoltaic power generation and agriculture. It involves combining crops with photovoltaic panels to maximize land use. In other words, the practice of growing crops underneath solar panels. Systems can be classified into two categories, namely open agrivoltaic systems and the closed agrivoltaic system. Stand-alone PV Systems can be divided into three categories: Without Battery, With Battery, and Hybrid PV Systems. Without Battery systems are Direct-Coupled systems, and With Battery systems may include Self-Regulating DC Systems or AC Systems with a charge controller for the battery and load. Agrivoltaic systems combine, on one plot of land, crops and dynamic photovoltaic panels in order to optimize the land's potential without damaging ecosystems. Solar energy is converted into thermal energy using a special kind of heat exchanger known as a solar collector. Solar collectors can be classified into two general categories: (i) non-concentrating and (ii) concentrating. An agrivoltaic system (AVS) offers a potential strategy for meeting global demands for renewable energy and sustainability by integrating photovoltaics and agriculture. Many empirical studies have installed facilities and cultivated actual crops, revealing that AVSs improve land use efficiency.Smart agro farm is a low cost solar powered soil and weather monitoring system which analyses the different soil properties and weather conditions and thereby developing a hi-tech smart farm set up for farmers.
  • asked a question related to Classification
Question
3 answers
To achieve the desired power quality, the causes of disturbances are required to be identified and adjusted through the detection and classification of various power quality disturbances.
Relevant answer
Answer
Refer my Question 👍
  • asked a question related to Classification
Question
3 answers
Hi everyone, I'm working on skin cancer classification and I've extracted the feature from three pre-trained CNN models and concatenated all the features. Finally a dense layer with softmax function was used for classification.
But I encountered constant validation accuracy with lower training accuracy.
How can I solve this problem please. Despite this, I have tried many optimizer functions with different learning rate, regularizer, early_stopping, dropout, and also I used image augmentation
Relevant answer
Answer
Maybe this answer help you a little , , addressing the challenge of constant validation accuracy and lower training accuracy in your skin cancer classification project requires a systematic approach of experimentation and optimization. By carefully considering data quality, preprocessing, architecture, hyperparameters, regularization, and various techniques, you can work towards improving your model's performance. Be prepared to iterate and fine-tune your strategies based on the results you observe, and keep a keen eye on the interplay between different factors to achieve the best possible outcome for your classification task
  • asked a question related to Classification
Question
6 answers
Which machine learning algorithm is specifically designed for binary classification problem?
Relevant answer
Answer
The machine learning algorithm specifically designed for binary classification problems is the Logistic Regression algorithm. Despite its name, logistic regression is primarily used for classification tasks where the outcome or dependent variable is binary (i.e., it can take one of two possible classes, such as 0 or 1, True or False, Yes or No, etc.).
Logistic regression works by modeling the probability that an instance belongs to a particular class. It uses a logistic function (sigmoid function) to map predicted values to probabilities between 0 and 1. Based on these probabilities, it assigns the instances to the most appropriate class, usually using a threshold (typically 0.5).
Though the name might suggest it's a regression algorithm, logistic regression is widely used for binary classification due to its simplicity, efficiency, and interpretability. For multiclass classification problems (more than two classes), extensions like "Multinomial Logistic Regression" or other algorithms like Support Vector Machines (SVM), Decision Trees, and Random Forests can be used.
  • asked a question related to Classification
Question
5 answers
Hello community,
I am new to Random Forest. I understand how it is trained with random selection of features in each split, and so on. In the end we have n_trees, each of which will give a different estimate.
All codes and tutorials and papers I read so far (were not many, I confess) get solely one output, the average in case of regression or the most frequent class in case of classification.
I am very much interested in the distribution of values that all the n_trees give. Is there a theoretical reason why one should NOT do this? Is it conceptually not meaningful somehow?
In any case, does someone know how to get those values, if I want? I didn't find how to do this with R party and I'm currently still migrating to Python SKLearn.
Thank you very much and best regards!
Relevant answer
Answer
Random Forest is an ensemble learning method that combines multiple decision trees to make predictions. Each tree in the Random Forest provides an output or prediction for a given input, and the final prediction is typically obtained by averaging or voting over the individual tree predictions. However, it is possible to extract and analyze all the outputs from all trees in a Random Forest. There is no theoretical reason why one couldn't do this, but it may not always be necessary or practical for several reasons:
  1. Ensemble Averaging for Stability: The primary reason for using ensemble methods like Random Forest is to improve the stability and generalization of the model. By averaging or voting over multiple trees, Random Forest reduces the impact of individual noisy or biased trees and provides a more robust and accurate prediction. Analyzing all outputs separately may not be as robust and may lead to overfitting or increased variance in the predictions.
  2. Computational Overhead: Random Forests often consist of a large number of decision trees. Analyzing all outputs from all trees can lead to a significant computational overhead, especially for large datasets and deep trees.
  3. Interpretability and Complexity: Random Forests are often used in scenarios where interpretability and simplicity are important. Analyzing all individual outputs from all trees could lead to increased complexity, making it harder to interpret the results.
  4. Consensus Information: By averaging the outputs, Random Forest provides a form of consensus information. Analyzing individual tree outputs may not provide additional insights and might even introduce more noise.
  5. Bias-Variance Tradeoff: Analyzing individual tree outputs might lead to a higher variance in predictions, potentially increasing the risk of overfitting and reducing the model's generalization ability.
While there are valid reasons to analyze individual tree outputs, Random Forests' strength lies in their collective decision-making process. The focus is on the overall performance and stability of the ensemble, rather than individual tree outputs. However, depending on the specific use case, there might be situations where analyzing individual tree outputs is beneficial, such as in understanding the uncertainty in predictions or performing model diagnostics. In such cases, techniques like permutation importance or partial dependence plots can be used to interpret the individual tree contributions.
  • asked a question related to Classification
Question
3 answers
Seeking insights from the research community: Does the imbalance of textual genres within corpora, when used as an explanatory variable rather than the response variable, affect the performance of logistic regression and classification models? I'm interested in understanding how the unequal distribution of word counts across genres might introduce biases and influence the accuracy of these machine learning algorithms. Any explanations or relevant details on mitigating strategies are appreciated. Thank you!
Relevant answer
Answer
Indeed, the whimsical dance of textual genres within corpora can sway the fate of logistic regression and classification models. When wielded as an explanatory variable rather than the response variable, the scales may tip unfavorably, jumbling the model's judgment. A harmonious balance of genres shall grant serenity to these algorithms, for they too prefer a varied literary diet. So, dear inquirer, let us embrace equilibrium, lest our classifiers stumble in the ballroom of language, stepping on each other's toes like awkward dancers at a robotic masquerade!
  • asked a question related to Classification
Question
14 answers
I need to know about the deep learning algorithms used in land cover classification and which one is best suited. Planning to use Sentinel-2 satellite images.
I also want to know about GAN in Landcover classification.
Relevant answer
Answer
The choice of the best deep learning algorithm for land cover classification depends on several factors, including the characteristics of the data, the complexity of the classification task, the available computational resources, and the desired level of accuracy. As of my last update in September 2021, some of the commonly used deep learning algorithms for land cover classification are:
  1. Convolutional Neural Networks (CNNs): CNNs are widely used for image-based tasks, including land cover classification. They are particularly effective at automatically learning spatial patterns and features from input images. CNNs have shown excellent performance in tasks like object recognition and image segmentation, making them well-suited for land cover classification where pixel-wise labeling is required.
  2. U-Net: U-Net is a variant of CNNs specifically designed for semantic segmentation tasks, where each pixel in the input image is assigned a class label. U-Net's architecture incorporates both downsampling and upsampling pathways, making it efficient at capturing spatial context and accurately delineating boundaries between different land cover classes.
  3. Recurrent Neural Networks (RNNs): RNNs are designed to process sequential data, and they can be useful for classifying time series data in land cover applications, such as analyzing vegetation changes over time. However, for most land cover classification tasks, CNN-based architectures tend to be more commonly used.
  4. Deep Belief Networks (DBNs): DBNs are generative models that have been used for land cover classification, especially in cases where unsupervised learning or feature learning is desired. However, they have been largely overshadowed by the success of CNNs in recent years.
  5. Transfer Learning with pre-trained models: Many deep learning models pre-trained on large-scale image datasets (e.g., ImageNet) can be fine-tuned for land cover classification tasks. Transfer learning allows leveraging the knowledge learned from these large datasets and applying it to land cover classification, even with limited labeled data.
  6. Attention-based models: Attention mechanisms have been introduced to deep learning architectures to focus on relevant parts of the input data. They can improve the performance of land cover classification by allowing the model to emphasize important spatial regions in the images.
Ultimately, the most suitable deep learning algorithm depends on the specific land cover classification task and the characteristics of the available data. It's essential to consider factors such as the size of the dataset, the spatial and spectral resolution of the imagery, and the presence of temporal data. Additionally, it's beneficial to experiment with different architectures, hyperparameters, and training strategies to determine the best-performing model for a given land cover classification project.
  • asked a question related to Classification
Question
8 answers
What are approaches to classify patent data using deep learning? (document, text, word, labels )
How patent classification using CNN, DNN, RNN ?
Is transfer learning is effective in patent classification?
Relevant answer
Answer
patent classification using deep learning is typically done:
  1. Data Preparation: The first step is to gather a large dataset of labeled patent documents, where each patent is associated with the correct classification label (class or subclass). The dataset needs to be well-prepared, cleaned, and formatted to be suitable for training the deep learning model.
  2. Text Representation: Patents are usually represented as text documents, and deep learning models require numerical input. To convert text into a numerical representation, techniques like word embeddings (e.g., Word2Vec, GloVe) or more advanced methods like transformer-based models (e.g., BERT, GPT) can be used. These methods capture the semantic meaning of words and their relationships within the patent text.
  3. Deep Learning Model: The core of the patent classification process is a deep learning model. Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and transformer-based models are commonly used architectures for this task. These models can handle the sequential nature of patent text and learn complex patterns within the data.
  4. Training: The labeled dataset is divided into training and validation sets. The deep learning model is then trained on the training set by adjusting its internal parameters iteratively to minimize the classification error. This process involves backpropagation and optimization algorithms (e.g., gradient descent) to update the model's weights.
  5. Evaluation: The performance of the trained deep learning model is assessed using the validation set. Metrics like accuracy, precision, recall, and F1 score are commonly used to evaluate the model's classification performance.
  6. Fine-tuning (Optional): Depending on the results from the evaluation, the model may be fine-tuned with different hyperparameters, architectures, or training techniques to achieve better performance.
  7. Deployment: Once the deep learning model demonstrates satisfactory classification performance, it can be used to automatically classify new, previously unseen patents into their appropriate classes or subclasses.
  • asked a question related to Classification
Question
2 answers
What is international patent classification?
Relevant answer
Answer
The International Patent Classification (IPC), established by the Strasbourg Agreement in 1971, is a hierarchical system of language-independent symbols for the classification of patents and utility models according to the different areas of technology to which they pertain.
Administered by the World Intellectual Property Organization (WIPO), the IPC divides technology into eight sections with approximately 70,000 subdivisions. Each subdivision is assigned a symbol consisting of Arabic numerals and Latin letters. This classification system is used by over 100 countries and forms an integral part of the patent documents.
The main purposes of the IPC are to establish an effective search tool for retrieving patent documents, to facilitate the sharing of patent examination results between offices, to assist in the preparation of industrial property statistics, and to aid in technology analysis. It is updated every five years to incorporate developments in technology.
  • asked a question related to Classification
Question
1 answer
Like what are the sexquioxide classification of laterite?
Relevant answer
Answer
The sexquioxide classification of laterite refers to the classification of laterite soils based on their content of sesquioxides, which are oxides with three atoms of oxygen and two metal atoms. Laterite is a soil and rock type that is rich in iron and aluminum, and is commonly found in hot and wet tropical areas. It forms through the intensive and prolonged weathering of the underlying parent rock, usually under conditions of high temperatures and heavy rainfall with alternating wet and dry periods. The sexquioxide classification of laterite is used to describe the composition of these soils, which can vary depending on factors such as the parent rock, the degree of weathering, and the local climate.
  • asked a question related to Classification
Question
1 answer
Hello everyone, I am very curious about the out of plane and in plane of SU8 photoresist in the material classification. What is the difference between them?
Relevant answer
Answer
Hello. I have searched the web for some relevant information that might help you.
SU8 photoresist is a high aspect ratio epoxy-based negative photoresist commonly used as structural material in lithographic fabrication1. It has some remarkable properties, such as a wide range of coating thicknesses, almost straight sidewalls, high chemical resistance, and good mechanical properties1.
The difference between out of plane and in plane of SU8 photoresist lies in the way the material behaves when subjected to stress. In the out-of-plane direction, the SU8 layer is unconstrained and shrinking can occur freely. A shrinkage of approximately 4.5% (std 0.25%) was measured over a post-exposure bake (PEB) of 1 h at 65°C1. In-plane, although some relaxation occurs, an internal stress is built up (~4 MPa for a PEB temperature of 95°C in a standard process1).
I hope this answer is helpful to you. If you have any further questions, please feel free to ask. 😊
  1. https://link.springer.com/referenceworkentry/10.1007/978-90-481-9751-4_360
  2. https://cns1.rc.fas.harvard.edu/facilities/docs/SOP031_r2_6_SU-8%20photolithography%20process.pdf
  3. https://mdpi-res.com/d_attachment/micromachines/micromachines-11-00317/article_deploy/micromachines-11-00317.pdf?version=1584539406
  • asked a question related to Classification
Question
1 answer
please share the physicochemical properties of mirogabalin besylate.
Relevant answer
Answer
I studied pharmacology as an undergraduate and would like to answer your question.
Mirogabalin belongs to BCS class 1 due to its high solubility and membrane permeability.
According to the pharmacological specifications for TARLIGE TABLETS written in Japanese, milogabalin besilate is soluble in methanol, ethanol, and water. It also undergoes membrane transport via amino acid transporters, which is thought to be the origin of its relatively high permeability.
Other physicochemical properties include its melting point of 169°C and it does not appear to be hygroscopicity.
  • asked a question related to Classification
Question
1 answer
As a researcher, image classification is an area where images are the primary data used in various domains such as agriculture, health, education, and technology. However, ethical considerations are a significant concern in this area. How should ethics be handled and taken into account?
Relevant answer
Answer
When using image classification models in real-world applications, several ethical considerations arise. Here are some of the key ethical considerations:
  1. Bias and Fairness: Image classification models can inherit biases from the training data, leading to unfair outcomes and potential discrimination. These biases can disproportionately affect certain groups based on factors such as race, gender, or age. It is crucial to ensure that the training data is diverse and representative of the population to mitigate bias and promote fairness.
  2. Privacy and Consent: Image classification models often require access to personal images or video data. Respecting individuals' privacy rights and obtaining informed consent is essential. It is necessary to clearly communicate the purpose, scope, and potential risks of data collection and usage. Transparency in data handling practices and providing options for individuals to opt out or have their data removed are crucial for maintaining ethical standards.
  3. Security and Protection: Image classification models may deal with sensitive or private images, such as medical or biometric data. It is imperative to implement robust security measures to protect this data from unauthorized access, breaches, or misuse. Adhering to industry best practices and regulations, such as encryption, access controls, and secure storage, helps ensure data protection.
  4. Accountability and Transparency: Developers and organizations utilizing image classification models have a responsibility to be accountable for their actions. This includes being transparent about the model's capabilities, limitations, and potential biases. Users should be informed about the decision-making process behind the model's predictions, enabling them to understand and challenge the outcomes if necessary.
  5. Impact on Society: Image classification models can have broad societal impacts. They can reinforce stereotypes, perpetuate biases, or contribute to social inequalities. Understanding and mitigating these impacts is crucial to ensure that the technology benefits all individuals and does not exacerbate existing disparities.
  6. Algorithmic Governance and Regulation: As image classification models become more prevalent and influential, there is a need for appropriate governance and regulation. Policymakers and regulatory bodies should ensure that ethical considerations, fairness, and accountability are addressed in the development, deployment, and use of image classification models.
  7. Human Oversight and Decision-Making: While image classification models can automate decision-making processes, it is important to maintain human oversight. Critical decisions should not be solely reliant on the model's predictions. Human judgment and intervention are necessary to address complex ethical dilemmas, interpret the context, and consider individual circumstances.
  • asked a question related to Classification
Question
6 answers
Why is agricultural diversification essential for sustainable livelihoods and what are the classification and characteristics of agriculture market in India?
Relevant answer
Answer
Diversification of crop production involves a shift from a single-cropping system to a multi-cropping system. Diversification involves a shift in cropping patterns from food grains to cash crops. The main aim is to promote a shift from subsistence farming to commercial farming. Therefore, expansion into other sectors is essential to provide supplementary gainful employment and in realizing higher levels of income for rural people to overcome poverty and other problems. Diversification helps mitigate the risk to you about such scenarios by choosing different investments and types of investments. Diversification doesn't guarantee investment returns or eliminate risk of loss including in a declining market. Diversification includes two aspects, one relates to diversification of crop production and the other relates to a shift of workforce from agriculture to other allied activities and non-agricultural sector. Diversification of crop production refers to a system of multiple cropping rather than mono cropping. It may also mean a shift from subsistence farming to commercial farming. It has the three advantages and lowers the risk of farmer on account of failure of monsoon. Agricultural diversification occurs when more species, plant varieties or animal breeds are added to a given farm or farming community and this may include landscape diversification different crops and cropping systems interspersed in space and time. Resource factors include a soil fertility status, rainfall and irrigation facility which influences the crop diversification. The agricultural diversification implies diversification of crop production and shifting of agricultural workforce to other allied activities such as livestock, poultry, fisheries, etc and non-agricultural sector. On the basis of their transactions, the agricultural markets can be classified as spot and forward markets. The spot markets undertake those transactions only in which the exchange is affected at the current prices; while in the forward markets, the commodities are traded for future delivery. The process of concentration, equalization and dispersion is a peculiar characteristic of agricultural marketing. To get products from growers into the hands of distant consumers involves three important and related processes which may be called concentration, equalization and dispersion. Agricultural marketing is a process that involves assembling, storage, processing, transportation, packaging, grading and distribution of different agricultural commodities across the country. The agricultural marketing system is a link between the farm and the non-farm sectors. Agricultural marketing is the study of all the activities, agencies, and policies involved in farmers procuring farm inputs and agricultural products moving from farms to consumers. The agricultural marketing system serves as a conduit between the agricultural and non-agricultural sectors Indian Agriculture Market Outlook. The Indian agriculture market reached a value of about USD 435.9 billion in 2022. The market is further expected to grow at a CAGR of nearly 4.9% between 2023 and 2028 to reach a value of approximately USD 580.82 billion by 2028. Agricultural marketing covers the services involved in moving an agricultural product from the farm to the consumer. These services involve the planning, organizing, directing and handling of agricultural produce in such a way as to satisfy farmers, intermediaries and consumers.
  • asked a question related to Classification
Question
14 answers
I am starting with AI, training, detection and classification of small objects like loosened nut, bolts etc on a railway track. What should be the ideal way to collect dataset ( images )?
1. Does background matters in pictures, shall I use multiple background
2. Shall all images come from same height but different angles, or varying height is also important.
3. Or top angle from various heights is more effective
And what should be the best way to train the model?
1. with any dataset, iteratively training the model, by manually correcting every incorrect inference
2. or using a very large dataset without iterations
Relevant answer
Answer
The dataset plays a crucial role in the performance of an AI model for the training, detection, and classification of small objects like loosened nuts and bolts on a railway track. Here's how the dataset affects model performance and considerations for collecting an ideal dataset:
  1. Quantity of Data: Having a sufficient quantity of data is important for training an AI model effectively. Adequate data ensures that the model can learn diverse patterns and variations of small objects, improving its generalization capabilities. Collecting a substantial number of images that capture various instances of loosened nuts, bolts, and other small objects is crucial.
  2. Quality and Diversity of Data: The dataset should be of high quality, with clear and well-captured images of the small objects of interest. It is important to include a wide variety of scenarios, lighting conditions, angles, and perspectives to ensure the model can handle different real-world situations. Including images from different railway tracks and environments can enhance the model's ability to generalize and detect objects accurately.
  3. Annotation and Labeling: Proper annotation and labeling of the dataset are essential. Each image should be annotated to indicate the presence and location of small objects like loosened nuts, bolts, etc. Accurate and detailed annotations provide the necessary ground truth for training the AI model and enable it to learn the specific characteristics and features of the objects.
  4. Balanced Classes: The dataset should have a balanced distribution of positive and negative examples. It is important to include a sufficient number of positive samples (images with the small objects) as well as negative samples (images without the small objects) to avoid biases and improve the model's ability to distinguish between classes accurately.
  5. Data Augmentation: Data augmentation techniques can help increase the diversity and size of the dataset by applying transformations such as rotation, scaling, flipping, and adding noise to the images. Data augmentation can help prevent overfitting, improve generalization, and enhance the model's ability to handle variations in real-world conditions.
  6. Data Collection Process: When collecting the dataset, consider the following factors:Ensure safety protocols are followed when capturing images on railway tracks. Capture images from different distances and perspectives to simulate real-world scenarios. Include images with varying levels of occlusion, clutter, and lighting conditions. Include positive examples that cover different types and degrees of loosened nuts, bolts, etc.
  7. Validation and Testing Set: It is important to split the dataset into training, validation, and testing sets. The validation set helps in tuning hyperparameters and monitoring the model's performance during training, while the testing set provides an unbiased evaluation of the model's performance after training.
Additionally, it is recommended to seek expert knowledge from railway track professionals or domain experts to guide the collection and annotation process effectively. Regular updates and additions to the dataset over time can further improve the model's performance as it encounters new variations and scenarios.
  • asked a question related to Classification
Question
5 answers
I want to annotate each gene in the Homo sapiens taxon with its respective GO terms and its hierarchical parent terms in the GO database. How can I systematically do that? While I am aware that the obo file contains information such as "is a," "part of," and "regulates," it lacks a comprehensive hierarchy from child GO terms to all their parent terms. Is there an existing method available to achieve this systematic annotation, or do I need to develop a custom script to extract this information from the obo file?
Relevant answer
Answer
Mohammad Shahbaz Khan Certainly! Although the data is currently presented in Gene Ontology (GO) format, I want to create a comprehensive graph that visualizes the entire information. Further, I intend to annotate each gene with its corresponding GO term, including all parent terms associated with each gene.
  • asked a question related to Classification
Question
8 answers
I want to solve imbalanced data issue for classification problem, which technique is probably more effective to solve it?
Relevant answer
Answer
Dear All,
The question posed by Viswapriya Elangovan is very complex, and there is a vast literature attempting to provide an answer.
Inès François makes valid points but does not directly address the specific question regarding new techniques to address (and overcome) the imbalanced class problem in Data Science.
On the other hand, Oger Amanuel's response is cryptic and not very helpful: dividing features into classes is a pointless operation, unless Oger meant something else and got confused. If confused with the terminology, what he suggests is redundant.
However, to clarify, it is not easy to determine when one is dealing with the problem of imbalanced classes in Data Mining. Typically, an imbalance index is calculated as the ratio between the elements of the positive class (minority) and the negative class (majority) (there are also other ways to calculate the index, such as using entropy). It is not known at what threshold of the index one can speak of an imbalanced class, but certainly, based on my experience, 85% and 15% do not pose a problem for the majority of machine learning algorithms, and Ines does not bring anything new to the table.
The approaches to overcome the imbalanced class problem (for example, if less than 1% of the data points in the dataset belong to the positive class) fundamentally remain two:
1) Data resampling: random undersampling (with various sub-samplings of the majority class, and then MCC can be used to determine which one to consider), oversampling with the creation of synthetic data, combining under and over.
2) Cost-sensitive: using a cost matrix during the learning phase.
Other techniques are less general and may depend on the nature of the problem and the data being analyzed.
  • asked a question related to Classification
Question
3 answers
Is there any clear references to classify the risk level of land subsidence based on annual subsidence rate? Such as 30mm/year or 50mm/year? Much thanks for your time.
Relevant answer
Answer
There are no universally agreed-upon specific standards for the classification of land subsidence levels based on annual subsidence rates. The classification and risk assessment of land subsidence often depend on various factors such as the local geological conditions, the magnitude and rate of subsidence, and the potential impact on human activities or infrastructure. Different regions or organizations may have their own criteria for classifying subsidence levels and assessing the associated risks. It is recommended to consult local geological surveys, engineering organizations, or relevant authorities for specific references and guidelines regarding the classification of land subsidence and its corresponding risk levels.
  • asked a question related to Classification
Question
1 answer
I would be thankful.
Relevant answer
Answer
If one searches for the title, one comes upon this link:
  • asked a question related to Classification
Question
5 answers
In Bio-Signals and Systems we are introduced with quite a number of Biosignals, but what are some classification methods for those signals? Is using CNN one of the classification methods for biomedical signals?
Relevant answer
Answer
There are several classification methods commonly used for biomedical signals. Some of them include:
  1. Support Vector Machines (SVM): SVM is a popular machine learning algorithm used for classification tasks. It works by finding an optimal hyperplane that separates different classes of data points.
  2. Random Forest: Random Forest is an ensemble learning method that combines multiple decision trees to make predictions. It can handle high-dimensional data and is often used for classification tasks in biomedical signal analysis.
  3. k-Nearest Neighbors (k-NN): k-NN is a simple yet effective classification algorithm. It classifies new data points based on the class labels of their k nearest neighbors in the feature space.
  4. Artificial Neural Networks (ANN): ANN, including deep learning architectures such as Convolutional Neural Networks (CNNs), are widely used for biomedical signal classification. CNNs can learn hierarchical features from raw data and have shown promising results in various biomedical signal analysis tasks.
  5. Hidden Markov Models (HMM): HMM is a statistical model widely used for analyzing sequential data. It has been applied to classify biomedical signals that exhibit temporal dependencies, such as electrocardiograms (ECG) and electroencephalograms (EEG).
  6. Decision Trees: Decision trees are simple yet powerful classification models. They build a tree-like model of decisions and their possible consequences based on features of the data.
  7. Ensemble Methods: Ensemble methods, such as AdaBoost and Gradient Boosting, combine multiple classifiers to improve classification accuracy. They can be used with various base classifiers, including those mentioned above.
It's worth noting that the choice of classification method depends on the specific characteristics of the biomedical signals, the size of the dataset, and the objectives of the analysis.
  • asked a question related to Classification
Question
1 answer
In the case of EMG, motor imagery, or other most of biosignal classification problems, the accuracy improves when more features are added. However, in the case of SSVEP signal classification, everyone is using only one method either MEC or CCA or FFT or PSD. Can we add more frequency domain features with MEC to further improve the results?
Relevant answer
Answer
The Minimum Energy Combination (MEC) spatial filter method can be combined with other techniques to improve SSVEP detection classification accuracy. One such technique is the Common Spatial Pattern (CSP), which is a well-known spatial filtering technique used for EEG signal processing.
The combination of MEC and CSP can be used to improve the classification accuracy of SSVEP detection. The MEC method is used to estimate the frequency components of the SSVEP signals, while the CSP method is used to improve the signal-to-noise ratio and separate the relevant signals from the irrelevant ones.
By combining the MEC and CSP methods, the resulting spatial filters can better extract the relevant SSVEP signals from the EEG data, leading to improved classification accuracy. This approach has been shown to be effective in several studies and is a promising direction for further improving SSVEP detection classification accuracy.
  • asked a question related to Classification
Question
3 answers
Hello everybody, I'm a master degree student. I'm working with 16S data on some environmental samples. After all the cleaning, denoising ecc... now I have an object that stores my sequences, their taxonomic classification, and a table of counts of ASV per sample linked to their taxonomic classification.
The question is, what should I do with the counts for assessing Diversity metrics? Should I transform them prior to the calculation of indexes, or i should transform them according to the index/distance i want to assess? Where can I find some resources linked to these problems and related other for study that out?
I know that these questions may be very simple ones, but I'm lost.
As far as I know there is no consensus on the statistical operation of transforming the data, but i cannot leave raw because of the compositionality of the datum.
Please help
Relevant answer
Answer
Assessing diversity metrics in 16S data is an important step in analyzing microbial communities. Handling count data in this context can be challenging due to the compositional nature of the data, as you mentioned. While there is no one-size-fits-all approach, there are several techniques and considerations you can explore. Here are some suggestions:
  1. Transformations for diversity metrics: The choice of transformation depends on the diversity metric you want to assess. Common transformations include rarefaction, normalization (e.g., by library size or cumulative sum scaling), or transformations that aim to address compositionality, such as log-ratio transformations (e.g., centered log-ratio, clr transformation) or Hellinger transformation. Different transformations may be more suitable for specific diversity metrics, so it's essential to consider the metric's assumptions and properties.
  2. Compositional data analysis (CoDA): Compositional data analysis provides a statistical framework to analyze and interpret compositional data. It accounts for the constrained nature of relative abundance data by working on transformed data. CoDA methods, such as ALDEx2 or ANCOM, can help identify differentially abundant features between groups while considering the compositional structure.
  3. Multivariate analyses: If you want to explore the overall community structure and relationships, multivariate techniques like principal component analysis (PCA), correspondence analysis (CA), or non-metric multidimensional scaling (NMDS) can be employed. It's advisable to perform these analyses on transformed data to mitigate the effects of compositionality.
  4. Research articles and resources: To delve deeper into the subject, you can refer to scientific articles and resources that discuss the statistical analysis of 16S data. Some useful references include: "Microbiome Analysis Methods" by Paul J. McMurdie and Susan Holmes. "A guide to statistical analysis in microbial ecology: a community-focused, living review of multivariate data analyses" by Egoitz Martínez-Costa et al. "Statistical analysis of microbiome data with R" by Yinglin Xia et al. "MicrobiomeSeq: An R package for analysis of microbial communities in an environmental context" by Paul McMurdie and Susan Holmes. These resources provide insights into various statistical approaches, transformations, and analysis techniques for 16S data.
Remember that there is ongoing research in the field, and best practices continue to evolve. It's important to critically evaluate the methods, consider the specific characteristics of your data, and consult with your advisor or peers with expertise in microbiome analysis to make informed decisions about data transformations and diversity metric assessment.
  • asked a question related to Classification
Question
4 answers
Dear community!
I am currently searching for a text dataset and maybe you can point me in the right direction.
I am looking for a dataset that possibly covers a medical problem (not necessarily), addresses (multilabel) document classification and contains some numeric values in the text. Ideally, the numeric values (e.g. measurements) have an influence on the classification labels (e.g. label is 'fever' when temperature in the text is above X). Of course, large data and many labels would be great, but I am thankful for interesting suggestions.
Thank you!
Roland
Relevant answer
Answer
Emergency departments (EDs) in hospitals are usually overcrowded by patients with various severity of health problems1. Prompt diagnosis and therapeutical decisions in such streaming environment might not be always based on the best available expert opinions. For example, patients with potential limb fractures may be treated by only examining the radiology images and prior to the availability of reports written by the radiologists. The manual reconciliation of the initial diagnosis with the formal radiology report usually occurs after the patient has been discharged from the ED2. As a result, timely follow-up of patients with abnormalities in their reports is a critical task.
Regards,
Shafagat
  • asked a question related to Classification
Question
8 answers
few short learning nowadays grooming for classification.
Relevant answer
Answer
Deep Learning and Few-Shot Learning are not necessarily different things. I.e., you can use pre-trained language models, which are deep neural networks, for few-shot learning. Traditionally, few-shot learning was approached by using statistical learners in combination with a strong feature set, and algorithms such as kNN. However, with conditional PLMs such as GPT or Llama, you can implement few-shot learning by prompting a model with some examples in addition to a task description. You can also use BERT sentence vectors in combination with kNN for few-shot learning.
  • asked a question related to Classification
Question
3 answers
Hi everyone,
Recently, I have read some papers relating to EEG based-neuromarketing. In which, the papers aim to classify customer preference (Like/Dislike a product) using participant's EEG signal. I think I do not really understand:
1. How to label the ground truth preference (Like or Dislike)? If we use self-report to label the data, so what is the meaning of EEG?
2. The first question leads to the second one: how can this be applied in real-world scenarios?
I find it very difficult to find an article that clearly explains those questions. I would greatly appreciate it if you could spare some time to help. Thank you so much!
Relevant answer
Answer
Aparna Sathya Murthy , Alireza Falakdin : Thank you so much for your answers.
  • asked a question related to Classification
Question
1 answer
What is the best way to classify attacks in IIoT? How can we differentiate this classification in IIoT and CPS?
Relevant answer
Answer
Dear Muhammad MUZAMIL Aslam,
in order to record and classify the uncertainties in the IIoT completely as possible, I would suggest developing a Threat Matrix for the IIoT.
The idea of such a Network Threat Matrix is shown in Figures 6.3-2 and 6.3-4 and at the address:
Best regards and much success
Anatol Badach
  • asked a question related to Classification
Question
2 answers
pour la détection de l’apnée du sommeil en utilisant Deep Learning est ce que je peux passer par classification des stades de sommeil puis la détection de l'apnée et comment faire çà l'output du modèle de la classification des stades sommeil comme entre pour le modèle de détection de l'apnée du sommeil pour chaque modèle j 'utilise différent dataset
Relevant answer
Answer
Oui, il est possible d'utiliser le deep learning pour la détection de l'apnée du sommeil en utilisant la classification des stades de sommeil. La classification des stades de sommeil est une tâche courante en analyse du sommeil, et le deep learning s'est avéré être une approche efficace pour résoudre ce type de problème.
La détection de l'apnée du sommeil implique généralement l'analyse de signaux physiologiques tels que l'électroencéphalogramme (EEG), l'électrocardiogramme (ECG) et l'électromyogramme (EMG). En utilisant des données d'enregistrements polysomnographiques (PSG), qui combinent plusieurs types de signaux, vous pouvez entraîner un modèle de deep learning pour reconnaître les différents stades de sommeil, tels que le sommeil léger, le sommeil profond, le sommeil paradoxal, etc.
Une fois que vous avez entraîné un modèle de classification des stades de sommeil, vous pouvez l'utiliser pour détecter l'apnée du sommeil en analysant les motifs caractéristiques dans les données. Par exemple, l'apnée du sommeil peut être associée à des interruptions de la respiration et à des changements dans les schémas respiratoires. En identifiant ces schémas anormaux dans les signaux physiologiques pendant les différents stades de sommeil, le modèle peut aider à détecter l'apnée du sommeil.
Il est important de noter que la détection de l'apnée du sommeil est un domaine complexe et qu'il existe différentes approches pour y parvenir. Le deep learning peut être utilisé comme une composante de ces approches, mais il est souvent combiné avec d'autres méthodes d'analyse du signal et de traitement du sommeil pour obtenir des résultats plus précis et fiables. Il est recommandé de consulter la littérature scientifique spécialisée et de collaborer avec des experts du domaine pour développer un système de détection de l'apnée du sommeil robuste et précis.
  • asked a question related to Classification
Question
4 answers
How can one classify the research based on some parameters.
Relevant answer
Answer
Dear Rahul Jain ,
The research is broadly classified into two main classes: 1. Fundamental or basic research and 2. Applied research. Basic and applied researches are generally of two kinds: normal research and revolutionary research.
Research is a logical and systematic search for new and useful information on a particular topic. Research is important both in scientific and nonscientific fields. In our life, new problems, events, phenomena, and processes occur every day. Practically, implementable solutions and suggestions are required for tackling new problems that arise. Scientists have to undertake research on them and find their causes, solutions, explanations, and applications.
Regards,
Shafagat
  • asked a question related to Classification
Question
4 answers
Hi, I am trying to tune the parameters of a classification problem. Can I use MSE as fitness function in PSO or is it only for regression problems? If yes, what fitness function can be used for classification problems?
Thanks
Relevant answer
Answer
Hi Arvin Fakhri Thank you for your response. I guess I was using the metric directly. Also, I tried using 1-f1_score and 1/f1 however, I am getting same results every time. I have tried increasing the no. of particles or iterations, but the results are still the same.
  • asked a question related to Classification
Question
3 answers
  • Describe the different methods used for implementing land capability classification and land suitability classification, including the use of remote sensing, GIS, and field surveys.
  • Analyze the strengths and weaknesses of each method.
  • Discuss how they can be combined to create more accurate and comprehensive land use assessments.
  • Evaluate the role of community participation and local knowledge in the implementation of these classification systems.
  • Discuss the potential benefits of incorporating indigenous knowledge into land use planning and management.
Relevant answer
Answer
I agree with Subhani Shaik that Land capability classification is a system of grouping soils primarily on the basis of their capability to produce common cultivated crops and pasture plants without deterioration over a long period of time. Capability Class Broadly, It is classified into two groups: Arable land classes and non-arable land. Land can also be classified on the basis of ownership as private land and community land. Private land is owned by individuals whereas, community land is owned by the community for common uses like collection of fodder, fruits, nuts or medicinal herbs. These community lands are also called common property resources. Suitability level of the six factors: Soil depth (a), Soil texture (b), Soil pH (c), Temperature (d), Rainfall (e) and Slope (f)suitability classes. Land Suitability Analysis for Rice Production: A GIS Based Multi-Criteria Decision Approach. Land capability may be defined as the ability of the land surface to support natural plant growth/ wildlife habitat or artificial crop growth/ human habitat. Thus, it indicates the type of land use. The reporting area is classified into the following ninecategories: Forests: This includes all lands classed as forest under any legal enactment dealing with forests or administered as forests, whether state-owned or private, and whether wooded or maintained as potential forest land. Lithology, soil physicochemical, topography and irrigation water. Suitability is determined through systematic, multi-factor analysis of the different aspects of the terrain. Model inputs include a variety of physical, cultural, and economic factors. The results are often displayed on a map that is used to highlight areas from high to low suitability. Method suitability is proven using three independent tests: accuracy validation, precision validation, and challenge organism recovery. Only quantitative methods need accuracy and precision validation parameters. Within Orders, land suitability classes reflect degrees of suitability. The classes are given Arabic numbers. The designations are: Class S1, Highly Suitable; Class S2, Moderately Suitable; Class S3, Marginally Suitable; Class N1, Marginally Not Suitable; and Class N2, Permanently Not Suitable. Suitability is a measure of how well the qualities of a land unit match the requirements of a particular form of land use. ‘Suitability is a measure of how well the qualities of a land unit match the requirements of a particular form of land use. Land capability is the ability of a piece of land to sustainably support a specific land use. If land is used beyond its capability, degradation is the consequence and land capability is based on the understanding that every component of land has its own particular capacity to provide ecosystem services.
  • asked a question related to Classification
Question
4 answers
can anyone able to download this paper
Task Offloading with Task Classification and Offloading Nodes Selection for MEC-Enabled IoV
Relevant answer
Answer
Sure Ms Aziz
Wish you the best
  • asked a question related to Classification
Question
1 answer
If a study examines the interaction effect of X1 (a dichotomous variable) and X2 (a dichotomous variable) on Y, the mediating variable M plays a mediating role in the interaction effect, and the moderating variable W changes the classification of the independent variable X1 (e.g., from dichotomous to trichotomous). How is this conceptual model figure drawn?
Relevant answer
Answer
Drawing a conceptual model figure can help to represent the relationships between variables in your study visually. Here is an example of how you could draw a conceptual model figure for the study you described:
  1. Start with a box in the centre of the figure that represents your dependent variable, Y.
  2. Draw arrows pointing to Y from two boxes representing your independent variables, X1 and X2.
  3. Draw a circle or oval around the arrow from X1 to Y to represent the interaction effect between X1 and X2.
  4. Add another box representing your mediating variable, M, and draw an arrow pointing from M to Y, passing through the interaction effect arrow.
  5. Add another box representing your moderating variable, W, and draw an arrow pointing from W to X1.
  6. Label each box with the name of the variable it represents.
  7. Label each arrow with the direction of the relationship (e.g., positive or negative) and any coefficients or statistical measures (e.g., β, p) you are testing.
Here is an example of what the conceptual model figure might look like:
luaCopy code
____
| |
| Y |
|____|
/ \
/ \
/ \
____ ____
| | | |
| X1 |----| X2 |
|____| |____|
| |
| |
| ____/ \____
| | |
|---| M |
|__________|
|
|
|
____
| |
| W |
|____|
Of course, your conceptual model figure's specific layout and design may vary depending on your preferences and the software you use to create it.
  • asked a question related to Classification
Question
1 answer
What is the meaning of the news, and what are its classifications?
Relevant answer
Answer
News is current significant event or happening in real world. It could be classified in various categories similar to newspapers . Ex. Business, Technology, Entertainment, Sports, etc.
  • asked a question related to Classification
Question
5 answers
What is the impact of varying the number of hidden layers in a deep neural network on its performance for a specific classification task, and how does this impact change when different activation functions are used?
Relevant answer
  • asked a question related to Classification
Question
2 answers
I want to do further research on aquatic microplastic classification and quantification. My target areas will be water, sediment, and plankton. I am in the initial stage of my work. I plan to prepare three review articles regarding microplastics within six months for a Scopus index journal. I am seeking collaborations with those in the same field having a similar interest in publication. My inbox is open for discussion. Authorship will be split equally. The corresponding author will be from my side.
Thanks for your time and consideration.
Thanks in advance.
Relevant answer
Answer
Microplastics are microscopic plastic particles smaller than 5mm in diameter that can affect marine life and the environment. Scientists are examining the distribution, quantity, and consequences of microplastics in the ocean. Microplastics have been found in all sections of the ocean, from the surface to the deep sea, and can be consumed by a broad variety of marine animals. Microplastics have also been found in studies to carry and release chemical pollutants, which may be dangerous to marine life and possibly even people. Researchers are utilising a variety of methodologies to better understand the effects of microplastics on the ocean and its ecosystems, including field surveys, laboratory trials, and modelling. They are also looking into possible solutions like reduction of plastic pollution and developing innovative technologies to remove the microplastics in the ocean.
  • asked a question related to Classification
Question
6 answers
I believe it is very difficult to discriminate slum area with a method other than object based since it has a very heterogeneous reflectance!!!!, but I am interested in pixel based classification. Is there any pixel based classification approach to discriminate slum area?
Relevant answer
Answer
  • asked a question related to Classification
Question
3 answers
I am trying to perform unsupervised Kmeans classification in ENVI 5.1 (also tried 5.2), but it results in an homogeneous image (completely dark or red!!!!). Does anyone knows the problem?
I am attaching the result!!!!
Relevant answer
  • asked a question related to Classification
Question
1 answer
The omnibus tests of model coefficients shows that the model is significant. However, the classification table under Block 1 shows the same ratio of observed to predicted values as Block 0. Does this mean that my independent variables do not contribute to the model? I've read online that this problem may arise due to rare occurrence of events. Is there an alternative way to do the classification tables or is it acceptable to report the findings of having same classification table for block 0 and block 1 due to rare occurrence of events? Thank you.
Relevant answer
Answer
I also met the same problem. Don't know the reason
  • asked a question related to Classification
Question
4 answers
I have a collection of sentences that is in an incorrect order. The system should output the correct order of the sentences. What would be the appropriate approach to this problem? Is it a good approach to embed each sentence into a vector and classify the sentence using multiclass classification (assuming the length of the collection is fixed)?
Please let me know if there can be other approaches.
  • asked a question related to Classification
Question
2 answers
Notani et al
Epstein et al.
Marx et al.
Glanzmann et al etc
Relevant answer
Answer
Efsun Somay Marx et al. classification and Notani et al. classification are two of the most frequently used osteoradionecrosis classification methods in clinical practice. The Marx et al. classification is a commonly used method based on osteoradionecrosis staging, whereas the Notani et al. classification concentrates on osteoradionecrosis location. The classification method used may be determined by the clinician's preferences as well as the particular instance being evaluated.
  • asked a question related to Classification
Question
1 answer
Hello qualitative software users
I am working on this project and I have linked files to cases that have attributes but could not find to select a code to select a certain attribute and see what all interviews with this attribute and this code have together. Is there a way to do this?
Relevant answer
Answer
Zeinab El Dirani Yes, you can use case categorization characteristics to sort codes in NVivo. The stages are as follows:
1. Navigate to the "Cases" section in the Navigation View and pick the case you want to work on.
2. Click the "Classifications" option in the "Case" page.
3. Choose the group to be used as a filter.
4. Select the value(s) to use as a filter by clicking the "Filter" option.
5. To implement the filter, click the "Apply" option.
You can now navigate to the "Nodes" subdirectory and pick the code you want to work with. NVivo will apply the filter to the chosen code immediately, so you will only see references that have both the selected code and the selected code.
  • asked a question related to Classification
Question
1 answer
Is there any standard to perform landform classification based on curvature and slope using specific values for small agricultural areas (10-20ha)? For instance, to define the following hillslope positions: summit, shoulder, backslope, footslope, and toeslope. Thanks.
Relevant answer
Answer
Yes, there should be information already posted on this website. Ill check for it. You may already have it (Three days ago you asked, and I just saw it) but I have seen references related. I'll look for more information and get back to you.
  • asked a question related to Classification
Question
4 answers
Which are the latest deep learning models for zero-shot speech classification ?
Relevant answer
Answer
  • asked a question related to Classification
Question
1 answer
please share if you have a thesis that contains a complete methodology for performing the HFACS method to identify unsafe acts? Thank
Relevant answer
Answer
The HFACS method is a five-level hierarchical model that identifies the underlying causes of human error, including unsafe acts, preconditions for unsafe acts, and organizational influences.
Here is a step-by-step methodology for performing the HFACS method to identify unsafe acts:
✓ Collect data: Collect all available data related to the incident or accident. This data may include witness statements, accident reports, maintenance records, air traffic control communications, and any other relevant documents.
✓ Identify the outcome: Identify the outcome of the incident or accident, such as an aircraft collision, runway excursion, or loss of control.
✓ Classify the unsafe acts: Classify the unsafe acts that contributed to the outcome using the HFACS framework. Unsafe acts can be further classified as skill-based errors, decision errors, or perceptual errors.
✓ Identify the preconditions for unsafe acts: Identify the preconditions for the unsafe acts that contributed to the outcome. Preconditions can include environmental factors, equipment design, and job demands.
✓ Identify the organizational influences: Identify the organizational influences that contributed to the outcome, such as policies, procedures, training, and communication.
✓ Analyze the data: Analyze the data collected in steps 1-5 to identify patterns and trends in the unsafe acts, preconditions, and organizational influences.
✓ Develop recommendations: Develop recommendations to address the identified unsafe acts, preconditions, and organizational influences. Recommendations should aim to prevent similar incidents from occurring in the future.
✓ Implement and evaluate recommendations: Implement the recommendations and evaluate their effectiveness. Make any necessary adjustments to the recommendations to ensure they are effective in preventing future incidents.
The HFACS method provides a systematic and structured approach to identifying and analyzing human error in aviation incidents and accidents. By identifying the underlying causes of human error, organizations can develop strategies to prevent similar incidents from occurring in the future.
  • asked a question related to Classification
Question
13 answers
My team and I are in the middle of a prioritization problem that involves 350 alternatives (see figure for context about alternatives) or so. I have used the AHP to support the decision-making process in the past with only 7 or 8 alternatives and it has worked perfectly.
I would like to know if the AHP has a limit on the number of alternatives, because consistency may become a problem as Dr. Saaty's method provides Random consistency Indexes for matrix sizes of up to 10.
I was thinking in distributing the 350 alternatives in groups of 10, according to an attribute or classification criteria, to be able to use the RI chart proposed by Dr. Saaty.
If there are other more adecuate multi-criteria analysis tools, or different approaches to calculate the RI for larger matrices, please let me know.
Greetings and thank you,
Relevant answer
Answer
Dear José de la Garza
I don’t think that AHP has a limit for alternatives, however, in your case, dealing with 350 alternatives involves a tremendous workload, and, if for whatever reasons and after you finish, you add or delete and alternative or a criterion, as ususually happens, you have to start all over again.
I would suggest not making pair-wise comparisons of criteria, but simply, the group may evaluate each criterion separately, and then finding the average. Consistency or lack of it, is a property of AHP, and in my opinion useless, since the DM may be forced to adjust something that he/she believed, and assuming that there must be transitivity with a 10 % of tolerance. And all of this trouble to gain what?
Nothing, because they can’t assume that the scenario in the real-world is transitive. Maybe it is, or may be not.
I believe that your group criteria is OK but short.
For instance, don’t you think that it is important a criterion that qualifies each supplier regarding compliance history in time and in quantities?
You can have a hint of it by researching the history of each supplier, and asking your competition. What about type and age of machinery? Are your potential suppliers metal foundries for Aluminium, Iron, Precision casting?) (I guess it since you talk about casting).
If it is so, it appears that in level 2 criteria, they may be related. For instance, I don’t think that you can address the manufacturing capability independently of product capacity and cost. A foundry with small capacity most probable will have higher production costs that a large one.
If this is the case, you are not allowed to use AHP, because in this method all criteria must be independent. This was specifically established by its creator, Dr. Thomas Saaty.
What I would suggest is computing weights independently and apply them to a decision matrix that responds to real issues, for instance production capacity in kgr/day, costs per unit, expertise, financial capacity, size of technical department, etc.
Once you have that you can applied methods like PROMETHEE, TOPSIS, ELECTRE,VIKOR, etc., to find the best supplier, or else a method that does not use weights, like SIMUS.
I hope it helps
  • asked a question related to Classification
Question
7 answers
Which are the latest deep learning models for zero-shot image classification ?
Relevant answer
Answer
Dear Titas De ,
Deep Learning has become an indispensable tool of Artificial Intelligence. Among its many applications, it is now commonly used to solve complex Computer Vision tasks through supervised learning.
Regards,
Shafagat
  • asked a question related to Classification
Question
5 answers
We are applying k-means clustering algorithm on an unlabeled data. Our aim is to at the end, pool a result that shows two possibility. Is it necessary then to carry out k-nn classification after the clustering?
Relevant answer
I think clustering is one way of classifying. Therefore, after clustering, some additional classification is not required.
  • asked a question related to Classification
Question
27 answers
Suppose we conveniently extended the standard concept of cellular automaton to include
graphs and state-spaces Q of any cardinality and that the transition function F belonged to a certain adequate notion of "(hyper)computable function". We call this a hyper-cellular automaton HCA.
Consider the postulate: the universe can be described by a HCA with transition function F.
We cannot escape the problem of the initial condition Q_0. In the Wolfram Classification random initial conditions are considered. Hence the expediency for some topology or measure on Q.
Q will include for instance the usual sheaves (principle bundles and connections) considered in the standard model. It will also include other aspects to account for quantum gravity, consciousness, emergent biological complexity, etc.
It is an empirical fact that this HCA must be WC4 "complex patterns of localised structures" in the Wolfram Classification.
A major problem is the goal of reverse engineering F is that we do not have evolutions for other initial conditions at our disposal neither for the universe nor for subsystems of the universe. For physics at least a lot of locality and invariance hypotheses come in to play to justify the universality of experimental conclusions. The chemistry we observe on earth must also be that of the most distant star.
For biology the situation is drastically different. My question is: how can biology go beyond being a merely descriptive science as contrasted with fundamental physics ?
Biology seems to be mainly a "reverse engineering" affair. But it is also important
to have detailed, mathematically precise models - perhaps using HCAs - that can be used to test hypotheses and perform simulations.
Molecular biology suggests a new paradigm for software-hardware, a fluid mobile computer with essentially interconnected parts. A key characteristic is that information operations are tied to material and energetic constraints.
Also we must focus on ecosystems (the analogue of the cell ? ) rather than individual species. What about the idea of a "natural internet" (via horizontal gene transfer, etc.) ?
Relevant answer
Answer
Dear Duan Xian Xiang,
Your words "The law of biology can be described by "program", but this "program" is not a computer program, computer program is only one of it, "program" or "calculation" includes psychological program, economic activity program, political activity program, gene program of organism, etc." are consistent with my review paper where are explained examples and methodologies used in biology and medicine to describe, simulate, and predict the behavior of biological systems.
The book you mentioned sounds interesting, but I do not know it.
According to my understanding, emergents and their mutual interactions are the best current methodological description of biological and many physical phenomena observed. In many cases, there exists no alternative to such descriptions, as demonstrated on medical applications in my review.
The review describing the basic principles of complex systems descriptions in biology:
"Complex Systems and Their Use in Medicine: Concepts, Methods and Bio-Medical Applications"
  • asked a question related to Classification
Question
9 answers
When using ML for classification tasks in disease diagnosis study, what level of accuracy can be considered enough, high, or a threshold/benchmark?
Relevant answer
Answer
I would like to suggest considering all the iterations for classification with respect to the epochs, batch size and the train-test splits. Considering not only the accuracy but also the Precision and F1 scores of the model, we can reduce the chances of selection of an overfitted model as mentioned by Inès François. Also, the classification technique matters for different datasets as it is different for text-based csv data, image data, etc. Overfitting and Skewness can be easily detected by the above-mentioned metrics by implementing a confusion matrix for each model.
Understanding Confusion Matrix, Precision-Recall, and F1-Score | by Pratheesh Shivaprasad | Towards Data Science
The above link might help you gain an insight on how to easily detect if a model is skewed.
Thank you.
  • asked a question related to Classification
Question
4 answers
Android malware classification using machine learning algorithm aim and the objectives?
Relevant answer
Answer
I have 2 research and one thesis in machine learning and deep learning. You can apply deep learning technique for your thesis... You can check my research papers from my profile for ideas about it.
  • asked a question related to Classification
Question
17 answers
Hi everyone,
I did a classification work with less number of samples (n=73). I compared different machine learning classifiers but two of them, SVM with poly kernel and Radial Basis Function kernel, produced an overall accuracy with 100%. Hence, is this finding acceptable? if unacceptable, why?
Relevant answer
Answer
Quality & diversity of dataset is more important parameter than dataset size. Try to find a standard dataset or use Regularization techniques on your dataset.
Hope it will help you.
  • asked a question related to Classification
Question
3 answers
The autonomic seizures (AS) are non recognized easly and it is not clear the origin area and usually without a ictal EEG positive. Someone has video Ictal EEG of AS?
Relevant answer
Answer
Hi! I have further information of this case, in case you are interested!
  • asked a question related to Classification
Question
7 answers
I'm doing research about deixis/deictic and its meanings in pragmatic. I didn't find any journals, books or ppt about pragmatic meanings in deixis, most of it is explained in semantic way, not pragmatic. Are pragmatic meanings in deixis same like speech act classifications? Or we can use implicature theories to determine those pragmatics meanings?
How to determine pragmatic meaning in this dialogue example?
Mother: What do you want for Christmas?
Daughter: I want a violin!
Father: Didn't you want a doll? You said that one week ago.
Daughter: Now I want violin, papa!
If the deixis of that dialogue is you, does that pragmatic meaning is asking the daughter what she wants for Christmas? Or it is semantic means?
Thank you in advance.
Relevant answer
Answer
Anna Rosa There are several resources accessible to learn more about pragmatic meanings in particular. Some probable sources for pragmatic meaning references include:
Journals include Linguistics and Philosophy, Pragmatics and Society, Pragmatics and Cognition, and others.
Works include Stephen C. Levinson's "Pragmatics," Jenny A. Thomas' "Meaning in Interaction: An Introduction to Pragmatics," and "The Handbook of Pragmatics," edited by Laurence R. Horn and Gregory Ward, among others.
Powerpoint slides: Several colleges make their course materials available to the public on their websites. To discover examples, try searching for "pragmatics course slides" or "pragmatics lecture slides".
Of course, these are just a few examples; there are several additional sites that may be beneficial. It is critical to ensure that the sources are reliable.
  • asked a question related to Classification
Question
2 answers
Someone can clarify the morphological classification of marine macroalgae used by Littler and Littler (1981).
Relevant answer
Answer
Sincere appreciation to you, Qamar Ul Islam.
Since I can't seem to find the original paper, I would be very appreciative if you could send it to me.
  • asked a question related to Classification
Question
4 answers
I have implemented a CNN for my image classification. Now I want to save the features from this model into a local CSV file. How can I do this? here is the example model:
model = tf.keras.models.Sequential([tf.keras.layers.Conv2D(16, (3,3), activation='relu', input_shape=(200,200,3)),
tf.keras.layers.MaxPool2D(2,2),
#
tf.keras.layers.Conv2D(32, (3,3), activation='relu'),
tf.keras.layers.MaxPool2D(2,2),
#
tf.keras.layers.Conv2D(64, (3,3), activation='relu'),
tf.keras.layers.MaxPool2D(2,2),
##
tf.keras.layers.Flatten(),
##
tf.keras.layers.Dense(1024, activation='relu'),
tf.keras.layers.Dense(512, activation='relu'),
##
tf.keras.layers.Dense(1, activation='sigmoid')
])
Relevant answer
Answer
You can extract the features from a Convolutional Neural Network (CNN) model by accessing the output of one of its intermediate layers, which represents the learned features. This can be done in several ways:
Get the activation values: You can use the Keras API's predict method to obtain the activation values of a particular layer in the model. For example, if you have a model model and you want to extract the activations from layer layer_name, you can use the following code:
layer_output = model.get_layer(layer_name).output
intermediate_model = Model(inputs=model.input, outputs=layer_output)
intermediate_output = intermediate_model.predict(input_data)
input_data is the input to the model and intermediate_output contains the activations from the specified layer.
Save the activations to a file: After obtaining them, you can save them to a local CSV file using the pandas library. Here's an example of how to do that:
import pandas as pd
df = pd.DataFrame(intermediate_output)
df.to_csv('features.csv', index=False)
This will save the activations to a CSV file named features.csv.
  • asked a question related to Classification
Question
1 answer
Hello,
I am conducting a latent transition analysis in Mplus, and I am examining the relationship between class membership, predictors, and outcomes. I am using a three-step process (Asparouhov & Muthén, 2014). During step 3, after I created model class assignment variables that I used as indicators of the latent classes, the inclusion of predictors in the model (i.e., sex and ethnicity) significantly changed the classification counts and proportions of groups at each time point. From what I understand, creating a modal class assignment variable and accounting for classification errors in the third step of the process should prevent covariates from significantly affecting class membership in the final model. I have read about alternative three-step approaches (e.g., Vermunt and Magidson, 2021) that include covariates that demonstrate DIF in the first step of the model and include all covariates (with and without DIF) in step three of the model. Is this an appropriate method, given that I am interested in the effect of sex and ethnicity on latent class membership, transition probabilities, and outcomes?
Is it possible that DIF could cause a significant change in class membership in the LTA? I have read the work by Masyn (2017), who developed a procedure to address DIF, but her analysis included individual items. I am using subscale scores as indicators of the latent classes. Is it possible to test for DIF using subscales, or does this need to be done at the item level? Also, is there any research that addresses DIF within the context of an LTA? I am a bit confused about whether DIF would be examined for each LPA before constructing the modal class variables to include in the LTA or if it would be addressed just in the LTA model.
I hope this makes sense. Any suggestions would be greatly appreciated.
Relevant answer
Answer
What I would recommend for categorical covariates such as sex and ethnicity is to examine their effects by means of multigroup analyses (KNOWNCLASS option in Mplus). That way, you can examine DIF / measurement non-equivalence across groups in addition to studying group differences in class sizes. With multigroup analysis, you can formally test whether your classes are defined in the same way across, for example, sex. You would see pretty clearly whether the classes are different across groups, which could explain why the classes change when you add these grouping variables as covariates in an overall (single-group) model.
Also, when classes change after adding covariates, this often indicates that the covariates have direct effects on one or more of the indicators, that is, effects above and beyond what is mediated through the latent class variable. For example, there could be particularly large gender differences on one of the indicators. Multistep approaches to examining covariates have the potential to mask such direct effects, and I'm not sure this is a good thing (because it could lead to bias). If there are differential effects of covariates on certain indicators, then I would think that we would want to know about these effects rather than masking them by "fixing" the classes based on an unconditional model first before introducing the covariates.