An optimal unsupervised text data segmentation 3prj_publication
This document summarizes a research paper that proposes an unsupervised text data segmentation model using genetic algorithms (OUTDSM) to improve clustering accuracy. The OUTDSM uses an encoding strategy, fitness function, and genetic operators to evolve optimal text clusters. Experimental results presented in the research paper demonstrate that OUTDSM can arrive at global optima and prevent stagnation at local optima due to its biologically diverse population. Key areas of related work discussed include text mining techniques, clustering methods, and prior uses of optimization algorithms like genetic algorithms for text mining tasks.
This document describes a proposed Optimal Frequent Patterns System (OFPS) that uses a genetic algorithm to discover optimal frequent patterns from transactional databases more efficiently. The OFPS is a three-fold system that first prepares data through cleaning, integration and transformation. It then constructs a Frequent Pattern Tree to discover frequent patterns. Finally, it applies a genetic algorithm to generate optimal frequent patterns, simulating biological evolution to find the best solutions. The proposed system aims to overcome limitations of conventional association rule mining approaches and efficiently discover optimal patterns from large, changing datasets.
ESTIMATION OF REGRESSION COEFFICIENTS USING GEOMETRIC MEAN OF SQUARED ERROR F...ijaia
Regression models and their statistical analyses is one of the most important tool used by scientists and practitioners. The aim of a regression model is to fit parametric functions to data. It is known that the true regression is unknown and specific methods are created and used strictly pertaining to the roblem. For the pioneering work to develop procedures for fitting functions, we refer to the work on the methods of least
absolute deviations, least squares deviations and minimax absolute deviations. Today’s widely celebrated
procedure of the method of least squares for function fitting is credited to the published works of Legendre and Gauss. However, the least squares based models in practice may fail to provide optimal results in nonGaussian situations especially when the errors follow distributions with the fat tails. In this paper an unorthodox method of estimating linear regression coefficients by minimising GMSE(geometric mean of squared errors) is explored. Though GMSE(geometric mean of squared errors) is used to compare models it is rarely used to obtain the coefficients. Such a method is tedious to handle due to the large number of roots obtained by minimisation of the loss function. This paper offers a way to tackle that problem.
Application is illustrated with the ‘Advertising’ dataset from ISLR and the obtained results are compared
with the results of the method of least squares for single index linear regression model.
Machine Learning for Chemistry: Representing and InterveningIchigaku Takigawa
Joint Symposium of Engineering & Information Science & WPI-ICReDD in Hokkaido University
Apr. 26 (Mon), 2021
https://www.icredd.hokudai.ac.jp/event/5430
Data mining , knowledge discovery is the process
of analyzing data from different perspectives and summarizing it
into useful information - information that can be used to increase
revenue, cuts costs, or both. Data mining software is one of a
number of analytical tools for analyzing data. It allows users to
analyze data from many different dimensions or angles, categorize
it, and summarize the relationships identified. Technically, data
mining is the process of finding correlations or patterns among
dozens of fields in large relational databases. The goal of
clustering is to determine the intrinsic grouping in a set of
unlabeled data. But how to decide what constitutes a good
clustering? It can be shown that there is no absolute “best”
criterion which would be independent of the final aim of the
clustering. Consequently, it is the user which must supply this
criterion, in such a way that the result of the clustering will suit
their needs.
For instance, we could be interested in finding
representatives for homogeneous groups (data reduction), in
finding “natural clusters” and describe their unknown properties
(“natural” data types), in finding useful and suitable groupings
(“useful” data classes) or in finding unusual data objects (outlier
detection).Of late, clustering techniques have been applied in the
areas which involve browsing the gathered data or in categorizing
the outcome provided by the search engines for the reply to the
query raised by the users. In this paper, we are providing a
comprehensive survey over the document clustering.
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology
Evaluating the efficiency of rule techniques for file classificationeSAT Journals
Abstract Text mining refers to the process of deriving high quality information from text. It is also known as knowledge discovery from text (KDT), deals with the machine supported analysis of text. It is used in various areas such as information retrieval, marketing, information extraction, natural language processing, document similarity, and so on. Document Similarity is one of the important techniques in text mining. In document similarity, the first and foremost step is to classify the files based on their category. In this research work, various classification rule techniques are used to classify the computer files based on their extensions. For example, the extension of computer files may be pdf, doc, ppt, xls, and so on. There are several algorithms for rule classifier such as decision table, JRip, Ridor, DTNB, NNge, PART, OneR and ZeroR. In this research work, three classification algorithms namely decision table, DTNB and OneR classifiers are used for performing classification of computer files based on their extension. The results produced by these algorithms are analyzed by using the performance factors classification accuracy and error rate. From the experimental results, DTNB proves to be more efficient than other two techniques. Index Terms: Data mining, Text mining, Classification, Decision table, DTNB, OneR
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
APPLICATION OF ARTIFICIAL NEURAL NETWORKS IN ESTIMATING PARTICIPATION IN ELEC...Zac Darcy
This document discusses using artificial neural networks to estimate voter participation rates in future elections in Iran. Specifically, it describes using a two-layer feed-forward neural network to predict voter turnout in the Kohgiluyeh and Boyer-Ahmad province with 91% accuracy. The neural network was trained on past electoral data from the province. The document also provides background on artificial neural networks and reviews their use in predicting outcomes in various domains, including economics, politics, tourism, the environment, and information technology.
A semantic framework and software design to enable the transparent integratio...Patricia Tavares Boralli
This document proposes a conceptual framework to unify representations of natural systems knowledge. The framework is based on separating the ontological nature of an object of study from the context of its observation. Each object is associated with a concept defined in an ontology and an observation context describing aspects like location and time. Models and data are treated as generic knowledge sources with a semantic type and observation context. This allows flexible integration and calculation of states across heterogeneous sources by composing their observation contexts and resolving semantic compatibility. The framework aims to simplify knowledge representation by abstracting away complexity related to data format and scale.
Iaetsd a survey on one class clusteringIaetsd Iaetsd
This document presents a new method for performing one-to-many data linkage called the One Class Clustering Tree (OCCT). The OCCT builds a tree structure with inner nodes representing features of the first dataset and leaves representing similar features of the second dataset. It uses splitting criteria and pruning methods to perform the data linkage more accurately than existing indexing techniques. The OCCT approach induces a decision tree using a splitting criteria and performs prepruning to determine which branches to trim. It then compares entities to match them between the two datasets and produces a final result.
Over the past decade, unprecedented progress in the development of neural networks influenced dozens of different industries, including weed recognition in the agro-industrial sector. The use of neural networks in agro-industrial activity in the task of recognizing cultivated crops is a new direction. The absence of any standards significantly complicates the understanding of the real situation of the use of the neural network in the agricultural sector. The manuscript presents the complete analysis of researches over the past 10 years on the use of neural networks for the classification and tracking of weeds due to neural networks. In particular, the analysis of the results of using various neural network algorithms for the task of classification and tracking was presented. As a result, we presented the recommendation for the use of neural networks in the tasks of recognizing a cultivated object and weeds. Using this standard can significantly improve the quality of research on this topic and simplify the analysis and understanding of any paper.
1) Machine learning can help rationalize the "experience and intuition" of chemical research by finding patterns and exceptions from large amounts of chemical data to predict new materials and phenomena.
2) While in theory chemical structures and properties can be described by Schrodinger's equation, it is impossible to solve for realistic systems, requiring approximations. Machine learning may help address this challenge.
3) Chemists have successfully created compounds with desired properties through "experience and intuition", which involves inductive reasoning from experiments rather than purely deductive logic, incorporating serendipitous findings.
A Novel Data mining Technique to Discover Patterns from Huge Text CorpusIJMER
Today, we have far more information than we can handle: from business transactions and scientific
data, to satellite pictures, text reports and military intelligence. Information retrieval is simply not enough
anymore for decision-making. Confronted with huge collections of data, we have now created new needs to
help us make better managerial choices. These needs are automatic summarization of data, extraction of the
"essence" of information stored, and the discovery of patterns in raw data. With this, Data mining with
inventory pattern came into existence and got popularized. Data mining finds these patterns and relationships
using data analysis tools and techniques to build models.
Subgraph relative frequency approach for extracting interesting substructurIAEME Publication
The document discusses a Subgraph Relative Frequency (SRF) approach for extracting interesting substructures from molecular data. SRF screens each frequent subgraph to determine if it is interesting based on its relative frequency. This is more efficient than the MISMOC approach, which requires calculating absolute frequencies. SRF was tested on a small molecular data set and found to perform satisfactorily and efficiently for classifying unknown molecules based on their subgraphs. The performance of SRF was comparable to MISMOC but with less computational complexity by using relative rather than absolute frequencies.
The interplay between data-driven and theory-driven methods for chemical scie...Ichigaku Takigawa
The 1st International Symposium on Human InformatiX
X-Dimensional Human Informatics and Biology
ATR, Kyoto, February 27-28, 2020
https://human-informatix.atr.jp
Correlation Coefficient Based Average Textual Similarity Model for Informatio...IOSR Journals
The document presents a proposed model for a textual similarity approach for information retrieval systems in wide area networks. It evaluates the performance of four similarity functions (Jaccard, Cosine, Dice, Overlap) using correlation coefficients. Three approaches are proposed: 1) Combining Cosine and Overlap similarity scores, which performed best. 2) Combining Cosine, Dice, and Overlap scores. 3) Combining all four similarity functions. The model is represented as a triangle where the vertices are the results from the three proposed approaches to measure textual similarity between retrieved documents.
A preliminary survey on optimized multiobjective metaheuristic methods for da...ijcsit
The present survey provides the state-of-the-art of research, copiously devoted to Evolutionary Approach
(EAs) for clustering exemplified with a diversity of evolutionary computations. The Survey provides a
nomenclature that highlights some aspects that are very important in the context of evolutionary data
clustering. The paper missions the clustering trade-offs branched out with wide-ranging Multi Objective
Evolutionary Approaches (MOEAs) methods. Finally, this study addresses the potential challenges of
MOEA design and data clustering, along with conclusions and recommendations for novice and
researchers by positioning most promising paths of future research.
Data Mining for Education
Ryan S.J.d. Baker, Carnegie Mellon University, Pittsburgh, Pennsylvania, USA
[email protected]
Article to appear as
Baker, R.S.J.d. (in press) Data Mining for Education. To appear in McGaw, B., Peterson, P.,
Baker, E. (Eds.) International Encyclopedia of Education (3rd edition). Oxford, UK: Elsevier.
This is a pre-print draft. Final article may involve minor changes and different formatting.
This PhD research proposal discusses using Bayesian inference methods for multi-target tracking in big data settings. The researcher proposes developing new stochastic MCMC algorithms that can scale to billions of data points by using small subsets of data in each iteration. This would make Bayesian methods computationally feasible for big data. The proposal outlines reviewing relevant literature, developing the theoretical foundations, and empirically validating new algorithms like sequential Monte Carlo on real-world problems to analyze text and user preferences at large scale.
This document discusses the potential for machine learning to accelerate scientific discovery by rationalizing the inductive process of generating hypotheses from data. It outlines two approaches in science - theory/hypothesis-driven modeling and data-driven modeling using machine learning. It argues that machine learning can help "rationalize" the intuitive, non-logical parts of the scientific process by using data to generate and test hypotheses. The document also discusses how machine learning may automate parts of the scientific method, from hypothesis generation to model building and experimentation, thereby amplifying a scientist's progress.
Searching in high dimensional spaces index structures for improving the perfo...unyil96
This document provides an overview of index structures for improving the performance of multimedia databases. It discusses how multimedia databases require content-based retrieval of similar objects, which is challenging due to the high-dimensional nature of feature spaces used to represent multimedia objects. The document summarizes the problems that arise from processing queries in high-dimensional spaces, known as the "curse of dimensionality", and provides an overview of index structure approaches that have been proposed to overcome these problems to efficiently process similarity queries in multimedia databases.
Ontology Based PMSE with Manifold PreferenceIJCERT
International journal from http://www.ijcert.org
IJCERT Standard on-line Journal
ISSN(Online):2349-7084,(An ISO 9001:2008 Certified Journal)
iso nicir csir
IJCERT (ISSN 2349–7084 (Online)) is approved by National Science Library (NSL), National Institute of Science Communication And Information Resources (NISCAIR), Council of Scientific and Industrial Research, New Delhi, India.
A Comprehensive Survey on Comparisons across Contextual Pre-Filtering, Contex...TELKOMNIKA JOURNAL
Recently, there has been growing interest in recommender systems (RS) and particularly in context-aware RS. Methods for generating context-aware recommendations are classified into pre-filtering, post-filtering and contextual modelling approaches. In this paper, we present the several novel approaches of the different variant of each of these three contextualization paradigms and present a complete survey on the state-of-the-art comparisons across them. We then identify the significant challenges that require being addressed by the current RS researchers, which will help academicians and practitioners in comparing these three approaches to select the best alternative according to their strategies.
A Semantic Retrieval System for Extracting Relationships from Biological Corpusijcsit
The World Wide Web holds a large size of different information. Sometimes while searching the World Wide Web, users always do not gain the type of information they expect. In the subject of information extraction, extracting semantic relationships between terms from documents become a challenge. This
paper proposes a system helps in retrieving documents based on the query expansion and tackles the extracting of semantic relationships from biological documents. This system retrieved documents that are relevant to the input terms then it extracts the existence of a relationship. In this system, we use Boolean
model and the pattern recognition which helps in determining the relevant documents and determining the place of the relationship in the biological document. The system constructs a term-relation table that accelerates the relation extracting part. The proposed method offers another usage of the system so the
researchers can use it to figure out the relationship between two biological terms through the available information in the biological documents. Also for the retrieved documents, the system measures the percentage of the precision and recall.
call for papers, research paper publishing, where to publish research paper, journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJEI, call for papers 2012,journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, research and review articles, engineering journal, International Journal of Engineering Inventions, hard copy of journal, hard copy of certificates, journal of engineering, online Submission, where to publish research paper, journal publishing, international journal, publishing a paper, hard copy journal, engineering journal
Predicting Material Properties Using Machine Learning for Accelerated Materia...Nikhil Sanjay Suryawanshi
The rapid prediction of material properties has become a pivotal factor in accelerating materials discovery and development, driven by advancements in machine learning and data-driven methodologies. This paper presents a novel system for predicting material properties using machine learning techniques, offering a scalable and efficient framework for exploring new materials with optimized properties. The system incorporates large datasets, feature engineering, and multiple machine learning models, such as Kernel Ridge Regression, Random Forest, and Neural Networks, to predict material properties like thermal conductivity, elastic modulus, and electronic bandgap. By integrating physics-based knowledge into machine learning models, the proposed system enhances the accuracy and interpretability of predictions. The results indicate that the system can significantly reduce the time and cost of material discovery while delivering high prediction accuracy. This is the potential approach to revolutionize materials science by enabling researchers to identify promising material candidates in silico, paving the way for breakthroughs in energy, electronics, and sustainable materials.
Machine Learning in Material Characterizationijtsrd
Machine learning has shown great potential applications in material science. It is widely used in material design, corrosion detection, material screening, new material discovery, and other fields of materials science. The majority of ML approaches in materials science is based on artificial neural networks ANNs . The use of ML and related techniques for materials design, development, and characterization has matured to a main stream field. This paper focuses on the applications of machine learning strategies for material characterization. Matthew N. O. Sadiku | Guddi K. Suman | Sarhan M. Musa "Machine Learning in Material Characterization" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-6 , October 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46392.pdf Paper URL : https://www.ijtsrd.com/engineering/electrical-engineering/46392/machine-learning-in-material-characterization/matthew-n-o-sadiku
Classifier Model using Artificial Neural NetworkAI Publications
This document summarizes a research paper that investigates using supervised instance selection (SIS) as a preprocessing step to improve the performance of artificial neural networks (ANNs) for classification tasks. SIS aims to select a subset of examples from the original dataset to enhance the accuracy of future classifications. The goal of applying SIS before ANNs is to provide a cleaner input dataset that handles noisy or redundant data better. The paper presents the architecture of feedforward neural networks and the backpropagation algorithm for training networks. It also discusses using mutual information-based feature selection as part of the SIS preprocessing approach.
11.software modules clustering an effective approach for reusabilityAlexander Decker
This document summarizes previous work on using clustering techniques for software module classification and reusability. It discusses hierarchical clustering and non-hierarchical clustering methods. Previous studies have used these techniques for software component classification, identifying reusable software modules, course clustering based on industry needs, mobile phone clustering based on attributes, and customer clustering based on electricity load. The document provides background on clustering analysis and its uses in various domains including software testing, pattern recognition, and software restructuring.
APPLICATION OF ARTIFICIAL NEURAL NETWORKS IN ESTIMATING PARTICIPATION IN ELEC...Zac Darcy
This document discusses using artificial neural networks to estimate voter participation rates in future elections in Iran. Specifically, it describes using a two-layer feed-forward neural network to predict voter turnout in the Kohgiluyeh and Boyer-Ahmad province with 91% accuracy. The neural network was trained on past electoral data from the province. The document also provides background on artificial neural networks and reviews their use in predicting outcomes in various domains, including economics, politics, tourism, the environment, and information technology.
A semantic framework and software design to enable the transparent integratio...Patricia Tavares Boralli
This document proposes a conceptual framework to unify representations of natural systems knowledge. The framework is based on separating the ontological nature of an object of study from the context of its observation. Each object is associated with a concept defined in an ontology and an observation context describing aspects like location and time. Models and data are treated as generic knowledge sources with a semantic type and observation context. This allows flexible integration and calculation of states across heterogeneous sources by composing their observation contexts and resolving semantic compatibility. The framework aims to simplify knowledge representation by abstracting away complexity related to data format and scale.
Iaetsd a survey on one class clusteringIaetsd Iaetsd
This document presents a new method for performing one-to-many data linkage called the One Class Clustering Tree (OCCT). The OCCT builds a tree structure with inner nodes representing features of the first dataset and leaves representing similar features of the second dataset. It uses splitting criteria and pruning methods to perform the data linkage more accurately than existing indexing techniques. The OCCT approach induces a decision tree using a splitting criteria and performs prepruning to determine which branches to trim. It then compares entities to match them between the two datasets and produces a final result.
Over the past decade, unprecedented progress in the development of neural networks influenced dozens of different industries, including weed recognition in the agro-industrial sector. The use of neural networks in agro-industrial activity in the task of recognizing cultivated crops is a new direction. The absence of any standards significantly complicates the understanding of the real situation of the use of the neural network in the agricultural sector. The manuscript presents the complete analysis of researches over the past 10 years on the use of neural networks for the classification and tracking of weeds due to neural networks. In particular, the analysis of the results of using various neural network algorithms for the task of classification and tracking was presented. As a result, we presented the recommendation for the use of neural networks in the tasks of recognizing a cultivated object and weeds. Using this standard can significantly improve the quality of research on this topic and simplify the analysis and understanding of any paper.
1) Machine learning can help rationalize the "experience and intuition" of chemical research by finding patterns and exceptions from large amounts of chemical data to predict new materials and phenomena.
2) While in theory chemical structures and properties can be described by Schrodinger's equation, it is impossible to solve for realistic systems, requiring approximations. Machine learning may help address this challenge.
3) Chemists have successfully created compounds with desired properties through "experience and intuition", which involves inductive reasoning from experiments rather than purely deductive logic, incorporating serendipitous findings.
A Novel Data mining Technique to Discover Patterns from Huge Text CorpusIJMER
Today, we have far more information than we can handle: from business transactions and scientific
data, to satellite pictures, text reports and military intelligence. Information retrieval is simply not enough
anymore for decision-making. Confronted with huge collections of data, we have now created new needs to
help us make better managerial choices. These needs are automatic summarization of data, extraction of the
"essence" of information stored, and the discovery of patterns in raw data. With this, Data mining with
inventory pattern came into existence and got popularized. Data mining finds these patterns and relationships
using data analysis tools and techniques to build models.
Subgraph relative frequency approach for extracting interesting substructurIAEME Publication
The document discusses a Subgraph Relative Frequency (SRF) approach for extracting interesting substructures from molecular data. SRF screens each frequent subgraph to determine if it is interesting based on its relative frequency. This is more efficient than the MISMOC approach, which requires calculating absolute frequencies. SRF was tested on a small molecular data set and found to perform satisfactorily and efficiently for classifying unknown molecules based on their subgraphs. The performance of SRF was comparable to MISMOC but with less computational complexity by using relative rather than absolute frequencies.
The interplay between data-driven and theory-driven methods for chemical scie...Ichigaku Takigawa
The 1st International Symposium on Human InformatiX
X-Dimensional Human Informatics and Biology
ATR, Kyoto, February 27-28, 2020
https://human-informatix.atr.jp
Correlation Coefficient Based Average Textual Similarity Model for Informatio...IOSR Journals
The document presents a proposed model for a textual similarity approach for information retrieval systems in wide area networks. It evaluates the performance of four similarity functions (Jaccard, Cosine, Dice, Overlap) using correlation coefficients. Three approaches are proposed: 1) Combining Cosine and Overlap similarity scores, which performed best. 2) Combining Cosine, Dice, and Overlap scores. 3) Combining all four similarity functions. The model is represented as a triangle where the vertices are the results from the three proposed approaches to measure textual similarity between retrieved documents.
A preliminary survey on optimized multiobjective metaheuristic methods for da...ijcsit
The present survey provides the state-of-the-art of research, copiously devoted to Evolutionary Approach
(EAs) for clustering exemplified with a diversity of evolutionary computations. The Survey provides a
nomenclature that highlights some aspects that are very important in the context of evolutionary data
clustering. The paper missions the clustering trade-offs branched out with wide-ranging Multi Objective
Evolutionary Approaches (MOEAs) methods. Finally, this study addresses the potential challenges of
MOEA design and data clustering, along with conclusions and recommendations for novice and
researchers by positioning most promising paths of future research.
Data Mining for Education
Ryan S.J.d. Baker, Carnegie Mellon University, Pittsburgh, Pennsylvania, USA
[email protected]
Article to appear as
Baker, R.S.J.d. (in press) Data Mining for Education. To appear in McGaw, B., Peterson, P.,
Baker, E. (Eds.) International Encyclopedia of Education (3rd edition). Oxford, UK: Elsevier.
This is a pre-print draft. Final article may involve minor changes and different formatting.
This PhD research proposal discusses using Bayesian inference methods for multi-target tracking in big data settings. The researcher proposes developing new stochastic MCMC algorithms that can scale to billions of data points by using small subsets of data in each iteration. This would make Bayesian methods computationally feasible for big data. The proposal outlines reviewing relevant literature, developing the theoretical foundations, and empirically validating new algorithms like sequential Monte Carlo on real-world problems to analyze text and user preferences at large scale.
This document discusses the potential for machine learning to accelerate scientific discovery by rationalizing the inductive process of generating hypotheses from data. It outlines two approaches in science - theory/hypothesis-driven modeling and data-driven modeling using machine learning. It argues that machine learning can help "rationalize" the intuitive, non-logical parts of the scientific process by using data to generate and test hypotheses. The document also discusses how machine learning may automate parts of the scientific method, from hypothesis generation to model building and experimentation, thereby amplifying a scientist's progress.
Searching in high dimensional spaces index structures for improving the perfo...unyil96
This document provides an overview of index structures for improving the performance of multimedia databases. It discusses how multimedia databases require content-based retrieval of similar objects, which is challenging due to the high-dimensional nature of feature spaces used to represent multimedia objects. The document summarizes the problems that arise from processing queries in high-dimensional spaces, known as the "curse of dimensionality", and provides an overview of index structure approaches that have been proposed to overcome these problems to efficiently process similarity queries in multimedia databases.
Ontology Based PMSE with Manifold PreferenceIJCERT
International journal from http://www.ijcert.org
IJCERT Standard on-line Journal
ISSN(Online):2349-7084,(An ISO 9001:2008 Certified Journal)
iso nicir csir
IJCERT (ISSN 2349–7084 (Online)) is approved by National Science Library (NSL), National Institute of Science Communication And Information Resources (NISCAIR), Council of Scientific and Industrial Research, New Delhi, India.
A Comprehensive Survey on Comparisons across Contextual Pre-Filtering, Contex...TELKOMNIKA JOURNAL
Recently, there has been growing interest in recommender systems (RS) and particularly in context-aware RS. Methods for generating context-aware recommendations are classified into pre-filtering, post-filtering and contextual modelling approaches. In this paper, we present the several novel approaches of the different variant of each of these three contextualization paradigms and present a complete survey on the state-of-the-art comparisons across them. We then identify the significant challenges that require being addressed by the current RS researchers, which will help academicians and practitioners in comparing these three approaches to select the best alternative according to their strategies.
A Semantic Retrieval System for Extracting Relationships from Biological Corpusijcsit
The World Wide Web holds a large size of different information. Sometimes while searching the World Wide Web, users always do not gain the type of information they expect. In the subject of information extraction, extracting semantic relationships between terms from documents become a challenge. This
paper proposes a system helps in retrieving documents based on the query expansion and tackles the extracting of semantic relationships from biological documents. This system retrieved documents that are relevant to the input terms then it extracts the existence of a relationship. In this system, we use Boolean
model and the pattern recognition which helps in determining the relevant documents and determining the place of the relationship in the biological document. The system constructs a term-relation table that accelerates the relation extracting part. The proposed method offers another usage of the system so the
researchers can use it to figure out the relationship between two biological terms through the available information in the biological documents. Also for the retrieved documents, the system measures the percentage of the precision and recall.
call for papers, research paper publishing, where to publish research paper, journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJEI, call for papers 2012,journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, research and review articles, engineering journal, International Journal of Engineering Inventions, hard copy of journal, hard copy of certificates, journal of engineering, online Submission, where to publish research paper, journal publishing, international journal, publishing a paper, hard copy journal, engineering journal
Predicting Material Properties Using Machine Learning for Accelerated Materia...Nikhil Sanjay Suryawanshi
The rapid prediction of material properties has become a pivotal factor in accelerating materials discovery and development, driven by advancements in machine learning and data-driven methodologies. This paper presents a novel system for predicting material properties using machine learning techniques, offering a scalable and efficient framework for exploring new materials with optimized properties. The system incorporates large datasets, feature engineering, and multiple machine learning models, such as Kernel Ridge Regression, Random Forest, and Neural Networks, to predict material properties like thermal conductivity, elastic modulus, and electronic bandgap. By integrating physics-based knowledge into machine learning models, the proposed system enhances the accuracy and interpretability of predictions. The results indicate that the system can significantly reduce the time and cost of material discovery while delivering high prediction accuracy. This is the potential approach to revolutionize materials science by enabling researchers to identify promising material candidates in silico, paving the way for breakthroughs in energy, electronics, and sustainable materials.
Machine Learning in Material Characterizationijtsrd
Machine learning has shown great potential applications in material science. It is widely used in material design, corrosion detection, material screening, new material discovery, and other fields of materials science. The majority of ML approaches in materials science is based on artificial neural networks ANNs . The use of ML and related techniques for materials design, development, and characterization has matured to a main stream field. This paper focuses on the applications of machine learning strategies for material characterization. Matthew N. O. Sadiku | Guddi K. Suman | Sarhan M. Musa "Machine Learning in Material Characterization" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-6 , October 2021, URL: https://www.ijtsrd.com/papers/ijtsrd46392.pdf Paper URL : https://www.ijtsrd.com/engineering/electrical-engineering/46392/machine-learning-in-material-characterization/matthew-n-o-sadiku
Classifier Model using Artificial Neural NetworkAI Publications
This document summarizes a research paper that investigates using supervised instance selection (SIS) as a preprocessing step to improve the performance of artificial neural networks (ANNs) for classification tasks. SIS aims to select a subset of examples from the original dataset to enhance the accuracy of future classifications. The goal of applying SIS before ANNs is to provide a cleaner input dataset that handles noisy or redundant data better. The paper presents the architecture of feedforward neural networks and the backpropagation algorithm for training networks. It also discusses using mutual information-based feature selection as part of the SIS preprocessing approach.
11.software modules clustering an effective approach for reusabilityAlexander Decker
This document summarizes previous work on using clustering techniques for software module classification and reusability. It discusses hierarchical clustering and non-hierarchical clustering methods. Previous studies have used these techniques for software component classification, identifying reusable software modules, course clustering based on industry needs, mobile phone clustering based on attributes, and customer clustering based on electricity load. The document provides background on clustering analysis and its uses in various domains including software testing, pattern recognition, and software restructuring.
Choosing allowability boundaries for describing objects in subject areasIAESIJAI
Anomaly detection is one of the most promising problems for study and can be used as independent units and preprocessing tools before solving any fundamental data mining problems. This article proposes a method for detecting specific errors with the involvement of experts from subject areas to fill knowledge. The proposed method about outliers hypothesizes that they locate closer to logical boundaries of intervals derived from pair features, and the interval ranges vary in different domains. We construct intervals leveraging pair feature values. While forming knowledge in a specific field, a domain specialist checks the logical allowability of objects based on the range of the intervals. If the objects are logical outliers, the specialist ignores or corrects them. We offer the general algorithm for the formation of the database based on the proposed method in the form of a pseudo-code, and we provide comparison results with existing methods.
LINEAR REGRESSION MODEL FOR KNOWLEDGE DISCOVERY IN ENGINEERING MATERIALScscpconf
Nowadays numerous interestingness measures have been proposed to disclose the relationships
of attributes in engineering materials database. However, it is still not clear when a measure is
truly elective in large data sets. So there is a need for a logically simple, systematic and
scientific method or mathematical tool to guide designers in selecting proper materials while
designing the new materials. In this paper, linear regression model is being proposed for
measuring correlated data and predicating the continues attribute values from the large
materials database. This method helps to find the relationships between two sub properties of
mechanical property of different types of materials and helps to predict the properties of
unknown materials. The method presenting here effectively satisfies for engineering materials
database, and shows the knowledge discovery from large volume of materials database.
Studying on regression analysis suggests that data mining techniques can contribute to the
investigation on materials informatics, and for discovering the knowledge in the materials
database, which make the manufacturing industries to hoard the waste of sampling the newly
materials
This document discusses improved K-means clustering techniques. It begins with an introduction to data mining and clustering. K-means clustering groups data objects into k clusters by minimizing distances between objects and cluster centers. However, K-means has limitations such as dependency on initialization. The document proposes a new clustering algorithm that uses an iterative relocation technique and distance determination approach to improve upon K-means clustering. It compares the computational complexity of K-means and K-medoids clustering algorithms.
Survey on MapReduce in Big Data Clustering using Machine Learning AlgorithmsIRJET Journal
This document summarizes research on using MapReduce techniques for big data clustering with machine learning algorithms. It discusses how traditional clustering algorithms do not scale well for large datasets. MapReduce allows distributed processing of large datasets in parallel. The document reviews several studies that implemented clustering algorithms like k-means using MapReduce on Hadoop. It found this improved efficiency and reduced complexity compared to traditional approaches. Faster processing of large datasets enables applications in areas like education and healthcare.
This document describes a classroom exercise where students developed deep neural networks to model and predict adsorption equilibrium data. The exercise introduced students to artificial intelligence and deep learning concepts. Students used MATLAB to create neural networks that modeled adsorption of acids by activated carbon at different temperatures, comparing results to theoretical models. The goals were to teach AI methodology, increase coding skills, and show neural networks can accurately model complex chemical engineering processes. Feedback confirmed students gained knowledge of machine learning terms and abilities to develop simple or sophisticated neural networks for modeling unit operations.
This paper discusses the several research methodologies that can
be used in Computer Science (CS) and Information Systems
(IS). The research methods vary according to the science
domain and project field. However a little of research
methodologies can be reasonable for Computer Science and
Information System.
The document discusses nonmetric similarity search problems in complex domains. It begins by defining similarity measuring and similarity search. Specifically, it defines similarity spaces, similarity functions that assign scores to object pairs, and two common similarity queries: range queries and k-nearest neighbor queries. The document then surveys domains that require nonmetric similarity functions for effective similarity search, and methods for efficient nonmetric similarity search.
On nonmetric similarity search problems in complex domainsunyil96
This document surveys the use of nonmetric similarity functions for efficient similarity search across complex domains. It begins by discussing the growth of digital data and need for content-based retrieval beyond text-based search. Similarity functions were traditionally metric, but increasingly complex data requires nonmetric functions. The document scopes the topic to context-free, static nonmetric functions and surveys domains using them along with techniques for efficient nonmetric similarity search, both exact and approximate. It aims to demonstrate the importance of nonmetric search across disciplines and review current methods.
This document summarizes a systematic literature review of 40 empirical studies on learning analytics and educational data mining from 2008-2013. The review aimed to document applied research approaches, identify strengths and weaknesses, and suggest opportunities for future research. Four major directions of LA/EDM empirical research were identified: 1) predicting student performance, 2) understanding student behavior, 3) improving educational systems, and 4) developing analytic methods/tools. The results highlighted the added value of LA/EDM in improving learning and informed decision making, but also identified opportunities to explore new technologies and research questions.
This document discusses using machine learning clustering algorithms to analyze stock market data. It compares the K-means, COBWEB, DBSCAN, EM and OPTICS clustering algorithms in the WEKA tool on a stock market dataset containing 420 instances and 6 attributes. The K-means algorithm had the best performance with the lowest error and fastest runtime. It clustered the data into 4 groups in 0.16 seconds. The COBWEB algorithm clustered the data into 107 groups in 27.88 seconds. The DBSCAN algorithm found 21 clusters in 3.97 seconds. The paper concludes that K-means is best suited for stock market data mining applications due to its simplicity and speed compared to other algorithms.
Data Science Demystified_ Journeying Through Insights and InnovationsVaishali Pal
In the digital age, data has emerged as one of the most valuable resources, driving decision-making processes across industries. Data science, the interdisciplinary field that extracts insights and knowledge from structured and unstructured data, plays a pivotal role in leveraging this resource. This section provides an overview of data science, its importance, and its applications in various domains.
A Survey and Comparative Study of Filter and Wrapper Feature Selection Techni...theijes
Feature selection is considered as a problem of global combinatorial optimization in machine learning, which reduces the number of features, removes irrelevant, noisy and redundant data. However, identification of useful features from hundreds or even thousands of related features is not an easy task. Selecting relevant genes from microarray data becomes even more challenging owing to the high dimensionality of features, multiclass categories involved and the usually small sample size. In order to improve the prediction accuracy and to avoid incomprehensibility due to the number of features different feature selection techniques can be implemented. This survey classifies and analyzes different approaches, aiming to not only provide a comprehensive presentation but also discuss challenges and various performance parameters. The techniques are generally classified into three; filter, wrapper and hybrid.
Data Mining Framework for Network Intrusion Detection using Efficient TechniquesIJAEMSJORNAL
The implementation measures the classification accuracy on benchmark datasets after combining SIS and ANNs. In order to put a number on the gains made by using SIS as a strategic tool in data mining, extensive experiments and analyses are carried out. The predicted results of this investigation will have implications for both theoretical and applied settings. Predictive models in a wide variety of disciplines may benefit from the enhanced classification accuracy enabled by SIS inside ANNs. An invaluable resource for scholars and practitioners in the fields of AI and data mining, this study adds to the continuing conversation about how to maximize the efficacy of machine learning methods.
The document discusses using artificial intelligence (AI) to accelerate materials innovation for clean energy applications. It outlines six elements needed for a Materials Acceleration Platform: 1) automated experimentation, 2) AI for materials discovery, 3) modular robotics for synthesis and characterization, 4) computational methods for inverse design, 5) bridging simulation length and time scales, and 6) data infrastructure. Examples of opportunities include using AI to bridge simulation scales, assist complex measurements, and enable automated materials design. The document argues that a cohesive infrastructure is needed to make effective use of AI, data, computation, and experiments for materials science.
Explaining GitHub Actions Failures with Large Language Models Challenges, In...ssuserb14185
GitHub Actions (GA) has become the de facto tool that developers use to automate software workflows, seamlessly building, testing, and deploying code. Yet when GA fails, it disrupts development, causing delays and driving up costs. Diagnosing failures becomes especially challenging because error logs are often long, complex and unstructured. Given these difficulties, this study explores the potential of large language models (LLMs) to generate correct, clear, concise, and actionable contextual descriptions (or summaries) for GA failures, focusing on developers’ perceptions of their feasibility and usefulness. Our results show that over 80% of developers rated LLM explanations positively in terms of correctness for simpler/small logs. Overall, our findings suggest that LLMs can feasibly assist developers in understanding common GA errors, thus, potentially reducing manual analysis. However, we also found that improved reasoning abilities are needed to support more complex CI/CD scenarios. For instance, less experienced developers tend to be more positive on the described context, while seasoned developers prefer concise summaries. Overall, our work offers key insights for researchers enhancing LLM reasoning, particularly in adapting explanations to user expertise.
https://arxiv.org/abs/2501.16495
What Do Contribution Guidelines Say About Software Testing? (MSR 2025)Andre Hora
Software testing plays a crucial role in the contribution process of open-source projects. For example, contributions introducing new features are expected to include tests, and contributions with tests are more likely to be accepted. Although most real-world projects require contributors to write tests, the specific testing practices communicated to contributors remain unclear. In this paper, we present an empirical study to understand better how software testing is approached in contribution guidelines. We analyze the guidelines of 200 Python and JavaScript open-source software projects. We find that 78% of the projects include some form of test documentation for contributors. Test documentation is located in multiple sources, including CONTRIBUTING files (58%), external documentation (24%), and README files (8%). Furthermore, test documentation commonly explains how to run tests (83.5%), but less often provides guidance on how to write tests (37%). It frequently covers unit tests (71%), but rarely addresses integration (20.5%) and end-to-end tests (15.5%). Other key testing aspects are also less frequently discussed: test coverage (25.5%) and mocking (9.5%). We conclude by discussing implications and future research.
Pixologic ZBrush Crack Plus Activation Key [Latest 2025] New Versionsaimabibi60507
Copy & Past Link👉👉
https://dr-up-community.info/
Pixologic ZBrush, now developed by Maxon, is a premier digital sculpting and painting software renowned for its ability to create highly detailed 3D models. Utilizing a unique "pixol" technology, ZBrush stores depth, lighting, and material information for each point on the screen, allowing artists to sculpt and paint with remarkable precision .
How Valletta helped healthcare SaaS to transform QA and compliance to grow wi...Egor Kaleynik
This case study explores how we partnered with a mid-sized U.S. healthcare SaaS provider to help them scale from a successful pilot phase to supporting over 10,000 users—while meeting strict HIPAA compliance requirements.
Faced with slow, manual testing cycles, frequent regression bugs, and looming audit risks, their growth was at risk. Their existing QA processes couldn’t keep up with the complexity of real-time biometric data handling, and earlier automation attempts had failed due to unreliable tools and fragmented workflows.
We stepped in to deliver a full QA and DevOps transformation. Our team replaced their fragile legacy tests with Testim’s self-healing automation, integrated Postman and OWASP ZAP into Jenkins pipelines for continuous API and security validation, and leveraged AWS Device Farm for real-device, region-specific compliance testing. Custom deployment scripts gave them control over rollouts without relying on heavy CI/CD infrastructure.
The result? Test cycle times were reduced from 3 days to just 8 hours, regression bugs dropped by 40%, and they passed their first HIPAA audit without issue—unlocking faster contract signings and enabling them to expand confidently. More than just a technical upgrade, this project embedded compliance into every phase of development, proving that SaaS providers in regulated industries can scale fast and stay secure.
This presentation explores code comprehension challenges in scientific programming based on a survey of 57 research scientists. It reveals that 57.9% of scientists have no formal training in writing readable code. Key findings highlight a "documentation paradox" where documentation is both the most common readability practice and the biggest challenge scientists face. The study identifies critical issues with naming conventions and code organization, noting that 100% of scientists agree readable code is essential for reproducible research. The research concludes with four key recommendations: expanding programming education for scientists, conducting targeted research on scientific code quality, developing specialized tools, and establishing clearer documentation guidelines for scientific software.
Presented at: The 33rd International Conference on Program Comprehension (ICPC '25)
Date of Conference: April 2025
Conference Location: Ottawa, Ontario, Canada
Preprint: https://arxiv.org/abs/2501.10037
PDF Reader Pro Crack Latest Version FREE Download 2025mu394968
🌍📱👉COPY LINK & PASTE ON GOOGLE https://dr-kain-geera.info/👈🌍
PDF Reader Pro is a software application, often referred to as an AI-powered PDF editor and converter, designed for viewing, editing, annotating, and managing PDF files. It supports various PDF functionalities like merging, splitting, converting, and protecting PDFs. Additionally, it can handle tasks such as creating fillable forms, adding digital signatures, and performing optical character recognition (OCR).
Copy & Paste On Google >>> https://dr-up-community.info/
EASEUS Partition Master Final with Crack and Key Download If you are looking for a powerful and easy-to-use disk partitioning software,
How to Batch Export Lotus Notes NSF Emails to Outlook PST Easily?steaveroggers
Migrating from Lotus Notes to Outlook can be a complex and time-consuming task, especially when dealing with large volumes of NSF emails. This presentation provides a complete guide on how to batch export Lotus Notes NSF emails to Outlook PST format quickly and securely. It highlights the challenges of manual methods, the benefits of using an automated tool, and introduces eSoftTools NSF to PST Converter Software — a reliable solution designed to handle bulk email migrations efficiently. Learn about the software’s key features, step-by-step export process, system requirements, and how it ensures 100% data accuracy and folder structure preservation during migration. Make your email transition smoother, safer, and faster with the right approach.
Read More:- https://www.esofttools.com/nsf-to-pst-converter.html
F-Secure Freedome VPN 2025 Crack Plus Activation New Versionsaimabibi60507
Copy & Past Link 👉👉
https://dr-up-community.info/
F-Secure Freedome VPN is a virtual private network service developed by F-Secure, a Finnish cybersecurity company. It offers features such as Wi-Fi protection, IP address masking, browsing protection, and a kill switch to enhance online privacy and security .
Designing AI-Powered APIs on Azure: Best Practices& ConsiderationsDinusha Kumarasiri
AI is transforming APIs, enabling smarter automation, enhanced decision-making, and seamless integrations. This presentation explores key design principles for AI-infused APIs on Azure, covering performance optimization, security best practices, scalability strategies, and responsible AI governance. Learn how to leverage Azure API Management, machine learning models, and cloud-native architectures to build robust, efficient, and intelligent API solutions
Societal challenges of AI: biases, multilinguism and sustainabilityJordi Cabot
Towards a fairer, inclusive and sustainable AI that works for everybody.
Reviewing the state of the art on these challenges and what we're doing at LIST to test current LLMs and help you select the one that works best for you
Solidworks Crack 2025 latest new + license codeaneelaramzan63
Copy & Paste On Google >>> https://dr-up-community.info/
The two main methods for installing standalone licenses of SOLIDWORKS are clean installation and parallel installation (the process is different ...
Disable your internet connection to prevent the software from performing online checks during installation
Download YouTube By Click 2025 Free Full Activatedsaniamalik72555
Copy & Past Link 👉👉
https://dr-up-community.info/
"YouTube by Click" likely refers to the ByClick Downloader software, a video downloading and conversion tool, specifically designed to download content from YouTube and other video platforms. It allows users to download YouTube videos for offline viewing and to convert them to different formats.
Interactive Odoo Dashboard for various business needs can provide users with dynamic, visually appealing dashboards tailored to their specific requirements. such a module that could support multiple dashboards for different aspects of a business
✅Visit And Buy Now : https://bit.ly/3VojWza
✅This Interactive Odoo dashboard module allow user to create their own odoo interactive dashboards for various purpose.
App download now :
Odoo 18 : https://bit.ly/3VojWza
Odoo 17 : https://bit.ly/4h9Z47G
Odoo 16 : https://bit.ly/3FJTEA4
Odoo 15 : https://bit.ly/3W7tsEB
Odoo 14 : https://bit.ly/3BqZDHg
Odoo 13 : https://bit.ly/3uNMF2t
Try Our website appointment booking odoo app : https://bit.ly/3SvNvgU
👉Want a Demo ?📧 [email protected]
➡️Contact us for Odoo ERP Set up : 091066 49361
👉Explore more apps: https://bit.ly/3oFIOCF
👉Want to know more : 🌐 https://www.axistechnolabs.com/
#odoo #odoo18 #odoo17 #odoo16 #odoo15 #odooapps #dashboards #dashboardsoftware #odooerp #odooimplementation #odoodashboardapp #bestodoodashboard #dashboardapp #odoodashboard #dashboardmodule #interactivedashboard #bestdashboard #dashboard #odootag #odooservices #odoonewfeatures #newappfeatures #odoodashboardapp #dynamicdashboard #odooapp #odooappstore #TopOdooApps #odooapp #odooexperience #odoodevelopment #businessdashboard #allinonedashboard #odooproducts
Douwan Crack 2025 new verson+ License codeaneelaramzan63
Copy & Paste On Google >>> https://dr-up-community.info/
Douwan Preactivated Crack Douwan Crack Free Download. Douwan is a comprehensive software solution designed for data management and analysis.
Avast Premium Security Crack FREE Latest Version 2025mu394968
🌍📱👉COPY LINK & PASTE ON GOOGLE https://dr-kain-geera.info/👈🌍
Avast Premium Security is a paid subscription service that provides comprehensive online security and privacy protection for multiple devices. It includes features like antivirus, firewall, ransomware protection, and website scanning, all designed to safeguard against a wide range of online threats, according to Avast.
Key features of Avast Premium Security:
Antivirus: Protects against viruses, malware, and other malicious software, according to Avast.
Firewall: Controls network traffic and blocks unauthorized access to your devices, as noted by All About Cookies.
Ransomware protection: Helps prevent ransomware attacks, which can encrypt your files and hold them hostage.
Website scanning: Checks websites for malicious content before you visit them, according to Avast.
Email Guardian: Scans your emails for suspicious attachments and phishing attempts.
Multi-device protection: Covers up to 10 devices, including Windows, Mac, Android, and iOS, as stated by 2GO Software.
Privacy features: Helps protect your personal data and online privacy.
In essence, Avast Premium Security provides a robust suite of tools to keep your devices and online activity safe and secure, according to Avast.