3th International Conference on Machine Learning Techniques and Data Science (MLDS 2022)

November 26 ~ 27, 2022, London, United Kingdom

Accepted Papers

Depression Detection using Machine and Deep Learning Models to Assess Mental Health of Social Media Users


Smita Ghosh1, Sneha Ghosh2, Diptaraj Sen2, Pramita Das3, 1Department of Mathematics and Computer Science, Santa Clara University, California, USA, 2Department of Computer Science and Engineering, University of Engineering andManagement, Kolkata, India, 3Department of Electronics and Communication Engineering, National Institute of Technology Durgapur, Durgapur, India

ABSTRACT

During the COVID-19 pandemic millions of people were af ected due to quarantine and restrictions. With more than half of the worlds population active on social media, people resorted to these platforms as their outlet for emotions. This led to researchers analysing content on social media to detect depression by studying the patterns of content posting. This paper focuses on finding a data-driven metric called ‘Happiness Factor’ of a user to assess their mental health. Various models were trained to classify a post as ‘depressed’. A user’s ‘Happiness Factor’ was calculated based on the nature of their posts. This metric identifies degrees of depression of a user. The results show the ef ectiveness of the classifier in identifying the depression level. Also, a Mental Health Awareness Resource System is proposed which recommends mental health awareness resources to users on their social media interface based on their ‘Happiness Factor’.

m

KEYWORDS

Depression Detection, Machine Learning, Deep Learning, Universal Sentence Encoder, Social Media.


Query Optimization meets Reinforcement Learning


Enamul Haque, David R. Cheriton School of Computer Science, University of Waterloo, Ontario, Canada

ABSTRACT

Query optimization is one of the most important tasks of relational database management systems that, if improved properly, can help achieve higher performance in terms of time and resources. Configuration tuning of diverse database instances in distributed systems and optimization of query workloads for cloud databases are also important performance indicators to meet demands of the growing user base globally. But these problems are already NP-Hard, and solutions depend mostly on heuristics-based approximation or randomized algorithms. Here, in this work we explore how advances in Reinforcement Learning (RL) is contributing to this branch of computer systems research.

KEYWORDS

Machine Learning, Database, Query Optimization, Reinforcement Learning, Deep Learning.


Prediction of Genetic Disorders using Machine Learning


Sadichchha Naik1, Amisha Panchal1, Disha Nevare1 and Dr. Chhaya Pawar2, 1Student, Department of Computer Engineering, Datta Meghe College of Engineering, Navi Mumbai, India, 2Asst. Professor, Department of Computer Engineering, Datta Meghe College of Engineering, Navi Mumbai, India

ABSTRACT

A genetic disorder is a health condition that is usually caused by mutations in DNA or changes in the number or overall structure of chromosomes. Several types of commonly-known diseases are related to hereditary gene mutations. Genetic testing aids patients in making important decisions in the prevention, treatment, or early detection of hereditary disorders. With increasing population, studies have shown that there has been an exponential increase in the number of genetic disorders. Genetic disorders impact not only the physical health, but also the psychological and social well-being of patients and their families. Genetic disorders have powerful effects on families. Like many chronic conditions, they may require continual attention and lack cures or treatments. Low awareness of the importance of genetic testing contributes to the increase in the incidence of hereditary disorders. Many children succumb to these disorders and it is extremely important that genetic testing be done during pregnancy. In that direction, the project aims to predict Genetic Disorder and Disorder Subclass using a Machine Learning Model trained from a medical dataset. The model being derived out of a predictor and two classifiers, shall predict the presence of genetic disorder and further specify the disorder and disorder subclass, if present.

KEYWORDS

Genetic disorder, Machine Learning, Medical dataset.


Mining Movement Patterns to Separate Rugby Super League Players into Groups


Victor Elijah Adeyemo1,2,3,4 Anna Palczewska1, Ben Jones2,3,4,5,6 Dan Weaving2,4, 1School of Built Environment, Engineering and Computing, Leeds Beckett University, Leeds, UK, 2Carnegie Applied Rugby Research (CARR) Centre, Carnegie School of Sport, Leeds Beckett University, Leeds, UK, 3England Performance Unit, Rugby Football League, Leeds, UK, 4Leeds Rhinos Rugby League Club, Leeds, UK, 5School of Science and Technology, University of New England, Armadale, Australia, 6Division of Exercise Science and Sports Medicine, Department of Human Biology, Rica, Cape Town, South Africa

ABSTRACT

The application of pattern mining algorithms to extract movement/event patterns from sports big data enables the extraction of on-field activities, provides context to such activities and enhances training specificity. Currently, aggregated physical performance indicators (e.g., total distance covered) are used to separate players into positional groups but these indicators do not capture the sequential nature of match activities. As there are various types of pattern mining algorithms, this study aims to identify which one discovers the best set of movement patterns (on-field activities) to separate players into two playing positions and utilize classification algorithms to find the most accurate separation. Three pattern mining frameworks were implemented to extract movement patterns and five machine learning classification algorithms to separate groups of players via a case study (i.e., two Elite Rugby League players playing positions). The pattern mining frameworks are Sequential Movement Pattern-mining (SMP), l-length Closed Contiguous (LCC) and Apriori Closed (APR). Five classifiers were fitted on tabular datasets (whose independent variables are set of movement patterns) to determine which type of movement patterns accurately separates the groups. Extracted “LCC” and “SMP” on-field activities shared a 0.179 Jaccard similarity score (18%) as both are consecutive patterns. Extracted “APR” on-field activities shared no significant similarity with both extracted “LCC” and “SMP” on-field activities because it mined non-consecutive patterns. Multi-layered Perceptron algorithm fitted on the dataset whose independent variables were the extracted “LCC” on-field activities achieved the highest accuracy of 91.02% ± 0.02 and precision, recall and F1 scores of 0.91. Therefore, we recommend the extraction of closed contiguous (consecutive) over non-consecutive patterns for separating groups of players.

KEYWORDS

Pattern Mining, Performance Analysis, Rugby League, Sports Analytics, Machine Learning.


An AI based solution to reduce food loss by predicting the expiry date of Perishable Food


Trewon Weerasooriya and Kishore Kumar, Faculty of Computing, Sri Lanka Institute of Information Technology Malabe, Sri Lanka

ABSTRACT

“Food Loss and Waste” is a growing issue for the ecology, economy, and social aspects of the world. The social, environmental, and economic factors are all negatively impacted by it. Food loss and waste are mostly caused by several factors, including poor planning, excessive production, and customer perception. This study focuses on the hospitality and food services industry’s two primary sub-sectors, specifically Sri Lanka’s hotels and restaurants. It develops a solution to eliminate the causes highlighted through the use of technologies like artificial intelligence and cloud services which eliminates or reduces food loss and waste that occurs during the processes of the food supply chain. Primarily this paper is based on a machine learning based system which focuses on prediction of expiry dates of perishable food and generating donation recommendations based on the predicted results. While implementing a platform to bridge the gap between food donors and food donation recipients. The collected datasets with the use of two machine learning models, revealed that the expiry date of perishable food items can be predicted with a mean squared error of 0.7651. Which means that, this model predicts expiry dates with a standard error which is less than a day.

KEYWORDS

Machine Learning, Artificial intelligence, Random Forest, XGBoost.


A Comprehensive Study on Machine Learning Methods to Increase the Prediction Accuracy of Classifiers and Reduce the Number of Medical Tests Required to Diagnose Alzheimers Disease


Md. Sharifur Rahman and Girijesh Prasad, School of Computing, Engineering and Intelligent Systems, Ulster University, Northern Ireland, UK

ABSTRACT

Alzheimers patients gradually lose their ability to think, behave, and interact with others. Medical history, laboratory tests, daily activities, and personality changes can all be used to diagnose the disorder. A series of time-consuming and expensive tests are used to diagnose the illness. The most effective way to identify Alzheimers disease is using a Random-forest classifier in this study, along with various other Machine Learning techniques. The main goal of this study is to fine-tune the classifier to detect illness with fewer tests while maintaining a reasonable disease discovery accuracy. We successfully identified the condition in almost 94% of cases using four of the thirty frequently utilized indicators.

KEYWORDS

Machine Learning, Accuracy, Precision, Recall, Classifier, Random Forest, Alzheimers.


SERP Evaluation and Model Interpretation Based on Behavior Sequence


Xiaoyu Dong and Shen Shen and Jinkang Jia and Yifan Wang, Baidu Inc, Beijing, China

ABSTRACT

SERP (Search Engine Result Page) quality evaluation plays a vital role in industrial practice. With the rapid iterations of search engine, traditional page-level metrics like click ratio, dwell time can no longer evaluate user experience on various templates of results. To promote evaluation accuracy, we implement Transformer to capture the sequential patterns from behavior sequence data. In recent studies, approaches focusing on modeling behavior sequences have emerged. Some studies concentrate on feature engineering by extracting subsequence patterns, others focus on end-to-end deep learning models. While widely used, these two methods both have drawbacks, either a risk of distortion of true subsequence patterns or difficulty for interpretation. Here we implement Transformer to give considerations to both completeness of sequential patterns and model interpretation. To find the best way of modeling behavior sequence data with continuous features, we adopt two embedding methods to predict SERP quality evaluation, and the second one achieves good promotion. What’s more, we develop a novel interpretation method for transformer models and demonstrate its ability to make interpretations for subsequence patterns.

KEYWORDS

Behavior sequence, Transformer, Model Interpretation.


Web Application and Internet Security Threats


Vaibhav Katiyar, University Institute of engineering and technology, CSJMU, Kanpur

ABSTRACT

Computer and network security are one of the most challenging topics in the Information Technology research community. Internet security is a significant subject that may affect a wide range of Internet users. People that use Internet to sell, buy and even to communicate needs their communications to be safe and secure. This paper is discussing the different aspects of Internet and networking security and weakness. Main elements of networking security techniques such as the firewalls, passwords, encryption, authentication and integrity are also discussed in this paper. The anatomy of a web applications attack and the attack techniques are also covered in details. The security of high-speed Internet as the growth of its use has stained the limits of existing network security measures. Therefore, other security defense techniques related to securing of high-speed Internet and computer security in the real world are studied as well such as, DNS, One-Time Password and defending the network as a whole.This paper is also surveyed the worm epidemics in the high- speed networks and their unprecedented rates spread.

KEYWORDS

Web application Security, Network Security, Protection tools, SQL Injection, Firewall, and Intrusion Detection System.


Cloud Oriented Virtual Fragmented Database


Ahmad Shukri Mohd Noor, Nur F. Mat Zain, Rabiei Mamat and Noor hafhizah A. Rahim, Faculty of Ocean Engineering Technology and Informatics, Universiti Malaysia Terengganu

ABSTRACT

Fragmentation is way to replicate the data into multiple servers to increase the availability and accessibility. The existing technique Hierarchical Replication Scheme (HRS), Read One-write-All (ROWA) and Branch Replication Scheme (BRS), these three techniques are the most common techniques that used to replicate the data but they have their own drawbacks such as communication cost that is high. The proposed technique in this research is Mater-slave replica technique which is uses to distribute the data into servers, so it enables the user to access it from any server in case one server fail. This research about creates cloud environment and replicate the data into virtual servers to enables the user to access the data at any location and in any time. VirtualBox used in this research to create windows guested in the hard disk of the local host that used these windows as virtual servers and stored the data of the master replica server. by using three virtual servers, a series of experiments will be performed, and the results will be compared to the existing techniques, then the findings of the experimental will achieve the replication consistency of the database.

KEYWORDS

Virtual Machine, Database fragmentation, Database Replication, Mater-salve replica, Data Availability.


Artificial Intelligence to Analyze Mineral Resource Cost Estimation using Ordinary Kriging and Schlumberger Methods


Paryati1 and Krit Salahddine2, 1Department of Tehchnic Engineering, UPN “Veteran” Yogyakarta, Indonesia, 2Ibn Zohr University, Agadir City, Morocco

ABSTRACT

Artificial intelligence to analyze the estimated cost of mineral resources, mostly done in mining areas, to estimate the mining area which is an element of mineral reserves in nature. So to determine the commodity and the estimated cost for the mining area, a study of andesite estimation and a study of technical factors that influence the development of mining processes and estimates at PT. Varia Usaha Beton Mandalika, East Lombok Regency, West Nusa Tenggara, Indonesia. Thus, with an exploratory study, a description of the condition of andesite mineral commodities will be obtained at the location of the area. One approach used to estimate andesite is to use the Schlumberger method and use ordinary kriging so that it can simplify and speed up the work process, especially when carrying out andesite activities in a very large mining area. Research conducted in the mining area owned by PT. Varia Usaha Beton Mandalika, East Lombok Regency, West Nusa Tenggara, Indonesia. Mining companies use the Schlumberger configuration method and ordinary kriging. The results obtained by andesite mineral resource estimates are very good and relevant and valid. This mining method is used in the production operation process of andesite resource estimation. The selection model is based on the type of commodity and the contours of the mining area in the form of sand and stone. This method has the ability to detect the presence of inhomogeneous rock layers by comparing the apparent resistivity value with changes in the MN/2 electrode distance and is able to detect the presence of inhomogeneous rock layers. So this method is very suitable for estimating andesite resources in mining areas.

KEYWORDS

Artificial Intelligence, Estimation, Schlumberger, Vertical Sounding, Ordinary Kriging.


Agent-based Modeling and Simulation of Complex Industrial Systems: Case Study of the Dehydration Process Noureddine


Seddari1,2, Sohaib Hamioud3, Abdelghani Bouras4, Sara Kerraoui1 and Nesrine Menai1, 1LICUS Laboratory, Department of Computer Science, Université 20 Août 1955-Skikda, Skikda 21000, Algeria, 2LIRE Laboratory, Abdelhamid Mehri-Constantine 2 University, Constantine 25000,Algeria, 3LISCO Laboratory, Computer Science Department, Badji-Mokhtar University,Annaba 23000, Algeria, 4Department of Industrial Engineering, College of Engineering, AlfaisalUniversity,Riyadh 11533, Saudi Arabia

ABSTRACT

Agent-based modeling and simulation (ABMS) is a new approach to modeling autonomous systems and interacting agents. This method is becoming more and more popular for its efficiency and simplicity. It constitutes an approach in the field of modeling complex systems. Indeed, ABMS offers, contrary to other types of simulations, the possibility of directly representing the simulated entities, their behaviors, and their interactions without having to recourse to mathematical equations. This work is a contribution in this sense, the goal is to propose an agent-based model to simulate an industrial system. The latter presents the problem of complexity, which can be mastered with the Multi-Agent Systems (MAS) approach. Our model is validated by a case study of the natural gas dehydration process. The latter is consolidated by a simulation made in the multi-agent platform JADE (Java Agent DEvelopment Framework).

KEYWORDS

Agent-based modeling and simulation (ABMS), Industrial system, Multi-Agent Systems (MAS), Multi-agent platform JADE.


Data Augmentation Using GANs


Dr Simant Prakoonwit1 and Emilija Strelcenia2, 1Department of Creative Technology, Bournemouth University, Bournemouth, England, 2Department of Creative Technology, Bournemouth University, Bournemouth, England

ABSTRACT

GANs have been found to be very successful in a number of applications, most notably image generation. However, there are still a number of challenges that need to be addressed in order to make GANs more widely applicable. In this literature review, we shall discuss various loss functions and variants of the GANs and their uses in the field of machine and deep learning that incorporates these improvements and can reliably be utilized for machine learning applications.

KEYWORDS

GANs, synthetic data, class imbalance, loss functions.


Towards Tuberculosis Incidence Time Series Forecasting in Colombian Regions: The Antioquia Case


Ivan Camilo Sanchez-Vega1 and Alvaro David Orjuela-Cañon2 and Carlos Enrique Awad Garcia3 and Erika Vergara4 and Maria Angelica Palencia5 and Andres L. Jutinico6, 1Mechanical, Electronics and Biomedical Engineering Faculty, Universidad Antonio Nariño, Bogota D.C., Colombia, 2School of Medicine and Health Sciences, Universidad del Rosario, Bogota D.C., Colombia, 3Subred Integrada de Servicios de Salud Centro Oriente, Bogota D.C., Colombia, 4Universidad Antonio Nariño, Bogota D.C., Colombia, 5Subred Integrada de Servicios de Salud Centro Oriente, Bogota D.C., Colombia, 6Mechanical, Electronics and Biomedical Engineering Faculty, Universidad Antonio Nariño, Bogota D.C., Colombia

ABSTRACT

Antioquia is a Colombian department where 6.7 million people live. Currently, it is the region of the country with the newest cases of tuberculosis reported in 2021, about 18.8%. In addition, the incidence rate of tuberculosis was 36.8 per 100,000 inhabitants. Public government health policy regarding tuberculosis should aim to prevent the uninfected community, in addition to detecting and treating people with tuberculosis. In this sense, the study of algorithms to predict the epidemic trend should be promoted. This work addresses the prediction of tuberculosis cases in Antioquia, considering data from the health surveillance system between 2007 and 2021. For the prediction, the Kalman filter and the autoregressive model are considered. The results show a better performance using the Kalman filter for the prediction of tuberculosis cases at six weeks.

KEYWORDS

Tuberculosis, Forecasting, Autoregressive Model, Kalman Filter, Performance.


CAVSS: A Commonsense Augmented Variational Sequence To Sequence Model For Language Generation


Mingxuan Liang1, Luchen Zhang2, Junwu Zhu1, Yang Li1 and Xiang Gao1, 1 College of Information Engineering, Yangzhou University, Yangzhou Jiangsu, China, 2 National Computer Network Emergency Response Technical Team/Coordination Center of China

ABSTRACT

Commonsense knowledge as external knowledge enhances the semantic understanding of the input sequences of the model and is of guidance to text generation models. In this paper, we propose a novel approach of incorporating commonsense knowledge for enhancing the performance of end-to-end text generation models. Firstly, given an input sequence and retrieving the relevant knowledge triples, the embedding of the commonsense knowledge and the context vector encoded in the encoder part are spliced for sampling. Then an autoregressive transformation is applied to the sampling to prevent the problem of too slow fitting of simple Gaussian distribution, and a new learning objective is designed in the training phase to make this transformed distribution fit the posterior distribution. In addition, we perform variational operations on the decoding part of the attention mechanism to weaken the attention strength and prevent reconstruction from playing a decisive role in generation while ignoring other modules.

KEYWORDS

Deep Learning, Commonsense Knowledge, Auto-Encoder, Generation.


Solving Security Problem in Access Control Systems


Anh Tuan Truong, Ho Chi Minh City University of Technology, VNU-HCM, Ho Chi Minh City, Vietnam

ABSTRACT

Access control model is one of the most widespread frameworks for controlling the access in an organization. Many extended features have been proposed to make the model increasingly well adapted to real world systems (e.g., the administration of rules in access control system). Although these extensions bring advantages to the system, the interaction between them may create a conflict in it (i.e., the security problem). In this paper, we discuss this problem as well as the automated security analysis solutions that have been proposed to solve the problem. We also propose an approach that uses the capability of a model checker to automatically analyse access control policies. An extensive experimentation shows that the proposed approach outperforms a state-of-the-art analysis technique.

KEYWORDS

CCS CONCEPTS, Security and Trust, Access Control System, Security Analysis.


Performance based Comparative Analysis of Naive Bayes Variations for Text Classification


Ali Abbas, Manish Jaiswal, Prajna Jha, Shreya Agarwal and Tanveer J. Siddiqui, Department of Electronics and Communication, University of Allahabad, Prayagraj, India.

ABSTRACT

With the high consumption of digital data, the problem of unorganized textual data created major challenges in today’s scenario. To overcome this issue of unorganized textual information we perform document classification which is divided into four major phases i.e., pre-processing, feature selection, model training, and model testing. In this paper, we have selected four types of feature vectors, three with weighting techniques and one without weighting. Performance has been tested on these four feature vectors with five variances of naïve bayes classifiers out of which two were not able to perform training due to the sparseness in the dataset and the rest three performed well. In the reported result of the experiment, F1-macro score and the accuracy of Complement Naïve Bayes using term frequency weighting scheme is 0.807901362 and 0.821163038 which outperform all the other feature sets with all the variance of naïve bayes classifiers. In terms of time consumption for training and testing again, the performance of Complement Naïve Bayes using term frequency weighting scheme found best.

KEYWORDS

Multinomial Naïve Bayes, Complement Naïve Bayes, Min-max scaling, tf-idf, machine learning, document classification.


Generic Question Classification for Dialogue Systems


Marine Troadec, Matthis Houl`es, and Philippe Blache, LPL-CNRS, ILCB, Aix-en-Provence, France

ABSTRACT

We present in this paper a new classification approach for identifying questions in dialogue systems. The difficulty in this task is first to be domain-independent, reusable whatever the dialogue application and second to be capable of a real time processing, in order to fit with the needs of reactivity in dialogue systems. The task is then different than that of question classification usually addressed in question answering systems. We propose in this paper a hierarchical classifier in two steps, filtering first question/no-question utterances and second the type of the question. Our method reaches a f-score of 98% for the first step and 97% for the second one, representing the state of the art for this task.

KEYWORDS

Question classification, Dialogue systems, Hierarchical classification.


Global Language Positioning System


Xiaohui Zou1 and Shunpeng Zou2, 1Searle Research Center, No. 100, Renshan Road, Hengqin, Guangdong 519000, China, 2Peng Ge Learning Center,Csanady utca 4/b, 1132 , Budapest, Hungary

ABSTRACT

The aim is to explore a new method of Global Language Positioning System, that is, natural language processing by using GPS sequencing positioning. The method is: first, it is pointed out that Susuer distinguishes language and speech although it is good for Indo -European language, but it is not enough for Chinese; then, in the process of Chinese information processing that can be divided into Yan and Yu as the formal language from Tasky and Calnap, as well as the way of combining a single element set of pure mathematics to combine the metal group set at all levels to build a broad -language formal information processing model that combines Chinese and arithmetic; finally, compare the broad - language bilingual treatment described herein, translated by Google Machines Translate as typical representatives, narrow bilingual processing, such as English -Chinese or Chinese -English machine translation, distinguish between complete matching and partial matching, and completely not matching as three types, and then through human -computer interaction and collaboration methods, the reasons why the centralized inspection is not matched, which provides a reliable basis for the improvement or optimization of the five links of classification, matching, translation, prediction and decision -making. As a result, it not only improved the quality of machine translation significantly, but also provided a brand - new natural language processing methods for further data mining, semantic identification, information extraction, knowledge processing and software modeling, and the improvement of NLP & AI technology quality, by using handling paradigm, namely: Global Language Positioning System. Its significance is: this new paradigm not only practically ef icient and precise, especially Chinese information processing, but also, theoretically, with the existing natural language processing methods, and it is not only the same function which can be made in three aspects of natural language understanding, expert knowledge expression, and software model recognition.

KEYWORDS

Natural Language Processing, Natural Language Understanding, NLP & AI , Susuer, Tasky and Calnap.


The Idea System for Semantic & NLP


Jiawen QIU1 and Yezhen ZHAO1 and Xiaohui ZOU2, 3, 1China Mobile Communications Group Guangdong Co., Ltd. Zhuhai Branch, China, 2Searle Research Center, No. 100, Renshan Road, Zhuhai, Guangdong 519000, China, 3Peking University Alumni Association, Beijing 100871, China

ABSTRACT

The purpose of this paper is to put forward the concept of idea system and try to design its knowledge representation principle on the basis of academic discussion on the formation principle of the conceptual relationship of human knowledge. This method is an idea-based knowledge modeling method. The steps are: first, clearly define the concept as the mapping from sign to meaning; further, clarify that both sign and meaning are just a subclass of idea; finally, determine the separation of the sign relation network and the meaning relation network in the conceptual idea relation network, and modeling and computational processing are unified with the idea relationship network. It is characterized by: the knowledge representation method and principle of the idea relation network. The result is that the idea system, which is a new knowledge modeling tool system, systematically introduces its typical embodiments in combination with word, formula, figure, table, and compares and analyzes the research results of related knowledge graphs. Its significance lies in that: it is significantly different from the traditional method, that is, the knowledge representation model that uses symbols instead of meanings to model. This paper establishes a new method and new principle of knowledge construction model that is assisted by symbols and directly modeled with ideas, which is a new method for cognitive computing. It provides a new way to express knowledge based on meaning rather than symbols.

KEYWORDS

Semantic & NLP, Idea System, Human-Machine Collaboration, Knowledge Graph, Knowledge Modeling, Artificial.


Ambiguity Detection in Requirements Classification Task using Fine-tuned Transformation Technique


Sadeen Alharbi, Department of Software Engineering, King Saud University, Saudi Arabia

ABSTRACT

The software requirement specification (SRS) document is essential in software development. This document influences all subsequent steps in software development. Nevertheless, requirements problems, such as insufficient or ambiguous specifications, can cause misunderstandings during the requirement analysis stage. This influences testing activities and increases the project’s duration and cost overrun risk. This paper represents an intuitive approach to detecting ambiguity in software requirements. The classifier should learn ambiguous features and characteristics extracted from the text on a training set and try to detect similar characteristics from a testing set. To achieve this, this study experimented with two main approaches. The first approach is feature extraction, which uses the hidden states as features and trains a support vector machine (SVM) classifier to assess software requirement ambiguity without modifying the pre-trained model. Unfortunately, this approach only identified 68% of the requirement ambiguity. The second approach is training an end-to-end model that updates the parameters of the pre-trained model. This approach enhanced the baseline results by 13%.

KEYWORDS

Requirements classification, NLP, Ambiguity.


Heterogenous Graph Network for Explainable Medical Event Identification


Fengyi Li and Hang Wang, Ant Group, Shanghai, China

ABSTRACT

The explainability is critical for claim-decision making in medical insurance scenario, which is a challenging task due to various disease. To remedy this dilemma, this paper sets forth a novel architecture of claim-decision making via exploring the historical medical clues and identifying explainable medical events. Specifically, we introduce a heterogeneous graph construction to nourish the information of medical events and conduct text classification. Then, a knowledge fusion mechanism consumes the dynamic relation between events to eliminate the error-prone impact of graph neural networks, and a new witness loss facilitates the consistency between the distributions over the values of graph edges. Our proposed method is evaluated on actual medical insurance claim cases. Improvements of 2.6% F1-Score for claim-decision making and 16.6% Jaccard Similarity Coefficient compared to other counterpart models indicate the superiority of our work.

KEYWORDS

Hierarchical Heterogeneous Graph, Explainability, Text Classification.


Impact of Blind and Automatic Image Classification in Improving Text Detection Performance of the Craft Algorithm


Clarisa V. Albarillo1 and Proceso L. Fernandez Jr2, 1College of Computer Science, Don Mariano Marcos Memorial State University, La Union, Philippines, 2Department of Information Systems & Computer Science, Ateneo de Manila University, Quezon City, Philippines

ABSTRACT

Text detection in natural scenes has been a significant and active research subject in computer vision and document analysis because of its wide range of application as evident by the emergence of the Robust Reading Competition. One of the algorithms which has good text detection performance in the said competition is the Character Region Awareness for Text Detection (CRAFT). Employing the ICDAR 2013 dataset, this study investigates the impact of automatic image classification and blind deconvolution as image pre-processing steps to further enhance the text detection performance of CRAFT. The proposed technique automatically classifies the scene images into two categories, blurry and non-blurry, by utilizing of a Laplacian operator with 100 as threshold. Prior to applying the CRAFT algorithm, images that are categorized as blurry are further pre-processed using blind deconvolution to reduce the blur. The results revealed that the proposed method significantly enhanced the detection performance of CRAFT, as demonstrated by its h-mean of 94.47% compared to the original 91.42% h-mean of CRAFT and even outperformed the top-ranked SenseTime, whose h-mean is 93.62%.

KEYWORDS

Blind Deconvolution, Computer Vision, Image Classification, Information Retrieval, Image Processing.


Dynamic aspect ratio


Njal Borch 1 and Ingar Arntzen, NORCE Norwegian Research Institute

ABSTRACT

Media is to a large extent consumed on devices that have non-standard aspect ratios, both physically or while rendering content.For example, social media platforms often favour 1:1 ratios, TVs 16:9, iPad tablets 4:3 or 3:4, most Androids 16:9 or 9:16, PCs 16:9 or 16:10 and web pages tend to use responsive design and can therefore have almost any aspect ratio In order to ensure good experiences, it is often therefore necessary to create multiple versions of content, where the content is cropped to a more suitable format. Creating multiple encoded version of the content is a static solution though, and there are good reasons for solving this dynamically on the client side. In this paper we make the case for a client side dynamic aspect ratio solution, present work on implementation and experimentation, and finally provide some insights into how such a system could be implemented and provided in real world systems. Our solution was tested on a few different pieces of content from NRK, both drama series and typical TV debates.

KEYWORDS

Dynamic aspect ratio, Focus track, Multi-device, client side, AI analysis.


PRAGAN: Progressive Recurrent Attention Gan with Pre-Trained Vit Discriminator for Single Image Deraining


Bingcai Wei, Liye Zhang and Di Wang

ABSTRACT

Images collected under bad weather conditions are not conducive to the development of visual tasks. To solve this problem, it is a trend to use a variety of neural networks. The ingenious integration of network structures can make full use of the powerful representation and fitting ability of deep learning to complete low-level visual tasks. In this study, we propose a generative adversarial network(GAN) containing multiple attention mechanisms for the image deraining task. Firstly, to our best knowledge, we propose a pre-trained vision transformer(ViT) which is used for the discriminator in our GAN for single image deraining. Secondly, we propose a neural network training method, which can use a small amount of data for training while maintaining promising results and reliable vision quality. A large number of experiments have proved the correctness and effectiveness of our methods.

KEYWORDS

Deep Learning, Image deraining, Neural Network, Vision Transformer.


Deep Learning for the Classification of the Injunction in French Oral Utterances


Asma BOUGRINE, Philippe RAVIER, Abdenour HACINE-GHARBI and Hanane OUACHOUR, Prisme Laboratory, University of Orleans, 12 rue de Blois, 45067 Orléans, France

ABSTRACT

The classification of the injunction in French oral speech is a difficult task. The logarithmic energy can be a good indicator and our goal is to validate the predominance of this prosodic feature using SVM and K-NN. When applied to our first dataset, this feature indeed showed the best classification rates (CR) of 82% for SVM and 71.42% for K-NN. However, the energy was not the relevant feature when applied to our second, heterogeneous and wild, dataset. In order to improve the classification rates, we applied LSTM and CNN networks. With LSTM, a CR=96.15% was found using 6 prosodic features with the first dataset against 64% with the second dataset. The CNN, a network capable of automatically extracting the most relevant features, gave a better result on the second dataset with CR=73% largely exceeding SVM, K-NN and LSTM.

KEYWORDS

Injunction classification, prosodic features, CNN, LSTM, wild oral dataset.


Fruit Type Classification using Deep Learning and Feature Fusion


Harmandeep Singh Gill1, Baljit Singh Khehra2, 1Principal, Mata Gujri Khalsa College(Jalandhar) Punjab-144801, 2Principal, BAM Khalsa College, Garhshankar (Hoshiarpur) Punjab

ABSTRACT

Machine and deep learning applications play a dominant role in the current scenario in the agriculture sector. To date, the classification of fruits using image features has attained the researcher’s attraction very much from the last few years. Fruit recognition and classification is an ill-posed problem due to the heterogeneous nature of fruits. In the proposed work, Convolution neural network (CNN), Recurrent Neural Network (RNN), and Long-short Term Memory (LSTM) deep learning methods are used to extract the optimal image features, and to select features after extraction, and finally, use extracted image features to classify the fruits. To evaluate the performance of the proposed approach, the Support vector machine (SVM) unsupervised learning method, Artificial neuro-fuzzy inference system (ANFIS), and Feed-forward neural network (FFNN) classification results are compared, and observed that the proposed fruit classification approach results are quite efficient and promising.

KEYWORDS

Image classification, Feature Extraction, Deep Learning, Feature Fusion, CNN, RNN, LSTM.


Head Detection in Top-View Fisheye Images using Radius-Aware Loss Function


Xuan Wei and Xiaobo Lu, School of Automation, Southeast University, Nanjing 210096, P.R China

ABSTRACT

People detection is commonly used in computer vision systems, particularly for video surveillance and passenger flow statistics. Unlike standard cameras, fisheye cameras offer a large field of view and reduce occlusions when mounted overhead. However, due to the orientation variation of people in fisheye images, head detection models suffer from severe distortion when applied to fisheye images captured by top-view fisheye cameras. In this work, we develop an end-to-end head detection method against complex situations in top-view fisheye images. The radius-aware loss function is designed, and the channel attention module is added to the model to make our detection method adapt to the fisheye distortion. For evaluation, we collected and annotated new fisheye-image datasets.

KEYWORDS

Head Detection, Fisheye Images, Neural Network.


Research on the Image Segmentation by Watershed Transforms


Sidi Ahmed Taouli, Department of Biomedical Genius, Faculty of technology, University Aboubekr-Belkaid, Tlemcen, Algeria.

ABSTRACT

The processing of medical images play an important role in diagnosis and detection of different pathologies. The mathematical morphology is very suitable for the pretreatment and segmentation of medical images, which present rich information content. In this work we presented a segmentation paradigm by Watershed preceded by a filtering to eliminate insignificant minima, a marking to remove unmarked minima, and finally we implemented a hierarchical segmentation using the mosaic image of the original image. In principle, watershed segmentation depends on ridges to perform a proper segmentation, a property that is often fulfilled in contour detection where the boundaries of the objects are expressed as ridges. Watershed is normally implemented by region growing, based on a set of markers to avoid over segmentation. The diversity of segmentation offers us several ways to segment the image, always we must look for the right method to get good results.

KEYWORDS

Image, Filtering, Mathematical Morphology, Watershed, Segmentation.


A Soft System Methodology Approach to Stakeholder Engagement in Water Sensitive Urban


Lina Ntomene Lukusa and Ulrike Rivett

ABSTRACT

Poor water management can increase the extreme pressure already faced by water scarcity. Water quality and quantity will continue to degrade unless water management is addressed holistically. A holistic approach to water management named Water Sensitive Urban Design (WSUD) has thus been created to facilitate the effective management of water. Traditionally, water management has employed a linear design approach, while WSUD requires a systematic, cyclical approach. In simple terms, WSUD assumes that everything is connected. Hence, it is critical for different stakeholders involved in WSUD to engage and reach a consensus on a solution. However, many stakeholders in WSUD have conflicting interests. Using the soft system methodology (SSM), developed by Peter Checkland, as a problem-solving method, decision-makers can understand this problematic situation from different world views. The SSM addresses ill and complex challenging situations involving human activities in a complex structured scenario. This paper demo nstrates how SSM can be applied to understand the complexity of stakeholder engagement in WSUD. The paper concludes that SSM is an adequate solution to understand a complex problem better and propose efficient solutions.

KEYWORDS

Co-design, ICT Platform, Soft Systems Methodology, Water Sensitive Urban Design.


A Soft System Methodology Approach to Stakeholder Engagement in Water Sensitive Urban


Lina Ntomene Lukusa and Ulrike Rivett

ABSTRACT

Poor water management can increase the extreme pressure already faced by water scarcity. Water quality and quantity will continue to degrade unless water management is addressed holistically. A holistic approach to water management named Water Sensitive Urban Design (WSUD) has thus been created to facilitate the effective management of water. Traditionally, water management has employed a linear design approach, while WSUD requires a systematic, cyclical approach. In simple terms, WSUD assumes that everything is connected. Hence, it is critical for different stakeholders involved in WSUD to engage and reach a consensus on a solution. However, many stakeholders in WSUD have conflicting interests. Using the soft system methodology (SSM), developed by Peter Checkland, as a problem-solving method, decision-makers can understand this problematic situation from different world views. The SSM addresses ill and complex challenging situations involving human activities in a complex structured scenario. This paper demo nstrates how SSM can be applied to understand the complexity of stakeholder engagement in WSUD. The paper concludes that SSM is an adequate solution to understand a complex problem better and propose efficient solutions.

KEYWORDS

Co-design, ICT Platform, Soft Systems Methodology, Water Sensitive Urban Design.


Indigenous Connection to Virtual Environments using Tribal Language Object Recognition as a Fuzzy Inference System


Kevin Shedlock and Thomas Chandler, School of Computer Science and Engineering, Victoria University of Wellington, New Zealand

ABSTRACT

For indigenous communities to engage with technology like virtual reality, being able to feel traditionally located inside the application is critical. Interactions with familiar language objects and methods in the real world need to correspond with experiences in the virtual world. Any deviation from tradition may result in a disconnection from the application. We investigate a method for embedding indigenous language during the processing of the virtual reality artefact. We argue that a virtual reality environment with strong tribal language and symbolic connections is important to alleviate any disconnect by tribal users of virtual reality. We ask; does the use of indigenous language objects improve the virtual reality experience? To explore this question, we quizzed indigenous Māori participants of Aotearoa, New Zealand. Then, implemented an indigenously organized process for evaluating a fuzzy inference system and report the results to better-understand the importance of language during the construction of virtual reality.

KEYWORDS

Indigenous Research, IT Artefact, Virtual Reality, Indigenous Language, Soft Computing.


Explainable AI and Interpretable Model for Insurance Premium Prediction


Umar Isa A and Anil Fernando, Department of Computer and Information Science, University of Strathclyde, UK

ABSTRACT

Traditional machine learning metrics such as precision and recall, accuracy, MSE, RMSE, etc. are not enough for a practitioner to be confident about the performance and dependability of their models. When unrelated features are included in the prediction process, machine learning models with greater than 99% accuracy score may fail to recognize particular classes from the dataset. We need to explain the model to domain specialists who require a human-understandable explanation in addition to ML professionals to establish trust in the prediction of our model. This is achieved by creating a model-independent, locally accurate explanation set that makes the primary models conclusions understandable to anyone, including those who are not experts in the insurance industry. For effective human interaction with machine learning systems, this model interpretability is crucial. In addition to completing and supporting set validations in model selection, it is crucial to explain individual predictions in order to gauge trust. In this study, we proposed the use of LIME and SHAP approaches to understand and explain a model developed using random forest regression to predict insurance premiums. The SHAP algorithms drawback, as seen in our experiments, is its lengthy computing time—in order to produce the findings, it must compute every possible combination. In addition, the experiments conducted were intended to focus on the models interpretability and explainability using LIME and SHAP, not the insurance premium charge prediction. Two experiments were conducted. Experiment one was conducted to interpret the random forest regression model using LIME techniques. In experiment two, we used the SHAP technique to interpret the model.

KEYWORDS

LIME, SHAP, Explainable AI, Random Forest, Machine Learning, Insurance premium.


Evaluation of Machines Learning Algorithms in Detection and Mitigation of DNS-Based Phishing Attacks


Kambey L. Kisambu1 and Eng. Gilberto Makero2, 1Msc, Cyber Security, Department of Computer Science, University of Dodoma, 2Network Engineer, e-Government Authority of Tanzania

ABSTRACT

DNS-based phishing attacks are among the major threats to Internet users that are difficult to defend against because they do not appear to be malicious in nature.Users have been primarytargets for these attacks that aim to steal sensitive information. The DNS protocol is one ofthe approaches that adversaries use to transfer stolen data outside the institutions networkusing various forms of DNS tunneling attacks. This study deals with evaluation of ML algorithms indetection of DNS-based phishing attacks for securing networks.It deeply evaluate the efficacy of the algorithms when integrated with other solution. The main classifiers used such as SVM, KNN, Logistic Regression and Naïve Bayes were evaluated using performance metrics namelyaccuracy, precision, recall and f-score. Based on the findings, the study proposed improvement for securing systems and networks against DNS-based phishingattacks using the best performing ML algorithm to keep pace with attacks evolution.

KEYWORDS

Malware, DNS-based, Phishing attacks, Machine learning, algorithms, DNS filters, DNS Poisoning, Detection, Mitigation techniques.


Radio Map Construction based on BERT for Fingerprint-based Indoor Positioning System


Zhuang Wang, Liye Zhang, Qun Kong, Cong Liu, Aikui Tian, Computer Science and Technology, Shandong University of Technology, Zibo, China

ABSTRACT

Due to the heavy workload of RSS collection and the complex indoor environment, the WLAN signal is easy to disappear. Therefore, it is time-consuming and laborious to build a WLAN fingerprint indoor positioning system based on Radio Map. To quickly deploy an indoor WLAN positioning system, the Bidirectional Encoder Representation from Transformers (BERT) model is used to populate the missing signal in the Radio Map to quickly build the Radio Map. Because the number of input data in the BERT model cannot exceed 512 words. Therefore, we redefine the model structure based on the original BERT model and fill the missing signals in the radio map in parallel. In addition, the sum of each fragment loss function is defined as the total loss function. The experimental results show that using the improved BERT model to fill the missing signals in Radio Map has higher accuracy and shorter time.

KEYWORDS

Indoor Positioning System, WLAN, Radio Map, BERT Model, Loss Function.


Decreasing Noise Effect on the MIMO OFDM Channel Estimation based on Information Theoretic Learning


Mohammad Reza pourmir1, Reza Monsefi1, Ghosheh Abed Hodtani2, 1Computer Department, Engineering Faculty Ferdowsi University of Mashhad (FUM), 2Electrical engineering Department, Engineering Faculty Ferdowsi University of Mashhad (FUM)

ABSTRACT

In this paper, we introduce a new method for multiple-input multiple-output (MIMO) OFDM channel estimation based on deep learning using the information theoretic learning criteria, as a development of previous works, leading to the low complexity in estimating non-linear and time varying channels. A deep network, applicable as auto-encoder, with complex weights is introduced and the information-theoretic optimization criteria are exploited to obtain optimal weights showing that the proposed method decreases effect of noise on model parameters.

KEYWORDS

Deep Learning, MIMO OFDM Channel Estimation, Information-theoretic Criteria.


Present and the Future of Artificial Intelligence in Software Testing


Nishtha Jaiswal1, Yatheshtha2, Rishi Prakash3, Paras Nath Barwal4, 1Department of E-Governance, CDAC, Noida, U.P., India, 2Department of CSE, Banasthali Vidhyapith, Jaipur, Rajasthan, India, 3Department of E-Governance, C-DAC, Noida, U.P., India, 4Department of E-Governance, C-DAC, Noida, U.P., India

ABSTRACT

Software Testing is an important phase in software development where the software is tested against certain parameters in order to assure its quality before reaching the end user. Testing itself involves a lot of stages. There are various testing software tools like Selenium, TestComplete, LambdaTest, etc. which have made this task easier by automating the test running and error reporting process. But, generating test cases, deciding which test case to run, repairing the tests, and acting on reported failure still remains repetitive and time-consuming tasks for testers and developers. Luckily, in a world that is rapidly getting automated through Artificial Intelligence (AI), testing has not remained untouched. The involvement of AI has automated many important activities in the testing cycle like modeling, test generation, and failure detection. This has brought decision-making power to the testing tools. With the help of regressive algorithms, visual testing and self-healing systems are now a reality. Their use is expected to grow in the near future, completely automating the testing process. The purpose of this paper is, therefore, to analyze the advantages AI has brought to software testing so far, the benefit of using an AI-enabled automated testing framework, and how this technology can evolve in the near future.

KEYWORDS

Software Testing, Selenium, Artificial Intelligence, Automated Testing, Self-Healing Systems.


SmartCarbs: An Intelligent Mobile Application to Assist Diet Control using Artificial Intelligence and Computer Vision


Olivia Lee1 and Yu Sun2, 1Portola High School, 1001 Cadence, Irvine, CA 92618, 2California State Polytechnic University, Pomona, CA, 91768, Irvine, CA 92

ABSTRACT

In today’s society, Type 2 diabetes is a prevalent disease that af ects hundreds of millions of people worldwide [11]. However, many people are unaware that they are diabetic or pre-diabetic, so they do not have access to the information to make better-informed nutritional choices that will suit their personal needs. In this paper, we designed an application that uses image classification to provide an estimate of the nutritional content of the selected food [12]. We applied our application to identify and list the nutritional content of multiple dif erent foods, then conducted a qualitative evaluation of the approach. The results show that this application will facilitate healthy eating and allow users to utilize the image classification predictions to make diabetes-friendly.

KEYWORDS

Machine Learning, Nutrition, Flutter, Image Classification.


An Object Detection and its Educational Effect on Aspiring Film Directors using Computer Vision and Machine Learning


Fengrong Han1 and Ang Li2, 1Veritas Collegiate Academy VA, 5225 Backlick Road Springfield,VA 22151, 2California State University, Long Beach, 1250 Bellflower Blvd, Long Beach, CA 90840

ABSTRACT

The issue that is being addressed is how those who wish to be involved in film production in the future can be more educated regarding what goes into making a movie. An application was created to help these people keep track of movie props and characters and where they are placed throughout a movie. This is performed by using videos or images as input for the application, and the application uses an object detection model from MediaPipe to determine what types of objects are in a scene and where exactly they are located in the scene [4]. Then, bounding boxes are placed around the objects with the classification of the object next to each bounding box, and the result is returned. The results of the experiment indicate that the application is consistent at detecting objects when the object is fully shown on the screen. However, when the application used the cropped images as input as opposed to the full version of the image, the application was not consistently accurate, as it was not able to recognize objects that were partially cut out of the picture. More work will be needed to improve both the object detection capabilities of partially obstructed objects and the overall user interface of the application [5].

KEYWORDS

Movies, Object Detection, MediaPipe, Application.


Fusion of Motion And Electromyographic Signals for Robust Hand Gesture Recognition Based on Heterogeneous Networks


Yongxiang Zou1, Long Cheng2, Lijun Han1 and Zhengwei Li2, 1School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China, 2State Key Laboratory of Management and Control for Complex Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China.

ABSTRACT

Hand gesture recognition has become a vital subject in the fields of human-computer interaction and rehabilitation assessment. This paper presented a multi-modal fusion system named MFHG, which used two heterogeneous networks to extract and fuse the deep features of the motion signals and the surface electromyography (sEMG) signals, respectively. To extract the features of the motion signals, a graph neural network named the cumulative graph attention (CGAT) model was first proposed to characterize the prior knowledge of motion coupling between finger joints. The CGAT model used the cumulative mechanism to fuse the early and late extracted features to improve motion-based hand gesture recognition. In addition, a convolutional neural network (CNN) model named TFHG was proposed to extract both the time-domain and frequency-domain information of the sEMG signals. To improve the performance of hand gesture recognition, the deep features from multiple modes were merged with the average layer, then the regularization items containing center loss and the mutual information loss were employed to enhance the robustness of this multi-modal system. Finally, a data set was built, which contained the multi-modal signals from seven subjects on different days, to verify the performance of the multi-modal system. The experimental results indicated that the MFHG can reach 99.96% and 92.46% accuracy on hand gesture recognition in the case of within-session and 92.46% cross-day, respectively.

KEYWORDS

Leap Motion, sEMG, Multi-modal, Graph Neural Network, Feature Regularization.