Search results for: problem-based learning approach
14360 Risk Assessment for International Investment: A Standardized Approach to Identify Risk, Risk Appetite, Risk Rating, Risk Treatment and Mitigation Plans
Authors: Pui Yong Leo, Normy Maziah Mohd Said
Abstract:
Change of global economy landscape and business environment has led to companies’ decision to go global and enter international markets. As the companies go beyond the comfort zone (i.e. investing in the home country), it is important to ensure a comprehensive risk assessment is carried out. This paper describes a standardized approach for international investment, ensuring identification of risk, risk appetite, risk rating, risk treatment and mitigation plans for respective international investment proposal. The standardized approach is divided into three (3) stages as follows: Stage 1 – Preliminary Risk profiling; with the objective to gauge exposure to countries and high level risk factors as first level assessment. Stage 2 – Risk Parameters; with the objective to define risk appetite for the international investment from the perspective of likelihood and impact. Stage 3 – Detailed Risk Assessments; with the objectives to assess in detail any triggered elements from Stage 1, and project specific risks. The final output will include the mitigation plans for the identified risks for the total investment. Example will be given in this paper to show how comprehensive risk assessment is carried out for an international investment in power energy sector.Keywords: international investment, mitigation plans, risk appetite, risk assessment
Procedia PDF Downloads 40414359 Improving the Students’ Writing Skill by Using Brainstorming Technique
Authors: M. Z. Abdul Rofiq Badril Rizal
Abstract:
This research is aimed to know the improvement of students’ English writing skill by using brainstorming technique. The technique used in writing is able to help the students’ difficulties in generating ideas and to lead the students to arrange the ideas well as well as to focus on the topic developed in writing. The research method used is classroom action research. The data sources of the research are an English teacher who acts as an observer and the students of class X.MIA5 consist of 35 students. The test result and observation are collected as the data in this research. Based on the research result in cycle one, the percentage of students who reach minimum accomplishment criteria (MAC) is 76.31%. It shows that the cycle must be continued to cycle two because the aim of the research has not accomplished, all of the students’ scores have not reached MAC yet. After continuing the research to cycle two and the weaknesses are improved, the process of teaching and learning runs better. At the test which is conducted in the end of learning process in cycle two, all of the students reach the minimum score and above 76 based on the minimum accomplishment criteria. It means the research has been successful and the percentage of students who reach minimum accomplishment criteria is 100%. Therefore, the writer concludes that brainstorming technique is able to improve the students’ English writing skill at the tenth grade of SMAN 2 Jember.Keywords: brainstorming technique, improving, writing skill, knowledge and innovation engineering
Procedia PDF Downloads 36914358 Autism Spectrum Disorder Classification Algorithm Using Multimodal Data Based on Graph Convolutional Network
Authors: Yuntao Liu, Lei Wang, Haoran Xia
Abstract:
Machine learning has shown extensive applications in the development of classification models for autism spectrum disorder (ASD) using neural image data. This paper proposes a fusion multi-modal classification network based on a graph neural network. First, the brain is segmented into 116 regions of interest using a medical segmentation template (AAL, Anatomical Automatic Labeling). The image features of sMRI and the signal features of fMRI are extracted, which build the node and edge embedding representations of the brain map. Then, we construct a dynamically updated brain map neural network and propose a method based on a dynamic brain map adjacency matrix update mechanism and learnable graph to further improve the accuracy of autism diagnosis and recognition results. Based on the Autism Brain Imaging Data Exchange I dataset(ABIDE I), we reached a prediction accuracy of 74% between ASD and TD subjects. Besides, to study the biomarkers that can help doctors analyze diseases and interpretability, we used the features by extracting the top five maximum and minimum ROI weights. This work provides a meaningful way for brain disorder identification.Keywords: autism spectrum disorder, brain map, supervised machine learning, graph network, multimodal data, model interpretability
Procedia PDF Downloads 7914357 Analyzing the Risk Based Approach in General Data Protection Regulation: Basic Challenges Connected with Adapting the Regulation
Authors: Natalia Kalinowska
Abstract:
The adoption of the General Data Protection Regulation, (GDPR) finished the four-year work of the European Commission in this area in the European Union. Considering far-reaching changes, which will be applied by GDPR, the European legislator envisaged two-year transitional period. Member states and companies have to prepare for a new regulation until 25 of May 2018. The idea, which becomes a new look at an attitude to data protection in the European Union is risk-based approach. So far, as a result of implementation of Directive 95/46/WE, in many European countries (including Poland) there have been adopted very particular regulations, specifying technical and organisational security measures e.g. Polish implementing rules indicate even how long password should be. According to the new approach from May 2018, controllers and processors will be obliged to apply security measures adequate to level of risk associated with specific data processing. The risk in GDPR should be interpreted as the likelihood of a breach of the rights and freedoms of the data subject. According to Recital 76, the likelihood and severity of the risk to the rights and freedoms of the data subject should be determined by reference to the nature, scope, context and purposes of the processing. GDPR does not indicate security measures which should be applied – in recitals there are only examples such as anonymization or encryption. It depends on a controller’s decision what type of security measures controller considered as sufficient and he will be responsible if these measures are not sufficient or if his identification of risk level is incorrect. Data protection regulation indicates few levels of risk. Recital 76 indicates risk and high risk, but some lawyers think, that there is one more category – low risk/now risk. Low risk/now risk data processing is a situation when it is unlikely to result in a risk to the rights and freedoms of natural persons. GDPR mentions types of data processing when a controller does not have to evaluate level of risk because it has been classified as „high risk” processing e.g. processing on a large scale of special categories of data, processing with using new technologies. The methodology will include analysis of legal regulations e.g. GDPR, the Polish Act on the Protection of personal data. Moreover: ICO Guidelines and articles concerning risk based approach in GDPR. The main conclusion is that an appropriate risk assessment is a key to keeping data safe and avoiding financial penalties. On the one hand, this approach seems to be more equitable, not only for controllers or processors but also for data subjects, but on the other hand, it increases controllers’ uncertainties in the assessment which could have a direct impact on incorrect data protection and potential responsibility for infringement of regulation.Keywords: general data protection regulation, personal data protection, privacy protection, risk based approach
Procedia PDF Downloads 25414356 GIS-Based Automatic Flight Planning of Camera-Equipped UAVs for Fire Emergency Response
Authors: Mohammed Sulaiman, Hexu Liu, Mohamed Binalhaj, William W. Liou, Osama Abudayyeh
Abstract:
Emerging technologies such as camera-equipped unmanned aerial vehicles (UAVs) are increasingly being applied in building fire rescue to provide real-time visualization and 3D reconstruction of the entire fireground. However, flight planning of camera-equipped UAVs is usually a manual process, which is not sufficient to fulfill the needs of emergency management. This research proposes a Geographic Information System (GIS)-based approach to automatic flight planning of camera-equipped UAVs for building fire emergency response. In this research, Haversine formula and lawn mowing patterns are employed to automate flight planning based on geometrical and spatial information from GIS. The resulting flight mission satisfies the requirements of 3D reconstruction purposes of the fireground, in consideration of flight execution safety and visibility of camera frames. The proposed approach is implemented within a GIS environment through an application programming interface. A case study is used to demonstrate the effectiveness of the proposed approach. The result shows that flight mission can be generated in a timely manner for application to fire emergency response.Keywords: GIS, camera-equipped UAVs, automatic flight planning, fire emergency response
Procedia PDF Downloads 13014355 The Profitability Management Mechanism of Leather Industry-Based on the Activity-Based Benefit Approach
Authors: Mei-Fang Wu, Shu-Li Wang, Tsung-Yueh Lu, Feng-Tsung Cheng
Abstract:
Strengthening core competitiveness is the main goal of enterprises in a fierce competitive environment. Accurate cost information is a great help for managers in dealing with operation strategies. This paper establishes a profitability management mechanism that applies the Activity-Based Benefit approach (ABBA) to solve the profitability for each customer from the market. ABBA provides financial and non-financial information for the operation, but also indicates what resources have expired in the operational process. The customer profit management model shows the level of profitability of each customer for the company. The empirical data were gathered from a case company operating in the leather industry in Taiwan. The research findings indicate that 30% of customers create little profit for the company as a result of asking for over 5% of sales discounts. Those customers ask for sales discount because of color differences of leather products. This paper provides a customer’s profitability evaluation mechanism to help enterprises to greatly improve operating effectiveness and promote operational activity efficiency and overall operation profitability.Keywords: activity-based benefit approach, customer profit analysis, leather industry, profitability management mechanism
Procedia PDF Downloads 31014354 Power-Aware Adaptive Coverage Control with Consensus Protocol
Authors: Mert Turanli, Hakan Temeltas
Abstract:
In this paper, we propose a new approach to coverage control problem by using adaptive coordination and power aware control laws. Nonholonomic mobile nodes position themselves suboptimally according to a time-varying density function using Centroidal Voronoi Tesellations. The Lyapunov stability analysis of the adaptive and decentralized approach is given. A linear consensus protocol is used to establish synchronization among the mobile nodes. Also, repulsive forces prevent nodes from collision. Simulation results show that by using power aware control laws, energy consumption of the nodes can be reduced.Keywords: power aware, coverage control, adaptive, consensus, nonholonomic, coordination
Procedia PDF Downloads 35614353 Learning Chinese Suprasegmentals for a Better Communicative Performance
Authors: Qi Wang
Abstract:
Chinese has become a powerful worldwide language and millions of learners are studying it all over the words. Chinese is a tone language with unique meaningful characters, which makes foreign learners master it with more difficulties. On the other hand, as each foreign language, the learners of Chinese first will learn the basic Chinese Sound Structure (the initials and finals, tones, Neutral Tone and Tone Sandhi). It’s quite common that in the following studies, teachers made a lot of efforts on drilling and error correcting, in order to help students to pronounce correctly, but ignored the training of suprasegmental features (e.g. stress, intonation). This paper analysed the oral data based on our graduation students (two-year program) from 2006-2013, presents the intonation pattern of our graduates to speak Chinese as second language -high and plain with heavy accents, without lexical stress, appropriate stop endings and intonation, which led to the misunderstanding in different real contexts of communications and the international official Chinese test, e.g. HSK (Chinese Proficiency Test), HSKK (HSK Speaking Test). This paper also demonstrated how the Chinese to use the suprasegmental features strategically in different functions and moods (declarative, interrogative, imperative, exclamatory and rhetorical intonations) in order to train the learners to achieve better Communicative Performance.Keywords: second language learning, suprasegmental, communication, HSK (Chinese Proficiency Test)
Procedia PDF Downloads 43914352 Service Information Integration Platform as Decision Making Tools for the Service Industry Supply Chain-Indonesia Service Integration Project
Authors: Haikal Achmad Thaha, Pujo Laksono, Dhamma Nibbana Putra
Abstract:
Customer service is one of the core interest in a service sector of a company, whether as the core business or as service part of the operation. Most of the time, the people and the previous research in service industry is focused on finding the best business model solution for the service sector, usually to decide between total in house customer service, outsourcing, or something in between. Conventionally, to take this decision is some important part of the management job, and this is a process that usually takes some time and staff effort, meanwhile market condition and overall company needs may change and cause loss of income and temporary disturbance in the companies operation . However, in this paper we have offer a new concept model to assist decision making process in service industry. This model will featured information platform as central tool to integrate service industry operation. The result is service information model which would ideally increase response time and effectivity of the decision making. it will also help service industry in switching the service solution system quickly through machine learning when the companies growth and the service solution needed are changing.Keywords: service industry, customer service, machine learning, decision making, information platform
Procedia PDF Downloads 62514351 Automatic Detection Of Diabetic Retinopathy
Authors: Zaoui Ismahene, Bahri Sidi Mohamed, Abbassa Nadira
Abstract:
Diabetic Retinopathy (DR) is a leading cause of vision impairment and blindness among individuals with diabetes. Early diagnosis is crucial for effective treatment, yet current diagnostic methods rely heavily on manual analysis of retinal images, which can be time-consuming and prone to subjectivity. This research proposes an automated system for the detection of DR using Jacobi wavelet-based feature extraction combined with Support Vector Machines (SVM) for classification. The integration of wavelet analysis with machine learning techniques aims to improve the accuracy, efficiency, and reliability of DR diagnosis. In this study, retinal images are preprocessed through normalization, resizing, and noise reduction to enhance the quality of the images. The Jacobi wavelet transform is then applied to extract both global and local features, effectively capturing subtle variations in retinal images that are indicative of DR. These extracted features are fed into an SVM classifier, known for its robustness in handling high-dimensional data and its ability to achieve high classification accuracy. The SVM classifier is optimized using parameter tuning to improve performance. The proposed methodology is evaluated using a comprehensive dataset of retinal images, encompassing a range of DR severity levels. The results show that the proposed system outperforms traditional wavelet-based methods, demonstrating significantly higher accuracy, sensitivity, and specificity in detecting DR. By leveraging the discriminative power of Jacobi wavelet features and the robustness of SVM, the system provides a promising solution for the automatic detection of DR, which could assist ophthalmologists in early diagnosis and intervention, ultimately improving patient outcomes. This research highlights the potential of combining wavelet-based image processing with machine learning for advancing automated medical diagnostics.Keywords: iabetic retinopathy (DR), Jacobi wavelets, machine learning, feature extraction, classification
Procedia PDF Downloads 1214350 Prediction of Survival Rate after Gastrointestinal Surgery Based on The New Japanese Association for Acute Medicine (JAAM Score) With Neural Network Classification Method
Authors: Ayu Nabila Kusuma Pradana, Aprinaldi Jasa Mantau, Tomohiko Akahoshi
Abstract:
The incidence of Disseminated intravascular coagulation (DIC) following gastrointestinal surgery has a poor prognosis. Therefore, it is important to determine the factors that can predict the prognosis of DIC. This study will investigate the factors that may influence the outcome of DIC in patients after gastrointestinal surgery. Eighty-one patients were admitted to the intensive care unit after gastrointestinal surgery in Kyushu University Hospital from 2003 to 2021. Acute DIC scores were estimated using the new Japanese Association for Acute Medicine (JAAM) score from before and after surgery from day 1, day 3, and day 7. Acute DIC scores will be compared with The Sequential Organ Failure Assessment (SOFA) score, platelet count, lactate level, and a variety of biochemical parameters. This study applied machine learning algorithms to predict the prognosis of DIC after gastrointestinal surgery. The results of this study are expected to be used as an indicator for evaluating patient prognosis so that it can increase life expectancy and reduce mortality from cases of DIC patients after gastrointestinal surgery.Keywords: the survival rate, gastrointestinal surgery, JAAM score, neural network, machine learning, disseminated intravascular coagulation (DIC)
Procedia PDF Downloads 26514349 Iranian Students’ and Teachers’ Perceptions of Effective Foreign Language Teaching
Authors: Mehrnoush Tajnia, Simin Sadeghi-Saeb
Abstract:
Students and teachers have different perceptions of effectiveness of instruction. Comparing students’ and teachers’ beliefs and finding the mismatches between them can increase L2 students’ satisfaction. Few studies have taken into account the beliefs of both students and teachers on different aspects of pedagogy and the effect of learners’ level of education and contexts on effective foreign language teacher practices. Therefore, the present study was conducted to compare students’ and teachers’ perceptions on effective foreign language teaching. A sample of 303 learners and 54 instructors from different private language institutes and universities participated in the study. A questionnaire was developed to elicit participants’ beliefs on effective foreign language teaching and learning. The analysis of the results revealed that: a) there is significant difference between the students’ beliefs about effective teacher practices and teachers’ belief, b) Class level influences students’ perception of effective foreign language teacher, d) There is a significant difference of opinion between those learners who study foreign languages at university and those who study foreign language in private institutes with respect to effective teacher practices. The present paper concludes that finding the gap between students’ and teachers’ beliefs would help both of the groups to enhance their learning and teaching.Keywords: effective teacher, effective teaching, students’ beliefs, teachers’ beliefs
Procedia PDF Downloads 32114348 Eco-Parcel As a Semi-Qualitative Approach to Support Environmental Impacts Assessments in Nature-Based Tourism Destinations
Authors: Halima Kilungu, Pantaleo, K. T. Munishi
Abstract:
Climate and land-cover change affect nature-based tourism (NBT) due to its attractions' close connection to natural environments and climate. Thus, knowledge of how each attraction reacts to the changing environments and devising simple yet science based approaches to respond to these changes from a tourism perspective in space and time is timely. Nevertheless, no specific approaches exist to address the knowledge gap. The eco-parcel approach is devised to address the gap and operationalized in Serengeti and Kilimanjaro National Parks: the most climate-sensitive NBT destinations in Africa. The approach is partly descriptive and has three simple steps: (1) to identify and define tourist attractions (i.e. biotic and abiotic attractions). This creates an important database of the most poorly kept information on attractions' types in NBT destinations. (2) To create a spatial and temporal link of each attraction and describe its characteristic environments (e.g. vegetation, soil, water and rock outcrops). This is the most limited attractions' information yet important as a proxy of changes in attractions. (3) To assess the importance of individual attractions for tourism based on tourists' preferences. This information enables an accurate assessment of the value of individual attractions for tourism. The importance of the eco-parcel approach is that it describes how each attraction emerges from and is connected to specific environments, which define its attractiveness in space and time. This information allows accurate assessment of the likely losses or gains of individual attractions when climate or environment changes in specific destinations and equips tourism stakeholders with informed responses.Keywords: climate change, environmental change, nature-based tourism, Serengeti National Park, Kilimanjaro National Park
Procedia PDF Downloads 12514347 Data Refinement Enhances The Accuracy of Short-Term Traffic Latency Prediction
Authors: Man Fung Ho, Lap So, Jiaqi Zhang, Yuheng Zhao, Huiyang Lu, Tat Shing Choi, K. Y. Michael Wong
Abstract:
Nowadays, a tremendous amount of data is available in the transportation system, enabling the development of various machine learning approaches to make short-term latency predictions. A natural question is then the choice of relevant information to enable accurate predictions. Using traffic data collected from the Taiwan Freeway System, we consider the prediction of short-term latency of a freeway segment with a length of 17 km covering 5 measurement points, each collecting vehicle-by-vehicle data through the electronic toll collection system. The processed data include the past latencies of the freeway segment with different time lags, the traffic conditions of the individual segments (the accumulations, the traffic fluxes, the entrance and exit rates), the total accumulations, and the weekday latency profiles obtained by Gaussian process regression of past data. We arrive at several important conclusions about how data should be refined to obtain accurate predictions, which have implications for future system-wide latency predictions. (1) We find that the prediction of median latency is much more accurate and meaningful than the prediction of average latency, as the latter is plagued by outliers. This is verified by machine-learning prediction using XGBoost that yields a 35% improvement in the mean square error of the 5-minute averaged latencies. (2) We find that the median latency of the segment 15 minutes ago is a very good baseline for performance comparison, and we have evidence that further improvement is achieved by machine learning approaches such as XGBoost and Long Short-Term Memory (LSTM). (3) By analyzing the feature importance score in XGBoost and calculating the mutual information between the inputs and the latencies to be predicted, we identify a sequence of inputs ranked in importance. It confirms that the past latencies are most informative of the predicted latencies, followed by the total accumulation, whereas inputs such as the entrance and exit rates are uninformative. It also confirms that the inputs are much less informative of the average latencies than the median latencies. (4) For predicting the latencies of segments composed of two or three sub-segments, summing up the predicted latencies of each sub-segment is more accurate than the one-step prediction of the whole segment, especially with the latency prediction of the downstream sub-segments trained to anticipate latencies several minutes ahead. The duration of the anticipation time is an increasing function of the traveling time of the upstream segment. The above findings have important implications to predicting the full set of latencies among the various locations in the freeway system.Keywords: data refinement, machine learning, mutual information, short-term latency prediction
Procedia PDF Downloads 17214346 Engaging the World Bank: Good Governance and Human Rights-Based Approaches
Authors: Lottie Lane
Abstract:
It is habitually assumed and stated that the World Bank should engage and comply with international human rights standards. However, the basis for holding the Bank to such standards is unclear. Most advocates of the idea invoke aspects of international law to argue that the Bank has existing obligations to act in compliance with human rights standards. The Bank itself, however, does not appear to accept such arguments, despite having endorsed the importance of human rights for a considerable length of time. A substantial challenge is that under the current international human rights law framework, the World Bank is considered a non-state actor, and as such, has no direct human rights obligations. In the absence of clear legal duties for the Bank, it is necessary to look at the tools available beyond the international human rights framework to encourage the Bank to comply with human rights standards. This article critically examines several bases for arguing that the Bank should comply and engage with human rights through its policies and practices. Drawing on the Bank’s own ‘good governance’ approach as well as the United Nations’ ‘human rights-based-approach’ to development, a new basis is suggested. First, the relationship between the World Bank and human rights is examined. Three perspectives are considered: (1) the legal position – what the status of the World Bank is under international human rights law, and whether it can be said to have existing legal human rights obligations; (2) the Bank’s own official position – how the Bank envisages its relationship with and role in the protection of human rights; and (3) the relationship between the Bank’s policies and practices and human rights (including how its attitudes are reflected in its policies and how the Bank’s operations impact human rights enjoyment in practice). Here, the article focuses on two examples – the (revised) 2016 Environmental and Social Safeguard Policies and the 2012 case-study regarding Gambella, Ethiopia. Both examples are widely considered missed opportunities for the Bank to actively engage with human rights. The analysis shows that however much pressure is placed on the Bank to improve its human rights footprint, it is extremely reluctant to do so explicitly, and the legal bases available are insufficient for requiring concrete, ex ante action by the Bank. Instead, the Bank’s own ‘good governance’ approach to development – which it has been advocating since the 1990s – can be relied upon. ‘Good governance’ has been used and applied by many actors in many contexts, receiving numerous different definitions. This article argues that human rights protection can now be considered a crucial component of good governance, at least in the context of development. In doing so, the article explains the relationship and interdependence between the two concepts, and provides three rationales for the Bank to take a ‘human rights-based approach’ to good governance. Ultimately, this article seeks to look beyond international human rights law and take a governance approach to provide a convincing basis upon which to argue that the World Bank should comply with human rights standards.Keywords: World Bank, international human rights law, good governance, human rights-based approach
Procedia PDF Downloads 36514345 Automatic Adult Age Estimation Using Deep Learning of the ResNeXt Model Based on CT Reconstruction Images of the Costal Cartilage
Authors: Ting Lu, Ya-Ru Diao, Fei Fan, Ye Xue, Lei Shi, Xian-e Tang, Meng-jun Zhan, Zhen-hua Deng
Abstract:
Accurate adult age estimation (AAE) is a significant and challenging task in forensic and archeology fields. Attempts have been made to explore optimal adult age metrics, and the rib is considered a potential age marker. The traditional way is to extract age-related features designed by experts from macroscopic or radiological images followed by classification or regression analysis. Those results still have not met the high-level requirements for practice, and the limitation of using feature design and manual extraction methods is loss of information since the features are likely not designed explicitly for extracting information relevant to age. Deep learning (DL) has recently garnered much interest in imaging learning and computer vision. It enables learning features that are important without a prior bias or hypothesis and could be supportive of AAE. This study aimed to develop DL models for AAE based on CT images and compare their performance to the manual visual scoring method. Chest CT data were reconstructed using volume rendering (VR). Retrospective data of 2500 patients aged 20.00-69.99 years were obtained between December 2019 and September 2021. Five-fold cross-validation was performed, and datasets were randomly split into training and validation sets in a 4:1 ratio for each fold. Before feeding the inputs into networks, all images were augmented with random rotation and vertical flip, normalized, and resized to 224×224 pixels. ResNeXt was chosen as the DL baseline due to its advantages of higher efficiency and accuracy in image classification. Mean absolute error (MAE) was the primary parameter. Independent data from 100 patients acquired between March and April 2022 were used as a test set. The manual method completely followed the prior study, which reported the lowest MAEs (5.31 in males and 6.72 in females) among similar studies. CT data and VR images were used. The radiation density of the first costal cartilage was recorded using CT data on the workstation. The osseous and calcified projections of the 1 to 7 costal cartilages were scored based on VR images using an eight-stage staging technique. According to the results of the prior study, the optimal models were the decision tree regression model in males and the stepwise multiple linear regression equation in females. Predicted ages of the test set were calculated separately using different models by sex. A total of 2600 patients (training and validation sets, mean age=45.19 years±14.20 [SD]; test set, mean age=46.57±9.66) were evaluated in this study. Of ResNeXt model training, MAEs were obtained with 3.95 in males and 3.65 in females. Based on the test set, DL achieved MAEs of 4.05 in males and 4.54 in females, which were far better than the MAEs of 8.90 and 6.42 respectively, for the manual method. Those results showed that the DL of the ResNeXt model outperformed the manual method in AAE based on CT reconstruction of the costal cartilage and the developed system may be a supportive tool for AAE.Keywords: forensic anthropology, age determination by the skeleton, costal cartilage, CT, deep learning
Procedia PDF Downloads 7714344 Analysis and Detection of Facial Expressions in Autism Spectrum Disorder People Using Machine Learning
Authors: Muhammad Maisam Abbas, Salman Tariq, Usama Riaz, Muhammad Tanveer, Humaira Abdul Ghafoor
Abstract:
Autism Spectrum Disorder (ASD) refers to a developmental disorder that impairs an individual's communication and interaction ability. Individuals feel difficult to read facial expressions while communicating or interacting. Facial Expression Recognition (FER) is a unique method of classifying basic human expressions, i.e., happiness, fear, surprise, sadness, disgust, neutral, and anger through static and dynamic sources. This paper conducts a comprehensive comparison and proposed optimal method for a continued research project—a system that can assist people who have Autism Spectrum Disorder (ASD) in recognizing facial expressions. Comparison has been conducted on three supervised learning algorithms EigenFace, FisherFace, and LBPH. The JAFFE, CK+, and TFEID (I&II) datasets have been used to train and test the algorithms. The results were then evaluated based on variance, standard deviation, and accuracy. The experiments showed that FisherFace has the highest accuracy for all datasets and is considered the best algorithm to be implemented in our system.Keywords: autism spectrum disorder, ASD, EigenFace, facial expression recognition, FisherFace, local binary pattern histogram, LBPH
Procedia PDF Downloads 18114343 Using Deep Learning in Lyme Disease Diagnosis
Authors: Teja Koduru
Abstract:
Untreated Lyme disease can lead to neurological, cardiac, and dermatological complications. Rapid diagnosis of the erythema migrans (EM) rash, a characteristic symptom of Lyme disease is therefore crucial to early diagnosis and treatment. In this study, we aim to utilize deep learning frameworks including Tensorflow and Keras to create deep convolutional neural networks (DCNN) to detect images of acute Lyme Disease from images of erythema migrans. This study uses a custom database of erythema migrans images of varying quality to train a DCNN capable of classifying images of EM rashes vs. non-EM rashes. Images from publicly available sources were mined to create an initial database. Machine-based removal of duplicate images was then performed, followed by a thorough examination of all images by a clinician. The resulting database was combined with images of confounding rashes and regular skin, resulting in a total of 683 images. This database was then used to create a DCNN with an accuracy of 93% when classifying images of rashes as EM vs. non EM. Finally, this model was converted into a web and mobile application to allow for rapid diagnosis of EM rashes by both patients and clinicians. This tool could be used for patient prescreening prior to treatment and lead to a lower mortality rate from Lyme disease.Keywords: Lyme, untreated Lyme, erythema migrans rash, EM rash
Procedia PDF Downloads 24614342 Automated Testing of Workshop Robot Behavior
Authors: Arne Hitzmann, Philipp Wentscher, Alexander Gabel, Reinhard Gerndt
Abstract:
Autonomous mobile robots can be found in a wide field of applications. Their types range from household robots over workshop robots to autonomous cars and many more. All of them undergo a number of testing steps during development, production and maintenance. This paper describes an approach to improve testing of robot behavior. It was inspired by the RoboCup @work competition that itself reflects a robotics benchmark for industrial robotics. There, scaled down versions of mobile industrial robots have to navigate through a workshop-like environment or operation area and have to perform tasks of manipulating and transporting work pieces. This paper will introduce an approach of automated vision-based testing of the behavior of the so called youBot robot, which is the most widely used robot platform in the RoboCup @work competition. The proposed system allows automated testing of multiple tries of the robot to perform a specific missions and it allows for the flexibility of the robot, e.g. selecting different paths between two tasks within a mission. The approach is based on a multi-camera setup using, off the shelf cameras and optical markers. It has been applied for test-driven development (TDD) and maintenance-like verification of the robot behavior and performance.Keywords: supervisory control, testing, markers, mono vision, automation
Procedia PDF Downloads 37914341 Semantic Data Schema Recognition
Authors: Aïcha Ben Salem, Faouzi Boufares, Sebastiao Correia
Abstract:
The subject covered in this paper aims at assisting the user in its quality approach. The goal is to better extract, mix, interpret and reuse data. It deals with the semantic schema recognition of a data source. This enables the extraction of data semantics from all the available information, inculding the data and the metadata. Firstly, it consists of categorizing the data by assigning it to a category and possibly a sub-category, and secondly, of establishing relations between columns and possibly discovering the semantics of the manipulated data source. These links detected between columns offer a better understanding of the source and the alternatives for correcting data. This approach allows automatic detection of a large number of syntactic and semantic anomalies.Keywords: schema recognition, semantic data profiling, meta-categorisation, semantic dependencies inter columns
Procedia PDF Downloads 42114340 Cakrawala Baca Transformation Model into Social Enterprise: A Benchmark Approach from Socentra Agro Mandiri (SAM) and Agritektur
Authors: Syafinatul Fitri
Abstract:
Cakrawala Baca is one of social organization in Indonesia that realize to transform its organization into social enterprise to create more sustainable organization that result more sustainable social impact. Cakrawala Baca implements voluntary system for its organization and it has passive social target. It funds its program by several fund rising activities that depend on donors or sponsor. Therefore social activity that held does not create sustainable social impact. It is different with social enterprise that usually more independent in funding its activity through social business and implement active social target and professional work for organization member. Therefore social enterprise can sustain its organization and then able to create sustainable social impact. Developing transformation model from social movement into social enterprise is the focus of this study. To achieve the aim of study, benchmark approach from successful social enterprise in Indonesia that has previously formed as social movement is employed. The benchmark is conducted through internal and external scanning that result the understanding of how they transformed into social enterprise. After understanding SAM and Agritektur transformation, transformation pattern is formulated based on their transformation similarities. This transformation pattern will be implemented to formulate the transformation plan for Cakrawala Baca to be a social enterprise.Keywords: social movement/social organization, non-profit organization (NPO), social enterprise, transformation, Benchmarks approach
Procedia PDF Downloads 51514339 The Role of Organizational Identity in Disaster Response, Recovery and Prevention: A Case Study of an Italian Multi-Utility Company
Authors: Shanshan Zhou, Massimo Battaglia
Abstract:
Identity plays a critical role when an organization faces disasters. Individuals reflect on their working identities and identify themselves with the group and the organization, which facilitate collective sensemaking under crisis situations and enable coordinated actions to respond to and recover from disasters. In addition, an organization’s identity links it to its regional community, which fosters the mobilization of resources and contributes to rapid recovery. However, identity is also problematic for disaster prevention because of its persistence. An organization’s ego-defenses system prohibits the rethink of its identity and a rigid identity obstructs disaster prevention. This research aims to tackle the ‘problem’ of identity by study in-depth a case of an Italian multi–utility which experienced the 2012 Northern Italy earthquakes. Collecting data from 11 interviews with top managers and key players in the local community and archived materials, we find that the earthquakes triggered the rethink of the organization’s identity, which got reinforced afterward. This research highlighted the importance of identity in disaster response and recovery. More importantly, it explored the solution of overcoming the barrier of ego-defense that is to transform the organization into a learning organization which constantly rethinks its identity.Keywords: community identity, disaster, identity, organizational learning
Procedia PDF Downloads 73714338 Supervisor Controller-Based Colored Petri Nets for Deadlock Control and Machine Failures in Automated Manufacturing Systems
Authors: Husam Kaid, Abdulrahman Al-Ahmari, Zhiwu Li
Abstract:
This paper develops a robust deadlock control technique for shared and unreliable resources in automated manufacturing systems (AMSs) based on structural analysis and colored Petri nets, which consists of three steps. The first step involves using strict minimal siphon control to create a live (deadlock-free) system that does not consider resource failure. The second step uses an approach based on colored Petri net, in which all monitors designed in the first step are merged into a single monitor. The third step addresses the deadlock control problems caused by resource failures. For all resource failures in the Petri net model a common recovery subnet based on colored petri net is proposed. The common recovery subnet is added to the obtained system at the second step to make the system reliable. The proposed approach is evaluated using an AMS from the literature. The results show that the proposed approach can be applied to an unreliable complex Petri net model, has a simpler structure and less computational complexity, and can obtain one common recovery subnet to model all resource failures.Keywords: automated manufacturing system, colored Petri net, deadlocks, siphon
Procedia PDF Downloads 13214337 Single Imputation for Audiograms
Authors: Sarah Beaver, Renee Bryce
Abstract:
Audiograms detect hearing impairment, but missing values pose problems. This work explores imputations in an attempt to improve accuracy. This work implements Linear Regression, Lasso, Linear Support Vector Regression, Bayesian Ridge, K Nearest Neighbors (KNN), and Random Forest machine learning techniques to impute audiogram frequencies ranging from 125Hz to 8000Hz. The data contains patients who had or were candidates for cochlear implants. Accuracy is compared across two different Nested Cross-Validation k values. Over 4000 audiograms were used from 800 unique patients. Additionally, training on data combines and compares left and right ear audiograms versus single ear side audiograms. The accuracy achieved using Root Mean Square Error (RMSE) values for the best models for Random Forest ranges from 4.74 to 6.37. The R\textsuperscript{2} values for the best models for Random Forest ranges from .91 to .96. The accuracy achieved using RMSE values for the best models for KNN ranges from 5.00 to 7.72. The R\textsuperscript{2} values for the best models for KNN ranges from .89 to .95. The best imputation models received R\textsuperscript{2} between .89 to .96 and RMSE values less than 8dB. We also show that the accuracy of classification predictive models performed better with our best imputation models versus constant imputations by a two percent increase.Keywords: machine learning, audiograms, data imputations, single imputations
Procedia PDF Downloads 8614336 Virtual Co-Creation Model in Hijab Fashion Industry: Business Model Approach
Authors: Lisandy A. Suryana, Lidia Mayangsari, Santi Novani
Abstract:
Creative industry in Indonesia become an important aspect of the economy. One of the sectors of creative industry which give the highest contribution toward Indonesia’s GDP is fashion sector. In line with the target of Indonesia in 2020 to be the qibla’ of moeslem fashion of the world, all of the stakeholders of the business ecosystem should collaborate. Rather than focus on the internal aspects of producer, external aspects such as customers, government, community, etc. become important to be involved in the ecosystem to support the development and sustainability of those fashion sector. Unfortunately, although Indonesia has the biggest moeslem population, the number of hijab business penetration only 10%. Therefore, this research aims to analyze and develop the virtual co-creation platform for hijab creative industry as the strategy to achieve sustainability and increase the market share. This preliminary research describes the main stakeholders in the hijab creative industry based on business model approach. This business model is adapted by considering the service science context, and the data is collected by using the qualitative approach especially in-depth interview. This business model shows the relationship between resource integration, value co-creation, the value proposition of the company, and also the financial aspect of the business.Keywords: value co-creation, Hijab Fashion Industry, creative industry, service business model, business model canvas
Procedia PDF Downloads 38414335 Exploring the Applications of Neural Networks in the Adaptive Learning Environment
Authors: Baladitya Swaika, Rahul Khatry
Abstract:
Computer Adaptive Tests (CATs) is one of the most efficient ways for testing the cognitive abilities of students. CATs are based on Item Response Theory (IRT) which is based on item selection and ability estimation using statistical methods of maximum information selection/selection from posterior and maximum-likelihood (ML)/maximum a posteriori (MAP) estimators respectively. This study aims at combining both classical and Bayesian approaches to IRT to create a dataset which is then fed to a neural network which automates the process of ability estimation and then comparing it to traditional CAT models designed using IRT. This study uses python as the base coding language, pymc for statistical modelling of the IRT and scikit-learn for neural network implementations. On creation of the model and on comparison, it is found that the Neural Network based model performs 7-10% worse than the IRT model for score estimations. Although performing poorly, compared to the IRT model, the neural network model can be beneficially used in back-ends for reducing time complexity as the IRT model would have to re-calculate the ability every-time it gets a request whereas the prediction from a neural network could be done in a single step for an existing trained Regressor. This study also proposes a new kind of framework whereby the neural network model could be used to incorporate feature sets, other than the normal IRT feature set and use a neural network’s capacity of learning unknown functions to give rise to better CAT models. Categorical features like test type, etc. could be learnt and incorporated in IRT functions with the help of techniques like logistic regression and can be used to learn functions and expressed as models which may not be trivial to be expressed via equations. This kind of a framework, when implemented would be highly advantageous in psychometrics and cognitive assessments. This study gives a brief overview as to how neural networks can be used in adaptive testing, not only by reducing time-complexity but also by being able to incorporate newer and better datasets which would eventually lead to higher quality testing.Keywords: computer adaptive tests, item response theory, machine learning, neural networks
Procedia PDF Downloads 18014334 Teacher Collaboration Impact on Bilingual Students’ Oral Communication Skills in Inclusive Contexts
Authors: Diana González, Marta Gràcia, Ana Luisa Adam-Alcocer
Abstract:
Incorporating digital tools into educational practices represents a valuable approach for enriching the quality of teachers' educational practices in oral competence and fostering improvements in student learning outcomes. This study aims to promote a collaborative and culturally sensitive approach to professional development between teachers and a speech therapist to enhance their self-awareness and reflection on high-quality educational practices that integrate school components to strengthen children’s oral communication and pragmatic skills. The study involved five bilingual teachers fluent in both English and Spanish, with three specializing in special education and two in general education. It focused on Spanish-English bilingual students, aged 3-6, who were experiencing speech delays or disorders in a New York City public school, with the collaboration of a speech therapist. Using EVALOE-DSS (Assessment Scale of Oral Language Teaching in the School Context - Decision Support System), teachers conducted self-assessments of their teaching practices, reflect and make-decisions throughout six classes from March to June, focusing on students' communicative competence across various activities. Concurrently, the speech therapist observed and evaluated six classes per teacher using EVALOE-DSS during the same period. Additionally, professional development meetings were held monthly between the speech therapist and teachers, centering on discussing classroom interactions, instructional strategies, and the progress of both teachers and students in their classes. Findings highlight the digital tool EVALOE-DSS's value in analyzing communication patterns and trends among bilingual children in inclusive settings. It helps in identifying improvement areas through teacher and speech therapist collaboration. After self-reflection meetings, teachers demonstrated increased awareness of student needs in oral language and pragmatic skills. They also exhibited enhanced utilization of strategies outlined in EVALOE-DSS, such as actively guiding and orienting students during oral language activities, promoting student-initiated communicative interactions, teaching students how to seek and provide information, and managing turn-taking to ensure inclusive participation. Teachers participating in the professional development program have shown positive progress in assessing their classes across all dimensions of the training tool, including instructional design, teacher conversation management, pupil conversation management, communicative functions, teacher strategies, and pupil communication functions. This includes aspects related to both teacher actions and child actions, particularly in child language development. This progress underscores the effectiveness of individual reflection (conducted weekly or biweekly using EVALOE-DSS) as well as collaborative reflection among teachers and the speech therapist during meetings. The EVALOE-SSD has proven effective in supporting teachers' self-reflection, decision-making, and classroom changes, leading to improved development of students' oral language and pragmatic skills. It has facilitated culturally sensitive evaluations of communication among bilingual children, cultivating collaboration between teachers and speech therapist to identify areas of growth. Participants in the professional development program demonstrated substantial progress across all dimensions assessed by EVALOE-DSS. This included improved management of pupil communication functions, implementation of effective teaching strategies, and better classroom dynamics. Regular reflection sessions using EVALOE-SSD supported continuous improvement in instructional practices, highlighting its role in fostering reflective teaching and enriching student learning experiences. Overall, EVALOE-DSS has proven invaluable for enhancing teaching effectiveness and promoting meaningful student interactions in diverse educational settings.Keywords: bilingual students, collaboration, culturally sensitive, oral communication skills, self-reflection
Procedia PDF Downloads 4114333 Model-Driven and Data-Driven Approaches for Crop Yield Prediction: Analysis and Comparison
Authors: Xiangtuo Chen, Paul-Henry Cournéde
Abstract:
Crop yield prediction is a paramount issue in agriculture. The main idea of this paper is to find out efficient way to predict the yield of corn based meteorological records. The prediction models used in this paper can be classified into model-driven approaches and data-driven approaches, according to the different modeling methodologies. The model-driven approaches are based on crop mechanistic modeling. They describe crop growth in interaction with their environment as dynamical systems. But the calibration process of the dynamic system comes up with much difficulty, because it turns out to be a multidimensional non-convex optimization problem. An original contribution of this paper is to propose a statistical methodology, Multi-Scenarios Parameters Estimation (MSPE), for the parametrization of potentially complex mechanistic models from a new type of datasets (climatic data, final yield in many situations). It is tested with CORNFLO, a crop model for maize growth. On the other hand, the data-driven approach for yield prediction is free of the complex biophysical process. But it has some strict requirements about the dataset. A second contribution of the paper is the comparison of these model-driven methods with classical data-driven methods. For this purpose, we consider two classes of regression methods, methods derived from linear regression (Ridge and Lasso Regression, Principal Components Regression or Partial Least Squares Regression) and machine learning methods (Random Forest, k-Nearest Neighbor, Artificial Neural Network and SVM regression). The dataset consists of 720 records of corn yield at county scale provided by the United States Department of Agriculture (USDA) and the associated climatic data. A 5-folds cross-validation process and two accuracy metrics: root mean square error of prediction(RMSEP), mean absolute error of prediction(MAEP) were used to evaluate the crop prediction capacity. The results show that among the data-driven approaches, Random Forest is the most robust and generally achieves the best prediction error (MAEP 4.27%). It also outperforms our model-driven approach (MAEP 6.11%). However, the method to calibrate the mechanistic model from dataset easy to access offers several side-perspectives. The mechanistic model can potentially help to underline the stresses suffered by the crop or to identify the biological parameters of interest for breeding purposes. For this reason, an interesting perspective is to combine these two types of approaches.Keywords: crop yield prediction, crop model, sensitivity analysis, paramater estimation, particle swarm optimization, random forest
Procedia PDF Downloads 23514332 Using Deep Learning Neural Networks and Candlestick Chart Representation to Predict Stock Market
Authors: Rosdyana Mangir Irawan Kusuma, Wei-Chun Kao, Ho-Thi Trang, Yu-Yen Ou, Kai-Lung Hua
Abstract:
Stock market prediction is still a challenging problem because there are many factors that affect the stock market price such as company news and performance, industry performance, investor sentiment, social media sentiment, and economic factors. This work explores the predictability in the stock market using deep convolutional network and candlestick charts. The outcome is utilized to design a decision support framework that can be used by traders to provide suggested indications of future stock price direction. We perform this work using various types of neural networks like convolutional neural network, residual network and visual geometry group network. From stock market historical data, we converted it to candlestick charts. Finally, these candlestick charts will be feed as input for training a convolutional neural network model. This convolutional neural network model will help us to analyze the patterns inside the candlestick chart and predict the future movements of the stock market. The effectiveness of our method is evaluated in stock market prediction with promising results; 92.2% and 92.1 % accuracy for Taiwan and Indonesian stock market dataset respectively.Keywords: candlestick chart, deep learning, neural network, stock market prediction
Procedia PDF Downloads 45614331 A Review on Parametric Optimization of Casting Processes Using Optimization Techniques
Authors: Bhrugesh Radadiya, Jaydeep Shah
Abstract:
In Indian foundry industry, there is a need of defect free casting with minimum production cost in short lead time. Casting defect is a very large issue in foundry shop which increases the rejection rate of casting and wastage of materials. The various parameters influences on casting process such as mold machine related parameters, green sand related parameters, cast metal related parameters, mold related parameters and shake out related parameters. The mold related parameters are most influences on casting defects in sand casting process. This paper review the casting produced by foundry with shrinkage and blow holes as a major defects was analyzed and identified that mold related parameters such as mold temperature, pouring temperature and runner size were not properly set in sand casting process. These parameters were optimized using different optimization techniques such as Taguchi method, Response surface methodology, Genetic algorithm and Teaching-learning based optimization algorithm. Finally, concluded that a Teaching-learning based optimization algorithm give better result than other optimization techniques.Keywords: casting defects, genetic algorithm, parametric optimization, Taguchi method, TLBO algorithm
Procedia PDF Downloads 732