Search results for: common platform for automated programming
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 8979

Search results for: common platform for automated programming

8379 Contrasted Mean and Median Models in Egyptian Stock Markets

Authors: Mai A. Ibrahim, Mohammed El-Beltagy, Motaz Khorshid

Abstract:

Emerging Markets return distributions have shown significance departure from normality were they are characterized by fatter tails relative to the normal distribution and exhibit levels of skewness and kurtosis that constitute a significant departure from normality. Therefore, the classical Markowitz Mean-Variance is not applicable for emerging markets since it assumes normally-distributed returns (with zero skewness and kurtosis) and a quadratic utility function. Moreover, the Markowitz mean-variance analysis can be used in cases of moderate non-normality and it still provides a good approximation of the expected utility, but it may be ineffective under large departure from normality. Higher moments models and median models have been suggested in the literature for asset allocation in this case. Higher moments models have been introduced to account for the insufficiency of the description of a portfolio by only its first two moments while the median model has been introduced as a robust statistic which is less affected by outliers than the mean. Tail risk measures such as Value-at Risk (VaR) and Conditional Value-at-Risk (CVaR) have been introduced instead of Variance to capture the effect of risk. In this research, higher moment models including the Mean-Variance-Skewness (MVS) and Mean-Variance-Skewness-Kurtosis (MVSK) are formulated as single-objective non-linear programming problems (NLP) and median models including the Median-Value at Risk (MedVaR) and Median-Mean Absolute Deviation (MedMAD) are formulated as a single-objective mixed-integer linear programming (MILP) problems. The higher moment models and median models are compared to some benchmark portfolios and tested on real financial data in the Egyptian main Index EGX30. The results show that all the median models outperform the higher moment models were they provide higher final wealth for the investor over the entire period of study. In addition, the results have confirmed the inapplicability of the classical Markowitz Mean-Variance to the Egyptian stock market as it resulted in very low realized profits.

Keywords: Egyptian stock exchange, emerging markets, higher moment models, median models, mixed-integer linear programming, non-linear programming

Procedia PDF Downloads 315
8378 Design of Semi-Automatic Vent and Flash Remover

Authors: Inba Blesso P., Senthil Kumar P.

Abstract:

The main consideration of any tire manufacturing process is wear resistance. One of the factors that cause tire wear is improper removal of vent and flash from the tire surface. The contact point between tyre surface and vent is highly supposed to wear. When the vehicle running at higher speed with heavy load, the tire vent and flash is wearing initially and it makes few of the tire surface material to wear along with it. Hence, provision must be given to efficient removal vent and flash thereby tire wear. Human efforts in trimming of tire vent results in time consuming and inaccurate output. Hence, this lead to the reduction in production rate and profit. Thus, the development of automated system can helps to attain minimum time consumption and provide a possible way to get the profitable production. Semi-automated system that employs Pneumatic actuators and sequencing circuits are focused in this study. By implementing this, one can achieve the accurate results with reduction in time and profitable output.

Keywords: tire manufacturing, pneumatic system, vent and flash removal, engineering and technology

Procedia PDF Downloads 381
8377 Optimization of Economic Order Quantity of Multi-Item Inventory Control Problem through Nonlinear Programming Technique

Authors: Prabha Rohatgi

Abstract:

To obtain an efficient control over a huge amount of inventory of drugs in pharmacy department of any hospital, generally, the medicines are categorized on the basis of their cost ‘ABC’ (Always Better Control), first and then categorize on the basis of their criticality ‘VED’ (Vital, Essential, desirable) for prioritization. About one-third of the annual expenditure of a hospital is spent on medicines. To minimize the inventory investment, the hospital management may like to keep the medicines inventory low, as medicines are perishable items. The main aim of each and every hospital is to provide better services to the patients under certain limited resources. To achieve the satisfactory level of health care services to outdoor patients, a hospital has to keep eye on the wastage of medicines because expiry date of medicines causes a great loss of money though it was limited and allocated for a particular period of time. The objectives of this study are to identify the categories of medicines requiring incentive managerial control. In this paper, to minimize the total inventory cost and the cost associated with the wastage of money due to expiry of medicines, an inventory control model is used as an estimation tool and then nonlinear programming technique is used under limited budget and fixed number of orders to be placed in a limited time period. Numerical computations have been given and shown that by using scientific methods in hospital services, we can give more effective way of inventory management under limited resources and can provide better health care services. The secondary data has been collected from a hospital to give empirical evidence.

Keywords: ABC-VED inventory classification, multi item inventory problem, nonlinear programming technique, optimization of EOQ

Procedia PDF Downloads 256
8376 Case Report: A Rare Case of Popliteal Artery Aneurysm Presenting with Foot Drop

Authors: John Yahng, Hansraj Riteesh Bookun

Abstract:

Popliteal artery aneurysms (PAAs) are the most common arterial aneurysm of the periphery. It is defined as focal dilation of the artery more than 50% of the normal vessel diameter which usually varies between 7 mm to 11 mm. The most common presentation for PAAs is claudication due to luminal stenosis secondary to mural thrombus or acute limb ischaemia due to occlusive thrombosis or distal thromboembolism. It is less common for patients to present with non-ischaemic symptoms secondary to mass effect and compression of adjacent structures, and of these, presentation with common peroneal nerve compression is particularly uncommon. We present a rare case of a 92-year-old female patient presenting with 4-month history of left foot drop with radiological evidence of common peroneal nerve compression secondary to PAA of 22 mm by 21mm in size. To the best of our knowledge, this is the smallest reported popliteal aneurysm presenting with foot drop. We also present the endovascular treatment option taken in our case.

Keywords: aneurysm, foot drop, peroneal nerve, popliteal

Procedia PDF Downloads 404
8375 Design and Implementation of Automated Car Anti-Collision System Device Using Distance Sensor

Authors: Mehrab Masayeed Habib, Tasneem Sanjana, Ahmed Amin Rumel

Abstract:

Automated car anti-collision system is a trending technology of science. A car anti-collision system is an automobile safety system. The aim of this paper was to describe designing a car anti-collision system device to reduce the severity of an accident. The purpose of this device is to prevent collision among cars and objects to reduce the accidental death of human. This project gives an overview of secure & smooth journey of car as well as the certainty of human life. This system is controlled by microcontroller PIC. Sharp distance sensor is used to detect any object within the danger range. A crystal oscillator is used to produce the oscillation and generates the clock pulse of the microcontroller. An LCD is used to give information about the safe distance and a buzzer is used as alarm. An actuator is used as automatic break and inside the actuator; there is a motor driver that runs the actuator. For coding ‘microC PRO for PIC’ was used and ’Proteus Design Suite version 8 Software’ was used for simulation.

Keywords: sharp distance sensor, microcontroller, MicroC PRO for PIC, proteus, actuator, automobile anti-collision system

Procedia PDF Downloads 474
8374 Development of an Autonomous Automated Guided Vehicle with Robot Manipulator under Robot Operation System Architecture

Authors: Jinsiang Shaw, Sheng-Xiang Xu

Abstract:

This paper presents the development of an autonomous automated guided vehicle (AGV) with a robot arm attached on top of it within the framework of robot operation system (ROS). ROS can provide libraries and tools, including hardware abstraction, device drivers, libraries, visualizers, message-passing, package management, etc. For this reason, this AGV can provide automatic navigation and parts transportation and pick-and-place task using robot arm for typical industrial production line use. More specifically, this AGV will be controlled by an on-board host computer running ROS software. Command signals for vehicle and robot arm control and measurement signals from various sensors are transferred to respective microcontrollers. Users can operate the AGV remotely through the TCP / IP protocol and perform SLAM (Simultaneous Localization and Mapping). An RGBD camera and LIDAR sensors are installed on the AGV, using these data to perceive the environment. For SLAM, Gmapping is used to construct the environment map by Rao-Blackwellized particle filter; and AMCL method (Adaptive Monte Carlo localization) is employed for mobile robot localization. In addition, current AGV position and orientation can be visualized by ROS toolkit. As for robot navigation and obstacle avoidance, A* for global path planning and dynamic window approach for local planning are implemented. The developed ROS AGV with a robot arm on it has been experimented in the university factory. A 2-D and 3-D map of the factory were successfully constructed by the SLAM method. Base on this map, robot navigation through the factory with and without dynamic obstacles are shown to perform well. Finally, pick-and-place of parts using robot arm and ensuing delivery in the factory by the mobile robot are also accomplished.

Keywords: automated guided vehicle, navigation, robot operation system, Simultaneous Localization and Mapping

Procedia PDF Downloads 150
8373 Using Facebook as an Alternative Learning Tools in Malaysian Higher Learning Institutions: A Structural Equation Modelling Approach

Authors: Ahasanul Haque, Abdullah Sarwar, Khaliq Ahmed

Abstract:

Networking is important among students to achieve better understanding. Social networking plays an important role in the education. Realizing its huge potential, various organizations, including institutions of higher learning have moved to the area of social networks to interact with their students especially through Facebook. Therefore, measuring the effectiveness of Facebook as a learning tool has become an area of interest to academicians and researchers. Therefore, this study tried to integrate and propose new theoretical and empirical evidences by linking the western idea of adopting Facebook as an alternative learning platform from a Malaysian perspective. This study, thus, aimed to fill a gap by being among the pioneering research that tries to study the effectiveness of adopting Facebook as a learning platform across other cultural settings, namely Malaysia. Structural equation modelling was employed for data analysis and hypothesis testing. This study findings have provided some insights that would likely affect students’ awareness towards using Facebook as an alternative learning platform in the Malaysian higher learning institutions. At the end, future direction is proposed.

Keywords: Learning Management Tool, social networking, education, Malaysia

Procedia PDF Downloads 424
8372 The Effective Use of the Network in the Distributed Storage

Authors: Mamouni Mohammed Dhiya Eddine

Abstract:

This work aims at studying the exploitation of high-speed networks of clusters for distributed storage. Parallel applications running on clusters require both high-performance communications between nodes and efficient access to the storage system. Many studies on network technologies led to the design of dedicated architectures for clusters with very fast communications between computing nodes. Efficient distributed storage in clusters has been essentially developed by adding parallelization mechanisms so that the server(s) may sustain an increased workload. In this work, we propose to improve the performance of distributed storage systems in clusters by efficiently using the underlying high-performance network to access distant storage systems. The main question we are addressing is: do high-speed networks of clusters fit the requirements of a transparent, efficient and high-performance access to remote storage? We show that storage requirements are very different from those of parallel computation. High-speed networks of clusters were designed to optimize communications between different nodes of a parallel application. We study their utilization in a very different context, storage in clusters, where client-server models are generally used to access remote storage (for instance NFS, PVFS or LUSTRE). Our experimental study based on the usage of the GM programming interface of MYRINET high-speed networks for distributed storage raised several interesting problems. Firstly, the specific memory utilization in the storage access system layers does not easily fit the traditional memory model of high-speed networks. Secondly, client-server models that are used for distributed storage have specific requirements on message control and event processing, which are not handled by existing interfaces. We propose different solutions to solve communication control problems at the filesystem level. We show that a modification of the network programming interface is required. Data transfer issues need an adaptation of the operating system. We detail several propositions for network programming interfaces which make their utilization easier in the context of distributed storage. The integration of a flexible processing of data transfer in the new programming interface MYRINET/MX is finally presented. Performance evaluations show that its usage in the context of both storage and other types of applications is easy and efficient.

Keywords: distributed storage, remote file access, cluster, high-speed network, MYRINET, zero-copy, memory registration, communication control, event notification, application programming interface

Procedia PDF Downloads 219
8371 GIS Based Project Management Information System for Infrastructure Projects

Authors: Riki Panchal, Debasis Sarkar

Abstract:

This paper describes the work done for the GIS-based project management for different infrastructure projects. It is a review paper which gives the idea of the trends in the construction project management and various models adopted for the betterment of the project planning and execution. Traditional scheduling and progress control techniques such as bar charts and the critical path method fail to provide information pertaining to the spatial aspects of a construction project. An integrated system was developed to represent construction progress not only in terms of a CPM schedule but also in terms of a graphical representation of the construction that is synchronized with the work schedule. Hence, it is suggested to work on the common platform from where all the data can be shared and analyzed.

Keywords: GIS, project management, integrated model, infrastructure project

Procedia PDF Downloads 519
8370 Water Quality Calculation and Management System

Authors: H. M. B. N Jayasinghe

Abstract:

The water is found almost everywhere on Earth. Water resources contain a lot of pollution. Some diseases can be spread through the water to the living beings. So to be clean water it should undergo a number of treatments necessary to make it drinkable. So it is must to have purification technology for the wastewater. So the waste water treatment plants act a major role in these issues. When considering the procedures taken after the water treatment process was always based on manual calculations and recordings. Water purification plants may interact with lots of manual processes. It means the process taking much time consuming. So the final evaluation and chemical, biological treatment process get delayed. So to prevent those types of drawbacks there are some computerized programmable calculation and analytical techniques going to be introduced to the laboratory staff. To solve this problem automated system will be a solution in which guarantees the rational selection. A decision support system is a way to model data and make quality decisions based upon it. It is widely used in the world for the various kind of process automation. Decision support systems that just collect data and organize it effectively are usually called passive models where they do not suggest a specific decision but only reveal information. This web base system is based on global positioning data adding facility with map location. Most worth feature is SMS and E-mail alert service to inform the appropriate person on a critical issue. The technological influence to the system is HTML, MySQL, PHP, and some other web developing technologies. Current issues in the computerized water chemistry analysis are not much deep in progress. For an example the swimming pool water quality calculator. The validity of the system has been verified by test running and comparison with an existing plant data. Automated system will make the life easier in productively and qualitatively.

Keywords: automated system, wastewater, purification technology, map location

Procedia PDF Downloads 247
8369 Author Profiling: Prediction of Learners’ Gender on a MOOC Platform Based on Learners’ Comments

Authors: Tahani Aljohani, Jialin Yu, Alexandra. I. Cristea

Abstract:

The more an educational system knows about a learner, the more personalised interaction it can provide, which leads to better learning. However, asking a learner directly is potentially disruptive, and often ignored by learners. Especially in the booming realm of MOOC Massive Online Learning platforms, only a very low percentage of users disclose demographic information about themselves. Thus, in this paper, we aim to predict learners’ demographic characteristics, by proposing an approach using linguistically motivated Deep Learning Architectures for Learner Profiling, particularly targeting gender prediction on a FutureLearn MOOC platform. Additionally, we tackle here the difficult problem of predicting the gender of learners based on their comments only – which are often available across MOOCs. The most common current approaches to text classification use the Long Short-Term Memory (LSTM) model, considering sentences as sequences. However, human language also has structures. In this research, rather than considering sentences as plain sequences, we hypothesise that higher semantic - and syntactic level sentence processing based on linguistics will render a richer representation. We thus evaluate, the traditional LSTM versus other bleeding edge models, which take into account syntactic structure, such as tree-structured LSTM, Stack-augmented Parser-Interpreter Neural Network (SPINN) and the Structure-Aware Tag Augmented model (SATA). Additionally, we explore using different word-level encoding functions. We have implemented these methods on Our MOOC dataset, which is the most performant one comparing with a public dataset on sentiment analysis that is further used as a cross-examining for the models' results.

Keywords: deep learning, data mining, gender predication, MOOCs

Procedia PDF Downloads 148
8368 Making Permanent Supportive Housing Work for Vulnerable Populations

Authors: Olayinka Ariba, Abe Oudshoorn, Steve Rolfe, Carrie Anne Marshall, Deanna Befus, Jason Gilliland, Miranda Crockett, Susana Caxaj, Sarah McLean, Amy Van Berkum, Natasha Thuemler

Abstract:

Background: Secure housing is a platform for health and well-being. Those who struggle with housing stability have complex life and health histories and often require some support services such as the provision of permanent supportive housing. Poor access to supportive resources creates an exacerbation of chronic homelessness, particularly affecting individuals who need immediate access to mental health and addiction supports. This paper presents the first phase of a three-part study examining how on-site support impacts housing stability for recently-re-housed persons. Method: This study utilized a community-based participatory research methodology. Twenty in-depth interviews were conducted with permanent supportive housing residents from a single-site dwelling. Interpretative description analysis was used to draw common themes and understand the experiences and challenges of housing support. Results: Three interconnected themes were identified: 1) Available and timely supports; 2) Affordability; and 3) Community, but with independence as desired. These interconnected components are helping residents transition from homelessness or long-term mental health inpatient care to live in the community. Despite some participant concerns about resident conflicts, staff availability, and affordability, this has been a welcome and successful move for most. Conclusion: Supportive housing is essential for successful tenancies as a platform for health and well-being among Canada’s most vulnerable and, from the perspective of persons recently re-housed, permanent supportive housing is a worthwhile investment.

Keywords: homelessness, supportive housing, rehoused, housing stability

Procedia PDF Downloads 106
8367 Building User Behavioral Models by Processing Web Logs and Clustering Mechanisms

Authors: Madhuka G. P. D. Udantha, Gihan V. Dias, Surangika Ranathunga

Abstract:

Today Websites contain very interesting applications. But there are only few methodologies to analyze User navigations through the Websites and formulating if the Website is put to correct use. The web logs are only used if some major attack or malfunctioning occurs. Web Logs contain lot interesting dealings on users in the system. Analyzing web logs has become a challenge due to the huge log volume. Finding interesting patterns is not as easy as it is due to size, distribution and importance of minor details of each log. Web logs contain very important data of user and site which are not been put to good use. Retrieving interesting information from logs gives an idea of what the users need, group users according to their various needs and improve site to build an effective and efficient site. The model we built is able to detect attacks or malfunctioning of the system and anomaly detection. Logs will be more complex as volume of traffic and the size and complexity of web site grows. Unsupervised techniques are used in this solution which is fully automated. Expert knowledge is only used in validation. In our approach first clean and purify the logs to bring them to a common platform with a standard format and structure. After cleaning module web session builder is executed. It outputs two files, Web Sessions file and Indexed URLs file. The Indexed URLs file contains the list of URLs accessed and their indices. Web Sessions file lists down the indices of each web session. Then DBSCAN and EM Algorithms are used iteratively and recursively to get the best clustering results of the web sessions. Using homogeneity, completeness, V-measure, intra and inter cluster distance and silhouette coefficient as parameters these algorithms self-evaluate themselves to input better parametric values to run the algorithms. If a cluster is found to be too large then micro-clustering is used. Using Cluster Signature Module the clusters are annotated with a unique signature called finger-print. In this module each cluster is fed to Associative Rule Learning Module. If it outputs confidence and support as value 1 for an access sequence it would be a potential signature for the cluster. Then the access sequence occurrences are checked in other clusters. If it is found to be unique for the cluster considered then the cluster is annotated with the signature. These signatures are used in anomaly detection, prevent cyber attacks, real-time dashboards that visualize users, accessing web pages, predict actions of users and various other applications in Finance, University Websites, News and Media Websites etc.

Keywords: anomaly detection, clustering, pattern recognition, web sessions

Procedia PDF Downloads 288
8366 MB-Slam: A Slam Framework for Construction Monitoring

Authors: Mojtaba Noghabaei, Khashayar Asadi, Kevin Han

Abstract:

Simultaneous Localization and Mapping (SLAM) technology has recently attracted the attention of construction companies for real-time performance monitoring. To effectively use SLAM for construction performance monitoring, SLAM results should be registered to a Building Information Models (BIM). Registring SLAM and BIM can provide essential insights for construction managers to identify construction deficiencies in real-time and ultimately reduce rework. Also, registering SLAM to BIM in real-time can boost the accuracy of SLAM since SLAM can use features from both images and 3d models. However, registering SLAM with the BIM in real-time is a challenge. In this study, a novel SLAM platform named Model-Based SLAM (MB-SLAM) is proposed, which not only provides automated registration of SLAM and BIM but also improves the localization accuracy of the SLAM system in real-time. This framework improves the accuracy of SLAM by aligning perspective features such as depth, vanishing points, and vanishing lines from the BIM to the SLAM system. This framework extracts depth features from a monocular camera’s image and improves the localization accuracy of the SLAM system through a real-time iterative process. Initially, SLAM can be used to calculate a rough camera pose for each keyframe. In the next step, each SLAM video sequence keyframe is registered to the BIM in real-time by aligning the keyframe’s perspective with the equivalent BIM view. The alignment method is based on perspective detection that estimates vanishing lines and points by detecting straight edges on images. This process will generate the associated BIM views from the keyframes' views. The calculated poses are later improved during a real-time gradient descent-based iteration method. Two case studies were presented to validate MB-SLAM. The validation process demonstrated promising results and accurately registered SLAM to BIM and significantly improved the SLAM’s localization accuracy. Besides, MB-SLAM achieved real-time performance in both indoor and outdoor environments. The proposed method can fully automate past studies and generate as-built models that are aligned with BIM. The main contribution of this study is a SLAM framework for both research and commercial usage, which aims to monitor construction progress and performance in a unified framework. Through this platform, users can improve the accuracy of the SLAM by providing a rough 3D model of the environment. MB-SLAM further boosts the application to practical usage of the SLAM.

Keywords: perspective alignment, progress monitoring, slam, stereo matching.

Procedia PDF Downloads 225
8365 Evaluation of Coastal Erosion in the Jurisdiction of the Municipalities of Puerto Colombia and Tubará, Atlántico – Colombia in Google Earth Engine with Landsat and Sentinel 2 Images

Authors: Francisco Reyes, Hector Ramirez

Abstract:

In the coastal zones are home to mangrove swamps, coral reefs, and seagrass ecosystems, which are the most biodiverse and fragile on the planet. These areas support a great diversity of marine life; they are also extraordinarily important for humans in the provision of food, water, wood, and other associated goods and services; they also contribute to climate regulation. The lack of an automated model that generates information on the dynamics of changes in coastlines and coastal erosion is identified as a central problem. Coastlines were determined from 1984 to 2020 on the Google Earth platform Engine from Landsat and Sentinel images, using the Normalized Differential Water Index (MNDWI) and Digital Shoreline Analysis System (DSAS) v5.0. Starting from the 2020 coastline, the 10-year prediction (Year 2031) was determined with the erosion of 238.32 hectares and an accretion of 181.96 hectares, while the 20-year prediction (Year 2041) will be presented an erosion of 544.04 hectares and an accretion of 133.94 hectares. The erosion and accretion of Playa Muelle in the municipality of Puerto Colombia were established, which will register the highest value of erosion. The coverage that presented the greatest change was that of artificialized Territories.

Keywords: coastline, coastal erosion, MNDWI, Google Earth Engine, Colombia

Procedia PDF Downloads 120
8364 Automated Method Time Measurement System for Redesigning Dynamic Facility Layout

Authors: Salam Alzubaidi, G. Fantoni, F. Failli, M. Frosolini

Abstract:

The dynamic facility layout problem is a really critical issue in the competitive industrial market; thus, solving this problem requires robust design and effective simulation systems. The sustainable simulation requires inputting reliable and accurate data into the system. So this paper describes an automated system integrated into the real environment to measure the duration of the material handling operations, collect the data in real-time, and determine the variances between the actual and estimated time schedule of the operations in order to update the simulation software and redesign the facility layout periodically. The automated method- time measurement system collects the real data through using Radio Frequency-Identification (RFID) and Internet of Things (IoT) technologies. Hence, attaching RFID- antenna reader and RFID tags enables the system to identify the location of the objects and gathering the time data. The real duration gathered will be manipulated by calculating the moving average duration of the material handling operations, choosing the shortest material handling path, and then updating the simulation software to redesign the facility layout accommodating with the shortest/real operation schedule. The periodic simulation in real-time is more sustainable and reliable than the simulation system relying on an analysis of historical data. The case study of this methodology is in cooperation with a workshop team for producing mechanical parts. Although there are some technical limitations, this methodology is promising, and it can be significantly useful in the redesigning of the manufacturing layout.

Keywords: dynamic facility layout problem, internet of things, method time measurement, radio frequency identification, simulation

Procedia PDF Downloads 120
8363 Duplicated Common Bile Duct: A Recipe for Injury

Authors: David Armany, Matthew Allaway, Preet Gosal, Senarath Edirimanne

Abstract:

A potentially devastating complication of routine laparoscopic cholecystectomy includes iatrogenic bile duct injuries, which represent a stable incidence rate of 0.3% over the past three decades. Whilst related to several relative risks such as surgeon experience and patient factors (older age, male sex), misinterpretation of biliary tree anatomy remains the most common cause, accounting for 80% of iatrogenic Common Bile Duct injuries. Whilst extremely rare, a duplicate common bile duct anomaly remains a potential variation to encounter during biliary surgery, with 30 recognised cases in the worldwide literature, of which type Vb accounts for 4. We report the case of a rare type Vb variation encountered during intra-operative laparoscopic cholecystectomy and confirmed on cholangiogram. To our knowledge, this is the first documented Type Vb case encountered in an Australian population. Given these anomalies are asymptomatic and can perpetuate iatrogenic common bile duct injuries, awareness of all subtypes is crucial. Irrevocably, preoperative Magnetic Resonance Cholangiopancreatography can help recognise these anomalies before the operating theatre; however, their widespread adoption is limited by expensive and availability.

Keywords: duplicated common bile duct, type Vb, cholecystitis, MRCP, cholangiogram, iatrogenic CBD

Procedia PDF Downloads 90
8362 Process Mining as an Ecosystem Platform to Mitigate a Deficiency of Processes Modelling

Authors: Yusra Abdulsalam Alqamati, Ahmed Alkilany

Abstract:

The teaching staff is a distinct group whose impact is on the educational process and which plays an important role in enhancing the quality of the academic education process. To improve the management effectiveness of the academy, the Teaching Staff Management System (TSMS) proposes that all teacher processes be digitized. Since the BPMN approach can accurately describe the processes, it lacks a clear picture of the process flow map, something that the process mining approach has, which is extracting information from event logs for discovery, monitoring, and model enhancement. Therefore, these two methodologies were combined to create the most accurate representation of system operations, the ability to extract data records and mining processes, recreate them in the form of a Petri net, and then generate them in a BPMN model for a more in-depth view of process flow. Additionally, the TSMS processes will be orchestrated to handle all requests in a guaranteed small-time manner thanks to the integration of the Google Cloud Platform (GCP), the BPM engine, and allowing business owners to take part throughout the entire TSMS project development lifecycle.

Keywords: process mining, BPM, business process model and notation, Petri net, teaching staff, Google Cloud Platform

Procedia PDF Downloads 142
8361 System and Method for Providing Web-Based Remote Application Service

Authors: Shuen-Tai Wang, Yu-Ching Lin, Hsi-Ya Chang

Abstract:

With the development of virtualization technologies, a new type of service named cloud computing service is produced. Cloud users usually encounter the problem of how to use the virtualized platform easily over the web without requiring the plug-in or installation of special software. The object of this paper is to develop a system and a method enabling process interfacing within an automation scenario for accessing remote application by using the web browser. To meet this challenge, we have devised a web-based interface that system has allowed to shift the GUI application from the traditional local environment to the cloud platform, which is stored on the remote virtual machine. We designed the sketch of web interface following the cloud virtualization concept that sought to enable communication and collaboration among users. We describe the design requirements of remote application technology and present implementation details of the web application and its associated components. We conclude that this effort has the potential to provide an elastic and resilience environment for several application services. Users no longer have to burden the system maintenances and reduce the overall cost of software licenses and hardware. Moreover, this remote application service represents the next step to the mobile workplace, and it lets user to use the remote application virtually from anywhere.

Keywords: virtualization technology, virtualized platform, web interface, remote application

Procedia PDF Downloads 289
8360 PsyVBot: Chatbot for Accurate Depression Diagnosis using Long Short-Term Memory and NLP

Authors: Thaveesha Dheerasekera, Dileeka Sandamali Alwis

Abstract:

The escalating prevalence of mental health issues, such as depression and suicidal ideation, is a matter of significant global concern. It is plausible that a variety of factors, such as life events, social isolation, and preexisting physiological or psychological health conditions, could instigate or exacerbate these conditions. Traditional approaches to diagnosing depression entail a considerable amount of time and necessitate the involvement of adept practitioners. This underscores the necessity for automated systems capable of promptly detecting and diagnosing symptoms of depression. The PsyVBot system employs sophisticated natural language processing and machine learning methodologies, including the use of the NLTK toolkit for dataset preprocessing and the utilization of a Long Short-Term Memory (LSTM) model. The PsyVBot exhibits a remarkable ability to diagnose depression with a 94% accuracy rate through the analysis of user input. Consequently, this resource proves to be efficacious for individuals, particularly those enrolled in academic institutions, who may encounter challenges pertaining to their psychological well-being. The PsyVBot employs a Long Short-Term Memory (LSTM) model that comprises a total of three layers, namely an embedding layer, an LSTM layer, and a dense layer. The stratification of these layers facilitates a precise examination of linguistic patterns that are associated with the condition of depression. The PsyVBot has the capability to accurately assess an individual's level of depression through the identification of linguistic and contextual cues. The task is achieved via a rigorous training regimen, which is executed by utilizing a dataset comprising information sourced from the subreddit r/SuicideWatch. The diverse data present in the dataset ensures precise and delicate identification of symptoms linked with depression, thereby guaranteeing accuracy. PsyVBot not only possesses diagnostic capabilities but also enhances the user experience through the utilization of audio outputs. This feature enables users to engage in more captivating and interactive interactions. The PsyVBot platform offers individuals the opportunity to conveniently diagnose mental health challenges through a confidential and user-friendly interface. Regarding the advancement of PsyVBot, maintaining user confidentiality and upholding ethical principles are of paramount significance. It is imperative to note that diligent efforts are undertaken to adhere to ethical standards, thereby safeguarding the confidentiality of user information and ensuring its security. Moreover, the chatbot fosters a conducive atmosphere that is supportive and compassionate, thereby promoting psychological welfare. In brief, PsyVBot is an automated conversational agent that utilizes an LSTM model to assess the level of depression in accordance with the input provided by the user. The demonstrated accuracy rate of 94% serves as a promising indication of the potential efficacy of employing natural language processing and machine learning techniques in tackling challenges associated with mental health. The reliability of PsyVBot is further improved by the fact that it makes use of the Reddit dataset and incorporates Natural Language Toolkit (NLTK) for preprocessing. PsyVBot represents a pioneering and user-centric solution that furnishes an easily accessible and confidential medium for seeking assistance. The present platform is offered as a modality to tackle the pervasive issue of depression and the contemplation of suicide.

Keywords: chatbot, depression diagnosis, LSTM model, natural language process

Procedia PDF Downloads 69
8359 Multi-Label Approach to Facilitate Test Automation Based on Historical Data

Authors: Warda Khan, Remo Lachmann, Adarsh S. Garakahally

Abstract:

The increasing complexity of software and its applicability in a wide range of industries, e.g., automotive, call for enhanced quality assurance techniques. Test automation is one option to tackle the prevailing challenges by supporting test engineers with fast, parallel, and repetitive test executions. A high degree of test automation allows for a shift from mundane (manual) testing tasks to a more analytical assessment of the software under test. However, a high initial investment of test resources is required to establish test automation, which is, in most cases, a limitation to the time constraints provided for quality assurance of complex software systems. Hence, a computer-aided creation of automated test cases is crucial to increase the benefit of test automation. This paper proposes the application of machine learning for the generation of automated test cases. It is based on supervised learning to analyze test specifications and existing test implementations. The analysis facilitates the identification of patterns between test steps and their implementation with test automation components. For the test case generation, this approach exploits historical data of test automation projects. The identified patterns are the foundation to predict the implementation of unknown test case specifications. Based on this support, a test engineer solely has to review and parameterize the test automation components instead of writing them manually, resulting in a significant time reduction for establishing test automation. Compared to other generation approaches, this ML-based solution can handle different writing styles, authors, application domains, and even languages. Furthermore, test automation tools require expert knowledge by means of programming skills, whereas this approach only requires historical data to generate test cases. The proposed solution is evaluated using various multi-label evaluation criteria (EC) and two small-sized real-world systems. The most prominent EC is ‘Subset Accuracy’. The promising results show an accuracy of at least 86% for test cases, where a 1:1 relationship (Multi-Class) between test step specification and test automation component exists. For complex multi-label problems, i.e., one test step can be implemented by several components, the prediction accuracy is still at 60%. It is better than the current state-of-the-art results. It is expected the prediction quality to increase for larger systems with respective historical data. Consequently, this technique facilitates the time reduction for establishing test automation and is thereby independent of the application domain and project. As a work in progress, the next steps are to investigate incremental and active learning as additions to increase the usability of this approach, e.g., in case labelled historical data is scarce.

Keywords: machine learning, multi-class, multi-label, supervised learning, test automation

Procedia PDF Downloads 132
8358 Low-Level Modeling for Optimal Train Routing and Scheduling in Busy Railway Stations

Authors: Quoc Khanh Dang, Thomas Bourdeaud’huy, Khaled Mesghouni, Armand Toguy´eni

Abstract:

This paper studies a train routing and scheduling problem for busy railway stations. Our objective is to allow trains to be routed in dense areas that are reaching saturation. Unlike traditional methods that allocate all resources to setup a route for a train and until the route is freed, our work focuses on the use of resources as trains progress through the railway node. This technique allows a larger number of trains to be routed simultaneously in a railway node and thus reduces their current saturation. To deal with this problem, this study proposes an abstract model and a mixed-integer linear programming formulation to solve it. The applicability of our method is illustrated on a didactic example.

Keywords: busy railway stations, mixed-integer linear programming, offline railway station management, train platforming, train routing, train scheduling

Procedia PDF Downloads 254
8357 The Integration and Automation of EDA Tools in an Integrated Circuit Design Environment

Authors: Rohaya Abdul Wahab, Raja Mohd Fuad Tengku Aziz, Nazaliza Othman, Sharifah Saleh, Nabihah Razali, Rozaimah Baharim, M. Hanif M. Nasir

Abstract:

This paper will discuss how EDA tools are integrated and automated in an Integrated Circuit Design Environment. Some of the problems face in our current environment is that users need to configure manually on the library paths, start-up files and project directories. Certain manual processes that happen between the users and applications can be automated but they must be transparent to the users. For example, the users can run the applications directly after login without knowing the library paths and start-up files locations. The solution to these problems is to automate the processes using standard configuration files which will benefit the users and EDA support. This paper will discuss how the implementation is done to automate the process using scripting languages such as Perl, Tcl, Scheme and Shell Script. These scripting tools are great assets for design engineers to build a robust and powerful design flow and this technique is widely used to integrate all the tools together.

Keywords: EDA tools, Integrated Circuits, scripting, integration, automation

Procedia PDF Downloads 324
8356 Development of the Web-Based Multimedia N-Screen Service System for Cross Platform

Authors: S. Bae, J. Shin, S. Lee

Abstract:

As the development of smart devices such as Smart TV, Smartphone, Tablet PC, Laptop, the interest in N-Screen Services that can be cross-linked with heterogeneous devices is increasing. N-Screen means User-centric services that can share and constantly watch multimedia contents anytime and anywhere. However, the existing N-Screen system has the limitation that N-Screen system has to implement the application for each platform and device to provide multimedia service. To overcome this limitation, Multimedia N-Screen Service System is proposed through the web, and it is independent of different environments. The combination of Web and cloud computing technologies from this study results in increasing efficiency and reduction in costs.

Keywords: N-screen, web, cloud, multimedia

Procedia PDF Downloads 301
8355 Brief Guide to Cloud-Based AI Prototyping: Key Insights from Selected Case Studies Using Google Cloud Platform

Authors: Kamellia Reshadi, Pranav Ragji, Theodoros Soldatos

Abstract:

Recent advancements in cloud computing and storage, along with rapid progress in artificial intelligence (AI), have transformed approaches to developing efficient, scalable applications. However, integrating AI with cloud computing poses challenges as these fields are often disjointed, and many advancements remain difficult to access, obscured in complex documentation or scattered across research reports. For this reason, we share experiences from prototype projects combining these technologies. Specifically, we focus on Google Cloud Platform (GCP) functionalities and describe vision and speech activities applied to labeling, subtitling, and urban traffic flow tasks. We describe challenges, pricing, architecture, and other key features, considering the goal of real-time performance. We hope our demonstrations provide not only essential guidelines for using these functionalities but also enable more similar approaches.

Keywords: artificial intelligence, cloud computing, real-time applications, case studies, knowledge management, research and development, text labeling, video annotation, urban traffic analysis, public safety, prototyping, Google Cloud Platform

Procedia PDF Downloads 12
8354 Applying Concurrent Development Process for the Web Using Aspect-Oriented Approach

Authors: Hiroaki Fukuda

Abstract:

This paper shows a concurrent development process for modern web application, called Rich Internet Application (RIA), and describes its effect using a non-trivial application development. In the last years, RIAs such as Ajax and Flex have become popular based mainly on high-speed network. RIA provides sophisticated interfaces and user experiences, therefore, the development of RIA requires two kinds of engineer: a developer who implements business logic, and a designer who designs interface and experiences. Although collaborative works are becoming important for the development of RIAs, shared resources such as source code make it difficult. For example, if a design of interface is modified after developers have finished business logic implementations, they need to repeat the same implementations, and also tests to verify application’s behavior. MVC architecture and Object-oriented programming (OOP) enables to dividing an application into modules such as interfaces and logic, however, developers and/or designers have to write pieces of code (e.g., event handlers) that make these modules work as an application. On the other hand, Aspect-oriented programming (AOP) is ex- pected to solve complexity of application software development nowadays. AOP provides methods to separate crosscutting concerns that are scattered pieces of code from primary concerns. In this paper, we provide a concurrent development process for RIAs by introducing AOP concept. This process makes it possible to reduce shared resources between developers and designers, therefore they can perform their tasks concurrently. In addition, we describe experiences of development for a practical application using our proposed development process to show its availability.

Keywords: aspect-oriented programming, concurrent, development process, rich internet application

Procedia PDF Downloads 301
8353 Energy Efficient Resource Allocation and Scheduling in Cloud Computing Platform

Authors: Shuen-Tai Wang, Ying-Chuan Chen, Yu-Ching Lin

Abstract:

There has been renewal of interest in the relation between Green IT and cloud computing in recent years. Cloud computing has to be a highly elastic environment which provides stable services to users. The growing use of cloud computing facilities has caused marked energy consumption, putting negative pressure on electricity cost of computing center or data center. Each year more and more network devices, storages and computers are purchased and put to use, but it is not just the number of computers that is driving energy consumption upward. We could foresee that the power consumption of cloud computing facilities will double, triple, or even more in the next decade. This paper aims at resource allocation and scheduling technologies that are short of or have not well developed yet to reduce energy utilization in cloud computing platform. In particular, our approach relies on recalling services dynamically onto appropriate amount of the machines according to user’s requirement and temporarily shutting down the machines after finish in order to conserve energy. We present initial work on integration of resource and power management system that focuses on reducing power consumption such that they suffice for meeting the minimizing quality of service required by the cloud computing platform.

Keywords: cloud computing, energy utilization, power consumption, resource allocation

Procedia PDF Downloads 339
8352 Understanding Psychological Distress and Protection Issues among Children Associated with Armed Groups

Authors: Grace Onubedo

Abstract:

The primary objective of this research study is to contribute to and deepen the understanding of the realities and conditions for which children recruited by violent extremist organisations in Nigeria live, as well as ascertain the state of their mental health following their reunification with either family or protection workers. The research is intended to contribute to a more focused child protection programming agenda for children associated with armed forces and groups in Nigeria and the wider conflict setting. The extent to which violence has affected the psychological well-being and mental health of children abducted and exposed to activities of Violent Extremist groups remains a largely empirical question. This research attempts to answer the following research questions with the aim of providing further evidences for informed programming: I. What are the demographic characteristics of children associated with armed groups? II. What is the state of their mental health? III. What is the relationship between their background and their mental health?

Keywords: counterterrorism, psychosocial support, psychological distress, children, armed groups

Procedia PDF Downloads 131
8351 GPU Accelerated Fractal Image Compression for Medical Imaging in Parallel Computing Platform

Authors: Md. Enamul Haque, Abdullah Al Kaisan, Mahmudur R. Saniat, Aminur Rahman

Abstract:

In this paper, we have implemented both sequential and parallel version of fractal image compression algorithms using CUDA (Compute Unified Device Architecture) programming model for parallelizing the program in Graphics Processing Unit for medical images, as they are highly similar within the image itself. There is several improvements in the implementation of the algorithm as well. Fractal image compression is based on the self similarity of an image, meaning an image having similarity in majority of the regions. We take this opportunity to implement the compression algorithm and monitor the effect of it using both parallel and sequential implementation. Fractal compression has the property of high compression rate and the dimensionless scheme. Compression scheme for fractal image is of two kinds, one is encoding and another is decoding. Encoding is very much computational expensive. On the other hand decoding is less computational. The application of fractal compression to medical images would allow obtaining much higher compression ratios. While the fractal magnification an inseparable feature of the fractal compression would be very useful in presenting the reconstructed image in a highly readable form. However, like all irreversible methods, the fractal compression is connected with the problem of information loss, which is especially troublesome in the medical imaging. A very time consuming encoding process, which can last even several hours, is another bothersome drawback of the fractal compression.

Keywords: accelerated GPU, CUDA, parallel computing, fractal image compression

Procedia PDF Downloads 336
8350 Eco-Design of Construction Industrial Park in China with Selection of Candidate Tenants

Authors: Yang Zhou, Kaijian Li, Guiwen Liu

Abstract:

Offsite construction is an innovative alternative to conventional site-based construction, with wide-ranging benefits. It requires building components, elements or modules were prefabricated and pre-assembly before installed into their final locations. To improve efficiency and achieve synergies, in recent years, construction companies were clustered into construction industrial parks (CIPs) in China. A CIP is a community of construction manufacturing and service businesses located together on a common property. Companies involved in industrial clusters can obtain environment and economic benefits by sharing resources and information in a given region. Therefore, the concept of industrial symbiosis (IS) can be applied to the traditional CIP to achieve sustainable industrial development or redevelopment through the implementation of eco-industrial parks (EIP). However, before designing a symbiosis network between companies in a CIP, candidate support tenants need to be selected to complement the existing construction companies. In this study, an access indicator system and a linear programming model are established to select candidate tenants in a CIP while satisfying the degree of connectivity among the enterprises in the CIP, minimizing the environmental impact, and maximizing the annualized profit of the CIP. The access indicator system comprises three primary indicators and fifteen secondary indicators, is proposed from the perspective of park-based level. The fifteen indicators are classified as three primary indicators including industrial symbiosis, environment performance and economic benefit, according to the three dimensions of sustainability (environment, economic and social dimensions) and the three R's of the environment (reduce, reuse and recycle). The linear programming model is a method to assess the satisfactoriness of all the indicators and to make an optimal multi-objective selection among candidate tenants. This method provides a practical tool for planners of a CIP in evaluating which among the candidate tenants would best complement existing anchor construction tenants. The reasonability and validity of the indicator system and the method is worth further study in the future.

Keywords: construction industrial park, China, industrial symbiosis, offsite construction, selection of support tenants

Procedia PDF Downloads 274