Search results for: magnetic domain
657 Inter-Annual Variations of Sea Surface Temperature in the Arabian Sea
Authors: K. S. Sreejith, C. Shaji
Abstract:
Though both Arabian Sea and its counterpart Bay of Bengal is forced primarily by the semi-annually reversing monsoons, the spatio-temporal variations of surface waters is very strong in the Arabian Sea as compared to the Bay of Bengal. This study focuses on the inter-annual variability of Sea Surface Temperature (SST) in the Arabian Sea by analysing ERSST dataset which covers 152 years of SST (January 1854 to December 2002) based on the ICOADS in situ observations. To capture the dominant SST oscillations and to understand the inter-annual SST variations at various local regions of the Arabian Sea, wavelet analysis was performed on this long time-series SST dataset. This tool is advantageous over other signal analysing tools like Fourier analysis, based on the fact that it unfolds a time-series data (signal) both in frequency and time domain. This technique makes it easier to determine dominant modes of variability and explain how those modes vary in time. The analysis revealed that pentadal SST oscillations predominate at most of the analysed local regions in the Arabian Sea. From the time information of wavelet analysis, it was interpreted that these cold and warm events of large amplitude occurred during the periods 1870-1890, 1890-1910, 1930-1950, 1980-1990 and 1990-2005. SST oscillations with peaks having period of ~ 2-4 years was found to be significant in the central and eastern regions of Arabian Sea. This indicates that the inter-annual SST variation in the Indian Ocean is affected by the El Niño-Southern Oscillation (ENSO) and Indian Ocean Dipole (IOD) events.Keywords: Arabian Sea, ICOADS, inter-annual variation, pentadal oscillation, SST, wavelet analysis
Procedia PDF Downloads 276656 Genome-Wide Identification and Characterization of MLO Family Genes in Pumpkin (Cucurbita maxima Duch.)
Authors: Khin Thanda Win, Chunying Zhang, Sanghyeob Lee
Abstract:
Mildew resistance locus o (Mlo), a plant-specific gene family with seven-transmembrane (TM), plays an important role in plant resistance to powdery mildew (PM). PM caused by Podosphaera xanthii is a widespread plant disease and probably represents the major fungal threat for many Cucurbits. The recent Cucurbita maxima genome sequence data provides an opportunity to identify and characterize the MLO gene family in this species. Total twenty genes (designated CmaMLO1 through CmaMLO20) have been identified by using an in silico cloning method with the MLO gene sequences of Cucumis sativus, Cucumis melo, Citrullus lanatus and Cucurbita pepo as probes. These CmaMLOs were evenly distributed on 15 chromosomes of 20 C. maxima chromosomes without any obvious clustering. Multiple sequence alignment showed that the common structural features of MLO gene family, such as TM domains, a calmodulin-binding domain and 30 important amino acid residues for MLO function, were well conserved. Phylogenetic analysis of the CmaMLO genes and other plant species reveals seven different clades (I through VII) and only clade IV is specific to monocots (rice, barley, and wheat). Phylogenetic and structural analyses provided preliminary evidence that five genes belonged to clade V could be the susceptibility genes which may play the importance role in PM resistance. This study is the first comprehensive report on MLO genes in C. maxima to our knowledge. These findings will facilitate the functional analysis of the MLOs related to PM susceptibility and are valuable resources for the development of disease resistance in pumpkin.Keywords: Mildew resistance locus o (Mlo), powdery mildew, phylogenetic relationship, susceptibility genes
Procedia PDF Downloads 181655 Multiscale Hub: An Open-Source Framework for Practical Atomistic-To-Continuum Coupling
Authors: Masoud Safdari, Jacob Fish
Abstract:
Despite vast amount of existing theoretical knowledge, the implementation of a universal multiscale modeling, analysis, and simulation software framework remains challenging. Existing multiscale software and solutions are often domain-specific, closed-source and mandate a high-level of experience and skills in both multiscale analysis and programming. Furthermore, tools currently existing for Atomistic-to-Continuum (AtC) multiscaling are developed with the assumptions such as accessibility of high-performance computing facilities to the users. These issues mentioned plus many other challenges have reduced the adoption of multiscale in academia and especially industry. In the current work, we introduce Multiscale Hub (MsHub), an effort towards making AtC more accessible through cloud services. As a joint effort between academia and industry, MsHub provides a universal web-enabled framework for practical multiscaling. Developed on top of universally acclaimed scientific programming language Python, the package currently provides an open-source, comprehensive, easy-to-use framework for AtC coupling. MsHub offers an easy to use interface to prominent molecular dynamics and multiphysics continuum mechanics packages such as LAMMPS and MFEM (a free, lightweight, scalable C++ library for finite element methods). In this work, we first report on the design philosophy of MsHub, challenges identified and issues faced regarding its implementation. MsHub takes the advantage of a comprehensive set of tools and algorithms developed for AtC that can be used for a variety of governing physics. We then briefly report key AtC algorithms implemented in MsHub. Finally, we conclude with a few examples illustrating the capabilities of the package and its future directions.Keywords: atomistic, continuum, coupling, multiscale
Procedia PDF Downloads 177654 Assessment of Platelet and Lymphocyte Interaction in Autoimmune Hyperthyroidism
Authors: Małgorzata Tomczyńska, Joanna Saluk-Bijak
Abstract:
Background: Graves’ disease is a frequent organ-specific autoimmune thyroid disease, which characterized by the presence of different kind autoantibodies, that, in most cases, act as agonists of the thyrotropin receptor, leading to hyperthyroidism. Role of platelets and lymphocytes can be modulated in the pathophysiology of thyroid autoimmune diseases. Interference in the physiology of platelets can lead to enhanced activity of these cells. Activated platelets can bind to circulating lymphocytes and to affect lymphocyte adhesion. Platelets and lymphocytes can regulate mutual functions. Therefore, the activation of T lymphocytes, as well as blood platelets, is associated with the development of inflammation and oxidative stress within the target tissue. The present study was performed to investigate a platelet-lymphocyte relation by assessing the degree of their mutual aggregation in whole blood of patients with Graves’ disease. Also, the purpose of this study was to examine the impact of platelet interaction on lymphocyte migration capacity. Methods: 30 patients with Graves’ disease were recruited in the study. The matched 30 healthy subjects were served as the control group. Immunophenotyping of lymphocytes was carried out by flow cytometry method. A CytoSelect™ Cell Migration Assay Kit was used to evaluate lymphocyte migration and adhesion to blood platelets. Visual assessment of lymphocyte-platelet aggregate morphology was done using confocal microscope after magnetic cell isolation by Miltenyi Biotec. Results: The migration and functional responses of lymphocytes to blood platelets were greater in the group of Graves’ disease patients compared with healthy controls. The group of Graves’ disease patients exhibited a reduced T lymphocyte and a higher B cell count compared with controls. Based on microscopic analysis, more platelet-lymphocyte aggregates were found in patients than in control. Conclusions: Studies have shown that in Graves' disease, lymphocytes show increased platelet affinity, more strongly migrating toward them, and forming mutual cellular conglomerates. This may be due to the increased activation of blood platelets in this disease.Keywords: blood platelets, cell migration, Graves’ disease, lymphocytes, lymphocyte-platelet aggregates
Procedia PDF Downloads 227653 Implementation of Free-Field Boundary Condition for 2D Site Response Analysis in OpenSees
Authors: M. Eskandarighadi, C. R. McGann
Abstract:
It is observed from past experiences of earthquakes that local site conditions can significantly affect the strong ground motion characteristics experience at the site. One-dimensional seismic site response analysis is the most common approach for investigating site response. This approach assumes that soil is homogeneous and infinitely extended in the horizontal direction. Therefore, tying side boundaries together is one way to model this behavior, as the wave passage is assumed to be only vertical. However, 1D analysis cannot capture the 2D nature of wave propagation, soil heterogeneity, and 2D soil profile with features such as inclined layer boundaries. In contrast, 2D seismic site response modeling can consider all of the mentioned factors to better understand local site effects on strong ground motions. 2D wave propagation and considering that the soil profile on the two sides of the model may not be identical clarifies the importance of a boundary condition on each side that can minimize the unwanted reflections from the edges of the model and input appropriate loading conditions. Ideally, the model size should be sufficiently large to minimize the wave reflection, however, due to computational limitations, increasing the model size is impractical in some cases. Another approach is to employ free-field boundary conditions that take into account the free-field motion that would exist far from the model domain and apply this to the sides of the model. This research focuses on implementing free-field boundary conditions in OpenSees for 2D site response analysisComparisons are made between 1D models and 2D models with various boundary conditions, and details and limitations of the developed free-field boundary modeling approach are discussed.Keywords: boundary condition, free-field, opensees, site response analysis, wave propagation
Procedia PDF Downloads 158652 Development of Medical Intelligent Process Model Using Ontology Based Technique
Authors: Emmanuel Chibuogu Asogwa, Tochukwu Sunday Belonwu
Abstract:
An urgent demand for creative solutions has been created by the rapid expansion of medical knowledge, the complexity of patient care, and the requirement for more precise decision-making. As a solution to this problem, the creation of a Medical Intelligent Process Model (MIPM) utilizing ontology-based appears as a promising way to overcome this obstacle and unleash the full potential of healthcare systems. The development of a Medical Intelligent Process Model (MIPM) using ontology-based techniques is motivated by a lack of quick access to relevant medical information and advanced tools for treatment planning and clinical decision-making, which ontology-based techniques can provide. The aim of this work is to develop a structured and knowledge-driven framework that leverages ontology, a formal representation of domain knowledge, to enhance various aspects of healthcare. Object-Oriented Analysis and Design Methodology (OOADM) were adopted in the design of the system as we desired to build a usable and evolvable application. For effective implementation of this work, we used the following materials/methods/tools: the medical dataset for the test of our model in this work was obtained from Kaggle. The ontology-based technique was used with Confusion Matrix, MySQL, Python, Hypertext Markup Language (HTML), Hypertext Preprocessor (PHP), Cascaded Style Sheet (CSS), JavaScript, Dreamweaver, and Fireworks. According to test results on the new system using Confusion Matrix, both the accuracy and overall effectiveness of the medical intelligent process significantly improved by 20% compared to the previous system. Therefore, using the model is recommended for healthcare professionals.Keywords: ontology-based, model, database, OOADM, healthcare
Procedia PDF Downloads 78651 Facile Fabrication of TiO₂NT/Fe₂O₃@Ag₂CO₃ Nanocomposite and Its Highly Efficient Visible Light Photocatalytic and Antibacterial Activity
Authors: Amal A. Al-Kahlawy, Heba H. El-Maghrabi
Abstract:
Due to the increasing need to environment protection in real time need to energize new materials are under extensive investigations. Between others, TiO2 nanotubes (TNTs) nanocomposite with iron oxide and silver carbonate, are promising alternatives as high-efficiency visible light photocatalyst due to their unique properties and their superior charge transport properties. Our efforts in this domain aim the construction of novel nanocomposite of TiO2NT/Fe2O3@Ag2CO3. The structure, surface morphology, chemical composition and optical properties were characterized by X-ray diffraction (XRD), Raman, Fourier-transform infrared spectroscopy (FTIR), scanning electron microscopy (SEM), energy dispersive X-ray spectrometer (EDS), transmission electron microscopy (TEM), selected area electron diffraction (SAED) and UV–vis diffuse reflectance spectroscopy (DRS). XRD results confirm the interaction of TiO2-NT with iron oxide. This novel nanocomposite shows remarkably enhanced performance for phenol compounds photodegradation. The experimental data shows a promising photocatalytic activity. In particular, a maximum value of 450 mg/g was removed within 60 min at solar light irradiation with degradation efficiency of 99.5%. The high photocatalytic activity of the nanocomposite is found to be related to the increased adsorption toward chemical species, enhanced light absorption and efficient charge separation and transfer. Finally, the designed TiO2NT/Fe2O3@Ag2CO3 nanocomposite has a great degree of sustainability and could has a potential application for the industrial treatment of wastewater containing toxic organic materials.Keywords: nanocomposite, photocatalyst, solar energy, titanium dioxide nanotubes
Procedia PDF Downloads 247650 Awareness among Medical Students and Faculty about Integration of Artifical Intelligence Literacy in Medical Curriculum
Authors: Fatima Faraz
Abstract:
BACKGROUND: While Artificial intelligence (AI) provides new opportunities across a wide variety of industries, healthcare is no exception. AI can lead to advancements in how the healthcare system functions and improves the quality of patient care. Developing countries like Pakistan are lagging in the implementation of AI-based solutions in healthcare. This demands increased knowledge and AI literacy among health care professionals. OBJECTIVES: To assess the level of awareness among medical students and faculty about AI in preparation for teaching AI basics and data science applications in clinical practice in an integrated medical curriculum. METHODS: An online 15-question semi-structured questionnaire, previously tested and validated, was delivered among participants through convenience sampling. The questionnaire composed of 3 parts: participant’s background knowledge, AI awareness, and attitudes toward AI applications in medicine. RESULTS: A total of 182 students and 39 faculty members from Rawalpindi Medical University, Pakistan, participated in the study. Only 26% of students and 46.2% of faculty members responded that they were aware of AI topics in clinical medicine. The major source of AI knowledge was social media (35.7%) for students and professional talks and colleagues (43.6%) for faculty members. 23.5% of participants answered that they personally had a basic understanding of AI. Students and faculty (60.1%) were interested in AI in patient care and teaching domain. These findings parallel similar published AI survey results. CONCLUSION: This survey concludes interest among students and faculty in AI developments and technology applications in healthcare. Further studies are required in order to correctly fit AI in the integrated modular curriculum of medical education.Keywords: medical education, data science, artificial intelligence, curriculum
Procedia PDF Downloads 101649 Exploring the Role of Building Information Modeling for Delivering Successful Construction Projects
Authors: Muhammad Abu Bakar Tariq
Abstract:
Construction industry plays a crucial role in the progress of societies and economies. Furthermore, construction projects have social as well as economic implications, thus, their success/failure have wider impacts. However, the industry is lagging behind in terms of efficiency and productivity. Building Information Modeling (BIM) is recognized as a revolutionary development in Architecture, Engineering and Construction (AEC) industry. There are numerous interest groups around the world providing definitions of BIM, proponents describing its advantages and opponents identifying challenges/barriers regarding adoption of BIM. This research is aimed at to determine what actually BIM is, along with its potential role in delivering successful construction projects. The methodology is critical analysis of secondary data sources i.e. information present in public domain, which include peer reviewed journal articles, industry and government reports, conference papers, books, case studies etc. It is discovered that clash detection and visualization are two major advantages of BIM. Clash detection option identifies clashes among structural, architectural and MEP designs before construction actually commences, which subsequently saves time as well as cost and ensures quality during execution phase of a project. Visualization is a powerful tool that facilitates in rapid decision-making in addition to communication and coordination among stakeholders throughout project’s life cycle. By eliminating inconsistencies that consume time besides cost during actual construction, improving collaboration among stakeholders throughout project’s life cycle, BIM can play a positive role to achieve efficiency and productivity that consequently deliver successful construction projects.Keywords: building information modeling, clash detection, construction project success, visualization
Procedia PDF Downloads 260648 Hand Symbol Recognition Using Canny Edge Algorithm and Convolutional Neural Network
Authors: Harshit Mittal, Neeraj Garg
Abstract:
Hand symbol recognition is a pivotal component in the domain of computer vision, with far-reaching applications spanning sign language interpretation, human-computer interaction, and accessibility. This research paper discusses the approach with the integration of the Canny Edge algorithm and convolutional neural network. The significance of this study lies in its potential to enhance communication and accessibility for individuals with hearing impairments or those engaged in gesture-based interactions with technology. In the experiment mentioned, the data is manually collected by the authors from the webcam using Python codes, to increase the dataset augmentation, is applied to original images, which makes the model more compatible and advanced. Further, the dataset of about 6000 coloured images distributed equally in 5 classes (i.e., 1, 2, 3, 4, 5) are pre-processed first to gray images and then by the Canny Edge algorithm with threshold 1 and 2 as 150 each. After successful data building, this data is trained on the Convolutional Neural Network model, giving accuracy: 0.97834, precision: 0.97841, recall: 0.9783, and F1 score: 0.97832. For user purposes, a block of codes is built in Python to enable a window for hand symbol recognition. This research, at its core, seeks to advance the field of computer vision by providing an advanced perspective on hand sign recognition. By leveraging the capabilities of the Canny Edge algorithm and convolutional neural network, this study contributes to the ongoing efforts to create more accurate, efficient, and accessible solutions for individuals with diverse communication needs.Keywords: hand symbol recognition, computer vision, Canny edge algorithm, convolutional neural network
Procedia PDF Downloads 65647 Diagnosis of Choledocholithiasis with Endosonography
Authors: A. Kachmazova, A. Shadiev, Y. Teterin, P. Yartcev
Abstract:
Introduction: Biliary calculi disease (LCS) still occupies the leading position among urgent diseases of the abdominal cavity, manifesting itself from asymptomatic course to life-threatening states. Nowadays arsenal of diagnostic methods for choledocholithiasis is quite wide: ultrasound, hepatobiliscintigraphy (HBSG), magnetic resonance imaging (MRI), endoscopic retrograde cholangiography (ERCP). Among them, transabdominal ultrasound (TA ultrasound) is the most accessible and routine diagnostic method. Nowadays ERCG is the "gold" standard in diagnosis and one-stage treatment of biliary tract obstruction. However, transpapillary techniques are accompanied by serious postoperative complications (postmanipulative pancreatitis (3-5%), endoscopic papillosphincterotomy bleeding (2%), cholangitis (1%)), the lethality being 0.4%. GBSG and MRI are also quite informative methods in the diagnosis of choledocholithiasis. Small size of concrements, their localization in intrapancreatic and retroduodenal part of common bile duct significantly reduces informativity of all diagnostic methods described above, that demands additional studying of this problem. Materials and Methods: 890 patients with the diagnosis of cholelithiasis (calculous cholecystitis) were admitted to the Sklifosovsky Scientific Research Institute of Hospital Medicine in the period from August, 2020 to June, 2021. Of them 115 people with mechanical jaundice caused by concrements in bile ducts. Results: Final EUS diagnosis was made in all patients (100,0%). In all patients in whom choledocholithiasis diagnosis was revealed or confirmed after EUS, ERCP was performed urgently (within two days from the moment of its detection) as the X-ray operation room was provided; it confirmed the presence of concrements. All stones were removed by lithoextraction using Dormia basket. The postoperative period in these patients had no complications. Conclusions: EUS is the most informative and safe diagnostic method, which allows to detect choledocholithiasis in patients with discrepancies between clinical-laboratory and instrumental methods of diagnosis in shortest time, that in its turn will help to decide promptly on the further tactics of patient treatment. We consider it reasonable to include EUS in the diagnostic algorithm for choledocholithiasis. Disclosure: Nothing to disclose.Keywords: endoscopic ultrasonography, choledocholithiasis, common bile duct, concrement, ERCP
Procedia PDF Downloads 85646 Possible Mechanism of DM2 Development in OSA Patients Mediated via Rev-Erb-Alpha and NPAS2 Proteins
Authors: Filip Franciszek Karuga, Szymon Turkiewicz, Marta Ditmer, Marcin Sochal, Piotr Białasiewicz, Agata Gabryelska
Abstract:
Circadian rhythm, an internal coordinator of physiological processes is composed of a set of semi-autonomous clocks. Clocks are regulated through the expression of circadian clock genes which form feedback loops, creating an oscillator. The primary loop consists of activators: CLOCK, BMAL1 and repressors: CRY, PER. CLOCK can be substituted by the Neuronal PAS Domain Protein 2 (NPAS2). Orphan nuclear receptor (REV-ERB-α) is a component of the secondary major loop, modulating the expression of BMAL1. Circadian clocks might be disrupted by the obstructive sleep apnea (OSA), which has also been associated with type II diabetes mellitus (DM2). Interestingly, studies suggest that dysregulation of NPAS2 and REV-ERB-α might contribute to the pathophysiology of DM2 as well. The goal of our study was to examine the role of NPAS2 and REV-ERB-α in DM2 in OSA patients. After examination of the clinical data, all participants underwent polysomnography (PSG) to assess their apnea-hypopnea index (AHI). Based on the acquired data participants were assigned to one of 3 groups: OSA (AHI>30, no DM2; n=17 for NPAS2 and 34 for REV-ERB-α), DM2 (AHI>30 + DM2; n=7 for NPAS2 and 15 for REV-ERB-α) and control group (AHI<5, no DM2; n=16 for NPAS2 and 31 for REV-ERB-α). ELISA immunoassay was performed to assess the serum protein level of REV-ERB-α and NPAS2. The only statistically significant difference between groups was observed in NPAS2 protein level (p=0.037). Post-hoc analysis showed significant differences between the OSA and the control group (p=0.017). AHI and NPAS2 level was significantly correlated (r=-0.478, p=0.002) in all groups. A significant correlation was observed between the REV-ERB-α level and sleep efficiency (r=0.617, p=0.005) as well as sleep maintenance efficiency (r=0.645, p=0.003) in the OSA group. We conclude, that NPAS2 is associated with OSA severity and might contribute to metabolic sequelae of this disease. REV-ERB-α on the other hand can influence sleep continuity and efficiency.Keywords: OSA, diabetes mellitus, endocrinology, chronobiology
Procedia PDF Downloads 155645 Recent Developments in the Application of Deep Learning to Stock Market Prediction
Authors: Shraddha Jain Sharma, Ratnalata Gupta
Abstract:
Predicting stock movements in the financial market is both difficult and rewarding. Analysts and academics are increasingly using advanced approaches such as machine learning techniques to anticipate stock price patterns, thanks to the expanding capacity of computing and the recent advent of graphics processing units and tensor processing units. Stock market prediction is a type of time series prediction that is incredibly difficult to do since stock prices are influenced by a variety of financial, socioeconomic, and political factors. Furthermore, even minor mistakes in stock market price forecasts can result in significant losses for companies that employ the findings of stock market price prediction for financial analysis and investment. Soft computing techniques are increasingly being employed for stock market prediction due to their better accuracy than traditional statistical methodologies. The proposed research looks at the need for soft computing techniques in stock market prediction, the numerous soft computing approaches that are important to the field, past work in the area with their prominent features, and the significant problems or issue domain that the area involves. For constructing a predictive model, the major focus is on neural networks and fuzzy logic. The stock market is extremely unpredictable, and it is unquestionably tough to correctly predict based on certain characteristics. This study provides a complete overview of the numerous strategies investigated for high accuracy prediction, with a focus on the most important characteristics.Keywords: stock market prediction, artificial intelligence, artificial neural networks, fuzzy logic, accuracy, deep learning, machine learning, stock price, trading volume
Procedia PDF Downloads 90644 Characterization of the Groundwater Aquifers at El Sadat City by Joint Inversion of VES and TEM Data
Authors: Usama Massoud, Abeer A. Kenawy, El-Said A. Ragab, Abbas M. Abbas, Heba M. El-Kosery
Abstract:
Vertical Electrical Sounding (VES) and Transient Electro Magnetic (TEM) survey have been applied for characterizing the groundwater aquifers at El Sadat industrial area. El-Sadat city is one of the most important industrial cities in Egypt. It has been constructed more than three decades ago at about 80 km northwest of Cairo along the Cairo–Alexandria desert road. Groundwater is the main source of water supplies required for domestic, municipal, and industrial activities in this area due to the lack of surface water sources. So, it is important to maintain this vital resource in order to sustain the development plans of this city. In this study, VES and TEM data were identically measured at 24 stations along three profiles trending NE–SW with the elongation of the study area. The measuring points were arranged in a grid like pattern with both inter-station spacing and line–line distance of about 2 km. After performing the necessary processing steps, the VES and TEM data sets were inverted individually to multi-layer models, followed by a joint inversion of both data sets. Joint inversion process has succeeded to overcome the model-equivalence problem encountered in the inversion of individual data set. Then, the joint models were used for the construction of a number of cross sections and contour maps showing the lateral and vertical distribution of the geo-electrical parameters in the subsurface medium. Interpretation of the obtained results and correlation with the available geological and hydrogeological information revealed TWO aquifer systems in the area. The shallow Pleistocene aquifer consists of sand and gravel saturated with fresh water and exhibits large thickness exceeding 200 m. The deep Pliocene aquifer is composed of clay and sand and shows low resistivity values. The water bearing layer of the Pleistocene aquifer and the upper surface of Pliocene aquifer are continuous and no structural features have cut this continuity through the investigated area.Keywords: El Sadat city, joint inversion, VES, TEM
Procedia PDF Downloads 370643 Cyber Security and Risk Assessment of the e-Banking Services
Authors: Aisha F. Bushager
Abstract:
Today we are more exposed than ever to cyber threats and attacks at personal, community, organizational, national, and international levels. More aspects of our lives are operating on computer networks simply because we are living in the fifth domain, which is called the Cyberspace. One of the most sensitive areas that are vulnerable to cyber threats and attacks is the Electronic Banking (e-Banking) area, where the banking sector is providing online banking services to its clients. To be able to obtain the clients trust and encourage them to practice e-Banking, also, to maintain the services provided by the banks and ensure safety, cyber security and risks control should be given a high priority in the e-banking area. The aim of the study is to carry out risk assessment on the e-banking services and determine the cyber threats, cyber attacks, and vulnerabilities that are facing the e-banking area specifically in the Kingdom of Bahrain. To collect relevant data, structured interviews were taken place with e-banking experts in different banks. Then, collected data where used as in input to the risk management framework provided by the National Institute of Standards and Technology (NIST), which was the model used in the study to assess the risks associated with e-banking services. The findings of the study showed that the cyber threats are commonly human errors, technical software or hardware failure, and hackers, on the other hand, the most common attacks facing the e-banking sector were phishing, malware attacks, and denial-of-service. The risks associated with the e-banking services were around the moderate level, however, more controls and countermeasures must be applied to maintain the moderate level of risks. The results of the study will help banks discover their vulnerabilities and maintain their online services, in addition, it will enhance the cyber security and contribute to the management and control of risks that are facing the e-banking sector.Keywords: cyber security, e-banking, risk assessment, threats identification
Procedia PDF Downloads 350642 The Relevance of Shared Cultural Leadership in the Survival of the Language and of the Francophone Culture in a Minority Language Environment
Authors: Lyne Chantal Boudreau, Claudine Auger, Arline Laforest
Abstract:
As an English-speaking country, Canada faces challenges in French-language education. During both editions of a provincial congress on education planned and conducted under shared cultural leadership, three organizers created a Francophone space where, for the first time in the province of New Brunswick (the only officially bilingual province in Canada), a group of stakeholders from the school, post-secondary and community sectors have succeeded in contributing to reflections on specific topics by sharing winning practices to meet the challenges of learning in a minority Francophone environment. Shared cultural leadership is a hybrid between theories of leadership styles in minority communities and theories of shared leadership. Through shared cultural leadership, the goal is simply to guide leadership and to set up all minority leaderships in minority context through shared leadership. This leadership style requires leaders to transition from a hierarchical to a horizontal approach, that is, to an approach where each individual is at the same level. In this exploratory research, it has been demonstrated that shared leadership exercised under the T-learning model best fosters the mobilization of all partners in advancing in-depth knowledge in a particular field while simultaneously allowing learning of the elements related to the domain in question. This session will present how it is possible to mobilize the whole community through leaders who continually develop their knowledge and skills in their specific field but also in related fields. Leaders in this style of management associated to shared cultural leadership acquire the ability to consider solutions to problems from a holistic perspective and to develop a collective power derived from the leadership of each and everyone in a space where all are rallied to promote the ultimate advancement of society.Keywords: education, minority context, shared leadership, t-leaning
Procedia PDF Downloads 247641 Multidimensional Inequality and Deprivation Among Tribal Communities of Andhra Pradesh, India
Authors: Sanjay Sinha, Mohd Umair Khan
Abstract:
The level of income inequality in India has been worrisome as the World Inequality Report termed it as a “poor and unequal country, with an affluent elite”. As important as income is to understand inequality and deprivation, it is just one dimension. But the historical roots and current realities of inequality and deprivation in India lies in many of the non-income dimensions such as housing, nutrition, education, agency, sense of inclusion etc. which are often ignored, especially in solution-oriented research. The level of inequality and deprivation among the tribal is one such case. There is a corpus of literature establishing that the tribal communities in India are disadvantageous on various grounds. Given their rural geography, issues of access and quality of basic facilities such as education and healthcare are often unaddressed. COVID-19 has further exacerbated this challenge and climate change will make it even more worrying. With this background, a succinct measurement tool at the village level is necessary to design short to medium-term actions with reference to risk mitigation for tribal communities. This research paper examines the level of inequality and deprivation among the tribal communities in the rural areas of Andhra Pradesh state of India using a Multidimensional Inequality and Deprivation Index based on the Alkire-Foster methodology. The methodology is theoretically grounded in the capability approach propounded by Amartya Sen, emphasizing on achieving the “beings and doings” (functionings) an individual reason to value. In the index, the authors have five domains, including Livelihood, Food Security, Education, Health and Housing and these domains are divided into sixteen indicators. This assessment is followed by domain-wise short-term and long-term solutions.Keywords: Andhra Pradesh, Alkire-Foster methodology, deprivation, inequality, multidimensionality, poverty, tribal
Procedia PDF Downloads 160640 Fast Bayesian Inference of Multivariate Block-Nearest Neighbor Gaussian Process (NNGP) Models for Large Data
Authors: Carlos Gonzales, Zaida Quiroz, Marcos Prates
Abstract:
Several spatial variables collected at the same location that share a common spatial distribution can be modeled simultaneously through a multivariate geostatistical model that takes into account the correlation between these variables and the spatial autocorrelation. The main goal of this model is to perform spatial prediction of these variables in the region of study. Here we focus on a geostatistical multivariate formulation that relies on sharing common spatial random effect terms. In particular, the first response variable can be modeled by a mean that incorporates a shared random spatial effect, while the other response variables depend on this shared spatial term, in addition to specific random spatial effects. Each spatial random effect is defined through a Gaussian process with a valid covariance function, but in order to improve the computational efficiency when the data are large, each Gaussian process is approximated to a Gaussian random Markov field (GRMF), specifically to the block nearest neighbor Gaussian process (Block-NNGP). This approach involves dividing the spatial domain into several dependent blocks under certain constraints, where the cross blocks allow capturing the spatial dependence on a large scale, while each individual block captures the spatial dependence on a smaller scale. The multivariate geostatistical model belongs to the class of Latent Gaussian Models; thus, to achieve fast Bayesian inference, it is used the integrated nested Laplace approximation (INLA) method. The good performance of the proposed model is shown through simulations and applications for massive data.Keywords: Block-NNGP, geostatistics, gaussian process, GRMF, INLA, multivariate models.
Procedia PDF Downloads 97639 Magnetic Resonance Imaging in Cochlear Implant Patients without Magnet Removal: A Safe and Effective Workflow Management Program
Authors: Yunhe Chen, Xinyun Liu, Qian Wang, Jianan Li
Abstract:
Background Cochlear implants (CIs) are currently the primary effective treatment for severe or profound sensorineural hearing loss. As China's population ages and the number of young children rises, the demand for MRI for CI patients is expected to increase. Methods Reviewed MRI cases of 25 CI patients between 2015 and 2024, assessed imaging auditory outcomes and adverse reactions. Use the adverse event record sheet and accompanying medication sheet to record follow-up measures. Results Most CI patients undergoing MRI may face risks such as artifacts, pain, redness, swelling, tissue damage, bleeding, and magnet displacement or demagnetization. Twenty-five CI patients in our hospital were reviewed. Seven patient underwent 3.0 T MR, the others underwent 1.5 T MR. The manufacturers are 18 cases in Austria, 5 cases in Australia and 2 cases in Nurotron. Among them, one patient with bilateral CI underwent 1.5 T MR examination after head pressure bandaging, and the left magnet was displaced (CI24RE Series, Australia). This patient underwent surgical replacement of the magnet under general anesthesia. Six days after the operation, the patient's feedback indicated that the performance of the cochlear implant was consistent with the previous results following the reactivation of the external device. Based on the experience of our hospital, we proposed the feasible management scheme of MRI examination procedure for CI patients. This plan should include a module for confirming MRI imaging parameters, informed consent, educational materials for patients, and other safety measures to ensure that patients receive imaging results safely and effectively, implify clinical. Conclusion As indications for both MRI and cochlear implantation expand,the number of MRI studies recommended for patients with cochlear implants will also increase. The process and management scheme proposed in this study can help to obtain imaging results safely and effectively, and reduce clinical stress.Keywords: cochlear implantation, MRI, magnet, displacement
Procedia PDF Downloads 14638 Dynamic Analysis of Mono-Pile: Spectral Element Method
Authors: Rishab Das, Arnab Banerjee, Bappaditya Manna
Abstract:
Mono-pile foundations are often used in soft soils in order to support heavy mega-structures, whereby often these deep footings may undergo dynamic excitation due to many causes like earthquake, wind or wave loads acting on the superstructure, blasting, and unbalanced machines, etc. A comprehensive analytical study is performed to study the dynamics of the mono-pile system embedded in cohesion-less soil. The soil is considered homogeneous and visco-elastic in nature and is analytically modeled using complex springs. Considering the N number of the elements of the pile, the final global stiffness matrix is obtained by using the theories of the spectral element matrix method. Further, statically condensing the intermediate internal nodes of the global stiffness matrix results to a smaller sub matrix containing the nodes experiencing the external translation and rotation, and the stiffness and damping functions (impedance functions) of the embedded piles are determined. Proper plots showing the variation of the real and imaginary parts of these impedance functions with the dimensionless frequency parameter are obtained. The plots obtained from this study are validated by that provided by Novak,1974. Further, the dynamic analysis of the resonator impregnated pile is proposed within this study. Moreover, with the aid of Wood's 1g laboratory scaling law, a proper scaled-down resonator-pile model is 3D printed using PLA material. Dynamic analysis of the scaled model is carried out in the time domain, whereby the lateral loads are imposed on the pile head. The response obtained from the sensors through the LabView software is compared with the proposed theoretical data.Keywords: mono-pile, visco-elastic, impedance, LabView
Procedia PDF Downloads 118637 Faster Pedestrian Recognition Using Deformable Part Models
Authors: Alessandro Preziosi, Antonio Prioletti, Luca Castangia
Abstract:
Deformable part models achieve high precision in pedestrian recognition, but all publicly available implementations are too slow for real-time applications. We implemented a deformable part model algorithm fast enough for real-time use by exploiting information about the camera position and orientation. This implementation is both faster and more precise than alternative DPM implementations. These results are obtained by computing convolutions in the frequency domain and using lookup tables to speed up feature computation. This approach is almost an order of magnitude faster than the reference DPM implementation, with no loss in precision. Knowing the position of the camera with respect to horizon it is also possible prune many hypotheses based on their size and location. The range of acceptable sizes and positions is set by looking at the statistical distribution of bounding boxes in labelled images. With this approach it is not needed to compute the entire feature pyramid: for example higher resolution features are only needed near the horizon. This results in an increase in mean average precision of 5% and an increase in speed by a factor of two. Furthermore, to reduce misdetections involving small pedestrians near the horizon, input images are supersampled near the horizon. Supersampling the image at 1.5 times the original scale, results in an increase in precision of about 4%. The implementation was tested against the public KITTI dataset, obtaining an 8% improvement in mean average precision over the best performing DPM-based method. By allowing for a small loss in precision computational time can be easily brought down to our target of 100ms per image, reaching a solution that is faster and still more precise than all publicly available DPM implementations.Keywords: autonomous vehicles, deformable part model, dpm, pedestrian detection, real time
Procedia PDF Downloads 281636 Computational Modelling of Epoxy-Graphene Composite Adhesive towards the Development of Cryosorption Pump
Authors: Ravi Verma
Abstract:
Cryosorption pump is the best solution to achieve clean, vibration free ultra-high vacuum. Furthermore, the operation of cryosorption pump is free from the influence of electric and magnetic fields. Due to these attributes, this pump is used in the space simulation chamber to create the ultra-high vacuum. The cryosorption pump comprises of three parts (a) panel which is cooled with the help of cryogen or cryocooler, (b) an adsorbent which is used to adsorb the gas molecules, (c) an epoxy which holds the adsorbent and the panel together thereby aiding in heat transfer from adsorbent to the panel. The performance of cryosorption pump depends on the temperature of the adsorbent and hence, on the thermal conductivity of the epoxy. Therefore we have made an attempt to increase the thermal conductivity of epoxy adhesive by mixing nano-sized graphene filler particles. The thermal conductivity of epoxy-graphene composite adhesive is measured with the help of indigenously developed experimental setup in the temperature range from 4.5 K to 7 K, which is generally the operating temperature range of cryosorption pump for efficiently pumping of hydrogen and helium gas. In this article, we have presented the experimental results of epoxy-graphene composite adhesive in the temperature range from 4.5 K to 7 K. We have also proposed an analytical heat conduction model to find the thermal conductivity of the composite. In this case, the filler particles, such as graphene, are randomly distributed in a base matrix of epoxy. The developed model considers the complete spatial random distribution of filler particles and this distribution is explained by Binomial distribution. The results obtained by the model have been compared with the experimental results as well as with the other established models. The developed model is able to predict the thermal conductivity in both isotropic regions as well as in anisotropic region over the required temperature range from 4.5 K to 7 K. Due to the non-empirical nature of the proposed model, it will be useful for the prediction of other properties of composite materials involving the filler in a base matrix. The present studies will aid in the understanding of low temperature heat transfer which in turn will be useful towards the development of high performance cryosorption pump.Keywords: composite adhesive, computational modelling, cryosorption pump, thermal conductivity
Procedia PDF Downloads 89635 Understanding Cyber Kill Chains: Optimal Allocation of Monitoring Resources Using Cooperative Game Theory
Authors: Roy. H. A. Lindelauf
Abstract:
Cyberattacks are complex processes consisting of multiple interwoven tasks conducted by a set of agents. Interdictions and defenses against such attacks often rely on cyber kill chain (CKC) models. A CKC is a framework that tries to capture the actions taken by a cyber attacker. There exists a growing body of literature on CKCs. Most of this work either a) describes the CKC with respect to one or more specific cyberattacks or b) discusses the tools and technologies used by the attacker at each stage of the CKC. Defenders, facing scarce resources, have to decide where to allocate their resources given the CKC and partial knowledge on the tools and techniques attackers use. In this presentation CKCs are analyzed through the lens of covert projects, i.e., interrelated tasks that have to be conducted by agents (human and/or computer) with the aim of going undetected. Various aspects of covert project models have been studied abundantly in the operations research and game theory domain, think of resource-limited interdiction actions that maximally delay completion times of a weapons project for instance. This presentation has investigated both cooperative and non-cooperative game theoretic covert project models and elucidated their relation to CKC modelling. To view a CKC as a covert project each step in the CKC is broken down into tasks and there are players of which each one is capable of executing a subset of the tasks. Additionally, task inter-dependencies are represented by a schedule. Using multi-glove cooperative games it is shown how a defender can optimize the allocation of his scarce resources (what, where and how to monitor) against an attacker scheduling a CKC. This study presents and compares several cooperative game theoretic solution concepts as metrics for assigning resources to the monitoring of agents.Keywords: cyber defense, cyber kill chain, game theory, information warfare techniques
Procedia PDF Downloads 140634 Regression Analysis in Estimating Stream-Flow and the Effect of Hierarchical Clustering Analysis: A Case Study in Euphrates-Tigris Basin
Authors: Goksel Ezgi Guzey, Bihrat Onoz
Abstract:
The scarcity of streamflow gauging stations and the increasing effects of global warming cause designing water management systems to be very difficult. This study is a significant contribution to assessing regional regression models for estimating streamflow. In this study, simulated meteorological data was related to the observed streamflow data from 1971 to 2020 for 33 stream gauging stations of the Euphrates-Tigris Basin. Ordinary least squares regression was used to predict flow for 2020-2100 with the simulated meteorological data. CORDEX- EURO and CORDEX-MENA domains were used with 0.11 and 0.22 grids, respectively, to estimate climate conditions under certain climate scenarios. Twelve meteorological variables simulated by two regional climate models, RCA4 and RegCM4, were used as independent variables in the ordinary least squares regression, where the observed streamflow was the dependent variable. The variability of streamflow was then calculated with 5-6 meteorological variables and watershed characteristics such as area and height prior to the application. Of the regression analysis of 31 stream gauging stations' data, the stations were subjected to a clustering analysis, which grouped the stations in two clusters in terms of their hydrometeorological properties. Two streamflow equations were found for the two clusters of stream gauging stations for every domain and every regional climate model, which increased the efficiency of streamflow estimation by a range of 10-15% for all the models. This study underlines the importance of homogeneity of a region in estimating streamflow not only in terms of the geographical location but also in terms of the meteorological characteristics of that region.Keywords: hydrology, streamflow estimation, climate change, hydrologic modeling, HBV, hydropower
Procedia PDF Downloads 129633 Prompt Design for Code Generation in Data Analysis Using Large Language Models
Authors: Lu Song Ma Li Zhi
Abstract:
With the rapid advancement of artificial intelligence technology, large language models (LLMs) have become a milestone in the field of natural language processing, demonstrating remarkable capabilities in semantic understanding, intelligent question answering, and text generation. These models are gradually penetrating various industries, particularly showcasing significant application potential in the data analysis domain. However, retraining or fine-tuning these models requires substantial computational resources and ample downstream task datasets, which poses a significant challenge for many enterprises and research institutions. Without modifying the internal parameters of the large models, prompt engineering techniques can rapidly adapt these models to new domains. This paper proposes a prompt design strategy aimed at leveraging the capabilities of large language models to automate the generation of data analysis code. By carefully designing prompts, data analysis requirements can be described in natural language, which the large language model can then understand and convert into executable data analysis code, thereby greatly enhancing the efficiency and convenience of data analysis. This strategy not only lowers the threshold for using large models but also significantly improves the accuracy and efficiency of data analysis. Our approach includes requirements for the precision of natural language descriptions, coverage of diverse data analysis needs, and mechanisms for immediate feedback and adjustment. Experimental results show that with this prompt design strategy, large language models perform exceptionally well in multiple data analysis tasks, generating high-quality code and significantly shortening the data analysis cycle. This method provides an efficient and convenient tool for the data analysis field and demonstrates the enormous potential of large language models in practical applications.Keywords: large language models, prompt design, data analysis, code generation
Procedia PDF Downloads 40632 Non-Invasive Data Extraction from Machine Display Units Using Video Analytics
Authors: Ravneet Kaur, Joydeep Acharya, Sudhanshu Gaur
Abstract:
Artificial Intelligence (AI) has the potential to transform manufacturing by improving shop floor processes such as production, maintenance and quality. However, industrial datasets are notoriously difficult to extract in a real-time, streaming fashion thus, negating potential AI benefits. The main example is some specialized industrial controllers that are operated by custom software which complicates the process of connecting them to an Information Technology (IT) based data acquisition network. Security concerns may also limit direct physical access to these controllers for data acquisition. To connect the Operational Technology (OT) data stored in these controllers to an AI application in a secure, reliable and available way, we propose a novel Industrial IoT (IIoT) solution in this paper. In this solution, we demonstrate how video cameras can be installed in a factory shop floor to continuously obtain images of the controller HMIs. We propose image pre-processing to segment the HMI into regions of streaming data and regions of fixed meta-data. We then evaluate the performance of multiple Optical Character Recognition (OCR) technologies such as Tesseract and Google vision to recognize the streaming data and test it for typical factory HMIs and realistic lighting conditions. Finally, we use the meta-data to match the OCR output with the temporal, domain-dependent context of the data to improve the accuracy of the output. Our IIoT solution enables reliable and efficient data extraction which will improve the performance of subsequent AI applications.Keywords: human machine interface, industrial internet of things, internet of things, optical character recognition, video analytics
Procedia PDF Downloads 109631 Abridging Pharmaceutical Analysis and Drug Discovery via LC-MS-TOF, NMR, in-silico Toxicity-Bioactivity Profiling for Therapeutic Purposing Zileuton Impurities: Need of Hour
Authors: Saurabh B. Ganorkar, Atul A. Shirkhedkar
Abstract:
The need for investigations protecting against toxic impurities though seems to be a primary requirement; the impurities which may prove non - toxic can be explored for their therapeutic potential if any to assist advanced drug discovery. The essential role of pharmaceutical analysis can thus be extended effectively to achieve it. The present study successfully achieved these objectives with characterization of major degradation products as impurities for Zileuton which has been used for to treat asthma since years. The forced degradation studies were performed to identify the potential degradation products using Ultra-fine Liquid-chromatography. Liquid-chromatography-Mass spectrometry (Time of Flight) and Proton Nuclear Magnetic Resonance Studies were utilized effectively to characterize the drug along with five major oxidative and hydrolytic degradation products (DP’s). The mass fragments were identified for Zileuton and path for the degradation was investigated. The characterized DP’s were subjected to In-Silico studies as XP Molecular Docking to compare the gain or loss in binding affinity with 5-Lipooxygenase enzyme. One of the impurity of was found to have the binding affinity more than the drug itself indicating for its potential to be more bioactive as better Antiasthmatic. The close structural resemblance has the ability to potentiate or reduce bioactivity and or toxicity. The chances of being active biologically at other sites cannot be denied and the same is achieved to some extent by predictions for probability of being active with Prediction of Activity Spectrum for Substances (PASS) The impurities found to be bio-active as Antineoplastic, Antiallergic, and inhibitors of Complement Factor D. The toxicological abilities as Ames-Mutagenicity, Carcinogenicity, Developmental Toxicity and Skin Irritancy were evaluated using Toxicity Prediction by Komputer Assisted Technology (TOPKAT). Two of the impurities were found to be non-toxic as compared to original drug Zileuton. As the drugs are purposed and repurposed effectively the impurities can also be; as they can have more binding affinity; less toxicity and better ability to be bio-active at other biological targets.Keywords: UFLC, LC-MS-TOF, NMR, Zileuton, impurities, toxicity, bio-activity
Procedia PDF Downloads 195630 Ionic Liquid and Chemical Denaturants Effects on the Fluorescence Properties of the Laccase
Authors: Othman Saoudi
Abstract:
In this work, we have interested in the investigation of the chemical denaturants and synthesized ionic liquids effects on the fluorescence properties of the laccase from Trametes versicolor. The fluorescence properties of the laccase result from the presence of Tryptophan, which has an aromatic core responsible for the absorption in ultra violet domain and the emission of the photons of fluorescence. The effect Pyrrolidinuim Formate ([pyrr][F]) and Morpholinium Formate ([morph][F]) ionic liquids on the laccase behavior for various volumetric fractions are studied. We have shown that the fluorescence spectrum relative to the [pyrr][F] presents a single band with a maximum around 340 nm and a secondary peak at 361 nm for a volumetric fraction of 20% v/v. For concentration superiors to 40%, the fluorescence intensity decreases and a displacement of the peaks toward higher wavelengths has occurred. For the [morph][F], the fluorescence spectrum showed a single band around 340 nm. The intensity of the principal peak decreases for concentration superiors to 20% v/v. From the plot representing the variation of the λₘₐₓ versus the volumetric concentration, we have determined the concentration of the half-transitions C1/2. These concentrations are equal to 42.62% and 40.91% v/v in the presence of [pyrr][F] and [morph][F] respectively. For the chemical denaturation, we have shown that the fluorescence intensity decreases with increasing denaturant concentrations where the maximum of the wavelength of emission shifts toward the higher wavelengths. We have also determined from the spectrum relative to the urea and GdmCl, the unfolding energy, ∆GD. The results show that the variation of the unfolding energy as a function of the denaturant concentrations varies according to the linear regression model. We have demonstrated also that the half-transitions C1/2 have occurred for urea and GdmCl denaturants concentrations around 3.06 and 3.17 M respectively.Keywords: laccase, fluorescence, ionic liquids, chemical denaturants
Procedia PDF Downloads 507629 A Rapid Prototyping Tool for Suspended Biofilm Growth Media
Authors: Erifyli Tsagkari, Stephanie Connelly, Zhaowei Liu, Andrew McBride, William Sloan
Abstract:
Biofilms play an essential role in treating water in biofiltration systems. The biofilm morphology and function are inextricably linked to the hydrodynamics of flow through a filter, and yet engineers rarely explicitly engineer this interaction. We develop a system that links computer simulation and 3-D printing to optimize and rapidly prototype filter media to optimize biofilm function with the hypothesis that biofilm function is intimately linked to the flow passing through the filter. A computational model that numerically solves the incompressible time-dependent Navier Stokes equations coupled to a model for biofilm growth and function is developed. The model is imbedded in an optimization algorithm that allows the model domain to adapt until criteria on biofilm functioning are met. This is applied to optimize the shape of filter media in a simple flow channel to promote biofilm formation. The computer code links directly to a 3-D printer, and this allows us to prototype the design rapidly. Its validity is tested in flow visualization experiments and by microscopy. As proof of concept, the code was constrained to explore a small range of potential filter media, where the medium acts as an obstacle in the flow that sheds a von Karman vortex street that was found to enhance the deposition of bacteria on surfaces downstream. The flow visualization and microscopy in the 3-D printed realization of the flow channel validated the predictions of the model and hence its potential as a design tool. Overall, it is shown that the combination of our computational model and the 3-D printing can be effectively used as a design tool to prototype filter media to optimize biofilm formation.Keywords: biofilm, biofilter, computational model, von karman vortices, 3-D printing.
Procedia PDF Downloads 142628 Evaluation of Model-Based Code Generation for Embedded Systems–Mature Approach for Development in Evolution
Authors: Nikolay P. Brayanov, Anna V. Stoynova
Abstract:
Model-based development approach is gaining more support and acceptance. Its higher abstraction level brings simplification of systems’ description that allows domain experts to do their best without particular knowledge in programming. The different levels of simulation support the rapid prototyping, verifying and validating the product even before it exists physically. Nowadays model-based approach is beneficial for modelling of complex embedded systems as well as a generation of code for many different hardware platforms. Moreover, it is possible to be applied in safety-relevant industries like automotive, which brings extra automation of the expensive device certification process and especially in the software qualification. Using it, some companies report about cost savings and quality improvements, but there are others claiming no major changes or even about cost increases. This publication demonstrates the level of maturity and autonomy of model-based approach for code generation. It is based on a real live automotive seat heater (ASH) module, developed using The Mathworks, Inc. tools. The model, created with Simulink, Stateflow and Matlab is used for automatic generation of C code with Embedded Coder. To prove the maturity of the process, Code generation advisor is used for automatic configuration. All additional configuration parameters are set to auto, when applicable, leaving the generation process to function autonomously. As a result of the investigation, the publication compares the quality of generated embedded code and a manually developed one. The measurements show that generally, the code generated by automatic approach is not worse than the manual one. A deeper analysis of the technical parameters enumerates the disadvantages, part of them identified as topics for our future work.Keywords: embedded code generation, embedded C code quality, embedded systems, model-based development
Procedia PDF Downloads 244