Search results for: pivot language translation approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 16551

Search results for: pivot language translation approach

12231 Sentiment Analysis of Ensemble-Based Classifiers for E-Mail Data

Authors: Muthukumarasamy Govindarajan

Abstract:

Detection of unwanted, unsolicited mails called spam from email is an interesting area of research. It is necessary to evaluate the performance of any new spam classifier using standard data sets. Recently, ensemble-based classifiers have gained popularity in this domain. In this research work, an efficient email filtering approach based on ensemble methods is addressed for developing an accurate and sensitive spam classifier. The proposed approach employs Naive Bayes (NB), Support Vector Machine (SVM) and Genetic Algorithm (GA) as base classifiers along with different ensemble methods. The experimental results show that the ensemble classifier was performing with accuracy greater than individual classifiers, and also hybrid model results are found to be better than the combined models for the e-mail dataset. The proposed ensemble-based classifiers turn out to be good in terms of classification accuracy, which is considered to be an important criterion for building a robust spam classifier.

Keywords: accuracy, arcing, bagging, genetic algorithm, Naive Bayes, sentiment mining, support vector machine

Procedia PDF Downloads 126
12230 Determination of the Phosphate Activated Glutaminase Localization in the Astrocyte Mitochondria Using Kinetic Approach

Authors: N. V. Kazmiruk, Y. R. Nartsissov

Abstract:

Phosphate activated glutaminase (GA, E.C. 3.5.1.2) plays a key role in glutamine/glutamate homeostasis in mammalian brain, catalyzing the hydrolytic deamidation of glutamine to glutamate and ammonium ions. GA is mainly localized in mitochondria, where it has the catalytically active form on the inner mitochondrial membrane (IMM) and the other soluble form, which is supposed to be dormant. At present time, the exact localization of the membrane glutaminase active site remains a controversial and an unresolved issue. The first hypothesis called c-side localization suggests that the catalytic site of GA faces the inter-membrane space and products of the deamidation reaction have immediate access to cytosolic metabolism. According to the alternative m-side localization hypothesis, GA orients to the matrix, making glutamate and ammonium available for the tricarboxylic acid cycle metabolism in mitochondria directly. In our study, we used a multi-compartment kinetic approach to simulate metabolism of glutamate and glutamine in the astrocytic cytosol and mitochondria. We used physiologically important ratio between the concentrations of glutamine inside the matrix of mitochondria [Glnₘᵢₜ] and glutamine in the cytosol [Glncyt] as a marker for precise functioning of the system. Since this ratio directly depends on the mitochondrial glutamine carrier (MGC) flow parameters, key observation was to investigate the dependence of the [Glnmit]/[Glncyt] ratio on the maximal velocity of MGC at different initial concentrations of mitochondrial glutamate. Another important task was to observe the similar dependence at different inhibition constants of the soluble GA. The simulation results confirmed the experimental c-side localization hypothesis, in which the glutaminase active site faces the outer surface of the IMM. Moreover, in the case of such localization of the enzyme, a 3-fold decrease in ammonium production was predicted.

Keywords: glutamate metabolism, glutaminase, kinetic approach, mitochondrial membrane, multi-compartment modeling

Procedia PDF Downloads 102
12229 Econophysical Approach on Predictability of Financial Crisis: The 2001 Crisis of Turkey and Argentina Case

Authors: Arzu K. Kamberli, Tolga Ulusoy

Abstract:

Technological developments and the resulting global communication have made the 21st century when large capitals are moved from one end to the other via a button. As a result, the flow of capital inflows has accelerated, and capital inflow has brought with it crisis-related infectiousness. Considering the irrational human behavior, the financial crisis in the world under the influence of the whole world has turned into the basic problem of the countries and increased the interest of the researchers in the reasons of the crisis and the period in which they lived. Therefore, the complex nature of the financial crises and its linearly unexplained structure have also been included in the new discipline, econophysics. As it is known, although financial crises have prediction mechanisms, there is no definite information. In this context, in this study, using the concept of electric field from the electrostatic part of physics, an early econophysical approach for global financial crises was studied. The aim is to define a model that can take place before the financial crises, identify financial fragility at an earlier stage and help public and private sector members, policy makers and economists with an econophysical approach. 2001 Turkey crisis has been assessed with data from Turkish Central Bank which is covered between 1992 to 2007, and for 2001 Argentina crisis, data was taken from IMF and the Central Bank of Argentina from 1997 to 2007. As an econophysical method, an analogy is used between the Gauss's law used in the calculation of the electric field and the forecasting of the financial crisis. The concept of Φ (Financial Flux) has been adopted for the pre-warning of the crisis by taking advantage of this analogy, which is based on currency movements and money mobility. For the first time used in this study Φ (Financial Flux) calculations obtained by the formula were analyzed by Matlab software, and in this context, in 2001 Turkey and Argentina Crisis for Φ (Financial Flux) crisis of values has been confirmed to give pre-warning.

Keywords: econophysics, financial crisis, Gauss's Law, physics

Procedia PDF Downloads 143
12228 Indexing and Incremental Approach Using Map Reduce Bipartite Graph (MRBG) for Mining Evolving Big Data

Authors: Adarsh Shroff

Abstract:

Big data is a collection of dataset so large and complex that it becomes difficult to process using data base management tools. To perform operations like search, analysis, visualization on big data by using data mining; which is the process of extraction of patterns or knowledge from large data set. In recent years, the data mining applications become stale and obsolete over time. Incremental processing is a promising approach to refreshing mining results. It utilizes previously saved states to avoid the expense of re-computation from scratch. This project uses i2MapReduce, an incremental processing extension to Map Reduce, the most widely used framework for mining big data. I2MapReduce performs key-value pair level incremental processing rather than task level re-computation, supports not only one-step computation but also more sophisticated iterative computation, which is widely used in data mining applications, and incorporates a set of novel techniques to reduce I/O overhead for accessing preserved fine-grain computation states. To optimize the mining results, evaluate i2MapReduce using a one-step algorithm and three iterative algorithms with diverse computation characteristics for efficient mining.

Keywords: big data, map reduce, incremental processing, iterative computation

Procedia PDF Downloads 330
12227 Production Line Layout Planning Based on Complexity Measurement

Authors: Guoliang Fan, Aiping Li, Nan Xie, Liyun Xu, Xuemei Liu

Abstract:

Mass customization production increases the difficulty of the production line layout planning. The material distribution process for variety of parts is very complex, which greatly increases the cost of material handling and logistics. In response to this problem, this paper presents an approach of production line layout planning based on complexity measurement. Firstly, by analyzing the influencing factors of equipment layout, the complexity model of production line is established by using information entropy theory. Then, the cost of the part logistics is derived considering different variety of parts. Furthermore, the function of optimization including two objectives of the lowest cost, and the least configuration complexity is built. Finally, the validity of the function is verified in a case study. The results show that the proposed approach may find the layout scheme with the lowest logistics cost and the least complexity. Optimized production line layout planning can effectively improve production efficiency and equipment utilization with lowest cost and complexity.

Keywords: production line, layout planning, complexity measurement, optimization, mass customization

Procedia PDF Downloads 375
12226 Agent/Group/Role Organizational Model to Simulate an Industrial Control System

Authors: Noureddine Seddari, Mohamed Belaoued, Salah Bougueroua

Abstract:

The modeling of complex systems is generally based on the decomposition of their components into sub-systems easier to handle. This division has to be made in a methodical way. In this paper, we introduce an industrial control system modeling and simulation based on the Multi-Agent System (MAS) methodology AALAADIN and more particularly the underlying conceptual model Agent/Group/Role (AGR). Indeed, in this division using AGR model, the overall system is decomposed into sub-systems in order to improve the understanding of regulation and control systems, and to simplify the implementation of the obtained agents and their groups, which are implemented using the Multi-Agents Development KIT (MAD-KIT) platform. This approach appears to us to be the most appropriate for modeling of this type of systems because, due to the use of MAS, it is possible to model real systems in which very complex behaviors emerge from relatively simple and local interactions between many different individuals, therefore a MAS is well adapted to describe a system from the standpoint of the activity of its components, that is to say when the behavior of the individuals is complex (difficult to describe with equations). The main aim of this approach is the take advantage of the performance, the scalability and the robustness that are intuitively provided by MAS.

Keywords: complex systems, modeling and simulation, industrial control system, MAS, AALAADIN, AGR, MAD-KIT

Procedia PDF Downloads 227
12225 Optimizing Telehealth Internet of Things Integration: A Sustainable Approach through Fog and Cloud Computing Platforms for Energy Efficiency

Authors: Yunyong Guo, Sudhakar Ganti, Bryan Guo

Abstract:

The swift proliferation of telehealth Internet of Things (IoT) devices has sparked concerns regarding energy consumption and the need for streamlined data processing. This paper presents an energy-efficient model that integrates telehealth IoT devices into a platform based on fog and cloud computing. This integrated system provides a sustainable and robust solution to address the challenges. Our model strategically utilizes fog computing as a localized data processing layer and leverages cloud computing for resource-intensive tasks, resulting in a significant reduction in overall energy consumption. The incorporation of adaptive energy-saving strategies further enhances the efficiency of our approach. Simulation analysis validates the effectiveness of our model in improving energy efficiency for telehealth IoT systems, particularly when integrated with localized fog nodes and both private and public cloud infrastructures. Subsequent research endeavors will concentrate on refining the energy-saving model, exploring additional functional enhancements, and assessing its broader applicability across various healthcare and industry sectors.

Keywords: energy-efficient, fog computing, IoT, telehealth

Procedia PDF Downloads 60
12224 Hybrid Control Mode Based on Multi-Sensor Information by Fuzzy Approach for Navigation Task of Autonomous Mobile Robot

Authors: Jonqlan Lin, C. Y. Tasi, K. H. Lin

Abstract:

This paper addresses the issue of the autonomous mobile robot (AMR) navigation task based on the hybrid control modes. The novel hybrid control mode, based on multi-sensors information by using the fuzzy approach, has been presented in this research. The system operates in real time, is robust, enables the robot to operate with imprecise knowledge, and takes into account the physical limitations of the environment in which the robot moves, obtaining satisfactory responses for a large number of different situations. An experiment is simulated and carried out with a pioneer mobile robot. From the experimental results, the effectiveness and usefulness of the proposed AMR obstacle avoidance and navigation scheme are confirmed. The experimental results show the feasibility, and the control system has improved the navigation accuracy. The implementation of the controller is robust, has a low execution time, and allows an easy design and tuning of the fuzzy knowledge base.

Keywords: autonomous mobile robot, obstacle avoidance, MEMS, hybrid control mode, navigation control

Procedia PDF Downloads 454
12223 The Use of Artificial Intelligence to Harmonization in the Lawmaking Process

Authors: Supriyadi, Andi Intan Purnamasari, Aminuddin Kasim, Sulbadana, Mohammad Reza

Abstract:

The development of the Industrial Revolution Era 4.0 brought a significant influence in the administration of countries in all parts of the world, including Indonesia, not only in the administration and economic sectors but the ways and methods of forming laws should also be adjusted. Until now, the process of making laws carried out by the Parliament with the Government still uses the classical method. The law-making process still uses manual methods, such as typing harmonization of regulations, so that it is not uncommon for errors to occur, such as writing errors, copying articles and so on, things that require a high level of accuracy and relying on inventory and harmonization carried out manually by humans. However, this method often creates several problems due to errors and inaccuracies on the part of officers who harmonize laws after discussion and approval; this has a very serious impact on the system of law formation in Indonesia. The use of artificial intelligence in the process of forming laws seems to be justified and becomes the answer in order to minimize the disharmony of various laws and regulations. This research is normative research using the Legislative Approach and the Conceptual Approach. This research focuses on the question of how to use Artificial Intelligence for Harmonization in the Lawmaking Process.

Keywords: artificial intelligence, harmonization, laws, intelligence

Procedia PDF Downloads 131
12222 Effective Public Health Communication: Vaccine Health Messaging with Aboriginal and Torres Strait Islander Peoples

Authors: Maria Karidakis, Barbara Kelly

Abstract:

The challenges precipitated by the advent of COVID-19 have brought to the fore the task governments and key stakeholders are faced with; ensuring public health communication is readily accessible to vulnerable populations. COVID-19 has presented challenges for the provision and reception of timely, accessible, and accurate health information pertaining to vaccine health messaging to Aboriginal and Torres Strait Islander peoples. The aim of this qualitative study was to explore strategies used by Aboriginal-led organisations to improve communication about COVID-19 and vaccination for their communities and to explore how these mediation and outreach strategies were received by community members. We interviewed 6 Aboriginal-led organisations and 15 community members from several states across Australian, and these interviews were analysed thematically. The findings suggest that effective public health communication is enhanced when aFirst nations-led response defines the governance that happens in First Nations communities. Pro-active and self-determining Aboriginal leadership and decision-making helps drive the response to counter a growing trend towards vaccine hesitancy. Other strategies include establishing partnerships with government departments and relevant non-governmental organisations to ensure services are implemented and culturally appropriate. The outcomes of this research will afford policymakers, stakeholders in healthcare, and cultural mediators the capacity to identify strengths and potential problems associated with pandemic health information and to subsequently implement creative and culturally specific solutions that go beyond the provision of written documentation via translation or interpreting. It will also enable governing bodies to adjust multilingual polices and to adopt mediation strategies that will improve information delivery and intercultural services on a national and international level.

Keywords: intercultural communication, qualitative, public health communication, COVID-19, pandemic, mediated communication, first nations people

Procedia PDF Downloads 144
12221 The Practices of Creative Tourism in Urban and Rural Areas at International Level

Authors: Isabel Freitas, Paula Remoaldo, Olga Matos, Ricardo Goja, Juliana Araujo, Vitor Ribeiro, Miguel Pereira

Abstract:

Several destinations have been experiencing a transition from a massified cultural tourism to a creative tourism approach. In this new segment of tourism, urban territories have been the focus for several decades. Urban studies on creative industries and initiatives have been taking place in big cities marginalizing small towns and more specifically rural areas. This paper envisages evaluating the differences between rural and urban institutions/platforms, mostly certified by the Creative Tourism Network, in what concerns the practices and initiatives in creative tourism worldwide. In the research carried out between March 2017 and March 2018, we had three levels of primary data and qualitative analysis: i) research on Google (web) by using several keywords like 'creative tourism initiatives', 'creative cities', 'best practices in creative tourism' (from March to August 2017). With the help of the certification of institutions/platforms by the Creative Tourism Network, 24 institutions were found and declared to be developing creative initiatives. It was decided to try to unravel the type of activities and some practices and initiatives carried out by these institutions and the analysis of the differences between rural and urban initiatives. A database of 20 items (e.g., institutions in charge of implementing the initiatives, year of implementation, site, activities developed, place of development, country of origin, type of partners chosen) was created for each institution/platform; ii) A deeper analysis was made on the websites’ information on the institutions (from September to December 2017). The type of professionals involved in the activities, the language used in the activities and the type of activity performed were some of the data analysed and iii) To complement these data, semi-structured interviews were done to representatives of the institutions, conducted mainly by Skype from July 2017 to April 2018. The interviews consisted of 17 questions. In the present paper, these interviews are used to complement the analysis of the same items. Some of the qualitative analysis was supported by the narratives of the leaders of the twenty-four institutions that were surveyed. The results indicate that creative tourism is more active and diverse in urban areas. Some more consolidated communication strategies and partnerships are needed for these activities to become economically more sustainable. The findings of this research provide researchers and practitioners with a better understanding of creative tourism and give some information of how creative tourism is developed in rural and urban areas, the gaps and lack of information, and all the possible directions towards the development of the creative tourism industry.

Keywords: creative tourism, practices of creative tourism, rural areas, urban areas

Procedia PDF Downloads 165
12220 Numerical Evaluation of the Flow Behavior inside the Scrubber Unit with Engine Exhaust Pipe

Authors: Kumaresh Selvakumar, Man Young Kim

Abstract:

A wet scrubber is an air pollution control device that removes particulate matter and acid gases from waste gas streams found in marine engine exhaust. If the flue gases in the exhaust is employed for CFD simulation, it makes the problem complicate due to the involvement of emissions. Owing to the fact, the scrubber system in this paper is handled with appropriate approach by designing with the flow properties of hot air and water droplet injections to evaluate the flow behavior inside the system. Since the wet scrubber has the capability of operating over wide range of mixture compositions, the current scrubber model with the designing approach doesn’t deviate from the actual behavior of the system. The scrubber design is constructed with engine exhaust pipe with the purpose of measuring the flow properties inside the scrubber by the influence of exhaust pipe characteristics. The flow properties are computed by the thermodynamic variables such as temperature and pressure with the flow velocity. In this work, numerical analyses have been conducted for the flow of fluid in the scrubber system through CFD technique.

Keywords: wet scrubber, water droplet injections, thermodynamic variables, CFD technique

Procedia PDF Downloads 328
12219 Linear Complementary Based Approach for Unilateral Frictional Contact between Wheel and Beam

Authors: Muskaan Sethi, Arnab Banerjee, Bappaditya Manna

Abstract:

The present paper aims to investigate a suitable contact between a wheel rolling over a flexible beam. A Linear Complementary (LCP) based approach has been adopted to simulate the contact dynamics for a rigid wheel traversing over a flexible Euler Bernoulli simply supported beam. The adopted methodology is suitable to incorporate the effect of frictional force acting at the wheel-beam interface. Moreover, the possibility of the generation of a gap between the two bodies has also been considered. The present method is based on a unilateral contact assumption which assumes that no penetration would occur when the two bodies come in contact. This assumption helps to predict the contact between wheels and beams in a more practical sense. The proposed methodology is validated with the previously published results and is found to be in good agreement. Further, this method is applied to simulate the contact between wheels and beams for various railway configurations. Moreover, different parametric studies are conducted to study the contact dynamics between the wheel and beam more thoroughly.

Keywords: contact dynamics, linear complementary problem, railway dynamics, unilateral contact

Procedia PDF Downloads 87
12218 Unraveling the Threads of Madness: Henry Russell’s 'The Maniac' as an Advocate for Deinstitutionalization in the Nineteenth Century

Authors: T. J. Laws-Nicola

Abstract:

Henry Russell was best known as a composer of more than 300 songs. Many of his compositions were popular for both their sentimental texts, as in ‘The Old Armchair,’ and those of a more political nature, such as ‘Woodsman, Spare That Tree!’ Indeed, Russell had written such songs of advocacy as those associated with abolitionism (‘The Slave Ship’) and environmentalism (‘Woodsman, Spare that Tree!’). ‘The Maniac’ is his only composition addressing the issue of institutionalization. The text is borrowed and adapted from the monodrama The Captive by M.G. ‘Monk’ Lewis. Through an analysis of form, harmony, melody, text, and thematic development and interactions between text and music we can approach a clearer understanding of ‘The Maniac’ and how the text and music interact. Select periodicals, such as The London Times, provide contemporary critical review for ‘The Maniac.’ Additional nineteenth century songs whose texts focus on madness and/or institutionalization will assist in building a stylistic and cultural context for ‘The Maniac.’ Through comparative analyses of ‘The Maniac’ with a body of songs that focus on similar topics, we can approach a clear understanding of the song as a vehicle for deinstitutionalization.

Keywords: 19th century song, institutionalization, M. G. Lewis, Henry Russell

Procedia PDF Downloads 517
12217 Inverse Polynomial Numerical Scheme for the Solution of Initial Value Problems in Ordinary Differential Equations

Authors: Ogunrinde Roseline Bosede

Abstract:

This paper presents the development, analysis and implementation of an inverse polynomial numerical method which is well suitable for solving initial value problems in first order ordinary differential equations with applications to sample problems. We also present some basic concepts and fundamental theories which are vital to the analysis of the scheme. We analyzed the consistency, convergence, and stability properties of the scheme. Numerical experiments were carried out and the results compared with the theoretical or exact solution and the algorithm was later coded using MATLAB programming language.

Keywords: differential equations, numerical, polynomial, initial value problem, differential equation

Procedia PDF Downloads 429
12216 Autonomous Flight Performance Improvement of Load-Carrying Unmanned Aerial Vehicles by Active Morphing

Authors: Tugrul Oktay, Mehmet Konar, Mohamed Abdallah Mohamed, Murat Aydin, Firat Sal, Murat Onay, Mustafa Soylak

Abstract:

In this paper, it is aimed to improve autonomous flight performance of a load-carrying (payload: 3 kg and total: 6kg) unmanned aerial vehicle (UAV) through active wing and horizontal tail active morphing and also integrated autopilot system parameters (i.e. P, I, D gains) and UAV parameters (i.e. extension ratios of wing and horizontal tail during flight) design. For this purpose, a loadcarrying UAV (i.e. ZANKA-II) is manufactured in Erciyes University, College of Aviation, Model Aircraft Laboratory is benefited. Optimum values of UAV parameters and autopilot parameters are obtained using a stochastic optimization method. Using this approach autonomous flight performance of UAV is substantially improved and also in some adverse weather conditions an opportunity for safe flight is satisfied. Active morphing and integrated design approach gives confidence, high performance and easy-utility request of UAV users.

Keywords: unmanned aerial vehicles, morphing, autopilots, autonomous performance

Procedia PDF Downloads 656
12215 The Basics of Cognitive Behavioral Family Therapy and the Treatment of Various Physical and Mental Diseases

Authors: Mahta Mohamadkashi

Abstract:

The family is the most important source of security and health for the people of the society, and at the same time, it is the main field of creating all kinds of social and psychological problems. On the one hand, a family is a natural group with many goals and roles that are important and necessary for all family members. On the other hand, the family is a strong and organized group that recruits the therapist because of the goals that are concealed in its policy and procedures. The relationship between the environment and the family background with mental illnesses has been the focus of various researchers for a long time, and the research and experiments that have been conducted to show that the functioning of the family is related to the mental health of the members of the family. Currently, several theoretical perspectives with different approaches seek to explain and resolve psychological problems and family conflicts that can be mentioned. This research aims to investigate "cognitive-behavioral family therapy" by using the "family therapy" research method which is included the descriptive-analytical method and the method of collecting library information, with special reliance on Persian and Latin books and articles. for considering one of the important approaches of family therapy that we are going which have been known as data and its conditions that also includes requirements and limitations. For this purpose, in the beginning, brief background and introduction about family and family therapy are going to describe, and then the basics of cognitive-behavioral family therapy and the implementation process and various techniques of this approach can go through a big discussion. After that, we will apply this approach in the treatment of various physical and mental diseases in the form of related research, and we will examine the ups and downs of the implementation procedures, limitations, and future directions in this field. In general, This study emphasizes the role of the family system in the occurrence of psychological diseases and disorders and also validates the role of the family system in the treatment of those diseases and disorders. Also, cognitive-behavioral family therapy has been approved as an effective treatment approach for a variety of mental disorders.

Keywords: cognitive-behavioral, family, family therapy, cognitive-behavioral family therapy

Procedia PDF Downloads 81
12214 Image Transform Based on Integral Equation-Wavelet Approach

Authors: Yuan Yan Tang, Lina Yang, Hong Li

Abstract:

Harmonic model is a very important approximation for the image transform. The harmanic model converts an image into arbitrary shape; however, this mode cannot be described by any fixed functions in mathematics. In fact, it is represented by partial differential equation (PDE) with boundary conditions. Therefore, to develop an efficient method to solve such a PDE is extremely significant in the image transform. In this paper, a novel Integral Equation-Wavelet based method is presented, which consists of three steps: (1) The partial differential equation is converted into boundary integral equation and representation by an indirect method. (2) The boundary integral equation and representation are changed to plane integral equation and representation by boundary measure formula. (3) The plane integral equation and representation are then solved by a method we call wavelet collocation. Our approach has two main advantages, the shape of an image is arbitrary and the program code is independent of the boundary. The performance of our method is evaluated by numerical experiments.

Keywords: harmonic model, partial differential equation (PDE), integral equation, integral representation, boundary measure formula, wavelet collocation

Procedia PDF Downloads 536
12213 On Stochastic Models for Fine-Scale Rainfall Based on Doubly Stochastic Poisson Processes

Authors: Nadarajah I. Ramesh

Abstract:

Much of the research on stochastic point process models for rainfall has focused on Poisson cluster models constructed from either the Neyman-Scott or Bartlett-Lewis processes. The doubly stochastic Poisson process provides a rich class of point process models, especially for fine-scale rainfall modelling. This paper provides an account of recent development on this topic and presents the results based on some of the fine-scale rainfall models constructed from this class of stochastic point processes. Amongst the literature on stochastic models for rainfall, greater emphasis has been placed on modelling rainfall data recorded at hourly or daily aggregation levels. Stochastic models for sub-hourly rainfall are equally important, as there is a need to reproduce rainfall time series at fine temporal resolutions in some hydrological applications. For example, the study of climate change impacts on hydrology and water management initiatives requires the availability of data at fine temporal resolutions. One approach to generating such rainfall data relies on the combination of an hourly stochastic rainfall simulator, together with a disaggregator making use of downscaling techniques. Recent work on this topic adopted a different approach by developing specialist stochastic point process models for fine-scale rainfall aimed at generating synthetic precipitation time series directly from the proposed stochastic model. One strand of this approach focused on developing a class of doubly stochastic Poisson process (DSPP) models for fine-scale rainfall to analyse data collected in the form of rainfall bucket tip time series. In this context, the arrival pattern of rain gauge bucket tip times N(t) is viewed as a DSPP whose rate of occurrence varies according to an unobserved finite state irreducible Markov process X(t). Since the likelihood function of this process can be obtained, by conditioning on the underlying Markov process X(t), the models were fitted with maximum likelihood methods. The proposed models were applied directly to the raw data collected by tipping-bucket rain gauges, thus avoiding the need to convert tip-times to rainfall depths prior to fitting the models. One advantage of this approach was that the use of maximum likelihood methods enables a more straightforward estimation of parameter uncertainty and comparison of sub-models of interest. Another strand of this approach employed the DSPP model for the arrivals of rain cells and attached a pulse or a cluster of pulses to each rain cell. Different mechanisms for the pattern of the pulse process were used to construct variants of this model. We present the results of these models when they were fitted to hourly and sub-hourly rainfall data. The results of our analysis suggest that the proposed class of stochastic models is capable of reproducing the fine-scale structure of the rainfall process, and hence provides a useful tool in hydrological modelling.

Keywords: fine-scale rainfall, maximum likelihood, point process, stochastic model

Procedia PDF Downloads 261
12212 Toward Automatic Chest CT Image Segmentation

Authors: Angely Sim Jia Wun, Sasa Arsovski

Abstract:

Numerous studies have been conducted on the segmentation of medical images. Segmenting the lungs is one of the common research topics in those studies. Our research stemmed from the lack of solutions for automatic bone, airway, and vessel segmentation, despite the existence of multiple lung segmentation techniques. Consequently, currently, available software tools used for medical image segmentation do not provide automatic lung, bone, airway, and vessel segmentation. This paper presents segmentation techniques along with an interactive software tool architecture for segmenting bone, lung, airway, and vessel tissues. Additionally, we propose a method for creating binary masks from automatically generated segments. The key contribution of our approach is the technique for automatic image thresholding using adjustable Hounsfield values and binary mask extraction. Generated binary masks can be successfully used as a training dataset for deep-learning solutions in medical image segmentation. In this paper, we also examine the current software tools used for medical image segmentation, discuss our approach, and identify its advantages.

Keywords: lung segmentation, binary masks, U-Net, medical software tools

Procedia PDF Downloads 79
12211 A New Approach to the Boom Welding Technique by Determining Seam Profile Tracking

Authors: Muciz Özcan, Mustafa Sacid Endiz, Veysel Alver

Abstract:

In this paper we present a new approach to the boom welding related to the mobile cranes manufacturing, implementing a new method in order to get homogeneous welding quality and reduced energy usage during booms production. We aim to get the realization of the same welding quality carried out on the boom in every region during the manufacturing process and to detect the possible welding errors whether they could be eliminated using laser sensors. We determine the position of the welding region directly through our system and with the help of the welding oscillator we are able to perform a proper boom welding. Errors that may occur in the welding process can be observed by monitoring and eliminated by means of an operator. The major modification in the production of the crane booms will be their form of the booms. Although conventionally, more than one welding is required to perform this process, with the suggested concept, only one particular welding is sufficient, which will be more energy and environment-friendly. Consequently, as only one welding is needed for the manufacturing of the boom, the particular welding quality becomes more essential. As a way to satisfy the welding quality, a welding manipulator was made and fabricated. By using this welding manipulator, the risks of involving dangerous gases formed during the welding process for the operator and the surroundings are diminished as much as possible.

Keywords: boom welding, seam tracking, energy saving, global warming

Procedia PDF Downloads 332
12210 Designed Purine Molecules and in-silico Evaluation of Aurora Kinase Inhibition in Breast Cancer

Authors: Pooja Kumari, Anandkumar Tengli

Abstract:

Aurora kinase enzyme, a protein on overexpression, leads to metastasis and is extremely important for women’s health in terms of prevention or treatment. While creating a targeted technique, the aim of the work is to design purine molecules that inhibit in aurora kinase enzyme and helps to suppress breast cancer. Purine molecules attached to an amino acid in DNA block protein synthesis or halt the replication and metastasis caused by the aurora kinase enzyme. Various protein related to the overexpression of aurora protein was docked with purine molecule using Biovia Drug Discovery, the perpetual software. Various parameters like X-ray crystallographic structure, presence of ligand, Ramachandran plot, resolution, etc., were taken into consideration for selecting the target protein. A higher negative binding scored molecule has been taken for simulation studies. According to the available research and computational analyses, purine compounds may be powerful enough to demonstrate a greater affinity for the aurora target. Despite being clinically effective now, purines were originally meant to fight breast cancer by inhibiting the aurora kinase enzyme. In in-silico studies, it is observed that purine compounds have a moderate to high potency compared to other molecules, and our research into the literature revealed that purine molecules have a lower risk of side effects. The research involves the design, synthesis, and identification of active purine molecules against breast cancer. Purines are structurally similar to the normal metabolites of adenine and guanine; hence interfere/compete with protein synthesis and suppress the abnormal proliferation of cells/tissues. As a result, purine target metastasis cells and stop the growth of kinase; purine derivatives bind with DNA and aurora protein which may stop the growth of protein or inhibits replication and stop metastasis of overexpressed aurora kinase enzyme.

Keywords: aurora kinases, in silico studies, medicinal chemistry, combination therapies, chronic cancer, clinical translation

Procedia PDF Downloads 72
12209 Experimental Activity on the Photovoltaic Effect

Authors: Salomão Manuel Francisco, Manuel António Salgueiro Da Silva, Bento Filipe Barreiras Pinto Cavadas, Teresa Monteiro Seixas

Abstract:

In bachelor's degrees in Physics Education framework in Angola, and to a certain extent, within the community of Portuguese language countries (CPLP), teaching methodologies rely heavily on theoretical memorization and mathematical demonstrations. This approach often discourages students, particularly the female population, as the reliance on theoretical mathematical demonstrations generates the perception of Physics as an arduous, challenging discipline. To address this challenge and recognize the value of practical application as an evaluative criterion of material truth, we propose a practical activity in Environmental Physics that will be shared with Angolan higher education teachers, who will receive full scaffolding and support from the authors. These teachers, adopting and developing similar activities in a classroom setting, will contribute to the environmental education framework as well. Additionally, this work aligns with different goals of UNESCO's 2030 agenda, namely, specifically, goals 4, 5, 7, 11, 13, and 17. The experimental activity developed in this work is centered around the demonstration of the photovoltaic effect and its application for renewable energy production. The first objective of the activity is to study the variation of electrical power supplied by a photovoltaic system (PV) to an electrical circuit as the angle of light incidence changes. Students can observe that the power supplied to the circuit is greater when light rays fall perpendicularly on the PV. However, as the angle of incidence increases, resulting in a larger area covered by the light rays, the power supplied to the circuit decreases due to lower irradiance. The second objective is to demonstrate that the power output can be maximized by adjusting the circuit load resistance at each irradiance value. In these two parts of the activity, students can analyze experimental data taking into account the irradiance law and the equivalent circuit description of a PV cell. Through detailed data analysis, students are also expected to assess the effects of temperature on PV efficiency degradation and the efficiency enhancement provided by light concentration mechanisms. As a third objective, students can explore how the color of incident light affects the PV output power, considering the quantum nature of light and its interaction with the PV system.

Keywords: experiments, irradiation law, physic teaching, photovoltaic effect

Procedia PDF Downloads 68
12208 Vehicle Activity Characterization Approach to Quantify On-Road Mobile Source Emissions

Authors: Hatem Abou-Senna, Essam Radwan

Abstract:

Transportation agencies and researchers in the past have estimated emissions using one average speed and volume on a long stretch of roadway. Other methods provided better accuracy utilizing annual average estimates. Travel demand models provided an intermediate level of detail through average daily volumes. Currently, higher accuracy can be established utilizing microscopic analyses by splitting the network links into sub-links and utilizing second-by-second trajectories to calculate emissions. The need to accurately quantify transportation-related emissions from vehicles is essential. This paper presents an examination of four different approaches to capture the environmental impacts of vehicular operations on a 10-mile stretch of Interstate 4 (I-4), an urban limited access highway in Orlando, Florida. First, (at the most basic level), emissions were estimated for the entire 10-mile section 'by hand' using one average traffic volume and average speed. Then, three advanced levels of detail were studied using VISSIM/MOVES to analyze smaller links: average speeds and volumes (AVG), second-by-second link drive schedules (LDS), and second-by-second operating mode distributions (OPMODE). This paper analyzes how the various approaches affect predicted emissions of CO, NOx, PM2.5, PM10, and CO2. The results demonstrate that obtaining precise and comprehensive operating mode distributions on a second-by-second basis provides more accurate emission estimates. Specifically, emission rates are highly sensitive to stop-and-go traffic and the associated driving cycles of acceleration, deceleration, and idling. Using the AVG or LDS approach may overestimate or underestimate emissions, respectively, compared to an operating mode distribution approach.

Keywords: limited access highways, MOVES, operating mode distribution (OPMODE), transportation emissions, vehicle specific power (VSP)

Procedia PDF Downloads 326
12207 Selective Circular Dichroism Sensor Based on the Generation of Quantum Dots for Cadmium Ion Detection

Authors: Pradthana Sianglam, Wittaya Ngeontae

Abstract:

A new approach for the fabrication of cadmium ion (Cd2+) sensor is demonstrated. The detection principle is based on the in-situ generation of cadmium sulfide quantum dots (CdS QDs) in the presence of chiral thiol containing compound and detection by the circular dichroism spectroscopy (CD). Basically, the generation of CdS QDs can be done in the presence of Cd2+, sulfide ion and suitable capping compounds. In addition, the strong CD signal can be recorded if the generated QDs possess chiral property (from chiral capping molecule). Thus, the degree of CD signal change depends on the number of the generated CdS QDs which can be related to the concentration of Cd2+ (excess of other components). In this work, we use the mixture of cysteamine (Cys) and L-Penicillamine (LPA) as the capping molecules. The strong CD signal can be observed when the solution contains sodium sulfide, Cys, LPA, and Cd2+. Moreover, the CD signal is linearly related to the concentration of Cd2+. This approach shows excellence selectivity towards the detection of Cd2+ when comparing to other cation. The proposed CD sensor provides low limit detection limits around 70 µM and can be used with real water samples with satisfactory results.

Keywords: circular dichroism sensor, quantum dots, enaniomer, in-situ generation, chemical sensor, heavy metal ion

Procedia PDF Downloads 354
12206 Development of High Strength Filler Consumables by Means of Calculations and Microstructural Characterization

Authors: S. Holly, R. Schnitzer, P. Haslberger, D. Zügner

Abstract:

The development of new filler consumables necessitates a high effort regarding samples and experiments to achieve the required mechanical properties and chemistry. In the scope of the development of a metal-cored wire with the target tensile strength of 1150 MPa and acceptable impact toughness, thermodynamic and kinetic calculations via MatCalc were used to reduce the experimental work and the resources required. Micro alloying elements were used to reach the high strength as an alternative approach compared to the conventional solid solution hardening. In order to understand the influence of different micro alloying elements in more detail, the influence of different elements on the precipitation behavior in the weld metal was evaluated. Investigations of the microstructure were made via atom probe and EBSD to understand the effect of micro alloying elements. The calculated results are in accordance with the results obtained by experiments and can be explained by the microstructural investigations. On the example of aluminium, the approach is exemplified and clarifies the efficient way of development.

Keywords: alloy development, high strength steel, MatCalc, metal-cored wire

Procedia PDF Downloads 224
12205 Market Solvency Capital Requirement Minimization: How Non-linear Solvers Provide Portfolios Complying with Solvency II Regulation

Authors: Abraham Castellanos, Christophe Durville, Sophie Echenim

Abstract:

In this article, a portfolio optimization problem is performed in a Solvency II context: it illustrates how advanced optimization techniques can help to tackle complex operational pain points around the monitoring, control, and stability of Solvency Capital Requirement (SCR). The market SCR of a portfolio is calculated as a combination of SCR sub-modules. These sub-modules are the results of stress-tests on interest rate, equity, property, credit and FX factors, as well as concentration on counter-parties. The market SCR is non convex and non differentiable, which does not make it a natural optimization criteria candidate. In the SCR formulation, correlations between sub-modules are fixed, whereas risk-driven portfolio allocation is usually driven by the dynamics of the actual correlations. Implementing a portfolio construction approach that is efficient on both a regulatory and economic standpoint is not straightforward. Moreover, the challenge for insurance portfolio managers is not only to achieve a minimal SCR to reduce non-invested capital but also to ensure stability of the SCR. Some optimizations have already been performed in the literature, simplifying the standard formula into a quadratic function. But to our knowledge, it is the first time that the standard formula of the market SCR is used in an optimization problem. Two solvers are combined: a bundle algorithm for convex non- differentiable problems, and a BFGS (Broyden-Fletcher-Goldfarb- Shanno)-SQP (Sequential Quadratic Programming) algorithm, to cope with non-convex cases. A market SCR minimization is then performed with historical data. This approach results in significant reduction of the capital requirement, compared to a classical Markowitz approach based on the historical volatility. A comparative analysis of different optimization models (equi-risk-contribution portfolio, minimizing volatility portfolio and minimizing value-at-risk portfolio) is performed and the impact of these strategies on risk measures including market SCR and its sub-modules is evaluated. A lack of diversification of market SCR is observed, specially for equities. This was expected since the market SCR strongly penalizes this type of financial instrument. It was shown that this direct effect of the regulation can be attenuated by implementing constraints in the optimization process or minimizing the market SCR together with the historical volatility, proving the interest of having a portfolio construction approach that can incorporate such features. The present results are further explained by the Market SCR modelling.

Keywords: financial risk, numerical optimization, portfolio management, solvency capital requirement

Procedia PDF Downloads 105
12204 Fracture Energy Corresponding to the Puncture/Cutting of Nitrile Rubber by Pointed Blades

Authors: Ennouri Triki, Toan Vu-Khanh

Abstract:

Resistance to combined puncture/cutting by pointed blades is an important property of gloves materials. The purpose of this study is to propose an approach derived from the fracture mechanics theory to calculate the fracture energy associated to the puncture/cutting of nitrile rubber. The proposed approach is also based on the application of a sample pre-strained during the puncture/cutting test in order to remove the contribution of friction. It was validated with two different pointed blade angles of 22.5° and 35°. Results show that the applied total fracture energy corresponding to puncture/cutting is controlled by three energies, one is the fracture energy or the intrinsic strength of the material, the other reflects the friction energy between a pointed blade and the material. For an applied pre-strain energy (or tearing energy) of high value, the friction energy is completely removed. Without friction, the total fracture energy is constant. In that case, the fracture contribution of the tearing energy is marginal. Growth of the crack is thus completely caused by the puncture/cutting by a pointed blade. Finally, results suggest that the value of the fracture energy corresponding to puncture/cutting by pointed blades is obtained at a frictional contribution of zero.

Keywords: elastomer, energy, fracture, friction, pointed blades

Procedia PDF Downloads 286
12203 An Attempt to Improve Student´s Understanding on Thermal Conductivity Using Thermal Cameras

Authors: Mariana Faria Brito Francisquini

Abstract:

Many thermal phenomena are present and play a substantial role in our daily lives. This presence makes the study of this area at both High School and University levels a very widely explored topic in the literature. However, a lot of important concepts to a meaningful understanding of the world are neglected at the expense of a traditional approach with senseless algebraic problems. In this work, we intend to show how the introduction of new technologies in the classroom, namely thermal cameras, can work in our favor to make a clearer understanding of many of these concepts, such as thermal conductivity. The use of thermal cameras in the classroom tends to diminish the everlasting abstractness in thermal phenomena as they enable us to visualize something that happens right before our eyes, yet we cannot see it. In our study, we will provide the same amount of heat to metallic cylindrical rods of the same length, but different materials in order to study the thermal conductivity of each one. In this sense, the thermal camera allows us to visualize the increase in temperature along each rod in real time enabling us to infer how heat is being transferred from one part of the rod to another. Therefore, we intend to show how this approach can contribute to the exposure of students to more enriching, intellectually prolific, scenarios than those provided by traditional approaches.

Keywords: teaching physics, thermal cameras, thermal conductivity, thermal physics

Procedia PDF Downloads 264
12202 Machine Learning Model to Predict TB Bacteria-Resistant Drugs from TB Isolates

Authors: Rosa Tsegaye Aga, Xuan Jiang, Pavel Vazquez Faci, Siqing Liu, Simon Rayner, Endalkachew Alemu, Markos Abebe

Abstract:

Tuberculosis (TB) is a major cause of disease globally. In most cases, TB is treatable and curable, but only with the proper treatment. There is a time when drug-resistant TB occurs when bacteria become resistant to the drugs that are used to treat TB. Current strategies to identify drug-resistant TB bacteria are laboratory-based, and it takes a longer time to identify the drug-resistant bacteria and treat the patient accordingly. But machine learning (ML) and data science approaches can offer new approaches to the problem. In this study, we propose to develop an ML-based model to predict the antibiotic resistance phenotypes of TB isolates in minutes and give the right treatment to the patient immediately. The study has been using the whole genome sequence (WGS) of TB isolates as training data that have been extracted from the NCBI repository and contain different countries’ samples to build the ML models. The reason that different countries’ samples have been included is to generalize the large group of TB isolates from different regions in the world. This supports the model to train different behaviors of the TB bacteria and makes the model robust. The model training has been considering three pieces of information that have been extracted from the WGS data to train the model. These are all variants that have been found within the candidate genes (F1), predetermined resistance-associated variants (F2), and only resistance-associated gene information for the particular drug. Two major datasets have been constructed using these three information. F1 and F2 information have been considered as two independent datasets, and the third information is used as a class to label the two datasets. Five machine learning algorithms have been considered to train the model. These are Support Vector Machine (SVM), Random forest (RF), Logistic regression (LR), Gradient Boosting, and Ada boost algorithms. The models have been trained on the datasets F1, F2, and F1F2 that is the F1 and the F2 dataset merged. Additionally, an ensemble approach has been used to train the model. The ensemble approach has been considered to run F1 and F2 datasets on gradient boosting algorithm and use the output as one dataset that is called F1F2 ensemble dataset and train a model using this dataset on the five algorithms. As the experiment shows, the ensemble approach model that has been trained on the Gradient Boosting algorithm outperformed the rest of the models. In conclusion, this study suggests the ensemble approach, that is, the RF + Gradient boosting model, to predict the antibiotic resistance phenotypes of TB isolates by outperforming the rest of the models.

Keywords: machine learning, MTB, WGS, drug resistant TB

Procedia PDF Downloads 32