Search results for: sequential
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 428

Search results for: sequential

188 Pushing the Boundary of Parallel Tractability for Ontology Materialization via Boolean Circuits

Authors: Zhangquan Zhou, Guilin Qi

Abstract:

Materialization is an important reasoning service for applications built on the Web Ontology Language (OWL). To make materialization efficient in practice, current research focuses on deciding tractability of an ontology language and designing parallel reasoning algorithms. However, some well-known large-scale ontologies, such as YAGO, have been shown to have good performance for parallel reasoning, but they are expressed in ontology languages that are not parallelly tractable, i.e., the reasoning is inherently sequential in the worst case. This motivates us to study the problem of parallel tractability of ontology materialization from a theoretical perspective. That is we aim to identify the ontologies for which materialization is parallelly tractable, i.e., in the NC complexity. Since the NC complexity is defined based on Boolean circuit that is widely used to investigate parallel computing problems, we first transform the problem of materialization to evaluation of Boolean circuits, and then study the problem of parallel tractability based on circuits. In this work, we focus on datalog rewritable ontology languages. We use Boolean circuits to identify two classes of datalog rewritable ontologies (called parallelly tractable classes) such that materialization over them is parallelly tractable. We further investigate the parallel tractability of materialization of a datalog rewritable OWL fragment DHL (Description Horn Logic). Based on the above results, we analyze real-world datasets and show that many ontologies expressed in DHL belong to the parallelly tractable classes.

Keywords: ontology materialization, parallel reasoning, datalog, Boolean circuit

Procedia PDF Downloads 245
187 Interval Bilevel Linear Fractional Programming

Authors: F. Hamidi, N. Amiri, H. Mishmast Nehi

Abstract:

The Bilevel Programming (BP) model has been presented for a decision making process that consists of two decision makers in a hierarchical structure. In fact, BP is a model for a static two person game (the leader player in the upper level and the follower player in the lower level) wherein each player tries to optimize his/her personal objective function under dependent constraints; this game is sequential and non-cooperative. The decision making variables are divided between the two players and one’s choice affects the other’s benefit and choices. In other words, BP consists of two nested optimization problems with two objective functions (upper and lower) where the constraint region of the upper level problem is implicitly determined by the lower level problem. In real cases, the coefficients of an optimization problem may not be precise, i.e. they may be interval. In this paper we develop an algorithm for solving interval bilevel linear fractional programming problems. That is to say, bilevel problems in which both objective functions are linear fractional, the coefficients are interval and the common constraint region is a polyhedron. From the original problem, the best and the worst bilevel linear fractional problems have been derived and then, using the extended Charnes and Cooper transformation, each fractional problem can be reduced to a linear problem. Then we can find the best and the worst optimal values of the leader objective function by two algorithms.

Keywords: best and worst optimal solutions, bilevel programming, fractional, interval coefficients

Procedia PDF Downloads 420
186 Airbnb, Hotel Industry and Optimum Strategies: Evidence from European Cities, Barcelona, London and Paris

Authors: Juan Pedro Aznar Alarcon, Josep Maria Sayeras Maspera

Abstract:

Airbnb and other similar platforms are offering a near substitute to the traditional accommodation service supplied by the hotel sector. In this context, hotels can try to compete by offering higher quality and additional services, which imply the need for new investments or try to compete by reducing prices. The theoretical model presented in this paper analyzes the best response using a sequential game theory model. The main conclusion is that due to the financial constraints that small and medium hotels have these hotels have reduced prices whereas hotels that belong to international groups or have an easy access to financial resources have increased their investment to increase the quality of the service provided. To check the validity of the theoretical model financial data from Barcelona, London and Paris hotels have been used analyzing profitability, quality of the service provided, the investment propensity and the evolution of the gross profit. The model and the empirical data provide the base for some industrial policy in the hospitality industry. To address the extra cost that small hotels in Europe have to face compared by bigger firms would help to improve the level of quality provided and to some extent have positive externalities in terms of job creation and an increasing added value for the industry.

Keywords: Airbnb, profitability, hospitality industry, game theory

Procedia PDF Downloads 321
185 Optimization of Technical and Technological Solutions for the Development of Offshore Hydrocarbon Fields in the Kaliningrad Region

Authors: Pavel Shcherban, Viktoria Ivanova, Alexander Neprokin, Vladislav Golovanov

Abstract:

Currently, LLC «Lukoil-Kaliningradmorneft» is implementing a comprehensive program for the development of offshore fields of the Kaliningrad region. This is largely associated with the depletion of the resource base of land in the region, as well as the positive results of geological investigation surrounding the Baltic Sea area and the data on the volume of hydrocarbon recovery from a single offshore field are working on the Kaliningrad region – D-6 «Kravtsovskoye».The article analyzes the main stages of the LLC «Lukoil-Kaliningradmorneft»’s development program for the development of the hydrocarbon resources of the region's shelf and suggests an optimization algorithm that allows managing a multi-criteria process of development of shelf deposits. The algorithm is formed on the basis of the problem of sequential decision making, which is a section of dynamic programming. Application of the algorithm during the consolidation of the initial data, the elaboration of project documentation, the further exploration and development of offshore fields will allow to optimize the complex of technical and technological solutions and increase the economic efficiency of the field development project implemented by LLC «Lukoil-Kaliningradmorneft».

Keywords: offshore fields of hydrocarbons of the Baltic Sea, development of offshore oil and gas fields, optimization of the field development scheme, solution of multicriteria tasks in oil and gas complex, quality management in oil and gas complex

Procedia PDF Downloads 177
184 Investigation of Cytotoxic Compounds in Ethyl Acetate and Chloroform Extracts of Nigella sativa Seeds by Sulforhodamine-B Assay-Guided Fractionation

Authors: Harshani Uggallage, Kapila D. Dissanayaka

Abstract:

A Sulforhodamine-B assay-guided fractionation on Nigella sativa seeds was conducted to determine the presence of cytotoxic compounds against human hepatoma (HepG2) cells. Initially, a freeze-dried sample of Nigella sativa seeds was sequentially extracted into solvents of increasing polarities. Crude extracts from the sequential extraction of Nigella sativa seeds in chloroform and ethyl acetate showed the highest cytotoxicity. The combined mixture of these two extracts was subjected to bioassay guided fractionation using a modified Kupchan method of partitioning, followed by Sephadex® LH-20 chromatography. This chromatographic separation process resulted in a column fraction with a convincing IC50 (half-maximal inhibitory concentration) value of 13.07µg/ml, which is considerable for developing therapeutic drug leads against human hepatoma. Reversed phase High-Performance Liquid Chromatography (HPLC) was finally conducted for the same column fraction, and the result indicates the presence of one or several main cytotoxic compounds against human HepG2 cells.

Keywords: cytotoxic compounds, half-maximal inhibitory concentration, high-performance liquid chromatography, human HepG2 cells, nigella sativa seeds, Sulforhodamine-B assay

Procedia PDF Downloads 365
183 Glycan Analyzer: Software to Annotate Glycan Structures from Exoglycosidase Experiments

Authors: Ian Walsh, Terry Nguyen-Khuong, Christopher H. Taron, Pauline M. Rudd

Abstract:

Glycoproteins and their covalently bonded glycans play critical roles in the immune system, cell communication, disease and disease prognosis. Ultra performance liquid chromatography (UPLC) coupled with mass spectrometry is conventionally used to qualitatively and quantitatively characterise glycan structures in a given sample. Exoglycosidases are enzymes that catalyze sequential removal of monosaccharides from the non-reducing end of glycans. They naturally have specificity for a particular type of sugar, its stereochemistry (α or β anomer) and its position of attachment to an adjacent sugar on the glycan. Thus, monitoring the peak movements (both in the UPLC and MS1) after application of exoglycosidases provides a unique and effective way to annotate sugars with high detail - i.e. differentiating positional and linkage isomers. Manual annotation of an exoglycosidase experiment is difficult and time consuming. As such, with increasing sample complexity and the number of exoglycosidases, the analysis could result in manually interpreting hundreds of peak movements. Recently, we have implemented pattern recognition software for automated interpretation of UPLC-MS1 exoglycosidase digestions. In this work, we explain the software, indicate how much time it will save and provide example usage showing the annotation of positional and linkage isomers in Immunoglobulin G, apolipoprotein J, and simple glycan standards.

Keywords: bioinformatics, automated glycan assignment, liquid chromatography, mass spectrometry

Procedia PDF Downloads 174
182 Character and Evolution of Electronic Waste: A Technologically Developing Country's Experience

Authors: Karen C. Olufokunbi, Odetunji A. Odejobi

Abstract:

The discourse of this paper is the examination of the generation, accumulation and growth of e-waste in a developing country. Images and other data about computer e-waste were collected using a digital camera, 290 copies of questionnaire and three structured interviews using Obafemi Awolowo University (OAU), Ile-Ife, Nigeria environment as a case study. The numerical data were analysed using R data analysis and process tool. Automata-based techniques and Petri net modeling tool were used to design and simulate a computational model for the recovery of saleable materials from e-waste. The R analysis showed that at a 95 percent confidence level, the computer equipment that will be disposed by 2020 will be 417 units. Compared to the 800 units in circulation in 2014, 50 percent of personal computer components will become e-waste. This indicates that personal computer components were in high demand due to their low costs and will be disposed more rapidly when replaced by new computer equipment Also, 57 percent of the respondents discarded their computer e-waste by throwing it into the garbage bin or by dumping it. The simulated model using Coloured Petri net modelling tool for the process showed that the e-waste dynamics is a forward sequential process in the form of a pipeline meaning that an e-waste recovery of saleable materials process occurs in identifiable discrete stages indicating that e-waste will continue to accumulate and grow in volume with time.

Keywords: Coloured Petri net, computational modelling, electronic waste, electronic waste process dynamics

Procedia PDF Downloads 141
181 Cognitive Benefits of Being Bilingual: The Effect of Language Learning on the Working Memory in Emerging Miao-Mandarin Juveniles in Rural Regions of China

Authors: Peien Ma

Abstract:

Bilingual effect/advantage theorized the positive effect of being bilingual on general cognitive abilities, but it was unknown which factors tend to modulate these bilingualism effects on working memory capacity. This study imposed empirical field research on a group of low-SES emerging bilinguals, Miao people, in the hill tribes of rural China to investigate whether bilingualism affected their verbal working memory performance. 20 Miao-Chinese bilinguals (13 girls and 7 boys with a mean age of 11.45, SD=1.67) and 20 Chinese monolingual peers (13 girls and 7 boys with a mean age of 11.6, SD=0.68) were recruited. These bilingual and monolingual juveniles, matched on age, sex, socioeconomic status, and educational status, completed a language background questionnaire and a standard forward and backward digit span test adapted from Wechsler Adult Intelligence Scale-Revised (WAIS-R). The results showed that bilinguals earned a significantly higher overall mean score of the task, suggesting the superiority of working memory ability over the monolinguals. And bilingual cognitive benefits were independent of proficiency levels in learners’ two languages. The results suggested that bilingualism enhances working memory in sequential bilinguals from low SES backgrounds and shed light on our understanding of the bilingual advantage from a psychological and social perspective.

Keywords: bilingual effects, heritage language, Miao/Hmong language Mandarin, working memory

Procedia PDF Downloads 124
180 Julia-Based Computational Tool for Composite System Reliability Assessment

Authors: Josif Figueroa, Kush Bubbar, Greg Young-Morris

Abstract:

The reliability evaluation of composite generation and bulk transmission systems is crucial for ensuring a reliable supply of electrical energy to significant system load points. However, evaluating adequacy indices using probabilistic methods like sequential Monte Carlo Simulation can be computationally expensive. Despite this, it is necessary when time-varying and interdependent resources, such as renewables and energy storage systems, are involved. Recent advances in solving power network optimization problems and parallel computing have improved runtime performance while maintaining solution accuracy. This work introduces CompositeSystems, an open-source Composite System Reliability Evaluation tool developed in Julia™, to address the current deficiencies of commercial and non-commercial tools. This work introduces its design, validation, and effectiveness, which includes analyzing two different formulations of the Optimal Power Flow problem. The simulations demonstrate excellent agreement with existing published studies while improving replicability and reproducibility. Overall, the proposed tool can provide valuable insights into the performance of transmission systems, making it an important addition to the existing toolbox for power system planning.

Keywords: open-source software, composite system reliability, optimization methods, Monte Carlo methods, optimal power flow

Procedia PDF Downloads 47
179 Modeling Stream Flow with Prediction Uncertainty by Using SWAT Hydrologic and RBNN Neural Network Models for Agricultural Watershed in India

Authors: Ajai Singh

Abstract:

Simulation of hydrological processes at the watershed outlet through modelling approach is essential for proper planning and implementation of appropriate soil conservation measures in Damodar Barakar catchment, Hazaribagh, India where soil erosion is a dominant problem. This study quantifies the parametric uncertainty involved in simulation of stream flow using Soil and Water Assessment Tool (SWAT), a watershed scale model and Radial Basis Neural Network (RBNN), an artificial neural network model. Both the models were calibrated and validated based on measured stream flow and quantification of the uncertainty in SWAT model output was assessed using ‘‘Sequential Uncertainty Fitting Algorithm’’ (SUFI-2). Though both the model predicted satisfactorily, but RBNN model performed better than SWAT with R2 and NSE values of 0.92 and 0.92 during training, and 0.71 and 0.70 during validation period, respectively. Comparison of the results of the two models also indicates a wider prediction interval for the results of the SWAT model. The values of P-factor related to each model shows that the percentage of observed stream flow values bracketed by the 95PPU in the RBNN model as 91% is higher than the P-factor in SWAT as 87%. In other words the RBNN model estimates the stream flow values more accurately and with less uncertainty. It could be stated that RBNN model based on simple input could be used for estimation of monthly stream flow, missing data, and testing the accuracy and performance of other models.

Keywords: SWAT, RBNN, SUFI 2, bootstrap technique, stream flow, simulation

Procedia PDF Downloads 331
178 Flashsonar or Echolocation Education: Expanding the Function of Hearing and Changing the Meaning of Blindness

Authors: Thomas, Daniel Tajo, Kish

Abstract:

Sight is primarily associated with the function of gathering and processing near and extended spatial information which is largely used to support self-determined interaction with the environment through self-directed movement and navigation. By contrast, hearing is primarily associated with the function of gathering and processing sequential information which may typically be used to support self-determined communication through the self-directed use of music and language. Blindness or the lack of vision is traditionally characterized by a lack of capacity to access spatial information which, in turn, is presumed to result in a lack of capacity for self-determined interaction with the environment due to limitations in self-directed movement and navigation. However, through a specific protocol of FlashSonar education developed by World Access for the Blind, the function of hearing can be expanded in blind people to carry out some of the functions normally associated with sight, that is to access and process near and extended spatial information to construct three-dimensional acoustic images of the environment. This perceptual education protocol results in a significant restoration in blind people of self-determined environmental interaction, movement, and navigational capacities normally attributed to vision - a new way to see. Thus, by expanding the function of hearing to process spatial information to restore self-determined movement, we are not only changing the meaning of blindness, and what it means to be blind, but we are also recasting the meaning of vision and what it is to see.

Keywords: echolocation, changing, sensory, function

Procedia PDF Downloads 130
177 Technology in the Calculation of People Health Level: Design of a Computational Tool

Authors: Sara Herrero Jaén, José María Santamaría García, María Lourdes Jiménez Rodríguez, Jorge Luis Gómez González, Adriana Cercas Duque, Alexandra González Aguna

Abstract:

Background: Health concept has evolved throughout history. The health level is determined by the own individual perception. It is a dynamic process over time so that you can see variations from one moment to the next. In this way, knowing the health of the patients you care for, will facilitate decision making in the treatment of care. Objective: To design a technological tool that calculates the people health level in a sequential way over time. Material and Methods: Deductive methodology through text analysis, extraction and logical knowledge formalization and education with expert group. Studying time: September 2015- actually. Results: A computational tool for the use of health personnel has been designed. It has 11 variables. Each variable can be given a value from 1 to 5, with 1 being the minimum value and 5 being the maximum value. By adding the result of the 11 variables we obtain a magnitude in a certain time, the health level of the person. The health calculator allows to represent people health level at a time, establishing temporal cuts being useful to determine the evolution of the individual over time. Conclusion: The Information and Communication Technologies (ICT) allow training and help in various disciplinary areas. It is important to highlight their relevance in the field of health. Based on the health formalization, care acts can be directed towards some of the propositional elements of the concept above. The care acts will modify the people health level. The health calculator allows the prioritization and prediction of different strategies of health care in hospital units.

Keywords: calculator, care, eHealth, health

Procedia PDF Downloads 237
176 An Exemption for Vertical Restraint Regarding Intellectual Property Licensing: Case Study of Thailand

Authors: Sanpetchuda Krutkrua, Suphawatchara Malanond

Abstract:

Throughout the history of Antitrust regimes in Thailand, Thailand has been trying to prevent collusive practices in the market through the amendments of the Trade Competition Act, and Thailand just passed the current Trade Competition Act of B.E. 2560 in 2017 of which several aspects of the law were amended in order to enhance the prevention of collusive outcome through both vertical trade restraints and horizontal trade restraints. An agreement is vertical when it involves arrangements that are in a complementary relationship. In Section 55 of the Act, any agreements to reduce the price, quantity, or quality of the goods, agreements to assign a sole retailer for the goods, and the agreement to impose conditions on the retailers are not allowed. However, Section 56 provides exemptions for the vertical relationship between the business operators, the franchise agreement, and the licensing agreement as long as such agreements do not surpass the necessity to do so, create monopolization, or affect the consumers in terms of price, quality, quantity, or options. The paper aims to explore the extent of the exemption under Section 56 and sequential regulations in terms of the vertical trade restraints regarding intellectual property licensing, and, at the same time, compare with the exemptions under the European Union competition law, and Singapore competition law. Comparative legal analysis with leading jurisdiction will illustrate the application of the newly enacted Thai Competition Act in terms of its enforcement in the global impact of IP rights, which, by nature are de jure or de facto international protection.

Keywords: antitrust, competition law, vertical restraint, intellectual property, IP licensing

Procedia PDF Downloads 134
175 Investigation of Crack Formation in Ordinary Reinforced Concrete Beams and in Beams Strengthened with Carbon Fiber Sheet: Theory and Experiment

Authors: Anton A. Bykov, Irina O. Glot, Igor N. Shardakov, Alexey P. Shestakov

Abstract:

This paper presents the results of experimental and theoretical investigations of the mechanisms of crack formation in reinforced concrete beams subjected to quasi-static bending. The boundary-value problem has been formulated in the framework of brittle fracture mechanics and has been solved by using the finite-element method. Numerical simulation of the vibrations of an uncracked beam and a beam with cracks of different size serves to determine the pattern of changes in the spectrum of eigenfrequencies observed during crack evolution. Experiments were performed on the sequential quasistatic four-point bending of the beam leading to the formation of cracks in concrete. At each loading stage, the beam was subjected to an impulse load to induce vibrations. Two stages of cracking were detected. At the first stage the conservative process of deformation is realized. The second stage is an active cracking, which is marked by a sharp change in eingenfrequencies. The boundary of a transition from one stage to another is well registered. The vibration behavior was examined for the beams strengthened by carbon-fiber sheet before loading and at the intermediate stage of loading after the grouting of initial cracks. The obtained results show that the vibrodiagnostic approach is an effective tool for monitoring of cracking and for assessing the quality of measures aimed at strengthening concrete structures.

Keywords: crack formation, experiment, mathematical modeling, reinforced concrete, vibrodiagnostics

Procedia PDF Downloads 274
174 To Estimate the Association between Visual Stress and Visual Perceptual Skills

Authors: Vijay Reena Durai, Krithica Srinivasan

Abstract:

Introduction: The two fundamental skills involved in the growth and wellbeing of any child can be categorized into visual motor and perceptual skills. Visual stress is a disorder which is characterized by visual discomfort, blurred vision, misspelling words, skipping lines, letters bunching together. There is a need to understand the deficits in perceptual skills among children with visual stress. Aim: To estimate the association between visual stress and visual perceptual skills Objective: To compare visual perceptual skills of children with and without visual stress Methodology: Children between 8 to 15 years of age participated in this cross-sectional study. All children with monocular visual acuity better than or equal to 6/6 were included. Visual perceptual skills were measured using test for visual perceptual skills (TVPS) tool. Reading speed was measured with the chosen colored overlay using Wilkins reading chart and pattern glare score was estimated using a 3cpd gratings. Visual stress was defined as change in reading speed of greater than or equal to 10% and a pattern glare score of greater than or equal to 4. Results: 252 children participated in this study and the male: female ratio of 3:2. Majority of the children preferred Magenta (28%) and Yellow (25%) colored overlay for reading. There was a significant difference between the two groups (MD=1.24±0.6) (p<0.04, 95% CI 0.01-2.43) only in the sequential memory skills. The prevalence of visual stress in this group was found to be 31% (n=78). Binary logistic regression showed that odds ratio of having poor visual perceptual skills was OR: 2.85 (95% CI 1.08-7.49) among children with visual stress. Conclusion: Children with visual stress are found to have three times poorer visual perceptual skills than children without visual stress.

Keywords: visual stress, visual perceptual skills, colored overlay, pattern glare

Procedia PDF Downloads 357
173 Partial Privatization, Control Rights of Large Shareholders and Privatized Shares Transfer: Evidence from Chinese State-Owned Listed Companies

Authors: Tingting Zhou

Abstract:

The partial privatization of state-owned enterprises (SOEs) is a dynamic process. The main features of this process lie in not only gradual and sequential privatizations, but also privatized shares transfer. For partially privatized SOEs, the introduction of private sector ownership is not the end of the story because the previously introduced private owners may choose to leave the SOEs by transferring the privatized shares after privatization, a process that is called “privatized shares transfer”. This paper investigates the determinants of privatized shares transfer from the perspective of large shareholders’ control rights. The results captures the fact that the higher control rights of large shareholders lead to more privatized shares transfer. After exploring the impacts of excessive control rights, the results provide evidence supporting the idea that firms with excessive numbers of directors, senior managers or supervisors who also have positions in the largest controlling shareholder’s entity are more likely to transfer privatized shares owned by private owners. In addition, the largest shareholders’ ownership also plays a role in privatized shares transfer. This evidence suggests that the large shareholders’ control rights should be limited to an appropriate range during the process of privatization, thereby giving private shareholders more opportunity to participate in the operation of firms, strengthen the state and enhance the competitiveness of state capital.

Keywords: control rights of large shareholders, partial privatization, privatized shares transfer, state-owned listed companies

Procedia PDF Downloads 264
172 Bringing the Confidence Intervals into Choropleth Mortality Map: An Example of Tainan, Taiwan

Authors: Tzu-Jung Tseng, Pei-Hsuen Han, Tsung-Hsueh Lu

Abstract:

Background: Choropleth mortality map is commonly used to identify areas with higher mortality risk. However, the use of choropleth map alone might result in the misinterpretation of differences in mortality rates between areas. Two areas with different color shades might not actually have a significant difference in mortality rates. The mortality rates estimated for an area with a small population would be less stable. We suggest of bringing the 95% confidence intervals (CI) into the choropleth mortality map to help users interpret the areal mortality rate difference more properly. Method: In the first choropleth mortality map, we used only three color to indicate standardized mortality ratio (SMR) for each district in Tainan, Taiwan. The red color denotes that the SMR of that district was significantly higher than the Tainan average; on the contrary, the green color suggests that the SMR of that district was significantly lower than the Tainan average. The yellow color indicates that the SMR of that district was not statistically significantly different from the Tainan average. In the second choropleth mortality map, we used traditional sequential color scheme (color ramp) for different SMR in 37 districts in Tainan City with bar chart of each SMR with 95% CI in which the users could examine if the line of 95% CI of SMR of two districts overlapped (nonsignificant difference). Results: The all-causes SMR of each district in Tainan for 2008 to 2013 ranged from 0.77 (95% CI 0.75 to 0.80) in East District to 1.39 Beimen (95% CI 1.25 to 1.52). In the first choropleth mortality map, only 16 of 37 districts had red color and 8 districts had green color. For different causes of death, the number of districts with red color differed. In the first choropleth mortality map we added a bar chart with line of 95% CI of SMR in each district, in which the users could visualize the SMR differences between districts. Conclusion: Through the use of 95% CI the users could interpret the aral mortality differences more properly.

Keywords: choropleth map, small area variation, standardized mortality ratio (SMR), Taiwan

Procedia PDF Downloads 294
171 Evaluation of DNA Oxidation and Chemical DNA Damage Using Electrochemiluminescent Enzyme/DNA Microfluidic Array

Authors: Itti Bist, Snehasis Bhakta, Di Jiang, Tia E. Keyes, Aaron Martin, Robert J. Forster, James F. Rusling

Abstract:

DNA damage from metabolites of lipophilic drugs and pollutants, generated by enzymes, represents a major toxicity pathway in humans. These metabolites can react with DNA to form either 8-oxo-7,8-dihydro-2-deoxyguanosine (8-oxodG), which is the oxidative product of DNA or covalent DNA adducts, both of which are genotoxic and hence considered important biomarkers to detect cancer in humans. Therefore, detecting reactions of metabolites with DNA is an effective approach for the safety assessment of new chemicals and drugs. Here we describe a novel electrochemiluminescent (ECL) sensor array which can detect DNA oxidation and chemical DNA damage in a single array, facilitating a more accurate diagnostic tool for genotoxicity screening. Layer-by-layer assembly of DNA and enzyme are assembled on the pyrolytic graphite array which is housed in a microfluidic device for sequential detection of two type of the DNA damages. Multiple enzyme reactions are run on test compounds using the array, generating toxic metabolites in situ. These metabolites react with DNA in the films to cause DNA oxidation and chemical DNA damage which are detected by ECL generating osmium compound and ruthenium polymer, respectively. The method is further validated by the formation of 8-oxodG and DNA adduct using similar films of DNA/enzyme on magnetic bead biocolloid reactors, hydrolyzing the DNA, and analyzing by liquid chromatography-mass spectrometry (LC-MS). Hence, this combined DNA/enzyme array/LC-MS approach can efficiently explore metabolic genotoxic pathways for drugs and environmental chemicals.

Keywords: biosensor, electrochemiluminescence, DNA damage, microfluidic array

Procedia PDF Downloads 339
170 Hand Gesture Recognition for Sign Language: A New Higher Order Fuzzy HMM Approach

Authors: Saad M. Darwish, Magda M. Madbouly, Murad B. Khorsheed

Abstract:

Sign Languages (SL) are the most accomplished forms of gestural communication. Therefore, their automatic analysis is a real challenge, which is interestingly implied to their lexical and syntactic organization levels. Hidden Markov models (HMM’s) have been used prominently and successfully in speech recognition and, more recently, in handwriting recognition. Consequently, they seem ideal for visual recognition of complex, structured hand gestures such as are found in sign language. In this paper, several results concerning static hand gesture recognition using an algorithm based on Type-2 Fuzzy HMM (T2FHMM) are presented. The features used as observables in the training as well as in the recognition phases are based on Singular Value Decomposition (SVD). SVD is an extension of Eigen decomposition to suit non-square matrices to reduce multi attribute hand gesture data to feature vectors. SVD optimally exposes the geometric structure of a matrix. In our approach, we replace the basic HMM arithmetic operators by some adequate Type-2 fuzzy operators that permits us to relax the additive constraint of probability measures. Therefore, T2FHMMs are able to handle both random and fuzzy uncertainties existing universally in the sequential data. Experimental results show that T2FHMMs can effectively handle noise and dialect uncertainties in hand signals besides a better classification performance than the classical HMMs. The recognition rate of the proposed system is 100% for uniform hand images and 86.21% for cluttered hand images.

Keywords: hand gesture recognition, hand detection, type-2 fuzzy logic, hidden Markov Model

Procedia PDF Downloads 435
169 Academic Achievement Differences in Grandiose and Vulnerable Narcissists and the Mediating Effects of Self-Esteem and Self-Efficacy

Authors: Amber Dummett, Efstathia Tzemou

Abstract:

Narcissism is a personality trait characterized by selfishness, entitlement, and superiority. Narcissism is split into two subtypes, grandiose narcissism (GN) and vulnerable narcissism (VN). Grandiose narcissists are extraverted and arrogant, while vulnerable narcissists are introverted and insecure. This study investigates the psychological mechanisms that lead to differences in academic achievement (AA) between grandiose and vulnerable narcissists, specifically the mediating effects of self-esteem and self-efficacy. While narcissism is considered to be a negative trait, one of the Dark Triads, GN, has been found to have some benefits; therefore, this study considers if better AA is one of them. Moreover, further research into VN is essential to fully compare and contrast it with GN. We hypothesize that grandiose narcissists achieve higher marks due to having high self-esteem and self-efficacy. In comparison, we hypothesize that vulnerable narcissists underperform and achieve lower marks due to having low self-esteem and self-efficacy. Two online surveys were distributed to undergraduate university students. The first was a collection of scales measuring the mentioned dimensions and semester one AA, and the second investigated end of year AA. Sequential mediation analyses were conducted using the gathered data. Our analysis shows that neither self-esteem nor self-efficacy mediates the relationship between GN and AA. GN positively predicts self-esteem but has no relationship with self-efficacy. Self-esteem does not mediate the relationship between VN and AA. VN has a negative indirect effect on AA via self-efficacy, and VN negatively predicts self-esteem. Self-efficacy positively predicts AA. GN does not affect AA through the mediation of self-esteem and then self-efficacy, and neither does VN in this way. Overall, having grandiose or vulnerable narcissistic traits does not affect students’ AA. However, being highly efficacious does lead to academic success; therefore, universities should employ methods to improve the self-efficacy of their students.

Keywords: academic achievement, grandiose narcissism, self-efficacy, self-esteem, vulnerable narcissism

Procedia PDF Downloads 154
168 Assessing the Perceptions toward the Impacts of Tourism in Poverty Alleviation: A Basis for Pro-Poor Tourism Policy in Sta. Lucia, Guimba, Nueva Ecija

Authors: Lady Salvador Purganan, Jojo M. Villamin, Noel L. Lansang

Abstract:

Tourism is a multifaceted but interdependent industry. This industry is composed of four major players, the public sector, the private sector, the local community, and the tourists. Each player has a vital role in the success of delivering high-quality tourism products and activities. There are various manifestations of positive economic outcomes that benefit the local community. Pro-poor tourism development approach has a great potential to serve as an avenue for capacity building leading to economic independence since natural attractions and cultural resources are assets that can be capitalized on, especially by the poor, because it is more accessible to them compared to financial resources. In the National Tourism Development Plan 2016-2022, specific mechanisms are not reflected to combat and lower poverty incidence through tourism. The researcher used the multidimensional poverty theory and sustainable tourism theory to formulate indicators in the research instrument and social exchange theory. The expected output of the study is to unlock opportunities, specifically in Brgy. Sta. Lucia, Guimba, Nueva Ecija, by crafting policies taking into utmost consideration local community involvement and participation in the process of tourism development which is essential in attaining inclusive growth and sustainability. This study will apply the sequential explanatory design mixed-method approach.

Keywords: pro-poor tourism, poverty alleviation, livelihood opportunities, tourism development plan

Procedia PDF Downloads 72
167 Text Emotion Recognition by Multi-Head Attention based Bidirectional LSTM Utilizing Multi-Level Classification

Authors: Vishwanath Pethri Kamath, Jayantha Gowda Sarapanahalli, Vishal Mishra, Siddhesh Balwant Bandgar

Abstract:

Recognition of emotional information is essential in any form of communication. Growing HCI (Human-Computer Interaction) in recent times indicates the importance of understanding of emotions expressed and becomes crucial for improving the system or the interaction itself. In this research work, textual data for emotion recognition is used. The text being the least expressive amongst the multimodal resources poses various challenges such as contextual information and also sequential nature of the language construction. In this research work, the proposal is made for a neural architecture to resolve not less than 8 emotions from textual data sources derived from multiple datasets using google pre-trained word2vec word embeddings and a Multi-head attention-based bidirectional LSTM model with a one-vs-all Multi-Level Classification. The emotions targeted in this research are Anger, Disgust, Fear, Guilt, Joy, Sadness, Shame, and Surprise. Textual data from multiple datasets were used for this research work such as ISEAR, Go Emotions, Affect datasets for creating the emotions’ dataset. Data samples overlap or conflicts were considered with careful preprocessing. Our results show a significant improvement with the modeling architecture and as good as 10 points improvement in recognizing some emotions.

Keywords: text emotion recognition, bidirectional LSTM, multi-head attention, multi-level classification, google word2vec word embeddings

Procedia PDF Downloads 150
166 Developing a Culturally Acceptable End of Life Survey (the VOICES-ESRD/Thai Questionnaire) for Evaluation Health Services Provision of Older Persons with End-Stage Renal Disease (ESRD) in Thailand

Authors: W. Pungchompoo, A. Richardson, L. Brindle

Abstract:

Background: The developing of a culturally acceptable end of life survey (the VOICES-ESRD/Thai questionnaire) is an essential instrument for evaluation health services provision of older persons with ESRD in Thailand. The focus of the questionnaire was on symptoms, symptom control and the health care needs of older people with ESRD who are managed without dialysis. Objective: The objective of this study was to develop and adapt VOICES to make it suitable for use in a population survey in Thailand. Methods: The mixed methods exploratory sequential design was focussed on modifying an instrument. Data collection: A cognitive interviewing technique was implemented, using two cycles of data collection with a sample of 10 bereaved carers and a prototype of the Thai VOICES questionnaire. Qualitative study was used to modify the developing a culturally acceptable end of life survey (the VOICES-ESRD/Thai questionnaire). Data analysis: The data were analysed by using content analysis. Results: The revisions to the prototype questionnaire were made. The results were used to adapt the VOICES questionnaire for use in a population-based survey with older ESRD patients in Thailand. Conclusions: A culturally specific questionnaire was generated during this second phase and issues with questionnaire design were rectified.

Keywords: VOICES-ESRD/Thai questionnaire, cognitive interviewing, end of life survey, health services provision, older persons with ESRD

Procedia PDF Downloads 264
165 Times2D: A Time-Frequency Method for Time Series Forecasting

Authors: Reza Nematirad, Anil Pahwa, Balasubramaniam Natarajan

Abstract:

Time series data consist of successive data points collected over a period of time. Accurate prediction of future values is essential for informed decision-making in several real-world applications, including electricity load demand forecasting, lifetime estimation of industrial machinery, traffic planning, weather prediction, and the stock market. Due to their critical relevance and wide application, there has been considerable interest in time series forecasting in recent years. However, the proliferation of sensors and IoT devices, real-time monitoring systems, and high-frequency trading data introduce significant intricate temporal variations, rapid changes, noise, and non-linearities, making time series forecasting more challenging. Classical methods such as Autoregressive integrated moving average (ARIMA) and Exponential Smoothing aim to extract pre-defined temporal variations, such as trends and seasonality. While these methods are effective for capturing well-defined seasonal patterns and trends, they often struggle with more complex, non-linear patterns present in real-world time series data. In recent years, deep learning has made significant contributions to time series forecasting. Recurrent Neural Networks (RNNs) and their variants, such as Long short-term memory (LSTMs) and Gated Recurrent Units (GRUs), have been widely adopted for modeling sequential data. However, they often suffer from the locality, making it difficult to capture local trends and rapid fluctuations. Convolutional Neural Networks (CNNs), particularly Temporal Convolutional Networks (TCNs), leverage convolutional layers to capture temporal dependencies by applying convolutional filters along the temporal dimension. Despite their advantages, TCNs struggle with capturing relationships between distant time points due to the locality of one-dimensional convolution kernels. Transformers have revolutionized time series forecasting with their powerful attention mechanisms, effectively capturing long-term dependencies and relationships between distant time points. However, the attention mechanism may struggle to discern dependencies directly from scattered time points due to intricate temporal patterns. Lastly, Multi-Layer Perceptrons (MLPs) have also been employed, with models like N-BEATS and LightTS demonstrating success. Despite this, MLPs often face high volatility and computational complexity challenges in long-horizon forecasting. To address intricate temporal variations in time series data, this study introduces Times2D, a novel framework that parallelly integrates 2D spectrogram and derivative heatmap techniques. The spectrogram focuses on the frequency domain, capturing periodicity, while the derivative patterns emphasize the time domain, highlighting sharp fluctuations and turning points. This 2D transformation enables the utilization of powerful computer vision techniques to capture various intricate temporal variations. To evaluate the performance of Times2D, extensive experiments were conducted on standard time series datasets and compared with various state-of-the-art algorithms, including DLinear (2023), TimesNet (2023), Non-stationary Transformer (2022), PatchTST (2023), N-HiTS (2023), Crossformer (2023), MICN (2023), LightTS (2022), FEDformer (2022), FiLM (2022), SCINet (2022a), Autoformer (2021), and Informer (2021) under the same modeling conditions. The initial results demonstrated that Times2D achieves consistent state-of-the-art performance in both short-term and long-term forecasting tasks. Furthermore, the generality of the Times2D framework allows it to be applied to various tasks such as time series imputation, clustering, classification, and anomaly detection, offering potential benefits in any domain that involves sequential data analysis.

Keywords: derivative patterns, spectrogram, time series forecasting, times2D, 2D representation

Procedia PDF Downloads 7
164 Direct Cost of Anesthesia in Traumatic Patients with Massive Bleeding: A Prospective Micro-Costing Study

Authors: Asamaporn Puetpaiboon, Sunisa Chatmongkolchart, Nalinee Kovitwanawong, Osaree Akaraborworn

Abstract:

Traumatic patients with massive bleeding require intensive resuscitation. The actual cost of anesthesia per case has never been clarified, so our study aimed to quantify the direct cost, and cost-to-charge ratio of anesthetic care in traumatic patients with intraoperative massive bleeding. This study was a prospective, observational, cost analysis study, conducted in Prince of Songkla University hospital, Thailand, with traumatic patients, of any mechanisms being recruited. Massive bleeding was defined as estimated blood loss of at least one blood volume in 24 hours, or a half of blood volume in 3 hours. The cost components were identified by the micro-costing method, and valued by the bottom-up approach. The direct cost was divided into 4 categories: the labor cost, the capital cost, the material cost and the cost of drugs. From September 2017 to August 2018, 10 patients with multiple injuries were included. Seven patients had motorcycle accidents, two patients fell from a height and another one was in a minibus accident. Two patients died on the operating table, and another two died within 48 hours. The median Sequential Organ Failure Assessment (SOFA) score was 8. The median intraoperative blood loss was 3,500 ml. The median direct cost, per case, was 250 United States Dollars (2017 exchange rate), and the cost-to-charge ratio was 0.53. In summary, the direct cost was nearly half of the hospital charge, for these traumatic patients with massive bleeding. However, our study did not analyze the indirect cost.

Keywords: cost, cost-to-charge ratio, micro-costing, trauma

Procedia PDF Downloads 122
163 Robust Design of a Ball Joint Considering Uncertainties

Authors: Bong-Su Sin, Jong-Kyu Kim, Se-Il Song, Kwon-Hee Lee

Abstract:

An automobile ball joint is a pivoting element used to allow rotational motion between the parts of the steering and suspension system. And it plays a role in smooth transmission of steering movement, also reduction in impact from the road surface. A ball joint is under various repeated loadings that may cause cracks and abrasion. This damages lead to safety problems of a car, as well as reducing the comfort of the driver's ride, and raise questions about the ball joint procedure and the whole durability of the suspension system. Accordingly, it is necessary to ensure the high durability and reliability of a ball joint. The structural responses of stiffness and pull-out strength were then calculated to check if the design satisfies the related requirements. The analysis was sequentially performed, following the caulking process. In this process, the deformation and stress results obtained from the analysis were saved. Sequential analysis has a strong advantage, in that it can be analyzed by considering the deformed shape and residual stress. The pull-out strength means the required force to pull the ball stud out from the ball joint assembly. The low pull-out strength can deteriorate the structural stability and safety performances. In this study, two design variables and two noise factors were set up. Two design variables were the diameter of a stud and the angle of a socket. And two noise factors were defined as the uncertainties of Young's modulus and yield stress of a seat. The DOE comprises 81 cases using these conditions. Robust design of a ball joint was performed using the DOE. The pull-out strength was generated from the uncertainties in the design variables and the design parameters. The purpose of robust design is to find the design with target response and smallest variation.

Keywords: ball joint, pull-out strength, robust design, design of experiments

Procedia PDF Downloads 393
162 Spatial Rank-Based High-Dimensional Monitoring through Random Projection

Authors: Chen Zhang, Nan Chen

Abstract:

High-dimensional process monitoring becomes increasingly important in many application domains, where usually the process distribution is unknown and much more complicated than the normal distribution, and the between-stream correlation can not be neglected. However, since the process dimension is generally much bigger than the reference sample size, most traditional nonparametric multivariate control charts fail in high-dimensional cases due to the curse of dimensionality. Furthermore, when the process goes out of control, the influenced variables are quite sparse compared with the whole dimension, which increases the detection difficulty. Targeting at these issues, this paper proposes a new nonparametric monitoring scheme for high-dimensional processes. This scheme first projects the high-dimensional process into several subprocesses using random projections for dimension reduction. Then, for every subprocess with the dimension much smaller than the reference sample size, a local nonparametric control chart is constructed based on the spatial rank test to detect changes in this subprocess. Finally, the results of all the local charts are fused together for decision. Furthermore, after an out-of-control (OC) alarm is triggered, a diagnostic framework is proposed. using the square-root LASSO. Numerical studies demonstrate that the chart has satisfactory detection power for sparse OC changes and robust performance for non-normally distributed data, The diagnostic framework is also effective to identify truly changed variables. Finally, a real-data example is presented to demonstrate the application of the proposed method.

Keywords: random projection, high-dimensional process control, spatial rank, sequential change detection

Procedia PDF Downloads 276
161 Evaluation of Liquid Fermentation Strategies to Obtain a Biofertilizer Based on Rhizobium sp.

Authors: Andres Diaz Garcia, Ana Maria Ceballos Rojas, Duvan Albeiro Millan Montano

Abstract:

This paper describes the initial technological development stages in the area of liquid fermentation required to reach the quantities of biomass of the biofertilizer microorganism Rhizobium sp. strain B02, for the application of the unitary stages downstream at laboratory scale. In the first stage, the adjustment and standardization of the fermentation process in conventional batch mode were carried out. In the second stage, various fed-batch and continuous fermentation strategies were evaluated in 10L-bioreactor in order to optimize the yields in concentration (Colony Forming Units/ml•h) and biomass (g/l•h), to make feasible the application of unit operations downstream of process. The growth kinetics, the evolution of dissolved oxygen and the pH profile generated in each of the strategies were monitored and used to make sequential adjustments. Once the fermentation was finished, the final concentration and viability of the obtained biomass were determined and performance parameters were calculated with the purpose of select the optimal operating conditions that significantly improved the baseline results. Under the conditions adjusted and standardized in batch mode, concentrations of 6.67E9 CFU/ml were reached after 27 hours of fermentation and a subsequent noticeable decrease was observed associated with a basification of the culture medium. By applying fed-batch and continuous strategies, significant increases in yields were achieved, but with similar concentration levels, which involved the design of several production scenarios based on the availability of equipment usage time and volume of required batch.

Keywords: biofertilizer, liquid fermentation, Rhizobium sp., standardization of processes

Procedia PDF Downloads 150
160 Quality Approaches for Mass-Produced Fashion: A Study in Malaysian Garment Manufacturing

Authors: N. J. M. Yusof, T. Sabir, J. McLoughlin

Abstract:

Garment manufacturing industry involves sequential processes that are subjected to uncontrollable variations. The industry depends on the skill of labour in handling the varieties of fabrics and accessories, machines, and also a complicated sewing operation. Due to these reasons, garment manufacturers created systems to monitor and control the product’s quality regularly by conducting quality approaches to minimize variation. The aims of this research were to ascertain the quality approaches deployed by Malaysian garment manufacturers in three key areas-quality systems and tools; quality control and types of inspection; sampling procedures chosen for garment inspection. The focus of this research also aimed to distinguish quality approaches used by companies that supplied the finished garments to both domestic and international markets. The feedback from each of company’s representatives was obtained using the online survey, which comprised of five sections and 44 questions on the organizational profile and quality approaches used in the garment industry. The results revealed that almost all companies had established their own mechanism of process control by conducting a series of quality inspection for daily production either it was formally been set up or vice versa. Quality inspection was the predominant quality control activity in the garment manufacturing and the level of complexity of these activities was substantially dictated by the customers. AQL-based sampling was utilized by companies dealing with the export market, whilst almost all the companies that only concentrated on the domestic market were comfortable using their own sampling procedures for garment inspection. This research provides an insight into the implementation of quality approaches that were perceived as important and useful in the garment manufacturing sector, which is truly labour-intensive.

Keywords: garment manufacturing, quality approaches, quality control, inspection, Acceptance Quality Limit (AQL), sampling

Procedia PDF Downloads 415
159 Bio Ethanol Production From the Co-Mixture of Jatropha Carcus L. Kernel Cake and Rice Straw

Authors: Felix U. Asoiro, Daniel I. Eleazar, Peter O. Offor

Abstract:

As a result of increasing energy demands, research in bioethanol has increased in recent years all through the world, in abide to partially or totally replace renewable energy supplies. The first and third generation feedstocks used for biofuel production have fundamental drawbacks. Waste rice straw and cake from second generation feedstock like Jatropha curcas l. kernel (JC) is seen as non-food feedstock and promising candidates for the industrial production of bioethanol. In this study, JC and rice husk (RH) wastes were characterized for proximate composition. Bioethanol was produced from the residual polysaccharides present in rice husk (RH) and Jatropha seed cake by sequential hydrolytic and fermentative processes at varying mixing proportions (50 g JC/50 g RH, 100 g JC/10 g RH, 100 g JC/20 g RH, 100 g JC/50 g RH, 100 g JC/100 g RH, 100 g JC/200 g RH and 200 g JC/100 g RH) and particle sizes (0.25, 0.5 and 1.00 mm). Mixing proportions and particle size significantly affected both bioethanol yield and some bioethanol properties. Bioethanol yield (%) increased with an increase in particle size. The highest bioethanol (8.67%) was produced at a mixing proportion of 100 g JC/50g RH at 0.25 mm particle size. The bioethanol had the lowest values of specific gravity and density of 1.25 and 0.92 g cm-3 and the highest values of 1.57 and 0.97 g cm-3 respectively. The highest values of viscosity (4.64 cSt) were obtained with 200 g JC/100 g RH, at 1.00 mm particle size. The maximum flash point and cloud point values were 139.9 oC and 23.7oC (100 g JC/200 g RH) at 1 mm and 0.5 mm particle sizes respectively. The maximum pour point value recorded was 3.85oC (100 g JC/50 g RH) at 1 mm particle size. The paper concludes that bioethanol can be recovered from JC and RH wastes. JC and RH blending proportions as well as particle sizes are important factors in bioethanol production.

Keywords: bioethanol, hydrolysis, Jatropha curcas l. kernel, rice husk, fermentation, proximate composition

Procedia PDF Downloads 68