Search results for: test case specification
1449 Optimizing Production Yield Through Process Parameter Tuning Using Deep Learning Models: A Case Study in Precision Manufacturing
Authors: Tolulope Aremu
Abstract:
This paper is based on the idea of using deep learning methodology for optimizing production yield by tuning a few key process parameters in a manufacturing environment. The study was explicitly on how to maximize production yield and minimize operational costs by utilizing advanced neural network models, specifically Long Short-Term Memory and Convolutional Neural Networks. These models were implemented using Python-based frameworks—TensorFlow and Keras. The targets of the research are the precision molding processes in which temperature ranges between 150°C and 220°C, the pressure ranges between 5 and 15 bar, and the material flow rate ranges between 10 and 50 kg/h, which are critical parameters that have a great effect on yield. A dataset of 1 million production cycles has been considered for five continuous years, where detailed logs are present showing the exact setting of parameters and yield output. The LSTM model would model time-dependent trends in production data, while CNN analyzed the spatial correlations between parameters. Models are designed in a supervised learning manner. For the model's loss, an MSE loss function is used, optimized through the Adam optimizer. After running a total of 100 training epochs, 95% accuracy was achieved by the models recommending optimal parameter configurations. Results indicated that with the use of RSM and DOE traditional methods, there was an increase in production yield of 12%. Besides, the error margin was reduced by 8%, hence consistent quality products from the deep learning models. The monetary value was annually around $2.5 million, the cost saved from material waste, energy consumption, and equipment wear resulting from the implementation of optimized process parameters. This system was deployed in an industrial production environment with the help of a hybrid cloud system: Microsoft Azure, for data storage, and the training and deployment of their models were performed on Google Cloud AI. The functionality of real-time monitoring of the process and automatic tuning of parameters depends on cloud infrastructure. To put it into perspective, deep learning models, especially those employing LSTM and CNN, optimize the production yield by fine-tuning process parameters. Future research will consider reinforcement learning with a view to achieving further enhancement of system autonomy and scalability across various manufacturing sectors.Keywords: production yield optimization, deep learning, tuning of process parameters, LSTM, CNN, precision manufacturing, TensorFlow, Keras, cloud infrastructure, cost saving
Procedia PDF Downloads 321448 The Transformation of Hot Spring Destinations in Taiwan in a Post-pandemic Future: Exploring the COVID-19 Impacts on Hot Spring Experiences and Resilience of Local Residents from a Posttraumatic Growth Perspective
Authors: Hsin-Hung Lin, Janet Chang, Te-Yi Chang, You-Sheng Huang
Abstract:
The natural and men-made disasters have become huge challenges for tourism destinations as well as emphasizing the fragility of the industry. Hot springs, among all destinations, are prone to disasters due to their dependence on natural resources and locations. After the COVID-19 outbreak, hot spring destinations have experienced not only the loss of businesses but also the psychological trauma. However, evidence has also shown that the impacts may not necessarily reduce the resilience for people but may be converted into posttraumatic growth. In Taiwan, a large proportion of hot springs are located in rural or indigenous areas. As a result, hot spring resources are associated with community cohesion for local residents. Yet prior research on hot spring destinations has mainly focused on visitors, whereas residents have been overlooked. More specifically, the relationship between hot springs resources and resident resilience in the face of the COVID-19 impacts remains unclear. To fulfill this knowledge gap, this paper aims to explore the COVID-19 impacts on residents’ hot spring experiences as well as individual and community resilience from the perspective of posttraumatic growth. A total of 315 residents of 13 hot spring destinations that are most popular in Taiwan were recruited. Online questionnaires were distributed over travel forums and social networks after the COVID-19. This paper subsequently used Partial Least Squares Structural Equation Modeling for data analysis as the technique offers significant advantages in addressing nonnormal data and small sample sizes. A preliminary test was conducted, and the results showed acceptable internal consistency and no serious common method variance. The path analysis demonstrated that the COVID-19 impacts strengthened residents’ perceptions of hot spring resources and experiences, implying that the pandemic had propelled the residents to visit hot springs for the healing benefits. In addition, the COVID-19 impacts significantly enhanced residents’ individual and community resilience, which indicates that the residents at hot springs are more resilient thanks to their awareness of external risks. Thirdly, residents’ individual resilience was positively associated with hot spring experiences, while community resilience was not affected by hot spring experiences. Such findings may suggest that hot spring experiences are more related to individual-level experiences and, consequently, have insignificant influence on community resilience. Finally, individual resilience was proved to be the most relevant factor that help foster community resilience. To conclude, the authorities may consider exploiting the hot spring resources so as to increase individual resilience for local residents. Such implications can be used as a reference for other post-disaster tourist destinations as well.As for future research, longitudinal studies with qualitative methods are suggested to better understand how the hot spring experiences have changed individuals and communities over the long term. It should be noted that the main subjects of this paper were focused on the hot spring communities in Taiwan. Therefore, the results cannot be generalized for all types of tourism destinations. That is, more diverse tourism destinations may be investigated to provide a broader perspective of post-disaster recovery.Keywords: community resilience, hot spring destinations, individual resilience, posttraumatic growth (PTG)
Procedia PDF Downloads 741447 Association between Obstetric Factors with Affected Areas of Health-Related Quality of Life of Pregnant Women
Authors: Cinthia G. P. Calou, Franz J. Antezana, Ana I. O. Nicolau, Eveliny S. Martins, Paula R. A. L. Soares, Glauberto S. Quirino, Dayanne R. Oliveira, Priscila S. Aquino, Régia C. M. B. Castro, Ana K. B. Pinheiro
Abstract:
Introduction: As an integral part of the health-disease process, gestation is a period in which the social insertion of women can influence, in a positive or negative way, the course of the pregnancy-puerperal cycle. Thus, evaluating the quality of life of this population can redirect the implementation of innovative practices in the quest to make them more effective and real for the promotion of a more humanized care. This study explores the associations between the obstetric factors with affected areas of health-related quality of life of pregnant women with habitual risk. Methods: This is a cross-sectional, quantitative study conducted in three public facilities and a private service that provides prenatal care in the city of Fortaleza, Ceara, Brazil. The sample consisted of 261 pregnant women who underwent low-risk prenatal care and were interviewed from September to November 2014. The collection instruments were a questionnaire containing socio-demographic and obstetric variables, in addition to the Brazilian version of the Mother scale Generated Index (MGI) characterized by being a specific and objective instrument, consisting of a single sheet and subdivided into three stages. It allows identifying the areas of life of the pregnant woman that are most affected, which could go unnoticed by the pre-formulated measurement instruments. The obstetric data, as well as the data concerning the application of the MGI scale, were compiled and analyzed through the statistical program Statistical Package for the Social Sciences (SPSS), version 20.0. After the compilation, a descriptive analysis was carried out. Then, associations were made between some variables. The tests applied were the Pearson Chi-Square and the Fisher's exact test. The odds ratio was also calculated. These associations were considered statistically significant when the p (probability) value was less than or equal to a level of 5% (α = 0.05) in the tests performed. Results: The variables that negatively reflected the quality of life of the pregnant women and presented a significant association with the polaciuria were: gestational age (p = 0.022) and parity (p = 0.048). Episodes of nausea and vomiting also showed significant with gestational age correlation (p = 0.0001). Evaluating the crossing of stress, we observed a significant association with parity (p = 0.0001). In turn, emotional lability revealed dependence on the variable type of delivery (p = 0.009). Conclusion: The health professionals involved in the assistance to the pregnant woman can understand how the process of gestation is experienced, considering all its peculiar transformations; to meet their individual needs, stimulating their autonomy and their power of choice, envisaging the achievement of a better quality of life related to health in the perspective of health promotion.Keywords: health-related quality of life, obstetric nursing, pregnant women, prenatal care
Procedia PDF Downloads 2931446 Variability and Stability of Bread and Durum Wheat for Phytic Acid Content
Authors: Gordana Branković, Vesna Dragičević, Dejan Dodig, Desimir Knežević, Srbislav Denčić, Gordana Šurlan-Momirović
Abstract:
Phytic acid is a major pool in the flux of phosphorus through agroecosystems and represents a sum equivalent to > 50% of all phosphorus fertilizer used annually. Nutrition rich in phytic acid can substantially decrease micronutrients apsorption as calcium, zink, iron, manganese, copper due to phytate salts excretion by human and non-ruminant animals as poultry, swine and fish, having in common very scarce phytase activity, and consequently the ability to digest and utilize phytic acid, thus phytic acid derived phosphorus in animal waste contributes to water pollution. The tested accessions consisted of 15 genotypes of bread wheat (Triticum aestivum L. ssp. vulgare) and of 15 genotypes of durum wheat (Triticum durum Desf.). The trials were sown at the three test sites in Serbia: Rimski Šančevi (RS) (45º19´51´´N; 19º50´59´´E), Zemun Polje (ZP) (44º52´N; 20º19´E) and Padinska Skela (PS) (44º57´N 20º26´E) during two vegetation seasons 2010-2011 and 2011-2012. The experimental design was randomized complete block design with four replications. The elementary plot consisted of 3 internal rows of 0.6 m2 area (3 × 0.2 m × 1 m). Grains were grinded with Laboratory Mill 120 Perten (“Perten”, Sweden) (particles size < 500 μm) and flour was used for the analysis. Phytic acid grain content was determined spectrophotometrically with the Shimadzu UV-1601 spectrophotometer (Shimadzu Corporation, Japan). Objectives of this study were to determine: i) variability and stability of the phytic acid content among selected genotypes of bread and durum wheat, ii) predominant source of variation regarding genotype (G), environment (E) and genotype × environment interaction (GEI) from the multi-environment trial, iii) influence of climatic variables on the GEI for the phytic acid content. Based on the analysis of variance it had been determined that the variation of phytic acid content was predominantly influenced by environment in durum wheat, while the GEI prevailed for the variation of the phytic acid content in bread wheat. Phytic acid content expressed on the dry mass basis was in the range 14.21-17.86 mg g-1 with the average of 16.05 mg g-1 for bread wheat and 14.63-16.78 mg g-1 with the average of 15.91 mg g-1 for durum wheat. Average-environment coordination view of the genotype by environment (GGE) biplot was used for the selection of the most desirable genotypes for breeding for low phytic acid content in the sense of good stability and lower level of phytic acid content. The most desirable genotypes of bread and durum wheat for breeding for phytic acid were Apache and 37EDUYT /07 No. 7849. Models of climatic factors in the highest percentage (> 91%) were useful in interpreting GEI for phytic acid content, and included relative humidity in June, sunshine hours in April, mean temperature in April and winter moisture reserves for genotypes of bread wheat, as well as precipitation in June and April, maximum temperature in April and mean temperature in June for genotypes of durum wheat.Keywords: genotype × environment interaction, phytic acid, stability, variability
Procedia PDF Downloads 3941445 Disconnect between Water, Sanitation and Hygiene Related Behaviours of Children in School and Family
Authors: Rehan Mohammad
Abstract:
Background: Improved Water, Sanitation and Hygiene (WASH) practices in schools ensure children’s health, well-being and cognitive performance. In India under various WASH interventions in schools, teachers, and other staff make every possible effort to educate children about personal hygiene, sanitation practices and harms of open defecation. However, once children get back to their families, they see other practicing inappropriate WASH behaviors, and they consequently start following them. This show disconnect between school behavior and family behavior, which needs to be bridged to achieve desired WASH outcomes. Aims and Objectives: The aim of this study is to assess the factors causing disconnect of WASH-related behaviors between school and the family of children. It also suggests behavior change interventions to bridge the gap. Methodology: The present study has chosen a mixed- method approach. Both quantitative and qualitative methods of data collection have been used in the present study. The purposive sampling for data collection has been chosen. The data have been collected from 20% children in each age group of 04-08 years and 09-12 years spread over three primary schools and 20% of households to which they belong to which is spread over three slum communities in south district of Delhi. Results: The present study shows that despite of several behavior change interventions at school level, children still practice inappropriate WASH behaviors due to disconnect between school and family behaviors. These behaviors show variation from one age group to another. The inappropriate WASH behaviors being practiced by children include open defecation, wrong disposal of garbage, not keeping personal hygiene, not practicing hand washing practices during critical junctures and not washing fruits and vegetables before eating. The present study has highlighted that 80% of children in the age group of 04-08 years still practice inappropriate WASH behaviors when they go back to their families after school whereas, this percentage has reduced to 40% in case of children in the age group 09-12 years. Present study uncovers association between school and family teaching which creates a huge gap between WASH-related behavioral practices. The study has established that children learn and de-learn the WASH behaviors due to the evident disconnect between behavior change interventions at schools and household level. The study has also made it clear that children understand the significance of appropriate WASH practices but owing to the disconnect the behaviors remain unsettled. The study proposes several behavior change interventions to sync the behaviors of children at school and family level to ensure children’s health, well-being and cognitive performance.Keywords: behavioral interventions, child health, family behavior, school behavior, WASH
Procedia PDF Downloads 1111444 Other Cancers in Patients With Head and Neck Cancer
Authors: Kim Kennedy, Daren Gibson, Stephanie Flukes, Chandra Diwakarla, Lisa Spalding, Leanne Pilkington, Andrew Redfern
Abstract:
Introduction: Head and neck cancers (HNC) are often associated with the development of non-HNC primaries, as the risk factors that predispose patients to HNC are often risk factors for other cancers. Aim: We sought to evaluate whether there was an increased risk of smoking and alcohol-related cancers and also other cancers in HNC patients and to evaluate whether there is a difference between the rates of non-HNC primaries in Aboriginal compared with non-Aboriginal HNC patients. Methods: We performed a retrospective cohort analysis of 320 HNC patients from a single center in Western Australia, identifying 80 Aboriginal and 240 non-Aboriginal patients matched on a 1:3 ratio by sites, histology, rurality, and age. We collected data on the patient characteristics, tumour features, treatments, outcomes, and past and subsequent HNCs and non-HNC primaries. Results: In the overall study population, there were 86 patients (26.9%) with a metachronous or synchronous non-HNC primary. Non-HNC primaries were actually significantly more common in the non-Aboriginal population compared with the Aboriginal population (30% vs. 17.5%, p=0.02); however, half of these were patients with cutaneous squamous or basal cell carcinomas (cSCC/BCC) only. When cSCC/BCCs were excluded, non-Aboriginal patients had a similar rate as Aboriginal patients (16.7% vs. 15%, p=0.73). There were clearly more cSCC/BCCs in non-Aboriginal patients compared with Aboriginal patients (16.7% vs. 2.5%, p=0.001) and more patients with melanoma (2.5% vs. 0%, p value not significant (p=NS). Rates of most cancers were similar between non-Aboriginal and Aboriginal patients, including prostate (2.9% vs. 3.8%), colorectal (2.9% vs. 2.5%), kidney (1.2% vs. 1.2%), and these rates appeared comparable to Australian Age Standardised Incidence Rates (ASIR) in the general community. Oesophageal cancer occurred at double the rate in Aboriginal patients (3.8%) compared with non-Aboriginal patients (1.7%), which was far in excess of ASIRs which estimated a lifetime risk of 0.59% in the general population. Interestingly lung cancer rates did not appear to be significantly increased in our cohort, with 2.5% of Aboriginal patients and 3.3% of non-Aboriginal patients having lung cancer, which is in line with ASIRs which estimates a lifetime risk of 5% (by age 85yo). Interestingly the rate of Glioma in the non-Aboriginal population was higher than the ASIR, with 0.8% of non-Aboriginal patients developing Glioma, with Australian averages predicting a 0.6% lifetime risk in the general population. As these are small numbers, this finding may well be due to chance. Unsurprisingly, second HNCs occurred at an increased incidence in our cohort, in 12.5% of Aboriginal patients and 11.2% of non-Aboriginal patients, compared to an ASIR of 17 cases per 100,000 persons, estimating a lifetime risk of 1.70%. Conclusions: Overall, 26.9% of patients had a non-HNC primary. When cSCC/BCCs were excluded, Aboriginal and non-Aboriginal patients had similar rates of non-HNC primaries, although non-Aboriginal patients had a significantly higher rate of cSCC/BCCs. Aboriginal patients had double the rate of oesophageal primaries; however, this was not statistically significant, possibly due to small case numbers.Keywords: head and neck cancer, synchronous and metachronous primaries, other primaries, Aboriginal
Procedia PDF Downloads 761443 Comparison of Parametric and Bayesian Survival Regression Models in Simulated and HIV Patient Antiretroviral Therapy Data: Case Study of Alamata Hospital, North Ethiopia
Authors: Zeytu G. Asfaw, Serkalem K. Abrha, Demisew G. Degefu
Abstract:
Background: HIV/AIDS remains a major public health problem in Ethiopia and heavily affecting people of productive and reproductive age. We aimed to compare the performance of Parametric Survival Analysis and Bayesian Survival Analysis using simulations and in a real dataset application focused on determining predictors of HIV patient survival. Methods: A Parametric Survival Models - Exponential, Weibull, Log-normal, Log-logistic, Gompertz and Generalized gamma distributions were considered. Simulation study was carried out with two different algorithms that were informative and noninformative priors. A retrospective cohort study was implemented for HIV infected patients under Highly Active Antiretroviral Therapy in Alamata General Hospital, North Ethiopia. Results: A total of 320 HIV patients were included in the study where 52.19% females and 47.81% males. According to Kaplan-Meier survival estimates for the two sex groups, females has shown better survival time in comparison with their male counterparts. The median survival time of HIV patients was 79 months. During the follow-up period 89 (27.81%) deaths and 231 (72.19%) censored individuals registered. The average baseline cluster of differentiation 4 (CD4) cells count for HIV/AIDS patients were 126.01 but after a three-year antiretroviral therapy follow-up the average cluster of differentiation 4 (CD4) cells counts were 305.74, which was quite encouraging. Age, functional status, tuberculosis screen, past opportunistic infection, baseline cluster of differentiation 4 (CD4) cells, World Health Organization clinical stage, sex, marital status, employment status, occupation type, baseline weight were found statistically significant factors for longer survival of HIV patients. The standard error of all covariate in Bayesian log-normal survival model is less than the classical one. Hence, Bayesian survival analysis showed better performance than classical parametric survival analysis, when subjective data analysis was performed by considering expert opinions and historical knowledge about the parameters. Conclusions: Thus, HIV/AIDS patient mortality rate could be reduced through timely antiretroviral therapy with special care on the potential factors. Moreover, Bayesian log-normal survival model was preferable than the classical log-normal survival model for determining predictors of HIV patients survival.Keywords: antiretroviral therapy (ART), Bayesian analysis, HIV, log-normal, parametric survival models
Procedia PDF Downloads 1961442 Time of Week Intensity Estimation from Interval Censored Data with Application to Police Patrol Planning
Authors: Jiahao Tian, Michael D. Porter
Abstract:
Law enforcement agencies are tasked with crime prevention and crime reduction under limited resources. Having an accurate temporal estimate of the crime rate would be valuable to achieve such a goal. However, estimation is usually complicated by the interval-censored nature of crime data. We cast the problem of intensity estimation as a Poisson regression using an EM algorithm to estimate the parameters. Two special penalties are added that provide smoothness over the time of day and day of the week. This approach presented here provides accurate intensity estimates and can also uncover day-of-week clusters that share the same intensity patterns. Anticipating where and when crimes might occur is a key element to successful policing strategies. However, this task is complicated by the presence of interval-censored data. The censored data refers to the type of data that the event time is only known to lie within an interval instead of being observed exactly. This type of data is prevailing in the field of criminology because of the absence of victims for certain types of crime. Despite its importance, the research in temporal analysis of crime has lagged behind the spatial component. Inspired by the success of solving crime-related problems with a statistical approach, we propose a statistical model for the temporal intensity estimation of crime with censored data. The model is built on Poisson regression and has special penalty terms added to the likelihood. An EM algorithm was derived to obtain maximum likelihood estimates, and the resulting model shows superior performance to the competing model. Our research is in line with the smart policing initiative (SPI) proposed by the Bureau Justice of Assistance (BJA) as an effort to support law enforcement agencies in building evidence-based, data-driven law enforcement tactics. The goal is to identify strategic approaches that are effective in crime prevention and reduction. In our case, we allow agencies to deploy their resources for a relatively short period of time to achieve the maximum level of crime reduction. By analyzing a particular area within cities where data are available, our proposed approach could not only provide an accurate estimate of intensities for the time unit considered but a time-variation crime incidence pattern. Both will be helpful in the allocation of limited resources by either improving the existing patrol plan with the understanding of the discovery of the day of week cluster or supporting extra resources available.Keywords: cluster detection, EM algorithm, interval censoring, intensity estimation
Procedia PDF Downloads 661441 Computational Characterization of Electronic Charge Transfer in Interfacial Phospholipid-Water Layers
Authors: Samira Baghbanbari, A. B. P. Lever, Payam S. Shabestari, Donald Weaver
Abstract:
Existing signal transmission models, although undoubtedly useful, have proven insufficient to explain the full complexity of information transfer within the central nervous system. The development of transformative models will necessitate a more comprehensive understanding of neuronal lipid membrane electrophysiology. Pursuant to this goal, the role of highly organized interfacial phospholipid-water layers emerges as a promising case study. A series of phospholipids in neural-glial gap junction interfaces as well as cholesterol molecules have been computationally modelled using high-performance density functional theory (DFT) calculations. Subsequent 'charge decomposition analysis' calculations have revealed a net transfer of charge from phospholipid orbitals through the organized interfacial water layer before ultimately finding its way to cholesterol acceptor molecules. The specific pathway of charge transfer from phospholipid via water layers towards cholesterol has been mapped in detail. Cholesterol is an essential membrane component that is overrepresented in neuronal membranes as compared to other mammalian cells; given this relative abundance, its apparent role as an electronic acceptor may prove to be a relevant factor in further signal transmission studies of the central nervous system. The timescales over which this electronic charge transfer occurs have also been evaluated by utilizing a system design that systematically increases the number of water molecules separating lipids and cholesterol. Memory loss through hydrogen-bonded networks in water can occur at femtosecond timescales, whereas existing action potential-based models are limited to micro or nanosecond scales. As such, the development of future models that attempt to explain faster timescale signal transmission in the central nervous system may benefit from our work, which provides additional information regarding fast timescale energy transfer mechanisms occurring through interfacial water. The study possesses a dataset that includes six distinct phospholipids and a collection of cholesterol. Ten optimized geometric characteristics (features) were employed to conduct binary classification through an artificial neural network (ANN), differentiating cholesterol from the various phospholipids. This stems from our understanding that all lipids within the first group function as electronic charge donors, while cholesterol serves as an electronic charge acceptor.Keywords: charge transfer, signal transmission, phospholipids, water layers, ANN
Procedia PDF Downloads 731440 An Improved Adaptive Dot-Shape Beamforming Algorithm Research on Frequency Diverse Array
Authors: Yanping Liao, Zenan Wu, Ruigang Zhao
Abstract:
Frequency diverse array (FDA) beamforming is a technology developed in recent years, and its antenna pattern has a unique angle-distance-dependent characteristic. However, the beam is always required to have strong concentration, high resolution and low sidelobe level to form the point-to-point interference in the concentrated set. In order to eliminate the angle-distance coupling of the traditional FDA and to make the beam energy more concentrated, this paper adopts a multi-carrier FDA structure based on proposed power exponential frequency offset to improve the array structure and frequency offset of the traditional FDA. The simulation results show that the beam pattern of the array can form a dot-shape beam with more concentrated energy, and its resolution and sidelobe level performance are improved. However, the covariance matrix of the signal in the traditional adaptive beamforming algorithm is estimated by the finite-time snapshot data. When the number of snapshots is limited, the algorithm has an underestimation problem, which leads to the estimation error of the covariance matrix to cause beam distortion, so that the output pattern cannot form a dot-shape beam. And it also has main lobe deviation and high sidelobe level problems in the case of limited snapshot. Aiming at these problems, an adaptive beamforming technique based on exponential correction for multi-carrier FDA is proposed to improve beamforming robustness. The steps are as follows: first, the beamforming of the multi-carrier FDA is formed under linear constrained minimum variance (LCMV) criteria. Then the eigenvalue decomposition of the covariance matrix is performed to obtain the diagonal matrix composed of the interference subspace, the noise subspace and the corresponding eigenvalues. Finally, the correction index is introduced to exponentially correct the small eigenvalues of the noise subspace, improve the divergence of small eigenvalues in the noise subspace, and improve the performance of beamforming. The theoretical analysis and simulation results show that the proposed algorithm can make the multi-carrier FDA form a dot-shape beam at limited snapshots, reduce the sidelobe level, improve the robustness of beamforming, and have better performance.Keywords: adaptive beamforming, correction index, limited snapshot, multi-carrier frequency diverse array, robust
Procedia PDF Downloads 1301439 The Connection between De Minimis Rule and the Effect on Trade
Authors: Pedro Mario Gonzalez Jimenez
Abstract:
The novelties introduced by the last Notice on agreements of minor importance tighten the application of the ‘De minimis’ safe harbour in the European Union. However, the undetermined legal concept of effect on trade between the Member States becomes importance at the same time. Therefore, the current analysis that the jurist should carry out in the European Union to determine if an agreement appreciably restrict competition under Article 101 of the Treaty on the Functioning of the European Union is double. Hence, it is necessary to know how to balance the significance in competition and the significance in effect on trade between the Member States. It is a crucial issue due to the negative delimitation of restriction of competition affects the positive one. The methodology of this research is rather simple. Beginning with a historical approach to the ‘De Minimis Rule’, their main problems and uncertainties will be found. So, after the analysis of normative documents and the jurisprudence of the Court of Justice of the European Union some proposals of ‘Lege ferenda’ will be offered. These proposals try to overcome the contradictions and questions that currently exist in the European Union as a consequence of the current legal regime of agreements of minor importance. The main findings of this research are the followings: Firstly, the effect on trade is another way to analyze the importance of an agreement different from the ‘De minimis rule’. In point of fact, this concept is singularly adapted to go through agreements that have as object the prevention, restriction or distortion of competition, as it is observed in the most famous European Union case-law. Thanks to the effect on trade, as long as the proper requirements are met there is no a restriction of competition under article 101 of the Treaty on the Functioning of the European Union, even if the agreement had an anti-competitive object. These requirements are an aggregate market share lower than 5% on any of the relevant markets affected by the agreement and turnover lower than 40 million of Euros. Secondly, as the Notice itself says ‘it is also intended to give guidance to the courts and competition authorities of the Member States in their application of Article 101 of the Treaty, but it has no binding force for them’. This reality makes possible the existence of different statements among the different Member States and a confusing perception of what a restriction of competition is. Ultimately, damage on trade between the Member States could be observed for this reason. The main conclusion is that the significant effect on trade between Member States is irrelevant in agreements that restrict competition because of their effects but crucial in agreements that restrict competition because of their object. Thus, the Member States should propose the incorporation of a similar concept in their legal orders in order to apply the content of the Notice. Otherwise, the significance of the restrictive agreement on competition would not be properly assessed.Keywords: De minimis rule, effect on trade, minor importance agreements, safe harbour
Procedia PDF Downloads 1821438 Detection and Identification of Antibiotic Resistant UPEC Using FTIR-Microscopy and Advanced Multivariate Analysis
Authors: Uraib Sharaha, Ahmad Salman, Eladio Rodriguez-Diaz, Elad Shufan, Klaris Riesenberg, Irving J. Bigio, Mahmoud Huleihel
Abstract:
Antimicrobial drugs have played an indispensable role in controlling illness and death associated with infectious diseases in animals and humans. However, the increasing resistance of bacteria to a broad spectrum of commonly used antibiotics has become a global healthcare problem. Many antibiotics had lost their effectiveness since the beginning of the antibiotic era because many bacteria have adapted defenses against these antibiotics. Rapid determination of antimicrobial susceptibility of a clinical isolate is often crucial for the optimal antimicrobial therapy of infected patients and in many cases can save lives. The conventional methods for susceptibility testing require the isolation of the pathogen from a clinical specimen by culturing on the appropriate media (this culturing stage lasts 24 h-first culturing). Then, chosen colonies are grown on media containing antibiotic(s), using micro-diffusion discs (second culturing time is also 24 h) in order to determine its bacterial susceptibility. Other methods, genotyping methods, E-test and automated methods were also developed for testing antimicrobial susceptibility. Most of these methods are expensive and time-consuming. Fourier transform infrared (FTIR) microscopy is rapid, safe, effective and low cost method that was widely and successfully used in different studies for the identification of various biological samples including bacteria; nonetheless, its true potential in routine clinical diagnosis has not yet been established. The new modern infrared (IR) spectrometers with high spectral resolution enable measuring unprecedented biochemical information from cells at the molecular level. Moreover, the development of new bioinformatics analyses combined with IR spectroscopy becomes a powerful technique, which enables the detection of structural changes associated with resistivity. The main goal of this study is to evaluate the potential of the FTIR microscopy in tandem with machine learning algorithms for rapid and reliable identification of bacterial susceptibility to antibiotics in time span of few minutes. The UTI E.coli bacterial samples, which were identified at the species level by MALDI-TOF and examined for their susceptibility by the routine assay (micro-diffusion discs), are obtained from the bacteriology laboratories in Soroka University Medical Center (SUMC). These samples were examined by FTIR microscopy and analyzed by advanced statistical methods. Our results, based on 700 E.coli samples, were promising and showed that by using infrared spectroscopic technique together with multivariate analysis, it is possible to classify the tested bacteria into sensitive and resistant with success rate higher than 90% for eight different antibiotics. Based on these preliminary results, it is worthwhile to continue developing the FTIR microscopy technique as a rapid and reliable method for identification antibiotic susceptibility.Keywords: antibiotics, E.coli, FTIR, multivariate analysis, susceptibility, UTI
Procedia PDF Downloads 1731437 Electrochemical Performance of Femtosecond Laser Structured Commercial Solid Oxide Fuel Cells Electrolyte
Authors: Mohamed A. Baba, Gazy Rodowan, Brigita Abakevičienė, Sigitas Tamulevičius, Bartlomiej Lemieszek, Sebastian Molin, Tomas Tamulevičius
Abstract:
Solid oxide fuel cells (SOFC) efficiently convert hydrogen to energy without producing any disturbances or contaminants. The core of the cell is electrolyte. For improving the performance of electrolyte-supported cells, it is desirable to extend the available exchange surface area by micro-structuring of the electrolyte with laser-based micromachining. This study investigated the electrochemical performance of cells micro machined using a femtosecond laser. Commercial ceramic SOFC (Elcogen, AS) with a total thickness of 400 μm was structured by 1030 nm wavelength Yb: KGW fs-laser Pharos (Light Conversion) using 100 kHz repetition frequency and 290 fs pulse length light by scanning with the galvanometer scanner (ScanLab) and focused with a f-Theta telecentric lens (SillOptics). The sample height was positioned using a motorized z-stage. The microstructures were formed using a laser spiral trepanning in Ni/YSZ anode supported membrane at the central part of the ceramic piece of 5.5 mm diameter at active area of the cell. All surface was drilled with 275 µm diameter holes spaced by 275 µm. The machining processes were carried out under ambient conditions. The microstructural effects of the femtosecond laser treatment on the electrolyte surface were investigated prior to the electrochemical characterisation using a scanning electron microscope (SEM) Quanta 200 FEG (FEI). The Novo control Alpha-A was used for electrochemical impedance spectroscopy on a symmetrical cell configuration with an excitation amplitude of 25 mV and a frequency range of 1 MHz to 0.1 Hz. The fuel cell characterization of the cell was examined on open flanges test setup by Fiaxell. Using nickel mesh on the anode side and au mesh on the cathode side, the cell was electrically linked. The cell was placed in a Kittec furnace with a Process IDentifier temperature controller. The wires were connected to a Solartron 1260/1287 frequency analyzer for the impedance and current-voltage characterization. In order to determine the impact of the anode's microstructure on the performance of the commercial cells, the acquired results were compared to cells with unstructured anode. Geometrical studies verified that the depth of the -holes increased linearly according to laser energy and scanning times. On the other hand, it reduced as the scanning speed increased. The electrochemical analysis demonstrates that the open circuit voltage OCV values of the two cells are equal. Further, the modified cell's initial slope reduces to 0.209 from 0.253 of the unmodified cell, revealing that the surface modification considerably decreases energy loss. Plus, the maximum power density for the cell with the microstructure and the reference cell respectively, are 1.45 and 1.16 Wcm⁻².Keywords: electrochemical performance, electrolyte-supported cells, laser micro-structuring, solid oxide fuel cells
Procedia PDF Downloads 691436 Building Brand Equity in a Stigmatised Market: A Cannabis Industry Case Study
Authors: Sibongile Masemola
Abstract:
In 2018, South Africa decriminalised recreational cannabis use and private cultivation, since then, cannabis businesses have been established to meet the demand. However, marketing activities remain limited in this industry, and businesses are unable to disseminate promotional messages, however, as a solution, firms can promote their brands and positioning instead of the actual product (Bick, 2015). Branding is essential to create differences among cannabis firms and to attract and keep customers (Abrahamsson, 2014). Building cannabis firms into brands can better position them in the mind of the consumer so that they become and remain competitive. The aim of this study was to explore how South African cannabis retailers can build brand equity in a stigmatised market, despite significant restrictions on marketing efforts. Keller’s (2001) customer-based brand equity (CBBE) model was used as the as the theoretical framework and explored how cannabis firms build their businesses into brands through developing their brand identity, meaning, performance, and relationships, and ultimately creating brand equity. The study employed a qualitative research method, using semi-structured in-depth interviews among 17 participants to gain insights from cannabis owners and marketers in the recreational cannabis environment. Most findings were presented according to the blocks of CBBE model. Furthermore, a conceptual framework named the stigma-based brand equity (SBBE) model was adapted from Keller’s CBBE model to include an additional building block that accounts for industry-specific characteristics unique to stigmatised markets. Findings revealed the pervasiveness of education and its significance to brand building in a stigmatised industry. Results also demonstrated the overall effect stigma has on businesses and their consumers due to the longstanding negative evaluations of cannabis. Hence, through stigma-bonding, brands can develop deep identity-related psychological bonds with their consumers that will potentially lead to strong brand resonance. This study aims to contribute business-relevant knowledge for firms operating in core-stigmatised markets under controlled marketing regulations by exploring how cannabis firms can build brand equity. Practically, this study presents recommendations for retailers in stigmatised markets on how to destigmatise, build brand identity, create brand meaning, elicit desired brand responses, and develop brand relationships – ultimately building brand equity.Keywords: branding, brand equity, cannabis, organisational stigma
Procedia PDF Downloads 1041435 Reflective Thinking and Experiential Learning – A Quasi-Experimental Quanti-Quali Response to Greater Diversification of Activities, Greater Integration of Student Profiles
Authors: Paulo Sérgio Ribeiro de Araújo Bogas
Abstract:
Although several studies have assumed (at least implicitly) that learners' approaches to learning develop into deeper approaches to higher education, there appears to be no clear theoretical basis for this assumption and no empirical evidence. As a scientific contribution to this discussion, a pedagogical intervention of a quasi-experimental nature was developed, with a mixed methodology, evaluating the intervention within a single curricular unit of Marketing, using cases based on real challenges of brands, business simulation, and customer projects. Primary and secondary experiences were incorporated in the intervention: the primary experiences are the experiential activities themselves; the secondary experiences result from the primary experience, such as reflection and discussion in work teams. A diversified learning relationship was encouraged through the various connections between the different members of the learning community. The present study concludes that in the same context, the student's responses can be described as students who reinforce the initial deep approach, students who maintain the initial deep approach level, and others who change from an emphasis on the deep approach to one closer to superficial. This typology did not always confirm studies reported in the literature, namely, whether the initial level of deep processing would influence the superficial and the opposite. The result of this investigation points to the inclusion of pedagogical and didactic activities that integrate different motivations and initial strategies, leading to the possible adoption of deep approaches to learning since it revealed statistically significant differences in the difference in the scores of the deep/superficial approach and the experiential level. In the case of real challenges, the categories of “attribution of meaning and meaning of studied” and the possibility of “contact with an aspirational context” for their future professional stand out. In this category, the dimensions of autonomy that will be required of them were also revealed when comparing the classroom context of real cases and the future professional context and the impact they may have on the world. Regarding the simulated practice, two categories of response stand out: on the one hand, the motivation associated with the possibility of measuring the results of the decisions taken, an awareness of oneself, and, on the other hand, the additional effort that this practice required for some of the students.Keywords: experiential learning, higher education, mixed methods, reflective learning, marketing
Procedia PDF Downloads 831434 Cost Based Analysis of Risk Stratification Tool for Prediction and Management of High Risk Choledocholithiasis Patients
Authors: Shreya Saxena
Abstract:
Background: Choledocholithiasis is a common complication of gallstone disease. Risk scoring systems exist to guide the need for further imaging or endoscopy in managing choledocholithiasis. We completed an audit to review the American Society for Gastrointestinal Endoscopy (ASGE) scoring system for prediction and management of choledocholithiasis against the current practice at a tertiary hospital to assess its utility in resource optimisation. We have now conducted a cost focused sub-analysis on patients categorized high-risk for choledocholithiasis according to the guidelines to determine any associated cost benefits. Method: Data collection from our prior audit was used to retrospectively identify thirteen patients considered high-risk for choledocholithiasis. Their ongoing management was mapped against the guidelines. Individual costs for the key investigations were obtained from our hospital financial data. Total cost for the different management pathways identified in clinical practice were calculated and compared against predicted costs associated with recommendations in the guidelines. We excluded the cost of laparoscopic cholecystectomy and considered a set figure for per day hospital admission related expenses. Results: Based on our previous audit data, we identified a77% positive predictive value for the ASGE risk stratification tool to determine patients at high-risk of choledocholithiasis. 47% (6/13) had an magnetic resonance cholangiopancreatography (MRCP) prior to endoscopic retrograde cholangiopancreatography (ERCP), whilst 53% (7/13) went straight for ERCP. The average length of stay in the hospital was 7 days, with an additional day and cost of £328.00 (£117 for ERCP) for patients awaiting an MRCP prior to ERCP. Per day hospital admission was valued at £838.69. When calculating total cost, we assumed all patients had admission bloods and ultrasound done as the gold standard. In doing an MRCP prior to ERCP, there was a 130% increase in cost incurred (£580.04 vs £252.04) per patient. When also considering hospital admission and the average length of stay, it was an additional £1166.69 per patient. We then calculated the exact costs incurred by the department, over a three-month period, for all patients, for key investigations or procedures done in the management of choledocholithiasis. This was compared to an estimate cost derived from the recommended pathways in the ASGE guidelines. Overall, 81% (£2048.45) saving was associated with following the guidelines compared to clinical practice. Conclusion: MRCP is the most expensive test associated with the diagnosis and management of choledocholithiasis. The ASGE guidelines recommend endoscopy without an MRCP in patients stratified as high-risk for choledocholithiasis. Our audit that focused on assessing the utility of the ASGE risk scoring system showed it to be relatively reliable for identifying high-risk patients. Our cost analysis has shown significant cost savings per patient and when considering the average length of stay associated with direct endoscopy rather than an additional MRCP. Part of this is also because of an increased average length of stay associated with waiting for an MRCP. The above data supports the ASGE guidelines for the management of high-risk for choledocholithiasis patients from a cost perspective. The only caveat is our small data set that may impact the validity of our average length of hospital stay figures and hence total cost calculations.Keywords: cost-analysis, choledocholithiasis, risk stratification tool, general surgery
Procedia PDF Downloads 981433 Prognostic Factors for Mortality and Duration of Admission in Malnourished Hospitalized, Elderly Patients: A Cross-Sectional Study
Authors: Christos E. Lampropoulos, Maria Konsta, Vicky Dradaki, Irini Dri, Tamta Sirbilatze, Ifigenia Apostolou, Christina Kordali, Konstantina Panouria, Kostas Argyros, Georgios Mavras
Abstract:
Malnutrition in hospitalized patients is related to increased morbidity and mortality. Purpose of our study was to assess nutritional status of hospitalized, elderly patients with various nutritional scores and to detect unfavorable prognostic factors, related to increased mortality and extended duration of admission. Methods: 150 patients (78 men, 72 women, mean age 80±8.2) were included in this cross-sectional study. Nutritional status was assessed by Mini Nutritional Assessment (MNA full, short-form), Malnutrition Universal Screening Tool (MUST) and short Nutritional Appetite Questionnaire (sNAQ). The following data were incorporated in analysis: Anthropometric and laboratory data, physical activity (International Physical Activity Questionnaires, IPAQ), smoking status, dietary habits and mediterranean diet (assessed by MedDiet score), cause and duration of current admission, medical history (co-morbidities, previous admissions). Primary endpoints were the mortality (from admission until 6 months afterwards) and duration of admission, compared to national guidelines for closed consolidated medical expenses. Mann-Whitney two-sample statistics or t-test was used for group comparisons and Spearman or Pearson coefficients for testing correlation between variables. Results: Normal nutrition was assessed in 54/150 (36%), 92/150 (61.3%) and in 106/150 (70.7%) of patients, according to full MNA, MUST and sNAQ questionnaires respectively. Mortality rate was 20.7% (31/150 patients). The patients who died until 6 months after admission had lower BMI (24±4.4 vs 26±4.8, p=0.04) and albumin levels (2.9±0.7 vs 3.4±0.7, p=0.002), significantly lower full MNA (14.5±7.3 vs 20.7±6, p<0.0001) and short-form MNA scores (7.3±4.2 vs 10.5±3.4, p=0.0002) compared to non-dead one. In contrast, the aforementioned patients had higher MUST (2.5±1.8 vs 0.5±1.02, p=<0.0001) and sNAQ scores (2.9±2.4 vs 1.1±1.3, p<0.0001). Additionally, they showed significantly lower MedDiet (23.5±4.3 vs 31.1±5.6, p<0.0001) and IPAQ scores (37.2±156.2 vs 516.5±1241.7, p<0.0001) compared to remaining one. These patients had extended hospitalization [5 (0-13) days vs 0 (-1-3) days, p=0.001]. Patients who admitted due to cancer depicted higher mortality rate (10/13, 77%), compared to those who admitted due to infections (12/73, 18%), stroke (4/15, 27%) or other causes (4/49, 8%) (p<0.0001). Extension of hospitalization was negatively correlated to both full (Spearman r=-0.35, p<0.0001) and short-form MNA (Spearman r=-0.33, p<0.0001) and positively correlated to MUST (Spearman r=0.34, p<0.0001) and sNAQ (Spearman r=0.3, p=0.0002). Additionally, the extension was inversely related to MedDiet score (Spearman r=-0.35, p<0.0001), IPAQ score (Spearman r=-0.34, p<0.0001), albumin levels (Pearson r=-0.36, p<0.0001), Ht (Pearson r=-0.2, p=0.02) and Hb (Pearson r=-0.18, p=0.02). Conclusion: A great proportion of elderly, hospitalized patients are malnourished or at risk of malnutrition. All nutritional scores, physical activity and albumin are significantly related to mortality and increased hospitalization.Keywords: dietary habits, duration of admission, malnutrition, prognostic factors for mortality
Procedia PDF Downloads 2891432 Investigating English Dominance in a Chinese-English Dual Language Program: Teachers' Language Use and Investment
Authors: Peizhu Liu
Abstract:
Dual language education, also known as immersion education, differs from traditional language programs that teach a second or foreign language as a subject. Instead, dual language programs adopt a content-based approach, using both a majority language (e.g., English, in the case of the United States) and a minority language (e.g., Spanish or Chinese) as a medium of instruction to teach math, science, and social studies. By granting each language of instruction equal status, dual language education seeks to educate not only meaningfully but equitably and to foster tolerance and appreciation of diversity, making it essential for immigrants, refugees, indigenous peoples, and other marginalized students. Despite the cognitive and academic benefits of dual language education, recent literature has revealed that English is disproportionately privileged across dual language programs. Scholars have expressed concerns about the unbalanced status of majority and minority languages in dual language education, as favoring English in this context may inadvertently reaffirm its dominance and moreover fail to serve the needs of children whose primary language is not English. Through a year-long study of a Chinese-English dual language program, the extensively disproportionate use of English has also been observed by the researcher. However, despite the fact that Chinese-English dual language programs are the second-most popular program type after Spanish in the United States, this issue remains underexplored in the existing literature on Chinese-English dual language education. In fact, the number of Chinese-English dual language programs being offered in the U.S. has grown rapidly, from 8 in 1988 to 331 as of 2023. Using Norton and Darvin's investment model theory, the current study investigates teachers' language use and investment in teaching Chinese and English in a Chinese-English dual language program at an urban public school in New York City. The program caters to a significant number of minority children from working-class families. Adopting an ethnographic and discourse analytic approach, this study seeks to understand language use dynamics in the program and how micro- and macro-factors, such as students' identity construction, parents' and teachers' language ideologies, and the capital associated with each language, influence teachers' investment in teaching Chinese and English. The research will help educators and policymakers understand the obstacles that stand in the way of the goal of dual language education—that is, the creation of a more inclusive classroom, which is achieved by regarding both languages of instruction as equally valuable resources. The implications for how to balance the use of the majority and minority languages will also be discussed.Keywords: dual language education, bilingual education, language immersion education, content-based language teaching
Procedia PDF Downloads 841431 Work Related Outcomes of Perceived Authentic Leadership: Moderating Role of Organizational Structures
Authors: Aisha Zubair, Anila Kamal
Abstract:
Leadership styles and practices greatly influence the organizational effectiveness and productivity. It also plays an important role in employees’ experiences of positive emotions at workplace and creative work behaviors. Authentic leadership as a newly emerging concept has been found as a significant predictor of various desirable work related outcomes. However, leadership practices and its work related outcomes, to a great extent, are determined by the very nature of the organizational structures (tall and flat). Tall organizations are characterized by multiple hierarchical layers with predominant vertical communication patterns, and narrow span of control; while flat organizations are featured by few layers of management employing both horizontal and vertical communication styles, and wide span of control. Therefore, the present study was undertaken to determine the work related outcomes of perceived authentic leadership; that is work related flow and creative work behavior among employees of flat and tall organizations. Moreover, it was also intended to determine the moderating role of organizational structure (flat and tall) in the relationship between perceived authentic leadership with work related flow and creative work behavior. In this regard, two types of companies have been considered; that is, banks as a form of tall organizational structure with multiple hierarchical structures while software companies have been considered as flat organizations with minimal layers of management. Respondents (N = 1180) were full time regular employees of marketing departments of banks (600) and software companies (580) including both men and women with age range of 22-52 years (M = 33.24; SD = 7.81). Confirmatory Factor Analysis yielded factor structures of measures of work related flow and creative work behavior in accordance to the theoretical models. However, model of authentic leadership exhibited variation in terms of two items which were not included in the final measure of the perceived authentic leadership. Results showed that perceived authentic leadership was positively associated with work related flow and creative work behavior. Likewise, work related flow was positively aligned with creative work behavior. Furthermore, type of organizational structure significantly moderated the relationship of perceived authentic leadership with work related flow and creative work behavior. Results of independent sample t-test showed that employees working in flat organization reflected better perceptions of authentic leadership; higher work related flow and elevated levels of creative work behavior as compared to those working in tall organizations. It was also found that employees with extended job experience and more job duration in the same organization displayed better perceptions of authentic leadership, reported more work related flow and augmented levels of creative work behavior. Findings of the present study distinctively highlighted the similarities as well as differences in the interactions of major constructs which function differentially in the context of tall (banks) and flat (software companies) organizations. Implications of the present study for employees and management as well as future recommendations were also discussed.Keywords: creative work behavior, organizational structure, perceived authentic leadership, work related flow
Procedia PDF Downloads 3891430 Geochemical Study of Natural Bitumen, Condensate and Gas Seeps from Sousse Area, Central Tunisia
Authors: Belhaj Mohamed, M. Saidi, N. Boucherab, N. Ouertani, I. Bouazizi, M. Ben Jrad
Abstract:
Natural hydrocarbon seepage has helped petroleum exploration as a direct indicator of gas and/or oil subsurface accumulations. Surface macro-seeps are generally an indication of a fault in an active Petroleum Seepage System belonging to a Total Petroleum System. This paper describes a case study in which multiple analytical techniques were used to identify and characterize trace petroleum-related hydrocarbons and other volatile organic compounds in groundwater samples collected from Sousse aquifer (Central Tunisia). The analytical techniques used for analyses of water samples included gas chromatography-mass spectrometry (GC-MS), capillary GC with flame-ionization detection, Compund Specific Isotope Analysis, Rock Eval Pyrolysis. The objective of the study was to confirm the presence of gasoline and other petroleum products or other volatile organic pollutants in those samples in order to assess the respective implication of each of the potentially responsible parties to the contamination of the aquifer. In addition, the degree of contamination at different depths in the aquifer was also of interest. The oil and gas seeps have been investigated using biomarker and stable carbon isotope analyses to perform oil-oil and oil-source rock correlations. The seepage gases are characterized by high CH4 content, very low δ13CCH4 values (-71,9 ‰) and high C1/C1–5 ratios (0.95–1.0), light deuterium–hydrogen isotope ratios (-198 ‰) and light δ13CC2 and δ13CCO2 values (-23,8‰ and-23,8‰ respectively) indicating a thermogenic origin with the contribution of the biogenic gas. An organic geochemistry study was carried out on the more ten oil seep samples. This study includes light hydrocarbon and biomarkers analyses (hopanes, steranes, n-alkanes, acyclic isoprenoids, and aromatic steroids) using GC and GC-MS. The studied samples show at least two distinct families, suggesting two different types of crude oil origins: the first oil seeps appears to be highly mature, showing evidence of chemical and/or biological degradation and was derived from a clay-rich source rock deposited in suboxic conditions. It has been sourced mainly by the lower Fahdene (Albian) source rocks. The second oil seeps was derived from a carbonate-rich source rock deposited in anoxic conditions, well correlated with the Bahloul (Cenomanian-Turonian) source rock.Keywords: biomarkers, oil and gas seeps, organic geochemistry, source rock
Procedia PDF Downloads 4431429 Integration Process and Analytic Interface of different Environmental Open Data Sets with Java/Oracle and R
Authors: Pavel H. Llamocca, Victoria Lopez
Abstract:
The main objective of our work is the comparative analysis of environmental data from Open Data bases, belonging to different governments. This means that you have to integrate data from various different sources. Nowadays, many governments have the intention of publishing thousands of data sets for people and organizations to use them. In this way, the quantity of applications based on Open Data is increasing. However each government has its own procedures to publish its data, and it causes a variety of formats of data sets because there are no international standards to specify the formats of the data sets from Open Data bases. Due to this variety of formats, we must build a data integration process that is able to put together all kind of formats. There are some software tools developed in order to give support to the integration process, e.g. Data Tamer, Data Wrangler. The problem with these tools is that they need data scientist interaction to take part in the integration process as a final step. In our case we don’t want to depend on a data scientist, because environmental data are usually similar and these processes can be automated by programming. The main idea of our tool is to build Hadoop procedures adapted to data sources per each government in order to achieve an automated integration. Our work focus in environment data like temperature, energy consumption, air quality, solar radiation, speeds of wind, etc. Since 2 years, the government of Madrid is publishing its Open Data bases relative to environment indicators in real time. In the same way, other governments have published Open Data sets relative to the environment (like Andalucia or Bilbao). But all of those data sets have different formats and our solution is able to integrate all of them, furthermore it allows the user to make and visualize some analysis over the real-time data. Once the integration task is done, all the data from any government has the same format and the analysis process can be initiated in a computational better way. So the tool presented in this work has two goals: 1. Integration process; and 2. Graphic and analytic interface. As a first approach, the integration process was developed using Java and Oracle and the graphic and analytic interface with Java (jsp). However, in order to open our software tool, as second approach, we also developed an implementation with R language as mature open source technology. R is a really powerful open source programming language that allows us to process and analyze a huge amount of data with high performance. There are also some R libraries for the building of a graphic interface like shiny. A performance comparison between both implementations was made and no significant differences were found. In addition, our work provides with an Official Real-Time Integrated Data Set about Environment Data in Spain to any developer in order that they can build their own applications.Keywords: open data, R language, data integration, environmental data
Procedia PDF Downloads 3151428 Linking Enhanced Resting-State Brain Connectivity with the Benefit of Desirable Difficulty to Motor Learning: A Functional Magnetic Resonance Imaging Study
Authors: Chien-Ho Lin, Ho-Ching Yang, Barbara Knowlton, Shin-Leh Huang, Ming-Chang Chiang
Abstract:
Practicing motor tasks arranged in an interleaved order (interleaved practice, or IP) generally leads to better learning than practicing tasks in a repetitive order (repetitive practice, or RP), an example of how desirable difficulty during practice benefits learning. Greater difficulty during practice, e.g. IP, is associated with greater brain activity measured by higher blood-oxygen-level dependent (BOLD) signal in functional magnetic resonance imaging (fMRI) in the sensorimotor areas of the brain. In this study resting-state fMRI was applied to investigate whether increase in resting-state brain connectivity immediately after practice predicts the benefit of desirable difficulty to motor learning. 26 healthy adults (11M/15F, age = 23.3±1.3 years) practiced two sets of three sequences arranged in a repetitive or an interleaved order over 2 days, followed by a retention test on Day 5 to evaluate learning. On each practice day, fMRI data were acquired in a resting state after practice. The resting-state fMRI data was decomposed using a group-level spatial independent component analysis (ICA), yielding 9 independent components (IC) matched to the precuneus network, primary visual networks (two ICs, denoted by I and II respectively), sensorimotor networks (two ICs, denoted by I and II respectively), the right and the left frontoparietal networks, occipito-temporal network, and the frontal network. A weighted resting-state functional connectivity (wRSFC) was then defined to incorporate information from within- and between-network brain connectivity. The within-network functional connectivity between a voxel and an IC was gauged by a z-score derived from the Fisher transformation of the IC map. The between-network connectivity was derived from the cross-correlation of time courses across all possible pairs of ICs, leading to a symmetric nc x nc matrix of cross-correlation coefficients, denoted by C = (pᵢⱼ). Here pᵢⱼ is the extremum of cross-correlation between ICs i and j; nc = 9 is the number of ICs. This component-wise cross-correlation matrix C was then projected to the voxel space, with the weights for each voxel set to the z-score that represents the above within-network functional connectivity. The wRSFC map incorporates the global characteristics of brain networks measured by the between-network connectivity, and the spatial information contained in the IC maps measured by the within-network connectivity. Pearson correlation analysis revealed that greater IP-minus-RP difference in wRSFC was positively correlated with the RP-minus-IP difference in the response time on Day 5, particularly in brain regions crucial for motor learning, such as the right dorsolateral prefrontal cortex (DLPFC), and the right premotor and supplementary motor cortices. This indicates that enhanced resting brain connectivity during the early phase of memory consolidation is associated with enhanced learning following interleaved practice, and as such wRSFC could be applied as a biomarker that measures the beneficial effects of desirable difficulty on motor sequence learning.Keywords: desirable difficulty, functional magnetic resonance imaging, independent component analysis, resting-state networks
Procedia PDF Downloads 2031427 Lean Implementation in a Nurse Practitioner Led Pediatric Primary Care Clinic: A Case Study
Authors: Lily Farris, Chantel E. Canessa, Rena Heathcote, Susan Shumay, Suzanna V. McRae, Alissa Collingridge, Minna K. Miller
Abstract:
Objective: To describe how the Lean approach can be applied to improve access, quality and safety of care in an ambulatory pediatric primary care setting. Background: Lean was originally developed by Toyota manufacturing in Japan, and subsequently adapted for use in the healthcare sector. Lean is a systematic approach, focused on identifying and reducing waste within organizational processes, improving patient-centered care and efficiency. Limited literature is available on the implementation of the Lean methodologies in a pediatric ambulatory care setting. Methods: A strategic continuous improvement event or Rapid Process Improvement Workshop (RPIW) was launched with the aim evaluating and structurally supporting clinic workflow, capacity building, sustainability, and ultimately improving access to care and enhancing the patient experience. The Lean process consists of five specific activities: Current state/process assessment (value stream map); development of a future state map (value stream map after waste reduction); identification, quantification and prioritization of the process improvement opportunities; implementation and evaluation of process changes; and audits to sustain the gains. Staff engagement is a critical component of the Lean process. Results: Through the implementation of the RPIW and shifting workload among the administrative team, four hours of wasted time moving between desks and doing work was eliminated from the Administrative Clerks role. To streamline clinic flow, the Nursing Assistants completed patient measurements and vitals for Nurse Practitioners, reducing patient wait times and adding value to the patients visit with the Nurse Practitioners. Additionally, through the Nurse Practitioners engagement in the Lean processes a need was recognized to articulate clinic vision, mission and the alignment of NP role and scope of practice with the agency and Ministry of Health strategic plan. Conclusions: Continuous improvement work in the Pediatric Primary Care NP Clinic has provided a unique opportunity to improve the quality of care delivered and has facilitated further alignment of the daily continuous improvement work with the strategic priorities of the Ministry of Health.Keywords: ambulatory care, lean, pediatric primary care, system efficiency
Procedia PDF Downloads 3001426 Community Engagement in Child Centered Space at Disaster Events: A Case Story of Sri Lanka
Authors: Wasantha Pushpakumara Hitihami Mudiyanselage
Abstract:
Since recent past, Sri Lanka is highly vulnerable to reoccurring climate shocks that severely impact the food security, loss of human & animal lives, destructions of human settlements, displacement of people and damaging properties. Hence, the Government of Sri Lanka has taken important steps towards strengthening legal and institutional arrangements for Disaster Risks management in the country in May 2005. Puttalam administrative district is one of the disaster prone districts in Sri Lanka which constantly face the devastating consequences of the increasing natural disasters annually. Therefore disaster risk management will be a timely intervention in the area to minimize the adverse impacts of the disasters. The few functioning Disaster Risk management networks do not take children’s specific needs and vulnerabilities during emergencies into account. The most affected children and their families were evacuated to the government schools and temples and it was observed that children were left to roaming around as their parents were busy queuing up for relief goods and other priorities. In this sense, VOICE understands that the community has vital role that has to be played in facing challenges of disaster management in the area. During and after the disaster, it was viewed that some children were having psychological disorders which could be impacted negatively to children well–being. Need of child friendly space at emergency is a must action in the area to turn away negative impact coming from the hazards. VOICE with the support of national & international communities have established safer places for the children (Child Centered Spaces – CCS) and their families at emergencies. Village religious venues and schools were selected and equipped with necessary materials to be used for the children at emergency. Materials such as tools, stationeries, play materials, which couldn’t be easily found in surrounding environment, were provided for CCS centers. Village animators, youth and elders were given comprehensive training on Disaster management and their role at CCS. They did the facilitation in keeping children without fear and stress at flooding occurred in 2015 as well as they were able to improve their skills when working with children. Flooding in 2016, the government agencies have taken service of these village animators at early stage of flooding to make all disaster-related recovery actions productively & efficiently. This mechanism is sustained at village level that can be used for disaster events.Keywords: child centered space, impacts, psychological disorders, village animators
Procedia PDF Downloads 1321425 Exploring the Potential of Mobile Learning in Distance Higher Education: A Case Study of the University of Jammu, Jammu, and Kashmir
Authors: Darshana Sharma
Abstract:
Distance Education has emerged as a viable alternative to serve the higher educational needs of the socially and economically disadvantaged people of the remote, rural areas of Jammu region. The University of Jammu is a National Accreditation, and Assessment Council accredited, A+ university and has been accorded graded autonomy by the University Grants Commission. It is a dual mode university offering academic programmes through the regular departments and through the Directorate of Distance Education. The Directorate of Distance Education, University of Jammu still uses printed study material as a mode of instructional delivery. The development of technologies has assured increased interaction and communication for distance learners throughout the distance open learning institutions. Though it is tempting and convenient to adopt technology already being used by others, it may not prove effective for the simple reason that two institutions may be unlike in some respect. The use of technology must be conceived in view of the needs of the learners; geographical socio-economic-cultural and technological contexts and financial, administrative and academic resources of the institution. Mobile learning (m-learning) is a novel approach to knowledge acquisition and dissemination and is gaining global attention. It has evolved as one of the useful channels of distance learning promoting interaction between learners and teachers. It is felt that the Directorate of Distance Education, University of Jammu also needs to adopt new technologies to provide more effective academic and information support to distance learners in order to keep them motivated and also to develop self-learning skills. The chief objective of the research on which this paper is based was to measure the opinion of the distance learners of the DDE, the University of Jammu about the merits of mobile learning. It also explores their preferences for implementing mobile learning. The survey research design of descriptive research has been used. The data was collected from 400 distance learners enrolled with undergraduate and post-graduate programmes using self-constructed questionnaire containing five-point Likert scale items arranging from strongly agree, agree, indifferent, disagree and strongly disagree. Percentages were used to analyze the data. The findings lead to conclude that mobile learning has a great potential for the DDE for reaching out to the rural, remotely located distance learners of the Jammu region and also to improve the teaching-learning environment. The paper also finds out the challenges in the implementation of mobile learning in the region and further makes suggestions for effective implementation of mobile learning in DDE, University of Jammu.Keywords: directorate of distance education, mobile learning, national accreditation and assessment council, university of Jammu
Procedia PDF Downloads 1231424 Old Houses for Tomorrow: Deliberating a Societal Need for Conserving Unprotected Heritage Houses in India
Authors: Protyoy Sen
Abstract:
Heritage conservation often holds different meanings and values for different people. To a cultural or architectural body it might be about protecting relics of the past, while for an government body or corporate it might be the value of the real estate which generates profits in terms of hospitality, tourism or some form of trade. But often, a significant proportion of the built fabric in our cities comprises of what usually does not come under the common lenses of collective heritage or conservation i.e. private houses. Standing ode to a bygone era of different communities, trades and practices that once inhabited the city, old private houses of certain architectural or historic character face the gravest challenges of heritage conservation. These – despite being significant to the heritage fabric of a city – neither get the social attention nor the financial aid for repair and periodic maintenance, that many monuments and public buildings do. The situation in India is no different. Private residences belonging to affluent families of an earlier time, today lie in varying degrees of neglect and dilapidation. With the growth of nuclear families, drastic change in people’s and expensive repairs of historic material fabric (amongst other reasons), houses of heritage value often become liabilities, and metaphorical to a white elephant in a poor man’s backyard. In a capitalistic setup that values time and money over everything, it is not reasonable that one justifies the conservation of individual / family assets solely through architectural, historical or cultural values. It is quite logical them, that the houseowner – in most cases, a layperson – must be made to understand of both tangible and intangible values in order to (1) take the trouble of the effort, resources and aid (if possible) to repair and maintain a house of heritage character and, (2) choose to invest into a building that today might’ve lost its practical relevance, over demolishing and building new. The question that still remains is – Why? If heritage conservation is to be seen as an economically viable and realistic building activity, it must shed its image of being an ‘elitist, cultural pursuit’ in the eyes of the common person. Through contextual studies of historic areas in Ahmedabad and Calcutta, reading of theoretical pieces on the subject and conversations with multiple stakeholders, this study intended to justify the act of heritage conservation to the common person – one who is assumed to have no particular sensitivity towards architectural or cultural value, and rather questions what these buildings tangibly bring to the table. The theoretical frameworks (taken from literature) are then tested through actual case studies in Indian cities, followed by an elaborate inference on the subject.Keywords: heritage values, heritage houses, private ownership, unprotected heritage
Procedia PDF Downloads 551423 Examination of the South African Fire Legislative Framework
Authors: Mokgadi Julia Ngoepe-Ntsoane
Abstract:
The article aims to make a case for a legislative framework for the fire sector in South Africa. Robust legislative framework is essential for empowering those with obligatory mandate within the sector. This article contributes to the body of knowledge in the field of policy reviews particularly with regards to the legal framework. It has been observed overtime that the scholarly contributions in this field are limited. Document analysis was the methodology selected for the investigation of the various legal frameworks existing in the country. It has been established that indeed the national legislation on the fire industry does not exist in South Africa. From the documents analysed, it was revealed that the sector is dominated by cartels who are exploiting the new entrants to the market particularly SMEs. It is evident that these cartels are monopolising the system as they have long been operating in the system turning it into self- owned entities. Commitment to addressing the challenges faced by fire services and creating a framework for the evolving role that fire brigade services are expected to execute in building safer and sustainable communities is vital. Legislation for the fire sector ought to be concluded with immediate effect. The outdated national fire legislation has necessitated the monopolisation and manipulation of the system by dominating organisations which cause a painful discrimination and exploitation of smaller service providers to enter the market for trading in that occupation. The barrier to entry bears long term negative effects on national priority areas such as employment creation, poverty, and others. This monopolisation and marginalisation practices by cartels in the sector calls for urgent attention by government because if left attended, it will leave a lot of people particularly women and youth being disadvantaged and frustrated. The downcast syndrome exercised within the fire sector has wreaked havoc and is devastating. This is caused by cartels that have been within the sector for some time, who know the strengths and weaknesses of processes, shortcuts, advantages and consequences of various actions. These people take advantage of new entrants to the sector who in turn find it difficult to manoeuvre, find the market dissonant and end up giving up their good ideas and intentions. There are many pieces of legislation which are industry specific such as housing, forestry, agriculture, health, security, environmental which are used to regulate systems within the institutions involved. Other regulations exist as bi-laws for guiding the management within the municipalities.Keywords: sustainable job creation, growth and development, transformation, risk management
Procedia PDF Downloads 1751422 Comparison of the Chest X-Ray and Computerized Tomography Scans Requested from the Emergency Department
Authors: Sahabettin Mete, Abdullah C. Hocagil, Hilal Hocagil, Volkan Ulker, Hasan C. Taskin
Abstract:
Objectives and Goals: An emergency department is a place where people can come for a multitude of reasons 24 hours a day. As it is an easy, accessible place, thanks to self-sacrificing people who work in emergency departments. But the workload and overcrowding of emergency departments are increasing day by day. Under these circumstances, it is important to choose a quick, easily accessible and effective test for diagnosis. This results in laboratory and imaging tests being more than 40% of all emergency department costs. Despite all of the technological advances in imaging methods and available computerized tomography (CT), chest X-ray, the older imaging method, has not lost its appeal and effectiveness for nearly all emergency physicians. Progress in imaging methods are very convenient, but physicians should consider the radiation dose, cost, and effectiveness, as well as imaging methods to be carefully selected and used. The aim of the study was to investigate the effectiveness of chest X-ray in immediate diagnosis against the advancing technology by comparing chest X-ray and chest CT scan results of the patients in the emergency department. Methods: Patients who applied to Bulent Ecevit University Faculty of Medicine’s emergency department were investigated retrospectively in between 1 September 2014 and 28 February 2015. Data were obtained via MIAMED (Clear Canvas Image Server v6.2, Toronto, Canada), information management system which patients’ files are saved electronically in the clinic, and were retrospectively scanned. The study included 199 patients who were 18 or older, had both chest X-ray and chest CT imaging. Chest X-ray images were evaluated by the emergency medicine senior assistant in the emergency department, and the findings were saved to the study form. CT findings were obtained from already reported data by radiology department in the clinic. Chest X-ray was evaluated with seven questions in terms of technique and dose adequacy. Patients’ age, gender, application complaints, comorbid diseases, vital signs, physical examination findings, diagnosis, chest X-ray findings and chest CT findings were evaluated. Data saved and statistical analyses have made via using SPSS 19.0 for Windows. And the value of p < 0.05 were accepted statistically significant. Results: 199 patients were included in the study. In 38,2% (n=76) of all patients were diagnosed with pneumonia and it was the most common diagnosis. The chest X-ray imaging technique was appropriate in patients with the rate of 31% (n=62) of all patients. There was not any statistically significant difference (p > 0.05) between both imaging methods (chest X-ray and chest CT) in terms of determining the rates of displacement of the trachea, pneumothorax, parenchymal consolidation, increased cardiothoracic ratio, lymphadenopathy, diaphragmatic hernia, free air levels in the abdomen (in sections including the image), pleural thickening, parenchymal cyst, parenchymal mass, parenchymal cavity, parenchymal atelectasis and bone fractures. Conclusions: When imaging findings, showing cases that needed to be quickly diagnosed, were investigated, chest X-ray and chest CT findings were matched at a high rate in patients with an appropriate imaging technique. However, chest X-rays, evaluated in the emergency department, were frequently taken with an inappropriate technique.Keywords: chest x-ray, chest computerized tomography, chest imaging, emergency department
Procedia PDF Downloads 1921421 Technological Tool-Use as an Online Learner Strategy in a Synchronous Speaking Task
Authors: J. Knight, E. Barberà
Abstract:
Language learning strategies have been defined as thoughts and actions, consciously chosen and operationalized by language learners, to help them in carrying out a multiplicity of tasks from the very outset of learning to the most advanced levels of target language performance. While research in the field of Second Language Acquisition has focused on ‘good’ language learners, the effectiveness of strategy-use and orchestration by effective learners in face-to-face classrooms much less research has attended to learner strategies in online contexts, particular strategies in relation to technological tool use which can be part of a task design. In addition, much research on learner strategies and strategy use has been explored focusing on cognitive, attitudinal and metacognitive behaviour with less research focusing on the social aspect of strategies. This study focuses on how learners mediate with a technological tool designed to support synchronous spoken interaction and how this shape their spoken interaction in the opening of their talk. A case study approach is used incorporating notions from communities of practice theory to analyse and understand learner strategies of dyads carrying out a role play task. The study employs analysis of transcripts of spoken interaction in the openings of the talk along with log files of tool use. The study draws on results of previous studies pertaining to the same tool as a form of triangulation. Findings show how learners gain pre-task planning time through technological tool control. The strategies involving learners’ choices to enter and exit the tool shape their spoken interaction qualitatively, with some cases demonstrating long silences whilst others appearing to start the pedagogical task immediately. Who/what learners orientate to in the openings of the talk: an audience (i.e. the teacher), each other and/or screen-based signifiers in the opening moments of the talk also becomes a focus. The study highlights how tool use as a social practice should be considered a learning strategy in online contexts whereby different usages may be understood in the light of the more usual asynchronous social practices of the online community. The teachers’ role in the community is also problematised as the evaluator of the practices of that community. Results are pertinent for task design for synchronous speaking tasks. The use of community of practice theory supports an understanding of strategy use that involves both metacognition alongside social context revealing how tool-use strategies may need to be orally (socially) negotiated by learners and may also differ from an online language community.Keywords: learner strategy, tool use, community of practice, speaking task
Procedia PDF Downloads 3421420 Creative Resolutions to Intercultural Conflicts: The Joint Effects of International Experience and Cultural Intelligence
Authors: Thomas Rockstuhl, Soon Ang, Kok Yee Ng, Linn Van Dyne
Abstract:
Intercultural interactions are often challenging and fraught with conflicts. To shed light on how to interact effectively across cultures, academics and practitioners alike have advanced a plethora of intercultural competence models. However, the majority of this work has emphasized distal outcomes, such as job performance and cultural adjustment, rather than proximal outcomes, such as how individuals resolve inevitable intercultural conflicts. As a consequence, the processes by which individuals negotiate challenging intercultural conflicts are not well understood. The current study advances theorizing on intercultural conflict resolution by exploring antecedents of how people resolve intercultural conflicts. To this end, we examine creativity – the generation of novel and useful ideas – in the context of resolving cultural conflicts in intercultural interactions. Based on the dual-identity theory of creativity, we propose that individuals with greater international experience will display greater creativity and that the relationship is accentuated by individual’s cultural intelligence. Two studies test these hypotheses. The first study comprises 84 senior university students, drawn from an international organizational behavior course. The second study replicates findings from the first study in a sample of 89 executives from eleven countries. Participants in both studies provided protocols of their strategies for resolving two intercultural conflicts, as depicted in two multimedia-vignettes of challenging intercultural work-related interactions. Two research assistants, trained in intercultural management but blind to the study hypotheses, coded all strategies for their novelty and usefulness following scoring procedures for creativity tasks. Participants also completed online surveys of demographic background information, including their international experience, and cultural intelligence. Hierarchical linear modeling showed that surprisingly, while international experience is positively associated with usefulness, it is unrelated to novelty. Further, a person’s cultural intelligence strengthens the positive effect of international experience on usefulness and mitigates the effect of international experience on novelty. Theoretically, our findings offer an important theoretical extension to the dual-identity theory of creativity by identifying cultural intelligence as an important individual difference moderator that qualifies the relationship between international experience and creative conflict resolution. In terms of novelty, individuals higher in cultural intelligence seem less susceptible to rigidity effects of international experiences. Perhaps they are more capable of assessing which aspects of culture are relevant and apply relevant experiences when they brainstorm novel ideas. For utility, individuals high in cultural intelligence are better able to leverage on their international experience to assess the viability of their ideas because their richer and more organized cultural knowledge structure allows them to assess possible options more efficiently and accurately. In sum, our findings suggest that cultural intelligence is an important and promising intercultural competence that fosters creative resolutions to intercultural conflicts. We hope that our findings stimulate future research on creativity and conflict resolution in intercultural contexts.Keywords: cultural Intelligence, intercultural conflict, intercultural creativity, international experience
Procedia PDF Downloads 148